Image Title

Search Results for Starburst:

Justin Borgman, Starburst & Ashwin Patil, Deloitte | AWS re:Invent 2022


 

(electronic music) (graphics whoosh) (graphics tinkle) >> Welcome to Las Vegas! It's theCUBE live at AWS re:Invent '22. Lisa Martin here with Dave Vellante. Dave, it is not only great to be back, but this re:Invent seems to be bigger than last year for sure. >> Oh, definitely. I'd say it's double last year. I'd say it's comparable to 2019. Maybe even a little bigger, I've heard it's the largest re:Invent ever. And we're going to talk data, one of our favorite topics. >> We're going to talk data products. We have some great guests. One of them is an alumni who's back with us. Justin Borgman, the CEO of Starburst, and Ashwin Patil also joins us, Principal AI and Data Engineering at Deloitte. Guys, welcome to the program. >> Thank you. >> Together: Thank you. >> Justin, define data products. Give us the scoop, what's goin' on with Starburst. But define data products and the value in it for organizations of productizing data. >> Mm-hmm. So, data products are curated data sets that are able to span across multiple data sets. And I think that's what's makes it particularly unique, is you can span across multiple data sources to create federated data products that allow you to really bring together the business value that you're seeking. And I think ultimately, what's driving the interest in data products is a desire to ultimately facilitate self-service consumption within the enterprise. I think that's the holy grail that we've all been building towards. And data products represents a framework for sort of how you would do that. >> So, monetization is not necessarily a criterion? >> Not necessarily. (Dave's voice drowns) >> But it could be. >> It could be. It can be internal data products or external data products. And in either case, it's really intended to facilitate easier discovery and consumption of data. >> Ashwin, bringing you into the conversation, talk about some of the revenue drivers that data products can help organizations to unlock. >> Sure. Like Justin said, there are internal and external revenue drivers. So internally, a lot of clients are focused around, hey, how do I make the most out of my modernization platform? So, a lot of them are thinking about what AI, what analytics, what can they run to drive consumption? And when you think about consumption, consumption typically requires data from across the enterprise, right? And data from the enterprise is sometimes fragmented in pieces, in places. So, we've gone from being data in too many places to now, data products, helping bring all of that together, and really aid, drive business decisions faster with more data and more accuracy, right? Externally, a lot of that has got to do with how the ecosystems are evolving for data products that use not only company data, but also the ecosystem data that includes customers, that include suppliers and vendors. >> I mean, conceptually, data products, you could say have been around a long time. When I think of financial services, I think that's always been a data product in a sense. But suddenly, there's a lot more conversation about it. There's data mesh, there's data fabric, we could talk about that too, but why do you think now it's coming to the fore again? >> Yeah, I mean, I think it's because historically, there's always been this disconnect between the people that understand data infrastructure, and the people who know the right questions to ask of the data. Generally, these have been two very distinct groups. And so, the interest in data mesh as you mentioned, and data products as a foundational element of it, is really centered around how do we bring these groups together? How do we get the people who know the data the best to participate in the process of creating data to be consumed? Ultimately, again, trying to facilitate greater self-service consumption. And I think that's the real beauty behind it. And I think increasingly, in today's world, people are realizing the data will always be decentralized to some degree. That notion of bringing everything together into one single database has never really been successfully achieved, and is probably even further from the truth at this point in time, given you've got data on-prem and multiple clouds, and multiple different systems. And so, data products and data mesh represents, again, a framework for you to sort of think about data that lives everywhere. >> We did a session this summer with (chuckles) Justin and I, and some others on the data lies. And that was one of the good ol' lies, right? There's a single source of truth. >> Justin: Right. >> And all that is, we've probably never been further from the single source of truth. But actually, you're suggesting that there's maybe multiple truths that the same data can support. Is that a right way to think about it? >> Yeah, exactly. And I think ultimately, you want a single point of access that gives you, at your fingertips, everything that your organization knows about its business today. And that's really what data products aims to do, is sort of curate that for you, and provide high quality data sets that you can trust, that you can now self-service to answer your business question. >> One of the things that, oh, go ahead. >> No, no, I was just going to say, I mean, if you pivot it from the way the usage of data has changed, right? Traditionally, IT has been in the business of providing data to the business users. Today, with more self-service being driven, we want business users to be the drivers of consumption, right? So if you take that backwards one step, it's basically saying, what data do I need to support my business needs, such that IT doesn't always have to get involved in providing that data, or providing the reports on top of that data? So, the data products concept, I think supports that thinking of business-led technology-enabled, or IT-enabled really well. >> Business led. One of the things that Adam Zelinsky talked with John Furrier about just a week or so ago in their pre re:Invent interview, was talking about the role of the data analyst going away. That everybody in an organization, regardless of function, will be able to eventually be a data analyst, and need to evaluate and analyze data for their roles. Talk about data products as a facilitator of that democratization. >> Yeah. We are seeing more and more the concept of citizen data scientists. We are seeing more and more citizens AI. What we are seeing is a general trend, as we move towards self-service, there is going to be a need for business users to be able to access data when they want, how they want, and merge data across the enterprise in ways that they haven't done before, right? Technology today, through products like data products, right, provides you the access to do that. And that's why we are going to see this movement of people of seeing people become more and more self-service oriented, where you're going to democratize the use of AI and analytics into the business users. >> Do you think, when you talk to a data analyst, by the way, about that, he or she will be like, yeah, mm, maybe, good luck with that. So, do ya think maybe there's a sort of an interim step? Because we've had these highly, ZeMac lays this out very well. We've had these highly-centralized, highly-specialized teams. The premise being, oh, that's less expensive. Perhaps data analysts, like functions, get put into the line of business. Do you see that as a bridge or a stepping stone? Because it feels like it's quite a distance between what a data analyst does today, and this nirvana that we talk about. What are your thoughts on that? >> Yeah, I mean, I think there's possibly a new role around a data product manager. Much the way you have product managers in the products you actually build to sell, you might need data product managers to help facilitate and curate the high quality data products that others can consume. And I think that becomes an interesting and important, a skill set. Much the way that data scientist was created as a occupation, if you will, maybe 10 years ago, when previously, those were statisticians, or other names. >> Right. A big risk that many clients are seeing around data products is, how do you drive governance? And to that, to the point that Justin's making, we are going to see that role evolve where governance in the world, where data products are getting democratized is going to become increasingly important in terms of how are data products being generated, how is the propensity of data products towards a more governed environment being managed? And that's going to continue to play an important role as data products evolve. >> Okay, so how do you guys fit, because you take ZeMac's four principles, domain ownership, data as product. And that creates two problems. Governance. (chuckles) Right? How do you automate, and self-service, infrastructure and automated governance. >> Yep. >> Tell us what role Starburst plays in solving all of those, but the latter two in particular. >> Yeah. Well, we're working on all four of those dimensions to some degree, but I think ultimately, what we're focused today is the governance piece, providing fine-grained access controls, which is so important, if you're going to have a single point of access, you better have a way of controlling who has access to what. But secondly, data products allows you to really abstract away or decouple where the data is stored from the business meaning of the data. And I think that's what's so key here is, if we're going to ultimately democratize data as we've talked about, we need to change the conversation from a very storage-centric world, like, oh, that table lives in this system or that system, or that system. And make it much more about the data, and the value that it represents. And I think that's what data products aims to do. >> What about data fabric? I have to say, I'm confused by data fabric. I read this, I feel like Gartner just threw it in there to muck it up. And say, no, no, we get to make up the terms, but I've read data mesh versus data fabric, is data fabric just more sort of the physical infrastructure? And data mesh is more of an organizational construct, or how do you see it? >> Yeah, I'm happy to take that or. So, I mean, to me, it's a little bit of potato potato. I think there are some subtle differences. Data fabric is a little bit more about data movement. Whereas, I think data mesh is a little bit more about accessing the data where it lies. But they're both trying to solve the similar problem, which is that we have data in a wide variety of different data sets. And for us to actually analyze it, we need to have a single view. >> Because Gartner hype cycle says data mesh is DOA- >> Justin: I know. >> Which I think is complete BS, I think is real. You talk to customers that are doing it, they're doing it on AWS, they're trying to extend it across clouds, I mean, it's a real trend. I mean, anyway, that's how I see it. >> Yeah. I feel the word data fabric many a times gets misused. Because when you think about the digitization movement that happened, started almost a decade ago, many companies tried to digitize or create digital twins of their systems into the data world, right? So, everything has an underlying data fabric that replicates what's happening transactionally, or otherwise in the real world. What data mesh does is creates structure that works complimentary to the data fabric, that then lends itself to data products, right? So to me, data products becomes a medium, which drives the connection between data mesh and data fabric into the real world for usage and consumption. >> You should write for Gartner. (all laugh) That's the best explanation I've heard. That made sense! >> That really did. That was excellent. So, when we think about any company these days has to be a data company, whether it's your grocery store, a gas station, a car dealer, what can companies do to start productizing their data, so that they can actually unlock new revenue streams, new routes to market? What are some steps and recommendations that you have? Justin, we'll start with you. >> Sure. I would say the first thing is find data that is ultimately valuable to the consumers within your business, and create a product of it. And the way you do that at Starburst is allow you to essentially create a view of your data that can span multiple data sources. So again, we're decoupling where the data lives. That might be a table that lives in a traditional data warehouse, a table that lives in an operational system like Mongo, a table that lives in a data lake. And you can actually join those together, and represent it as a view, and now make it easily consumable. And so, the end user doesn't need to know, did that live in a data warehouse, an operational database, or a data lake? I'm just accessing that. And I think that's a great, easy way to start in your journey. Because I think if you absorb all the elements of data mesh at once, it can feel overwhelming. And I think that's a great way to start. >> Irrespective of physical location. >> Yes. >> Right? >> Precisely. Yep, multicloud, hybrid cloud, you name it. >> And when you think about the broader landscape, right? For the traditionally, companies that only looked at internal data as a way of driving business decisions. More and more, as things evolve into industry, clouds, or ecosystem data, and companies start going beyond their four walls in terms of the data that they manage or the data that they use to make decisions, I think data products are going to play more and more an important part in that construct where you don't govern all the data that our entities within that ecosystem will govern parts of their data, but that data lives together in the form of data products that are governed somewhat centrally. I mean, kind of like a blockchain system, but not really. >> Justin, for our folks here, as we kind of wrap the segment here, what's the bumper sticker for Starburst, and how you're helping organizations to really be able to build data products that add value to their organization? >> I would say analytics anywhere. Our core ethos is, we want to give you the ability to access data wherever it lives, and understand your business holistically. And our query engine allows you to do that from a query perspective, and data products allows you to bring that up a level and make it consumable. >> Make it consumable. Ashwin, last question for you, here we are, day one of re:Invent, loads of people behind us. Tomorrow all the great keynotes kick up. What are you hoping to take away from re:Invent '22? >> Well, I'm hoping to understand how all of these different entities that are represented here connect with each other, right? And to me, Starburst is an important player in terms of how do you drive connectivity. And to me, as we help plans from a Deloitte perspective, drive that business value, connectivity across all of the technology players is extremely important part. So, integration across those technology players is what I'm trying to get from re:Invent here. >> And so, you guys do, you're dot connectors. (Ashwin chuckles) >> Exactly, excellent. Guys, thank you so much for joining David and me on the program tonight. We appreciate your insights, your time, and probably the best explanation of data fabric versus data mesh. (Justin chuckles) And data products that we've maybe ever had on the show! We appreciate your time, thank you. >> Together: Thank you- >> Thanks, guys. >> All right. For our guests and Dave Vellante, I'm Lisa Martin, you're watching theCUBE, the leader in enterprise and emerging tech coverage. (electronic music)

Published Date : Nov 29 2022

SUMMARY :

Dave, it is not only great to be back, I've heard it's the Justin Borgman, the CEO of Starburst, and the value in it for that are able to span really intended to facilitate into the conversation, And data from the enterprise coming to the fore again? And so, the interest in data mesh and some others on the data lies. And all that is, we've And I think ultimately, you want data do I need to support One of the things that Adam Zelinsky and merge data across the enterprise into the line of business. in the products you And that's going to continue And that creates two problems. all of those, but the data products aims to do. And data mesh is more of an about accessing the data where it lies. You talk to customers that are doing it, and data fabric into the real world That's the best explanation I've heard. recommendations that you have? And the way you do that cloud, you name it. in terms of the data that they manage the ability to access Tomorrow all the great keynotes kick up. And to me, as we help plans And so, you guys do, And data products that we've the leader in enterprise

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Dave VellantePERSON

0.99+

JustinPERSON

0.99+

Lisa MartinPERSON

0.99+

Dave VellantePERSON

0.99+

Justin BorgmanPERSON

0.99+

DavePERSON

0.99+

Adam ZelinskyPERSON

0.99+

Las VegasLOCATION

0.99+

StarburstORGANIZATION

0.99+

GartnerORGANIZATION

0.99+

2019DATE

0.99+

John FurrierPERSON

0.99+

DeloitteORGANIZATION

0.99+

AWSORGANIZATION

0.99+

TomorrowDATE

0.99+

Ashwin PatilPERSON

0.99+

twoQUANTITY

0.99+

last yearDATE

0.99+

AshwinPERSON

0.99+

TodayDATE

0.98+

MongoORGANIZATION

0.98+

OneQUANTITY

0.98+

two very distinct groupsQUANTITY

0.98+

two problemsQUANTITY

0.98+

tonightDATE

0.97+

single sourceQUANTITY

0.97+

bothQUANTITY

0.97+

10 years agoDATE

0.96+

todayDATE

0.96+

single pointQUANTITY

0.96+

fourQUANTITY

0.95+

this summerDATE

0.94+

a decade agoDATE

0.93+

one stepQUANTITY

0.92+

one single databaseQUANTITY

0.91+

secondlyQUANTITY

0.89+

StarburstTITLE

0.87+

ZeMacORGANIZATION

0.86+

oneQUANTITY

0.84+

doubleQUANTITY

0.84+

single viewQUANTITY

0.83+

reEVENT

0.83+

first thingQUANTITY

0.81+

re:InventEVENT

0.8+

re:Invent '22EVENT

0.78+

a week or so agoDATE

0.72+

four principlesQUANTITY

0.68+

Invent 2022EVENT

0.67+

InventEVENT

0.57+

CEOPERSON

0.56+

onceQUANTITY

0.53+

Lie 3, Today’s Modern Data Stack Is Modern | Starburst


 

(energetic music) >> Okay, we're back with Justin Borgman, CEO of Starburst, Richard Jarvis is the CTO of EMIS Health, and Teresa Tung is the cloud first technologist from Accenture. We're on to lie number three. And that is the claim that today's "Modern Data Stack" is actually modern. So (chuckles), I guess that's the lie. Or, is that it's not modern. Justin, what do you say? >> Yeah, I think new isn't modern. Right? I think it's the new data stack. It's the cloud data stack, but that doesn't necessarily mean it's modern. I think a lot of the components actually, are exactly the same as what we've had for 40 years. Rather than Teradata, you have Snowflake. Rather than Informatica, you have Fivetran. So, it's the same general stack, just, y'know, a cloud version of it. And I think a lot of the challenges that have plagued us for 40 years still maintain. >> So, let me come back to you Justin. Okay, but there are differences, right? You can scale. You can throw resources at the problem. You can separate compute from storage. You really, there's a lot of money being thrown at that by venture capitalists, and Snowflake you mentioned, its competitors. So that's different. Is it not? Is that not at least an aspect of modern dial it up, dial it down? So what do you say to that? >> Well, it is. It's certainly taking, y'know what the cloud offers and taking advantage of that. But it's important to note that the cloud data warehouses out there are really just separating their compute from their storage. So it's allowing them to scale up and down, but your data's still stored in a proprietary format. You're still locked in. You still have to ingest the data to get it even prepared for analysis. So a lot of the same structural constraints that exist with the old enterprise data warehouse model on-preem still exist. Just yes, a little bit more elastic now because the cloud offers that. >> So Teresa, let me go to you, 'cause you have cloud-first in your title. So, what's say you to this conversation? >> Well, even the cloud providers are looking towards more of a cloud continuum, right? So the centralized cloud as we know it, maybe data lake, data warehouse in the central place, that's not even how the cloud providers are looking at it. They have use query services. Every provider has one that really expands those queries to be beyond a single location. And if we look at a lot of where our- the future goes, right? That's going to very much fall the same thing. There was going to be more edge. There's going to be more on-premise, because of data sovereignty, data gravity, because you're working with different parts of the business that have already made major cloud investments in different cloud providers, right? So, there's a lot of reasons why the modern, I guess, the next modern generation of the data stack needs to be much more federated. >> Okay, so Richard, how do you deal with this? You've obviously got, you know, the technical debt, the existing infrastructure, it's on the books. You don't want to just throw it out. A lot of conversation about modernizing applications, which a lot of times is, you know, of microservices layer on top of legacy apps. How do you think about the Modern Data Stack? >> Well, I think probably the first thing to say is that the stack really has to include the processes and people around the data as well is all well and good changing the technology. But if you don't modernize how people use that technology, then you're not going to be able to, to scale because just 'cause you can scale CPU and storage doesn't mean you can get more people to use your data to generate you more value for the business. And so what we've been looking at is really changing in very much aligned to data products and, and data mesh. How do you enable more people to consume the service and have the stack respond in a way that keeps costs low? Because that's important for our customers consuming this data but also allows people to occasionally run enormous queries and then tick along with smaller ones when required. And it's a good job we did because during COVID all of a sudden we had enormous pressures on our data platform to answer really important life threatening queries. And if we couldn't scale both our data stack and our teams we wouldn't have been able to answer those as quickly as we had. So I think the stack needs to support a scalable business not just the technology itself. >> Well thank you for that. So Justin let's, let's try to break down what the critical aspects are of the modern data stack. So you think about the past, you know, five seven years cloud obviously has given a different pricing model. Derisked experimentation, you know that we talked about the ability to scale up scale down, but it's, I'm taking away that that's not enough. Based on what Richard just said, the modern data stack has to serve the business and enable the business to build data products. I buy that. I'm you a big fan of the data mesh concepts, even though we're early days. So what are the critical aspects if you had to think about you know, the, maybe putting some guardrails and definitions around the modern data stack, what does that look like? What are some of the attributes and, and principles there >> Of how it should look like or, or how >> Yeah. What it should be? >> Yeah. Yeah. Well, I think, you know, in, in Theresa mentioned this in in a previous segment about the data warehouse is not necessarily going to disappear. It just becomes one node, one element of the overall data mesh. And I certainly agree with that. So by no means, are we suggesting that, you know Snowflake or what Redshift or whatever cloud data warehouse you may be using is going to disappear, but it's it's not going to become the end all be all. It's not the, the central single source of truth. And I think that's the paradigm shift that needs to occur. And I think it's also worth noting that those who were the early adopters of the modern data stack were primarily digital, native born in the cloud young companies who had the benefit of of idealism. They had the benefit of starting with a clean slate that does not reflect the vast majority of enterprises. And even those companies, as they grow up, mature out of that ideal state, they go by a business. Now they've got something on another cloud provider that has a different data stack and they have to deal with that heterogeneity that is just change and change is a part of life. And so I think there is an element here that is almost philosophical. It's like, do you believe in an absolute ideal where I can just fit everything into one place or do I believe in reality? And I think the far more pragmatic approach is really what data mesh represents. So to answer your question directly, I think it's adding you know, the ability to access data that lives outside of the data warehouse, maybe living in open data formats in a data lake or accessing operational systems as well. Maybe you want to directly access data that lives in an Oracle database or a Mongo database or, or what have you. So creating that flexibility to really future proof yourself from the inevitable change that you will you won't encounter over time. >> So thank you. So Theresa, based on what Justin just said, I I might take away there is it's inclusive whether it's a data mart, data hub, data lake, data warehouse, just a node on the mesh. Okay. I get that. Does that include Theresa on, on Preem data? Obviously it has to. What are you seeing in terms of the ability to, to take that data mesh concept on Preem I mean most implementations I've seen and data mesh, frankly really aren't, you know adhering to the philosophy there. Maybe, maybe it's data lake and maybe it's using glue. You look at what JPMC is doing, HelloFresh, a lot of stuff happening on the AWS cloud in that, you know, closed stack, if you will. What's the answer to that Theresa? >> I mean, I think it's a killer case for data mesh. The fact that you have valuable data sources on Preem, and then yet you still want to modernize and take the best of cloud. Cloud is still, like we mentioned, there's a lot of great reasons for it around the economics and the way ability to tap into the innovation that the cloud providers are giving around data and AI architecture. It's an easy button. So the mesh allows you to have the best of both world. You can start using the data products on Preem, or in the existing systems that are working already. It's meaningful for the business. At the same time, you can modernize the ones that make business sense because it needs better performance. It needs, you know, something that is, is cheaper or or maybe just tapping into better analytics to get better insights, right? So you're going to be able to stretch and really have the best of both worlds. That, again, going back to Richard's point, that is meaningful by the business. Not everything has to have that one size fits all set a tool. >> Okay. Thank you. So Richard, you know, talking about data as product wonder if we could give us your perspectives here what are the advantages of treating data as a product? What, what role do data products have in the modern data stack? We talk about monetizing data. What are your thoughts on data products? >> So for us, one of the most important data products that we've been creating is taking data that is healthcare data across a wide variety of different settings. So information about patients, demographics about their their treatment, about their medications and so on, and taking that into a standards format that can be utilized by a wide variety of different researchers because misinterpreting that data or having the data not presented in the way that the user is expecting means that you generate the wrong insight and in any business that's clearly not a desirable outcome but when that insight is so critical as it might be in healthcare or some security settings you really have to have gone to the trouble of understanding the data, presenting it in a format that everyone can clearly agree on. And then letting people consume in a very structured managed way, even if that data comes from a variety of different sources in the first place. And so our data product journey has really begun by standardizing data across a number of different silos through the data mesh. So we can present out both internally and through the right governance externally to, to researchers. >> So that data product through whatever APIs is is accessible, it's discoverable, but it's obviously got to be governed as well. You mentioned appropriately provided to internally. >> Yeah. >> But also, you know, external folks as well. So the, so you've, you've architected that capability today? >> We have and because the data is standard it can generate value much more quickly and we can be sure of the security and value that that's providing, because the data product isn't just about formatting the data into the correct tables, it's understanding what it means to redact the data or to remove certain rows from it or to interpret what a date actually means. Is it the start of the contract or the start of the treatment or the date of birth of a patient? These things can be lost in the data storage without having the proper product management around the data to say in a very clear business context what does this data mean, and what does it mean to process this data for a particular use case. >> Yeah, it makes sense. It's got the context. If the, if the domains on the data, you know you got to cut through a lot of the, the centralized teams, the technical teams that that data agnostic, they don't really have that context. All right, let's end. Justin. How does Starburst fit into this modern data stack? Bring us home. >> Yeah. So I think for us it's really providing our customers with, you know the flexibility to operate and analyze data that lives in a wide variety of different systems. Ultimately giving them that optionality, you know and optionality provides the ability to reduce costs store more in a data lake rather than data warehouse. It provides the ability for the fastest time to insight to access the data directly where it lives. And ultimately with this concept of data products that we've now, you know incorporated into our offering as well you can really create and, and curate, you know data as a product to be shared and consumed. So we're trying to help enable the data mesh, you know model and make that an appropriate compliment to you know, the modern data stack that people have today. >> Excellent. Hey, I want to thank Justin, Teresa, and Richard for joining us today. You guys are great. Big believers in the in the data mesh concept, and I think, you know we're seeing the future of data architecture. So thank you. Now, remember, all these conversations are going to be available on the cube.net for on demand viewing. You can also go to starburst.io. They have some great content on the website and they host some really thought provoking interviews and they have awesome resources. Lots of data mesh conversations over there and really good stuff in, in the resource section. So check that out. Thanks for watching the "Data Doesn't Lie... or Does It?" made possible by Starburst data. This is Dave Vellante for the Cube, and we'll see you next time. (upbeat music)

Published Date : Aug 22 2022

SUMMARY :

And that is the claim It's the cloud data stack, So, let me come back to you Justin. that the cloud data warehouses out there So Teresa, let me go to you, So the centralized cloud as we know it, it's on the books. the first thing to say is of the modern data stack. from the inevitable change that you will What's the answer to that Theresa? So the mesh allows you to in the modern data stack? or having the data not presented So that data product But also, you know, around the data to say in a on the data, you know enable the data mesh, you know in the data mesh concept,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RichardPERSON

0.99+

Teresa TungPERSON

0.99+

JustinPERSON

0.99+

TeresaPERSON

0.99+

Dave VellantePERSON

0.99+

Justin BorgmanPERSON

0.99+

Richard JarvisPERSON

0.99+

40 yearsQUANTITY

0.99+

TheresaPERSON

0.99+

StarburstORGANIZATION

0.99+

JPMCORGANIZATION

0.99+

AWSORGANIZATION

0.99+

InformaticaORGANIZATION

0.99+

AccentureORGANIZATION

0.99+

both worldsQUANTITY

0.99+

todayDATE

0.99+

EMIS HealthORGANIZATION

0.99+

first technologistQUANTITY

0.98+

one elementQUANTITY

0.98+

bothQUANTITY

0.98+

first thingQUANTITY

0.98+

five seven yearsQUANTITY

0.98+

oneQUANTITY

0.97+

TeradataORGANIZATION

0.97+

OracleORGANIZATION

0.97+

cube.netOTHER

0.96+

MongoORGANIZATION

0.95+

one sizeQUANTITY

0.93+

CubeORGANIZATION

0.92+

PreemTITLE

0.92+

both worldQUANTITY

0.91+

one placeQUANTITY

0.91+

Today’sTITLE

0.89+

FivetranORGANIZATION

0.86+

Data Doesn't Lie... or Does It?TITLE

0.86+

single locationQUANTITY

0.85+

HelloFreshORGANIZATION

0.84+

first placeQUANTITY

0.83+

CEOPERSON

0.83+

LieTITLE

0.82+

single sourceQUANTITY

0.79+

firstQUANTITY

0.75+

one nodeQUANTITY

0.72+

SnowflakeORGANIZATION

0.66+

SnowflakeTITLE

0.66+

threeQUANTITY

0.59+

CTOPERSON

0.53+

Data StackTITLE

0.53+

RedshiftTITLE

0.52+

starburst.ioOTHER

0.48+

COVIDTITLE

0.37+

Lie 2, An Open Source Based Platform Cannot Give You Performance and Control | Starburst


 

>>We're back with Jess Borgman of Starburst and Richard Jarvis of EVAs health. Okay. We're gonna get into lie. Number two, and that is this an open source based platform cannot give you the performance and control that you can get with a proprietary system. Is that a lie? Justin, the enterprise data warehouse has been pretty dominant and has evolved and matured. Its stack has mature over the years. Why is it not the default platform for data? >>Yeah, well, I think that's become a lie over time. So I, I think, you know, if we go back 10 or 12 years ago with the advent of the first data lake really around Hudu, that probably was true that you couldn't get the performance that you needed to run fast, interactive, SQL queries in a data lake. Now a lot's changed in 10 or 12 years. I remember in the very early days, people would say, you'll, you'll never get performance because you need to be column. You need to store data in a column format. And then, you know, column formats were introduced to, to data lake. You have Parque ORC file in aro that were created to ultimately deliver performance out of that. So, okay. We got, you know, largely over the performance hurdle, you know, more recently people will say, well, you don't have the ability to do updates and deletes like a traditional data warehouse. >>And now we've got the creation of new data formats, again, like iceberg and Delta and hoote that do allow for updates and delete. So I think the data lake has continued to mature. And I remember a quote from, you know, Kurt Monash many years ago where he said, you know, it takes six or seven years to build a functional database. I think that's that's right. And now we've had almost a decade go by. So, you know, these technologies have matured to really deliver very, very close to the same level performance and functionality of, of cloud data warehouses. So I think the, the reality is that's become a lie and now we have large giant hyperscale internet companies that, you know, don't have the traditional data warehouse at all. They do all of their analytics in a data lake. So I think we've, we've proven that it's very much possible today. >>Thank you for that. And so Richard, talk about your perspective as a practitioner in terms of what open brings you versus, I mean, the clothes is it's open as a moving target. I remember Unix used to be open systems and so it's, it is an evolving, you know, spectrum, but, but from your perspective, what does open give you that you can't get from a proprietary system where you are fearful of in a proprietary system? >>I, I suppose for me open buys us the ability to be unsure about the future, because one thing that's always true about technology is it evolves in a, a direction, slightly different to what people expect and what you don't want to end up done is backed itself into a corner that then prevents it from innovating. So if you have chosen the technology and you've stored trillions of records in that technology and suddenly a new way of processing or machine learning comes out, you wanna be able to take advantage your competitive edge might depend upon it. And so I suppose for us, we acknowledge that we don't have perfect vision of what the future might be. And so by backing open storage technologies, we can apply a number of different technologies to the processing of that data. And that gives us the ability to remain relevant, innovate on our data storage. And we have bought our way out of the, any performance concerns because we can use cloud scale infrastructure to scale up and scale down as we need. And so we don't have the concerns that we don't have enough hardware today to process what we want to do, want to achieve. We can just scale up when we need it and scale back down. So open source has really allowed us to maintain the being at the cutting edge. >>So Jess, let me play devil's advocate here a little bit, and I've talked to JAK about this and you know, obviously her vision is there's an open source that, that data mesh is open source, an open source tooling, and it's not a proprietary, you know, you're not gonna buy a data mesh. You're gonna build it with, with open source toolings and, and vendors like you are gonna support it, but come back to sort of today, you can get to market with a proprietary solution faster. I'm gonna make that statement. You tell me if it's a lie and then you can say, okay, we support Apache iceberg. We're gonna support open source tooling, take a company like VMware, not really in the data business, but how, the way they embraced Kubernetes and, and you know, every new open source thing that comes along, they say, we do that too. Why can't proprietary systems do that and be as effective? >>Yeah, well I think at least with the, within the data landscape saying that you can access open data formats like iceberg or, or others is, is a bit dis disingenuous because really what you're selling to your customer is a certain degree of performance, a certain SLA, and you know, those cloud data warehouses that can reach beyond their own proprietary storage drop all the performance that they were able to provide. So it is, it reminds me kind of, of, again, going back 10 or 12 years ago when everybody had a connector to hit and that they thought that was the solution, right? But the reality was, you know, a connector was not the same as running workloads in hit back then. And I think similarly, you know, being able to connect to an external table that lives in an open data format, you know, you're, you're not going to give it the performance that your customers are accustomed to. And at the end of the day, they're always going to be predisposed. They're always going to be incentivized to get that data ingested into the data warehouse, cuz that's where they have control. And you know, the bottom line is the database industry has really been built around vendor lockin. I mean, from the start, how, how many people love Oracle today, but our customers, nonetheless, I think, you know, lockin is, is, is part of this industry. And I think that's really what we're trying to change with open data formats. >>Well, it's interesting remind of when I, you know, I see the, the gas price, the TSR gas price I, I drive up and then I say, oh, that's the cash price credit card. I gotta pay 20 cents more, but okay. But so the, the argument then, so let me, let me come back to you, Justin. So what's wrong with saying, Hey, we support open data formats, but yeah, you're gonna get better performance if you, if you, you keep it into our closed system, are you saying that long term that's gonna come back and bite you cuz you're gonna end up, you mentioned Oracle, you mentioned Teradata. Yeah. That's by, by implication, you're saying that's where snowflake customers are headed. >>Yeah, absolutely. I think this is a movie that, you know, we've all seen before. At least those of us who've been in the industry long enough to, to see this movie play over a couple times. So I do think that's the future. And I think, you know, I loved what Richard said. I actually wrote it down. Cause I thought it was an amazing quote. He said, it buys us the ability to be unsure of the future. That that pretty much says it all the, the future is unknowable and the reality is using open data formats. You remain interoperable with any technology you want to utilize. If you want to use spark to train a machine learning model and you wanna use Starbust to query via sequel, that's totally cool. They can both work off the same exact, you know, data, data sets by contrast, if you're, you know, focused on a proprietary model, then you're kind of locked in again to that model. I think the same applies to data, sharing to data products, to a wide variety of, of aspects of the data landscape that a proprietary approach kind of closes you and, and locks you in. >>So I, I would say this Richard, I'd love to get your thoughts on it. Cause I talked to a lot of Oracle customers, not as many te data customers there, but, but a lot of Oracle customers and they, you know, they'll admit yeah, you know, the Jammin us on price and the license cost, but we do get value out of it. And so my question to you, Richard, is, is do the, let's call it data warehouse systems or the proprietary systems. Are they gonna deliver a greater ROI sooner? And is that in allure of, of that customers, you know, are attracted to, or can open platforms deliver as fast an ROI? >>I think the answer to that is it can depend a bit. It depends on your business's skillset. So we are lucky that we have a number of proprietary teams that work in databases that provide our operational data capability. And we have teams of analytics and big data experts who can work with open data sets and open data formats. And so for those different teams, they can get to an ROI more quickly with different technologies for the business though, we can't do better for our operational data stores than proprietary databases. Today we can back off very tight SLAs to them. We can demonstrate reliability from millions of hours of those databases being run at enterprise scale, but for an analytics workload where increasing our business is growing in that direction, we can't do better than open data formats with cloud-based data mesh type technologies. And so it's not a simple answer. That one will always be the right answer for our business. We definitely have times when proprietary databases provide a capability that we couldn't easily represent or replicate with open technologies. >>Yeah. Richard, stay with you. You mentioned, you know, you know, some things before that, that strike me, you know, the data brick snowflake, you know, thing is always a lot of fun for analysts like me. You've got data bricks coming at it. Richard, you mentioned you have a lot of rockstar, data engineers, data bricks coming at it from a data engineering heritage. You get snowflake coming at it from an analytics heritage. Those two worlds are, are colliding people like PJI Mohan said, you know what? I think it's actually harder to play in the data engineering. So IE, it's easier to for data engineering world to go into the analytics world versus the reverse, but thinking about up and coming engineers and developers preparing for this future of data engineering and data analytics, how, how should they be thinking about the future? What, what's your advice to those young people? >>So I think I'd probably fall back on general programming skill sets. So the advice that I saw years ago was if you have open source technologies, the pythons and Javas on your CV, you command a 20% pay, hike over people who can only do proprietary programming languages. And I think that's true of data technologies as well. And from a business point of view, that makes sense. I'd rather spend the money that I save on proprietary licenses on better engineers, because they can provide more value to the business that can innovate us beyond our competitors. So I think I would my advice to people who are starting here or trying to build teams to capitalize on data assets is begin with open license, free capabilities because they're very cheap to experiment with. And they generate a lot of interest from people who want to join you as a business. And you can make them very successful early, early doors with, with your analytics journey. >>It's interesting. Again, analysts like myself, we do a lot of TCO work and have over the last 20 plus years and in the world of Oracle, you know, normally it's the staff, that's the biggest nut in total cost of ownership, not an Oracle. It's the it's the license cost is by far the biggest component in the, in the blame pie. All right, Justin, help us close out this segment. We've been talking about this sort of data mesh open, closed snowflake data bricks. Where does Starburst sort of as this engine for the data lake data lake house, the data warehouse, it, it fit in this, in this world. >>Yeah. So our view on how the future ultimately unfolds is we think that data lakes will be a natural center of gravity for a lot of the reasons that we described open data formats, lowest total cost of ownership, because you get to choose the cheapest storage available to you. Maybe that's S3 or Azure data lake storage or Google cloud storage, or maybe it's on-prem object storage that you bought at a, at a really good price. So ultimately storing a lot of data in a data lake makes a lot of sense, but I think what makes our perspective unique is we still don't think you're gonna get everything there either. We think that basically centralization of all your data assets is just an impossible endeavor. And so you wanna be able to access data that lives outside of the lake as well. So we kind of think of the lake as maybe the biggest place by volume in terms of how much data you have, but to, to have comprehensive analytics and to truly understand your business and understanding holistically, you need to be able to go access other data sources as well. And so that's the role that we wanna play is to be a single point of access for our customers, provide the right level of fine grained access controls so that the right people have access to the right data and ultimately make it easy to discover and consume via, you know, the creation of data products as well. >>Great. Okay. Thanks guys. Right after this quick break, we're gonna be back to debate whether the cloud data model that we see emerging and the so-called modern data stack is really modern or is it the same wine new bottle when it comes to data architectures, you're watching the cube, the leader in enterprise and emerging tech coverage.

Published Date : Aug 22 2022

SUMMARY :

give you the performance and control that you can get with a proprietary We got, you know, largely over the performance hurdle, you know, more recently people will say, And I remember a quote from, you know, Kurt Monash many years ago where he said, you know, it is an evolving, you know, spectrum, but, but from your perspective, in a, a direction, slightly different to what people expect and what you don't want to end up So Jess, let me play devil's advocate here a little bit, and I've talked to JAK about this and you know, And I think similarly, you know, being able to connect to an external table that lives in an open data format, Well, it's interesting remind of when I, you know, I see the, the gas price, the TSR gas price And I think, you know, I loved what Richard said. you know, the Jammin us on price and the license cost, but we do get value out And so for those different teams, they can get to an you know, the data brick snowflake, you know, thing is always a lot of fun for analysts like me. So the advice that I saw years ago was if you have open source technologies, years and in the world of Oracle, you know, normally it's the staff, to discover and consume via, you know, the creation of data products as well. data model that we see emerging and the so-called modern data stack is

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jess BorgmanPERSON

0.99+

RichardPERSON

0.99+

20 centsQUANTITY

0.99+

sixQUANTITY

0.99+

JustinPERSON

0.99+

Richard JarvisPERSON

0.99+

OracleORGANIZATION

0.99+

Kurt MonashPERSON

0.99+

20%QUANTITY

0.99+

JessPERSON

0.99+

pythonsTITLE

0.99+

seven yearsQUANTITY

0.99+

TodayDATE

0.99+

JavasTITLE

0.99+

TeradataORGANIZATION

0.99+

VMwareORGANIZATION

0.98+

millionsQUANTITY

0.98+

EVAsORGANIZATION

0.98+

JAKPERSON

0.98+

StarburstORGANIZATION

0.98+

bothQUANTITY

0.97+

10DATE

0.97+

12 years agoDATE

0.97+

StarbustTITLE

0.96+

todayDATE

0.95+

Apache icebergORGANIZATION

0.94+

GoogleORGANIZATION

0.93+

12 yearsQUANTITY

0.92+

single pointQUANTITY

0.92+

two worldsQUANTITY

0.92+

10QUANTITY

0.91+

HuduLOCATION

0.91+

UnixTITLE

0.9+

one thingQUANTITY

0.87+

trillions of recordsQUANTITY

0.83+

first data lakeQUANTITY

0.82+

StarburstTITLE

0.8+

PJIORGANIZATION

0.79+

years agoDATE

0.76+

IETITLE

0.75+

Lie 2TITLE

0.72+

many years agoDATE

0.72+

over a couple timesQUANTITY

0.7+

TCOORGANIZATION

0.7+

ParqueORGANIZATION

0.67+

Number twoQUANTITY

0.64+

KubernetesORGANIZATION

0.59+

a decadeQUANTITY

0.58+

plus yearsDATE

0.57+

AzureTITLE

0.57+

S3TITLE

0.55+

DeltaTITLE

0.54+

20QUANTITY

0.49+

lastDATE

0.48+

MohanPERSON

0.44+

ORCORGANIZATION

0.27+

Lie 1, The Most Effective Data Architecture Is Centralized | Starburst


 

(bright upbeat music) >> In 2011, early Facebook employee and Cloudera co-founder Jeff Hammerbacher famously said, "The best minds of my generation are thinking about how to get people to click on ads, and that sucks!" Let's face it. More than a decade later, organizations continue to be frustrated with how difficult it is to get value from data and build a truly agile and data-driven enterprise. What does that even mean, you ask? Well, it means that everyone in the organization has the data they need when they need it in a context that's relevant to advance the mission of an organization. Now, that could mean cutting costs, could mean increasing profits, driving productivity, saving lives, accelerating drug discovery, making better diagnoses, solving supply chain problems, predicting weather disasters, simplifying processes, and thousands of other examples where data can completely transform people's lives beyond manipulating internet users to behave a certain way. We've heard the prognostications about the possibilities of data before and in fairness we've made progress, but the hard truth is the original promises of master data management, enterprise data warehouses, data marts, data hubs, and yes even data lakes were broken and left us wanting for more. Welcome to The Data Doesn't Lie... Or Does It? A series of conversations produced by theCUBE and made possible by Starburst Data. I'm your host, Dave Vellante, and joining me today are three industry experts. Justin Borgman is the co-founder and CEO of Starburst, Richard Jarvis is the CTO at EMIS Health, and Teresa Tung is cloud first technologist at Accenture. Today, we're going to have a candid discussion that will expose the unfulfilled, and yes, broken promises of a data past. We'll expose data lies: big lies, little lies, white lies, and hidden truths. And we'll challenge, age old data conventions and bust some data myths. We're debating questions like is the demise of a single source of truth inevitable? Will the data warehouse ever have feature parity with the data lake or vice versa? Is the so-called modern data stack simply centralization in the cloud, AKA the old guards model in new cloud close? How can organizations rethink their data architectures and regimes to realize the true promises of data? Can and will an open ecosystem deliver on these promises in our lifetimes? We're spanning much of the Western world today. Richard is in the UK, Teresa is on the West Coast, and Justin is in Massachusetts with me. I'm in theCUBE studios, about 30 miles outside of Boston. Folks, welcome to the program. Thanks for coming on. >> Thanks for having us. >> Okay, let's get right into it. You're very welcome. Now, here's the first lie. The most effective data architecture is one that is centralized with a team of data specialists serving various lines of business. What do you think Justin? >> Yeah, definitely a lie. My first startup was a company called Hadapt, which was an early SQL engine for IDU that was acquired by Teradata. And when I got to Teradata, of course, Teradata is the pioneer of that central enterprise data warehouse model. One of the things that I found fascinating was that not one of their customers had actually lived up to that vision of centralizing all of their data into one place. They all had data silos. They all had data in different systems. They had data on prem, data in the cloud. Those companies were acquiring other companies and inheriting their data architecture. So despite being the industry leader for 40 years, not one of their customers truly had everything in one place. So I think definitely history has proven that to be a lie. >> So Richard, from a practitioner's point of view, what are your thoughts? I mean, there's a lot of pressure to cut cost, keep things centralized, serve the business as best as possible from that standpoint. What does your experience show? >> Yeah, I mean, I think I would echo Justin's experience really that we as a business have grown up through acquisition, through storing data in different places sometimes to do information governance in different ways to store data in a platform that's close to data experts people who really understand healthcare data from pharmacies or from doctors. And so, although if you were starting from a greenfield site and you were building something brand new, you might be able to centralize all the data and all of the tooling and teams in one place. The reality is that businesses just don't grow up like that. And it's just really impossible to get that academic perfection of storing everything in one place. >> Teresa, I feel like Sarbanes-Oxley have kind of saved the data warehouse, right? (laughs) You actually did have to have a single version of the truth for certain financial data, but really for some of those other use cases I mentioned, I do feel like the industry has kind of let us down. What's your take on this? Where does it make sense to have that sort of centralized approach versus where does it make sense to maybe decentralize? >> I think you got to have centralized governance, right? So from the central team, for things like Sarbanes-Oxley, for things like security, for certain very core data sets having a centralized set of roles, responsibilities to really QA, right? To serve as a design authority for your entire data estate, just like you might with security, but how it's implemented has to be distributed. Otherwise, you're not going to be able to scale, right? So being able to have different parts of the business really make the right data investments for their needs. And then ultimately, you're going to collaborate with your partners. So partners that are not within the company, right? External partners. We're going to see a lot more data sharing and model creation. And so you're definitely going to be decentralized. >> So Justin, you guys last, jeez, I think it was about a year ago, had a session on data mesh. It was a great program. You invited Zhamak Dehghani. Of course, she's the creator of the data mesh. One of our fundamental premises is that you've got this hyper specialized team that you've got to go through if you want anything. But at the same time, these individuals actually become a bottleneck, even though they're some of the most talented people in the organization. So I guess, a question for you Richard. How do you deal with that? Do you organize so that there are a few sort of rock stars that build cubes and the like or have you had any success in sort of decentralizing with your constituencies that data model? >> Yeah. So we absolutely have got rockstar data scientists and data guardians, if you like. People who understand what it means to use this data, particularly the data that we use at EMIS is very private, it's healthcare information. And some of the rules and regulations around using the data are very complex and strict. So we have to have people who understand the usage of the data, then people who understand how to build models, how to process the data effectively. And you can think of them like consultants to the wider business because a pharmacist might not understand how to structure a SQL query, but they do understand how they want to process medication information to improve patient lives. And so that becomes a consulting type experience from a set of rock stars to help a more decentralized business who needs to understand the data and to generate some valuable output. >> Justin, what do you say to a customer or prospect that says, "Look, Justin. I got a centralized team and that's the most cost effective way to serve the business. Otherwise, I got duplication." What do you say to that? >> Well, I would argue it's probably not the most cost effective, and the reason being really twofold. I think, first of all, when you are deploying a enterprise data warehouse model, the data warehouse itself is very expensive, generally speaking. And so you're putting all of your most valuable data in the hands of one vendor who now has tremendous leverage over you for many, many years to come. I think that's the story at Oracle or Teradata or other proprietary database systems. But the other aspect I think is that the reality is those central data warehouse teams, as much as they are experts in the technology, they don't necessarily understand the data itself. And this is one of the core tenets of data mesh that Zhamak writes about is this idea of the domain owners actually know the data the best. And so by not only acknowledging that data is generally decentralized, and to your earlier point about Sarbanes-Oxley, maybe saving the data warehouse, I would argue maybe GDPR and data sovereignty will destroy it because data has to be decentralized for those laws to be compliant. But I think the reality is the data mesh model basically says data's decentralized and we're going to turn that into an asset rather than a liability. And we're going to turn that into an asset by empowering the people that know the data the best to participate in the process of curating and creating data products for consumption. So I think when you think about it that way, you're going to get higher quality data and faster time to insight, which is ultimately going to drive more revenue for your business and reduce costs. So I think that that's the way I see the two models comparing and contrasting. >> So do you think the demise of the data warehouse is inevitable? Teresa, you work with a lot of clients. They're not just going to rip and replace their existing infrastructure. Maybe they're going to build on top of it, but what does that mean? Does that mean the EDW just becomes less and less valuable over time or it's maybe just isolated to specific use cases? What's your take on that? >> Listen, I still would love all my data within a data warehouse. I would love it mastered, would love it owned by a central team, right? I think that's still what I would love to have. That's just not the reality, right? The investment to actually migrate and keep that up to date, I would say it's a losing battle. Like we've been trying to do it for a long time. Nobody has the budgets and then data changes, right? There's going to be a new technology that's going to emerge that we're going to want to tap into. There's going to be not enough investment to bring all the legacy, but still very useful systems into that centralized view. So you keep the data warehouse. I think it's a very, very valuable, very high performance tool for what it's there for, but you could have this new mesh layer that still takes advantage of the things I mentioned: the data products in the systems that are meaningful today, and the data products that actually might span a number of systems. Maybe either those that either source systems with the domains that know it best, or the consumer-based systems or products that need to be packaged in a way that'd be really meaningful for that end user, right? Each of those are useful for a different part of the business and making sure that the mesh actually allows you to use all of them. >> So, Richard, let me ask you. Take Zhamak's principles back to those. You got the domain ownership and data as product. Okay, great. Sounds good. But it creates what I would argue are two challenges: self-serve infrastructure, let's park that for a second, and then in your industry, one of the most regulated, most sensitive, computational governance. How do you automate and ensure federated governance in that mesh model that Teresa was just talking about? >> Well, it absolutely depends on some of the tooling and processes that you put in place around those tools to centralize the security and the governance of the data. And I think although a data warehouse makes that very simple 'cause it's a single tool, it's not impossible with some of the data mesh technologies that are available. And so what we've done at EMIS is we have a single security layer that sits on top of our data mesh, which means that no matter which user is accessing which data source, we go through a well audited, well understood security layer. That means that we know exactly who's got access to which data field, which data tables. And then everything that they do is audited in a very kind of standard way regardless of the underlying data storage technology. So for me, although storing the data in one place might not be possible, understanding where your source of truth is and securing that in a common way is still a valuable approach, and you can do it without having to bring all that data into a single bucket so that it's all in one place. And so having done that and investing quite heavily in making that possible has paid dividends in terms of giving wider access to the platform, and ensuring that only data that's available under GDPR and other regulations is being used by the data users. >> Yeah. So Justin, we always talk about data democratization, and up until recently, they really haven't been line of sight as to how to get there, but do you have anything to add to this because you're essentially doing analytic queries with data that's all dispersed all over. How are you seeing your customers handle this challenge? >> Yeah, I mean, I think data products is a really interesting aspect of the answer to that. It allows you to, again, leverage the data domain owners, the people who know the data the best, to create data as a product ultimately to be consumed. And we try to represent that in our product as effectively, almost eCommerce like experience where you go and discover and look for the data products that have been created in your organization, and then you can start to consume them as you'd like. And so really trying to build on that notion of data democratization and self-service, and making it very easy to discover and start to use with whatever BI tool you may like or even just running SQL queries yourself. >> Okay guys, grab a sip of water. After the short break, we'll be back to debate whether proprietary or open platforms are the best path to the future of data excellence. Keep it right there. (bright upbeat music)

Published Date : Aug 22 2022

SUMMARY :

has the data they need when they need it Now, here's the first lie. has proven that to be a lie. of pressure to cut cost, and all of the tooling have kind of saved the data So from the central team, for that build cubes and the like and to generate some valuable output. and that's the most cost effective way is that the reality is those of the data warehouse is inevitable? and making sure that the mesh one of the most regulated, most sensitive, and processes that you put as to how to get there, aspect of the answer to that. or open platforms are the best path

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

RichardPERSON

0.99+

Justin BorgmanPERSON

0.99+

JustinPERSON

0.99+

Richard JarvisPERSON

0.99+

Teresa TungPERSON

0.99+

Jeff HammerbacherPERSON

0.99+

TeresaPERSON

0.99+

TeradataORGANIZATION

0.99+

OracleORGANIZATION

0.99+

MassachusettsLOCATION

0.99+

Zhamak DehghaniPERSON

0.99+

UKLOCATION

0.99+

2011DATE

0.99+

two challengesQUANTITY

0.99+

HadaptORGANIZATION

0.99+

40 yearsQUANTITY

0.99+

StarburstORGANIZATION

0.99+

two modelsQUANTITY

0.99+

thousandsQUANTITY

0.99+

BostonLOCATION

0.99+

FacebookORGANIZATION

0.99+

Sarbanes-OxleyORGANIZATION

0.99+

EachQUANTITY

0.99+

first lieQUANTITY

0.99+

AccentureORGANIZATION

0.99+

GDPRTITLE

0.99+

TodayDATE

0.98+

todayDATE

0.98+

SQLTITLE

0.98+

Starburst DataORGANIZATION

0.98+

EMIS HealthORGANIZATION

0.98+

ClouderaORGANIZATION

0.98+

oneQUANTITY

0.98+

first startupQUANTITY

0.98+

one placeQUANTITY

0.98+

about 30 milesQUANTITY

0.98+

OneQUANTITY

0.97+

More than a decade laterDATE

0.97+

EMISORGANIZATION

0.97+

single bucketQUANTITY

0.97+

first technologistQUANTITY

0.96+

three industry expertsQUANTITY

0.96+

single toolQUANTITY

0.96+

single versionQUANTITY

0.94+

ZhamakPERSON

0.92+

theCUBEORGANIZATION

0.91+

single sourceQUANTITY

0.9+

West CoastLOCATION

0.87+

one vendorQUANTITY

0.84+

single security layerQUANTITY

0.81+

about a year agoDATE

0.75+

IDUORGANIZATION

0.68+

IsTITLE

0.65+

a secondQUANTITY

0.64+

EDWORGANIZATION

0.57+

examplesQUANTITY

0.55+

echoCOMMERCIAL_ITEM

0.54+

twofoldQUANTITY

0.5+

LieTITLE

0.35+

Starburst The Data Lies FULL V2b


 

>>In 2011, early Facebook employee and Cloudera co-founder Jeff Ocker famously said the best minds of my generation are thinking about how to get people to click on ads. And that sucks. Let's face it more than a decade later organizations continue to be frustrated with how difficult it is to get value from data and build a truly agile data-driven enterprise. What does that even mean? You ask? Well, it means that everyone in the organization has the data they need when they need it. In a context that's relevant to advance the mission of an organization. Now that could mean cutting cost could mean increasing profits, driving productivity, saving lives, accelerating drug discovery, making better diagnoses, solving, supply chain problems, predicting weather disasters, simplifying processes, and thousands of other examples where data can completely transform people's lives beyond manipulating internet users to behave a certain way. We've heard the prognostications about the possibilities of data before and in fairness we've made progress, but the hard truth is the original promises of master data management, enterprise data, warehouses, data marts, data hubs, and yes, even data lakes were broken and left us wanting from more welcome to the data doesn't lie, or doesn't a series of conversations produced by the cube and made possible by Starburst data. >>I'm your host, Dave Lanta and joining me today are three industry experts. Justin Borgman is this co-founder and CEO of Starburst. Richard Jarvis is the CTO at EMI health and Theresa tongue is cloud first technologist at Accenture. Today we're gonna have a candid discussion that will expose the unfulfilled and yes, broken promises of a data past we'll expose data lies, big lies, little lies, white lies, and hidden truths. And we'll challenge, age old data conventions and bust some data myths. We're debating questions like is the demise of a single source of truth. Inevitable will the data warehouse ever have featured parody with the data lake or vice versa is the so-called modern data stack, simply centralization in the cloud, AKA the old guards model in new cloud close. How can organizations rethink their data architectures and regimes to realize the true promises of data can and will and open ecosystem deliver on these promises in our lifetimes, we're spanning much of the Western world today. Richard is in the UK. Teresa is on the west coast and Justin is in Massachusetts with me. I'm in the cube studios about 30 miles outside of Boston folks. Welcome to the program. Thanks for coming on. Thanks for having us. Let's get right into it. You're very welcome. Now here's the first lie. The most effective data architecture is one that is centralized with a team of data specialists serving various lines of business. What do you think Justin? >>Yeah, definitely a lie. My first startup was a company called hit adapt, which was an early SQL engine for hit that was acquired by Teradata. And when I got to Teradata, of course, Teradata is the pioneer of that central enterprise data warehouse model. One of the things that I found fascinating was that not one of their customers had actually lived up to that vision of centralizing all of their data into one place. They all had data silos. They all had data in different systems. They had data on prem data in the cloud. You know, those companies were acquiring other companies and inheriting their data architecture. So, you know, despite being the industry leader for 40 years, not one of their customers truly had everything in one place. So I think definitely history has proven that to be a lie. >>So Richard, from a practitioner's point of view, you know, what, what are your thoughts? I mean, there, there's a lot of pressure to cut cost, keep things centralized, you know, serve the business as best as possible from that standpoint. What, what is your experience show? >>Yeah, I mean, I think I would echo Justin's experience really that we, as a business have grown up through acquisition, through storing data in different places sometimes to do information governance in different ways to store data in, in a platform that's close to data experts, people who really understand healthcare data from pharmacies or from, from doctors. And so, although if you were starting from a Greenfield site and you were building something brand new, you might be able to centralize all the data and all of the tooling and teams in one place. The reality is that that businesses just don't grow up like that. And, and it's just really impossible to get that academic perfection of, of storing everything in one place. >>Y you know, Theresa, I feel like Sarbanes Oxley kinda saved the data warehouse, you know, right. You actually did have to have a single version of the truth for certain financial data, but really for those, some of those other use cases, I, I mentioned, I, I do feel like the industry has kinda let us down. What's your take on this? Where does it make sense to have that sort of centralized approach versus where does it make sense to maybe decentralized? >>I, I think you gotta have centralized governance, right? So from the central team, for things like star Oxley, for things like security for certainly very core data sets, having a centralized set of roles, responsibilities to really QA, right. To serve as a design authority for your entire data estate, just like you might with security, but how it's implemented has to be distributed. Otherwise you're not gonna be able to scale. Right? So being able to have different parts of the business really make the right data investments for their needs. And then ultimately you're gonna collaborate with your partners. So partners that are not within the company, right. External partners, we're gonna see a lot more data sharing and model creation. And so you're definitely going to be decentralized. >>So, you know, Justin, you guys last, geez, I think it was about a year ago, had a session on, on data mesh. It was a great program. You invited Jamma, Dani, of course, she's the creator of the data mesh. And her one of our fundamental premises is that you've got this hyper specialized team that you've gotta go through. And if you want anything, but at the same time, these, these individuals actually become a bottleneck, even though they're some of the most talented people in the organization. So I guess question for you, Richard, how do you deal with that? Do you, do you organize so that there are a few sort of rock stars that, that, you know, build cubes and, and the like, and, and, and, or have you had any success in sort of decentralizing with, you know, your, your constituencies, that data model? >>Yeah. So, so we absolutely have got rockstar, data scientists and data guardians. If you like people who understand what it means to use this data, particularly as the data that we use at emos is very private it's healthcare information. And some of the, the rules and regulations around using the data are very complex and, and strict. So we have to have people who understand the usage of the data, then people who understand how to build models, how to process the data effectively. And you can think of them like consultants to the wider business, because a pharmacist might not understand how to structure a SQL query, but they do understand how they want to process medication information to improve patient lives. And so that becomes a, a consulting type experience from a, a set of rock stars to help a, a more decentralized business who needs to, to understand the data and to generate some valuable output. >>Justin, what do you say to a, to a customer or prospect that says, look, Justin, I'm gonna, I got a centralized team and that's the most cost effective way to serve the business. Otherwise I got, I got duplication. What do you say to that? >>Well, I, I would argue it's probably not the most cost effective and, and the reason being really twofold. I think, first of all, when you are deploying a enterprise data warehouse model, the, the data warehouse itself is very expensive, generally speaking. And so you're putting all of your most valuable data in the hands of one vendor who now has tremendous leverage over you, you know, for many, many years to come. I think that's the story at Oracle or Terra data or other proprietary database systems. But the other aspect I think is that the reality is those central data warehouse teams is as much as they are experts in the technology. They don't necessarily understand the data itself. And this is one of the core tenants of data mash that that jam writes about is this idea of the domain owners actually know the data the best. >>And so by, you know, not only acknowledging that data is generally decentralized and to your earlier point about SAR, brain Oxley, maybe saving the data warehouse, I would argue maybe GDPR and data sovereignty will destroy it because data has to be decentralized for, for those laws to be compliant. But I think the reality is, you know, the data mesh model basically says, data's decentralized, and we're gonna turn that into an asset rather than a liability. And we're gonna turn that into an asset by empowering the people that know the data, the best to participate in the process of, you know, curating and creating data products for, for consumption. So I think when you think about it, that way, you're going to get higher quality data and faster time to insight, which is ultimately going to drive more revenue for your business and reduce costs. So I think that that's the way I see the two, the two models comparing and contrasting. >>So do you think the demise of the data warehouse is inevitable? I mean, I mean, you know, there Theresa you work with a lot of clients, they're not just gonna rip and replace their existing infrastructure. Maybe they're gonna build on top of it, but what does that mean? Does that mean the E D w just becomes, you know, less and less valuable over time, or it's maybe just isolated to specific use cases. What's your take on that? >>Listen, I still would love all my data within a data warehouse would love it. Mastered would love it owned by essential team. Right? I think that's still what I would love to have. That's just not the reality, right? The investment to actually migrate and keep that up to date. I would say it's a losing battle. Like we've been trying to do it for a long time. Nobody has the budgets and then data changes, right? There's gonna be a new technology. That's gonna emerge that we're gonna wanna tap into. There's going to be not enough investment to bring all the legacy, but still very useful systems into that centralized view. So you keep the data warehouse. I think it's a very, very valuable, very high performance tool for what it's there for, but you could have this, you know, new mesh layer that still takes advantage of the things. I mentioned, the data products in the systems that are meaningful today and the data products that actually might span a number of systems, maybe either those that either source systems for the domains that know it best, or the consumer based systems and products that need to be packaged in a way that be really meaningful for that end user, right? Each of those are useful for a different part of the business and making sure that the mesh actually allows you to use all of them. >>So, Richard, let me ask you, you take, take Gemma's principles back to those. You got to, you know, domain ownership and, and, and data as product. Okay, great. Sounds good. But it creates what I would argue are two, you know, challenges, self-serve infrastructure let's park that for a second. And then in your industry, the one of the high, most regulated, most sensitive computational governance, how do you automate and ensure federated governance in that mesh model that Theresa was just talking about? >>Well, it absolutely depends on some of the tooling and processes that you put in place around those tools to be, to centralize the security and the governance of the data. And I think, although a data warehouse makes that very simple, cause it's a single tool, it's not impossible with some of the data mesh technologies that are available. And so what we've done at emus is we have a single security layer that sits on top of our data match, which means that no matter which user is accessing, which data source, we go through a well audited well understood security layer. That means that we know exactly who's got access to which data field, which data tables. And then everything that they do is, is audited in a very kind of standard way, regardless of the underlying data storage technology. So for me, although storing the data in one place might not be possible understanding where your source of truth is and securing that in a common way is still a valuable approach and you can do it without having to bring all that data into a single bucket so that it's all in one place. And, and so having done that and investing quite heavily in making that possible has paid dividends in terms of giving wider access to the platform and ensuring that only data that's available under GDPR and other regulations is being used by, by the data users. >>Yeah. So Justin, I mean, Democrat, we always talk about data democratization and you know, up until recently, they really haven't been line of sight as to how to get there. But do you have anything to add to this because you're essentially taking, you know, do an analytic queries and with data that's all dispersed all over the, how are you seeing your customers handle this, this challenge? >>Yeah. I mean, I think data products is a really interesting aspect of the answer to that. It allows you to, again, leverage the data domain owners, people know the data, the best to, to create, you know, data as a product ultimately to be consumed. And we try to represent that in our product as effectively a almost eCommerce like experience where you go and discover and look for the data products that have been created in your organization. And then you can start to consume them as, as you'd like. And so really trying to build on that notion of, you know, data democratization and self-service, and making it very easy to discover and, and start to use with whatever BI tool you, you may like, or even just running, you know, SQL queries yourself, >>Okay. G guys grab a sip of water. After this short break, we'll be back to debate whether proprietary or open platforms are the best path to the future of data excellence, keep it right there. >>Your company has more data than ever, and more people trying to understand it, but there's a problem. Your data is stored across multiple systems. It's hard to access and that delays analytics and ultimately decisions. The old method of moving all of your data into a single source of truth is slow and definitely not built for the volume of data we have today or where we are headed while your data engineers spent over half their time, moving data, your analysts and data scientists are left, waiting, feeling frustrated, unproductive, and unable to move the needle for your business. But what if you could spend less time moving or copying data? What if your data consumers could analyze all your data quickly? >>Starburst helps your teams run fast queries on any data source. We help you create a single point of access to your data, no matter where it's stored. And we support high concurrency, we solve for speed and scale, whether it's fast, SQL queries on your data lake or faster queries across multiple data sets, Starburst helps your teams run analytics anywhere you can't afford to wait for data to be available. Your team has questions that need answers. Now with Starburst, the wait is over. You'll have faster access to data with enterprise level security, easy connectivity, and 24 7 support from experts, organizations like Zolando Comcast and FINRA rely on Starburst to move their businesses forward. Contact our Trino experts to get started. >>We're back with Jess Borgman of Starburst and Richard Jarvis of EVAs health. Okay, we're gonna get to lie. Number two, and that is this an open source based platform cannot give you the performance and control that you can get with a proprietary system. Is that a lie? Justin, the enterprise data warehouse has been pretty dominant and has evolved and matured. Its stack has mature over the years. Why is it not the default platform for data? >>Yeah, well, I think that's become a lie over time. So I, I think, you know, if we go back 10 or 12 years ago with the advent of the first data lake really around Hudu, that probably was true that you couldn't get the performance that you needed to run fast, interactive, SQL queries in a data lake. Now a lot's changed in 10 or 12 years. I remember in the very early days, people would say, you you'll never get performance because you need to be column there. You need to store data in a column format. And then, you know, column formats we're introduced to, to data apes, you have Parque ORC file in aro that were created to ultimately deliver performance out of that. So, okay. We got, you know, largely over the performance hurdle, you know, more recently people will say, well, you don't have the ability to do updates and deletes like a traditional data warehouse. >>And now we've got the creation of new data formats, again like iceberg and Delta and Hodi that do allow for updates and delete. So I think the data lake has continued to mature. And I remember a, a quote from, you know, Kurt Monash many years ago where he said, you know, know it takes six or seven years to build a functional database. I think that's that's right. And now we've had almost a decade go by. So, you know, these technologies have matured to really deliver very, very close to the same level performance and functionality of, of cloud data warehouses. So I think the, the reality is that's become a line and now we have large giant hyperscale internet companies that, you know, don't have the traditional data warehouse at all. They do all of their analytics in a data lake. So I think we've, we've proven that it's very much possible today. >>Thank you for that. And so Richard, talk about your perspective as a practitioner in terms of what open brings you versus, I mean, look closed is it's open as a moving target. I remember Unix used to be open systems and so it's, it is an evolving, you know, spectrum, but, but from your perspective, what does open give you that you can't get from a proprietary system where you are fearful of in a proprietary system? >>I, I suppose for me open buys us the ability to be unsure about the future, because one thing that's always true about technology is it evolves in a, a direction, slightly different to what people expect. And what you don't want to end up is done is backed itself into a corner that then prevents it from innovating. So if you have chosen a technology and you've stored trillions of records in that technology and suddenly a new way of processing or machine learning comes out, you wanna be able to take advantage and your competitive edge might depend upon it. And so I suppose for us, we acknowledge that we don't have perfect vision of what the future might be. And so by backing open storage technologies, we can apply a number of different technologies to the processing of that data. And that gives us the ability to remain relevant, innovate on our data storage. And we have bought our way out of the, any performance concerns because we can use cloud scale infrastructure to scale up and scale down as we need. And so we don't have the concerns that we don't have enough hardware today to process what we want to do, want to achieve. We can just scale up when we need it and scale back down. So open source has really allowed us to maintain the being at the cutting edge. >>So Jess, let me play devil's advocate here a little bit, and I've talked to Shaak about this and you know, obviously her vision is there's an open source that, that the data meshes open source, an open source tooling, and it's not a proprietary, you know, you're not gonna buy a data mesh. You're gonna build it with, with open source toolings and, and vendors like you are gonna support it, but to come back to sort of today, you can get to market with a proprietary solution faster. I'm gonna make that statement. You tell me if it's a lie and then you can say, okay, we support Apache iceberg. We're gonna support open source tooling, take a company like VMware, not really in the data business, but how, the way they embraced Kubernetes and, and you know, every new open source thing that comes along, they say, we do that too. Why can't proprietary systems do that and be as effective? >>Yeah, well, I think at least with the, within the data landscape saying that you can access open data formats like iceberg or, or others is, is a bit dis disingenuous because really what you're selling to your customer is a certain degree of performance, a certain SLA, and you know, those cloud data warehouses that can reach beyond their own proprietary storage drop all the performance that they were able to provide. So it is, it reminds me kind of, of, again, going back 10 or 12 years ago when everybody had a connector to Haddo and that they thought that was the solution, right? But the reality was, you know, a connector was not the same as running workloads in Haddo back then. And I think similarly, you know, being able to connect to an external table that lives in an open data format, you know, you're, you're not going to give it the performance that your customers are accustomed to. And at the end of the day, they're always going to be predisposed. They're always going to be incentivized to get that data ingested into the data warehouse, cuz that's where they have control. And you know, the bottom line is the database industry has really been built around vendor lockin. I mean, from the start, how, how many people love Oracle today, but our customers, nonetheless, I think, you know, lockin is, is, is part of this industry. And I think that's really what we're trying to change with open data formats. >>Well, that's interesting reminded when I, you know, I see the, the gas price, the tees or gas price I, I drive up and then I say, oh, that's the cash price credit card. I gotta pay 20 cents more, but okay. But so the, the argument then, so let me, let me come back to you, Justin. So what's wrong with saying, Hey, we support open data formats, but yeah, you're gonna get better performance if you, if you keep it into our closed system, are you saying that long term that's gonna come back and bite you cuz you're gonna end up, you mentioned Oracle, you mentioned Teradata. Yeah. That's by, by implication, you're saying that's where snowflake customers are headed. >>Yeah, absolutely. I think this is a movie that, you know, we've all seen before. At least those of us who've been in the industry long enough to, to see this movie play over a couple times. So I do think that's the future. And I think, you know, I loved what Richard said. I actually wrote it down. Cause I thought it was an amazing quote. He said, it buys us the ability to be unsure of the future. Th that that pretty much says it all the, the future is unknowable and the reality is using open data formats. You remain interoperable with any technology you want to utilize. If you want to use spark to train a machine learning model and you want to use Starbust to query via sequel, that's totally cool. They can both work off the same exact, you know, data, data sets by contrast, if you're, you know, focused on a proprietary model, then you're kind of locked in again to that model. I think the same applies to data, sharing to data products, to a wide variety of, of aspects of the data landscape that a proprietary approach kind of closes you in and locks you in. >>So I, I would say this Richard, I'd love to get your thoughts on it. Cause I talked to a lot of Oracle customers, not as many te data customers, but, but a lot of Oracle customers and they, you know, they'll admit, yeah, you know, they're jamming us on price and the license cost they give, but we do get value out of it. And so my question to you, Richard, is, is do the, let's call it data warehouse systems or the proprietary systems. Are they gonna deliver a greater ROI sooner? And is that in allure of, of that customers, you know, are attracted to, or can open platforms deliver as fast in ROI? >>I think the answer to that is it can depend a bit. It depends on your businesses skillset. So we are lucky that we have a number of proprietary teams that work in databases that provide our operational data capability. And we have teams of analytics and big data experts who can work with open data sets and open data formats. And so for those different teams, they can get to an ROI more quickly with different technologies for the business though, we can't do better for our operational data stores than proprietary databases. Today we can back off very tight SLAs to them. We can demonstrate reliability from millions of hours of those databases being run at enterprise scale, but for an analytics workload where increasing our business is growing in that direction, we can't do better than open data formats with cloud-based data mesh type technologies. And so it's not a simple answer. That one will always be the right answer for our business. We definitely have times when proprietary databases provide a capability that we couldn't easily represent or replicate with open technologies. >>Yeah. Richard, stay with you. You mentioned, you know, you know, some things before that, that strike me, you know, the data brick snowflake, you know, thing is, oh, is a lot of fun for analysts like me. You've got data bricks coming at it. Richard, you mentioned you have a lot of rockstar, data engineers, data bricks coming at it from a data engineering heritage. You get snowflake coming at it from an analytics heritage. Those two worlds are, are colliding people like PJI Mohan said, you know what? I think it's actually harder to play in the data engineering. So I E it's easier to for data engineering world to go into the analytics world versus the reverse, but thinking about up and coming engineers and developers preparing for this future of data engineering and data analytics, how, how should they be thinking about the future? What, what's your advice to those young people? >>So I think I'd probably fall back on general programming skill sets. So the advice that I saw years ago was if you have open source technologies, the pythons and Javas on your CV, you commander 20% pay, hike over people who can only do proprietary programming languages. And I think that's true of data technologies as well. And from a business point of view, that makes sense. I'd rather spend the money that I save on proprietary licenses on better engineers, because they can provide more value to the business that can innovate us beyond our competitors. So I think I would my advice to people who are starting here or trying to build teams to capitalize on data assets is begin with open license, free capabilities, because they're very cheap to experiment with. And they generate a lot of interest from people who want to join you as a business. And you can make them very successful early, early doors with, with your analytics journey. >>It's interesting. Again, analysts like myself, we do a lot of TCO work and have over the last 20 plus years. And in world of Oracle, you know, normally it's the staff, that's the biggest nut in total cost of ownership, not an Oracle. It's the it's the license cost is by far the biggest component in the, in the blame pie. All right, Justin, help us close out this segment. We've been talking about this sort of data mesh open, closed snowflake data bricks. Where does Starburst sort of as this engine for the data lake data lake house, the data warehouse fit in this, in this world? >>Yeah. So our view on how the future ultimately unfolds is we think that data lakes will be a natural center of gravity for a lot of the reasons that we described open data formats, lowest total cost of ownership, because you get to choose the cheapest storage available to you. Maybe that's S3 or Azure data lake storage, or Google cloud storage, or maybe it's on-prem object storage that you bought at a, at a really good price. So ultimately storing a lot of data in a deal lake makes a lot of sense, but I think what makes our perspective unique is we still don't think you're gonna get everything there either. We think that basically centralization of all your data assets is just an impossible endeavor. And so you wanna be able to access data that lives outside of the lake as well. So we kind of think of the lake as maybe the biggest place by volume in terms of how much data you have, but to, to have comprehensive analytics and to truly understand your business and understand it holistically, you need to be able to go access other data sources as well. And so that's the role that we wanna play is to be a single point of access for our customers, provide the right level of fine grained access controls so that the right people have access to the right data and ultimately make it easy to discover and consume via, you know, the creation of data products as well. >>Great. Okay. Thanks guys. Right after this quick break, we're gonna be back to debate whether the cloud data model that we see emerging and the so-called modern data stack is really modern, or is it the same wine new bottle? When it comes to data architectures, you're watching the cube, the leader in enterprise and emerging tech coverage. >>Your data is capable of producing incredible results, but data consumers are often left in the dark without fast access to the data they need. Starers makes your data visible from wherever it lives. Your company is acquiring more data in more places, more rapidly than ever to rely solely on a data centralization strategy. Whether it's in a lake or a warehouse is unrealistic. A single source of truth approach is no longer viable, but disconnected data silos are often left untapped. We need a new approach. One that embraces distributed data. One that enables fast and secure access to any of your data from anywhere with Starburst, you'll have the fastest query engine for the data lake that allows you to connect and analyze your disparate data sources no matter where they live Starburst provides the foundational technology required for you to build towards the vision of a decentralized data mesh Starburst enterprise and Starburst galaxy offer enterprise ready, connectivity, interoperability, and security features for multiple regions, multiple clouds and everchanging global regulatory requirements. The data is yours. And with Starburst, you can perform analytics anywhere in light of your world. >>Okay. We're back with Justin Boardman. CEO of Starbust Richard Jarvis is the CTO of EMI health and Theresa tongue is the cloud first technologist from Accenture. We're on July number three. And that is the claim that today's modern data stack is actually modern. So I guess that's the lie it's it is it's is that it's not modern. Justin, what do you say? >>Yeah. I mean, I think new isn't modern, right? I think it's the, it's the new data stack. It's the cloud data stack, but that doesn't necessarily mean it's modern. I think a lot of the components actually are exactly the same as what we've had for 40 years, rather than Terra data. You have snowflake rather than Informatica you have five trend. So it's the same general stack, just, you know, a cloud version of it. And I think a lot of the challenges that it plagued us for 40 years still maintain. >>So lemme come back to you just, but okay. But, but there are differences, right? I mean, you can scale, you can throw resources at the problem. You can separate compute from storage. You really, you know, there's a lot of money being thrown at that by venture capitalists and snowflake, you mentioned it's competitors. So that's different. Is it not, is that not at least an aspect of, of modern dial it up, dial it down. So what, what do you say to that? >>Well, it, it is, it's certainly taking, you know, what the cloud offers and taking advantage of that, but it's important to note that the cloud data warehouses out there are really just separating their compute from their storage. So it's allowing them to scale up and down, but your data still stored in a proprietary format. You're still locked in. You still have to ingest the data to get it even prepared for analysis. So a lot of the same sort of structural constraints that exist with the old enterprise data warehouse model OnPrem still exist just yes, a little bit more elastic now because the cloud offers that. >>So Theresa, let me go to you cuz you have cloud first in your, in your, your title. So what's what say you to this conversation? >>Well, even the cloud providers are looking towards more of a cloud continuum, right? So the centralized cloud, as we know it, maybe data lake data warehouse in the central place, that's not even how the cloud providers are looking at it. They have news query services. Every provider has one that really expands those queries to be beyond a single location. And if we look at a lot of where our, the future goes, right, that that's gonna very much fall the same thing. There was gonna be more edge. There's gonna be more on premise because of data sovereignty, data gravity, because you're working with different parts of the business that have already made major cloud investments in different cloud providers. Right? So there's a lot of reasons why the modern, I guess, the next modern generation of the data staff needs to be much more federated. >>Okay. So Richard, how do you deal with this? You you've obviously got, you know, the technical debt, the existing infrastructure it's on the books. You don't wanna just throw it out. A lot of, lot of conversation about modernizing applications, which a lot of times is a, you know, a microservices layer on top of leg legacy apps. How do you think about the modern data stack? >>Well, I think probably the first thing to say is that the stack really has to include the processes and people around the data as well is all well and good changing the technology. But if you don't modernize how people use that technology, then you're not going to be able to, to scale because just cuz you can scale CPU and storage doesn't mean you can get more people to use your data, to generate you more, more value for the business. And so what we've been looking at is really changing in very much aligned to data products and, and data mesh. How do you enable more people to consume the service and have the stack respond in a way that keeps costs low? Because that's important for our customers consuming this data, but also allows people to occasionally run enormous queries and then tick along with smaller ones when required. And it's a good job we did because during COVID all of a sudden we had enormous pressures on our data platform to answer really important life threatening queries. And if we couldn't scale both our data stack and our teams, we wouldn't have been able to answer those as quickly as we had. So I think the stack needs to support a scalable business, not just the technology itself. >>Well thank you for that. So Justin let's, let's try to break down what the critical aspects are of the modern data stack. So you think about the past, you know, five, seven years cloud obviously has given a different pricing model. De-risked experimentation, you know that we talked about the ability to scale up scale down, but it's, I'm, I'm taking away that that's not enough based on what Richard just said. The modern data stack has to serve the business and enable the business to build data products. I, I buy that. I'm a big fan of the data mesh concepts, even though we're early days. So what are the critical aspects if you had to think about, you know, paying, maybe putting some guardrails and definitions around the modern data stack, what does that look like? What are some of the attributes and, and principles there >>Of, of how it should look like or, or how >>It's yeah. What it should be. >>Yeah. Yeah. Well, I think, you know, in, in Theresa mentioned this in, in a previous segment about the data warehouse is not necessarily going to disappear. It just becomes one node, one element of the overall data mesh. And I, I certainly agree with that. So by no means, are we suggesting that, you know, snowflake or Redshift or whatever cloud data warehouse you may be using is going to disappear, but it's, it's not going to become the end all be all. It's not the, the central single source of truth. And I think that's the paradigm shift that needs to occur. And I think it's also worth noting that those who were the early adopters of the modern data stack were primarily digital, native born in the cloud young companies who had the benefit of, of idealism. They had the benefit of it was starting with a clean slate that does not reflect the vast majority of enterprises. >>And even those companies, as they grow up mature out of that ideal state, they go buy a business. Now they've got something on another cloud provider that has a different data stack and they have to deal with that heterogeneity that is just change and change is a part of life. And so I think there is an element here that is almost philosophical. It's like, do you believe in an absolute ideal where I can just fit everything into one place or do I believe in reality? And I think the far more pragmatic approach is really what data mesh represents. So to answer your question directly, I think it's adding, you know, the ability to access data that lives outside of the data warehouse, maybe living in open data formats in a data lake or accessing operational systems as well. Maybe you want to directly access data that lives in an Oracle database or a Mongo database or, or what have you. So creating that flexibility to really Futureproof yourself from the inevitable change that you will, you won't encounter over time. >>So thank you. So there, based on what Justin just said, I, my takeaway there is it's inclusive, whether it's a data Mar data hub, data lake data warehouse, it's a, just a node on the mesh. Okay. I get that. Does that include there on Preem data? O obviously it has to, what are you seeing in terms of the ability to, to take that data mesh concept on Preem? I mean, most implementations I've seen in data mesh, frankly really aren't, you know, adhering to the philosophy. They're maybe, maybe it's data lake and maybe it's using glue. You look at what JPMC is doing. Hello, fresh, a lot of stuff happening on the AWS cloud in that, you know, closed stack, if you will. What's the answer to that Theresa? >>I mean, I, I think it's a killer case for data. Me, the fact that you have valuable data sources, OnPrem, and then yet you still wanna modernize and take the best of cloud cloud is still, like we mentioned, there's a lot of great reasons for it around the economics and the way ability to tap into the innovation that the cloud providers are giving around data and AI architecture. It's an easy button. So the mesh allows you to have the best of both worlds. You can start using the data products on-prem or in the existing systems that are working already. It's meaningful for the business. At the same time, you can modernize the ones that make business sense because it needs better performance. It needs, you know, something that is, is cheaper or, or maybe just tap into better analytics to get better insights, right? So you're gonna be able to stretch and really have the best of both worlds. That, again, going back to Richard's point, that is meaningful by the business. Not everything has to have that one size fits all set a tool. >>Okay. Thank you. So Richard, you know, talking about data as product, wonder if we could give us your perspectives here, what are the advantages of treating data as a product? What, what role do data products have in the modern data stack? We talk about monetizing data. What are your thoughts on data products? >>So for us, one of the most important data products that we've been creating is taking data that is healthcare data across a wide variety of different settings. So information about patients' demographics about their, their treatment, about their medications and so on, and taking that into a standards format that can be utilized by a wide variety of different researchers because misinterpreting that data or having the data not presented in the way that the user is expecting means that you generate the wrong insight. And in any business, that's clearly not a desirable outcome, but when that insight is so critical, as it might be in healthcare or some security settings, you really have to have gone to the trouble of understanding the data, presenting it in a format that everyone can clearly agree on. And then letting people consume in a very structured, managed way, even if that data comes from a variety of different sources in, in, in the first place. And so our data product journey has really begun by standardizing data across a number of different silos through the data mesh. So we can present out both internally and through the right governance externally to, to researchers. >>So that data product through whatever APIs is, is accessible, it's discoverable, but it's obviously gotta be governed as well. You mentioned you, you appropriately provided to internally. Yeah. But also, you know, external folks as well. So the, so you've, you've architected that capability today >>We have, and because the data is standard, it can generate value much more quickly and we can be sure of the security and, and, and value that that's providing because the data product isn't just about formatting the data into the correct tables, it's understanding what it means to redact the data or to remove certain rows from it or to interpret what a date actually means. Is it the start of the contract or the start of the treatment or the date of birth of a patient? These things can be lost in the data storage without having the proper product management around the data to say in a very clear business context, what does this data mean? And what does it mean to process this data for a particular use case? >>Yeah, it makes sense. It's got the context. If the, if the domains own the data, you, you gotta cut through a lot of the, the, the centralized teams, the technical teams that, that data agnostic, they don't really have that context. All right. Let's send Justin, how does Starburst fit into this modern data stack? Bring us home. >>Yeah. So I think for us, it's really providing our customers with, you know, the flexibility to operate and analyze data that lives in a wide variety of different systems. Ultimately giving them that optionality, you know, and optionality provides the ability to reduce costs, store more in a data lake rather than data warehouse. It provides the ability for the fastest time to insight to access the data directly where it lives. And ultimately with this concept of data products that we've now, you know, incorporated into our offering as well, you can really create and, and curate, you know, data as a product to be shared and consumed. So we're trying to help enable the data mesh, you know, model and make that an appropriate compliment to, you know, the, the, the modern data stack that people have today. >>Excellent. Hey, I wanna thank Justin Theresa and Richard for joining us today. You guys are great. I big believers in the, in the data mesh concept, and I think, you know, we're seeing the future of data architecture. So thank you. Now, remember, all these conversations are gonna be available on the cube.net for on-demand viewing. You can also go to starburst.io. They have some great content on the website and they host some really thought provoking interviews and, and, and they have awesome resources, lots of data mesh conversations over there, and really good stuff in, in the resource section. So check that out. Thanks for watching the data doesn't lie or does it made possible by Starburst data? This is Dave Valante for the cube, and we'll see you next time. >>The explosion of data sources has forced organizations to modernize their systems and architecture and come to terms with one size does not fit all for data management today. Your teams are constantly moving and copying data, which requires time management. And in some cases, double paying for compute resources. Instead, what if you could access all your data anywhere using the BI tools and SQL skills your users already have. And what if this also included enterprise security and fast performance with Starburst enterprise, you can provide your data consumers with a single point of secure access to all of your data, no matter where it lives with features like strict, fine grained, access control, end to end data encryption and data masking Starburst meets the security standards of the largest companies. Starburst enterprise can easily be deployed anywhere and managed with insights where data teams holistically view their clusters operation and query execution. So they can reach meaningful business decisions faster, all this with the support of the largest team of Trino experts in the world, delivering fully tested stable releases and available to support you 24 7 to unlock the value in all of your data. You need a solution that easily fits with what you have today and can adapt to your architecture. Tomorrow. Starbust enterprise gives you the fastest path from big data to better decisions, cuz your team can't afford to wait. Trino was created to empower analytics anywhere and Starburst enterprise was created to give you the enterprise grade performance, connectivity, security management, and support your company needs organizations like Zolando Comcast and FINRA rely on Starburst to move their businesses forward. Contact us to get started.

Published Date : Aug 22 2022

SUMMARY :

famously said the best minds of my generation are thinking about how to get people to the data warehouse ever have featured parody with the data lake or vice versa is So, you know, despite being the industry leader for 40 years, not one of their customers truly had So Richard, from a practitioner's point of view, you know, what, what are your thoughts? although if you were starting from a Greenfield site and you were building something brand new, Y you know, Theresa, I feel like Sarbanes Oxley kinda saved the data warehouse, I, I think you gotta have centralized governance, right? So, you know, Justin, you guys last, geez, I think it was about a year ago, had a session on, And you can think of them Justin, what do you say to a, to a customer or prospect that says, look, Justin, I'm gonna, you know, for many, many years to come. But I think the reality is, you know, the data mesh model basically says, I mean, you know, there Theresa you work with a lot of clients, they're not just gonna rip and replace their existing that the mesh actually allows you to use all of them. But it creates what I would argue are two, you know, Well, it absolutely depends on some of the tooling and processes that you put in place around those do an analytic queries and with data that's all dispersed all over the, how are you seeing your the best to, to create, you know, data as a product ultimately to be consumed. open platforms are the best path to the future of data But what if you could spend less you create a single point of access to your data, no matter where it's stored. give you the performance and control that you can get with a proprietary system. I remember in the very early days, people would say, you you'll never get performance because And I remember a, a quote from, you know, Kurt Monash many years ago where he said, you know, know it takes six or seven it is an evolving, you know, spectrum, but, but from your perspective, And what you don't want to end up So Jess, let me play devil's advocate here a little bit, and I've talked to Shaak about this and you know, And I think similarly, you know, being able to connect to an external table that lives in an open data format, Well, that's interesting reminded when I, you know, I see the, the gas price, And I think, you know, I loved what Richard said. not as many te data customers, but, but a lot of Oracle customers and they, you know, And so for those different teams, they can get to an ROI more quickly with different technologies that strike me, you know, the data brick snowflake, you know, thing is, oh, is a lot of fun for analysts So the advice that I saw years ago was if you have open source technologies, And in world of Oracle, you know, normally it's the staff, easy to discover and consume via, you know, the creation of data products as well. really modern, or is it the same wine new bottle? And with Starburst, you can perform analytics anywhere in light of your world. And that is the claim that today's So it's the same general stack, just, you know, a cloud version of it. So lemme come back to you just, but okay. So a lot of the same sort of structural constraints that exist with So Theresa, let me go to you cuz you have cloud first in your, in your, the data staff needs to be much more federated. you know, a microservices layer on top of leg legacy apps. So I think the stack needs to support a scalable So you think about the past, you know, five, seven years cloud obviously has given What it should be. And I think that's the paradigm shift that needs to occur. data that lives outside of the data warehouse, maybe living in open data formats in a data lake seen in data mesh, frankly really aren't, you know, adhering to So the mesh allows you to have the best of both worlds. So Richard, you know, talking about data as product, wonder if we could give us your perspectives is expecting means that you generate the wrong insight. But also, you know, around the data to say in a very clear business context, It's got the context. And ultimately with this concept of data products that we've now, you know, incorporated into our offering as well, This is Dave Valante for the cube, and we'll see you next time. You need a solution that easily fits with what you have today and can adapt

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RichardPERSON

0.99+

Dave LantaPERSON

0.99+

Jess BorgmanPERSON

0.99+

JustinPERSON

0.99+

TheresaPERSON

0.99+

Justin BorgmanPERSON

0.99+

TeresaPERSON

0.99+

Jeff OckerPERSON

0.99+

Richard JarvisPERSON

0.99+

Dave ValantePERSON

0.99+

Justin BoardmanPERSON

0.99+

sixQUANTITY

0.99+

DaniPERSON

0.99+

MassachusettsLOCATION

0.99+

20 centsQUANTITY

0.99+

TeradataORGANIZATION

0.99+

OracleORGANIZATION

0.99+

JammaPERSON

0.99+

UKLOCATION

0.99+

FINRAORGANIZATION

0.99+

40 yearsQUANTITY

0.99+

Kurt MonashPERSON

0.99+

20%QUANTITY

0.99+

twoQUANTITY

0.99+

fiveQUANTITY

0.99+

JessPERSON

0.99+

2011DATE

0.99+

StarburstORGANIZATION

0.99+

10QUANTITY

0.99+

AccentureORGANIZATION

0.99+

seven yearsQUANTITY

0.99+

thousandsQUANTITY

0.99+

pythonsTITLE

0.99+

BostonLOCATION

0.99+

GDPRTITLE

0.99+

TodayDATE

0.99+

two modelsQUANTITY

0.99+

Zolando ComcastORGANIZATION

0.99+

GemmaPERSON

0.99+

StarbustORGANIZATION

0.99+

JPMCORGANIZATION

0.99+

FacebookORGANIZATION

0.99+

JavasTITLE

0.99+

todayDATE

0.99+

AWSORGANIZATION

0.99+

millionsQUANTITY

0.99+

first lieQUANTITY

0.99+

10DATE

0.99+

12 yearsQUANTITY

0.99+

one placeQUANTITY

0.99+

TomorrowDATE

0.99+

Starburst The Data Lies FULL V1


 

>>In 2011, early Facebook employee and Cloudera co-founder Jeff Ocker famously said the best minds of my generation are thinking about how to get people to click on ads. And that sucks. Let's face it more than a decade later organizations continue to be frustrated with how difficult it is to get value from data and build a truly agile data-driven enterprise. What does that even mean? You ask? Well, it means that everyone in the organization has the data they need when they need it. In a context that's relevant to advance the mission of an organization. Now that could mean cutting cost could mean increasing profits, driving productivity, saving lives, accelerating drug discovery, making better diagnoses, solving, supply chain problems, predicting weather disasters, simplifying processes, and thousands of other examples where data can completely transform people's lives beyond manipulating internet users to behave a certain way. We've heard the prognostications about the possibilities of data before and in fairness we've made progress, but the hard truth is the original promises of master data management, enterprise data, warehouses, data marts, data hubs, and yes, even data lakes were broken and left us wanting from more welcome to the data doesn't lie, or doesn't a series of conversations produced by the cube and made possible by Starburst data. >>I'm your host, Dave Lanta and joining me today are three industry experts. Justin Borgman is this co-founder and CEO of Starburst. Richard Jarvis is the CTO at EMI health and Theresa tongue is cloud first technologist at Accenture. Today we're gonna have a candid discussion that will expose the unfulfilled and yes, broken promises of a data past we'll expose data lies, big lies, little lies, white lies, and hidden truths. And we'll challenge, age old data conventions and bust some data myths. We're debating questions like is the demise of a single source of truth. Inevitable will the data warehouse ever have featured parody with the data lake or vice versa is the so-called modern data stack, simply centralization in the cloud, AKA the old guards model in new cloud close. How can organizations rethink their data architectures and regimes to realize the true promises of data can and will and open ecosystem deliver on these promises in our lifetimes, we're spanning much of the Western world today. Richard is in the UK. Teresa is on the west coast and Justin is in Massachusetts with me. I'm in the cube studios about 30 miles outside of Boston folks. Welcome to the program. Thanks for coming on. Thanks for having us. Let's get right into it. You're very welcome. Now here's the first lie. The most effective data architecture is one that is centralized with a team of data specialists serving various lines of business. What do you think Justin? >>Yeah, definitely a lie. My first startup was a company called hit adapt, which was an early SQL engine for hit that was acquired by Teradata. And when I got to Teradata, of course, Teradata is the pioneer of that central enterprise data warehouse model. One of the things that I found fascinating was that not one of their customers had actually lived up to that vision of centralizing all of their data into one place. They all had data silos. They all had data in different systems. They had data on prem data in the cloud. You know, those companies were acquiring other companies and inheriting their data architecture. So, you know, despite being the industry leader for 40 years, not one of their customers truly had everything in one place. So I think definitely history has proven that to be a lie. >>So Richard, from a practitioner's point of view, you know, what, what are your thoughts? I mean, there, there's a lot of pressure to cut cost, keep things centralized, you know, serve the business as best as possible from that standpoint. What, what is your experience show? >>Yeah, I mean, I think I would echo Justin's experience really that we, as a business have grown up through acquisition, through storing data in different places sometimes to do information governance in different ways to store data in, in a platform that's close to data experts, people who really understand healthcare data from pharmacies or from, from doctors. And so, although if you were starting from a Greenfield site and you were building something brand new, you might be able to centralize all the data and all of the tooling and teams in one place. The reality is that that businesses just don't grow up like that. And, and it's just really impossible to get that academic perfection of, of storing everything in one place. >>Y you know, Theresa, I feel like Sarbanes Oxley kinda saved the data warehouse, you know, right. You actually did have to have a single version of the truth for certain financial data, but really for those, some of those other use cases, I, I mentioned, I, I do feel like the industry has kinda let us down. What's your take on this? Where does it make sense to have that sort of centralized approach versus where does it make sense to maybe decentralized? >>I, I think you gotta have centralized governance, right? So from the central team, for things like star Oxley, for things like security for certainly very core data sets, having a centralized set of roles, responsibilities to really QA, right. To serve as a design authority for your entire data estate, just like you might with security, but how it's implemented has to be distributed. Otherwise you're not gonna be able to scale. Right? So being able to have different parts of the business really make the right data investments for their needs. And then ultimately you're gonna collaborate with your partners. So partners that are not within the company, right. External partners, we're gonna see a lot more data sharing and model creation. And so you're definitely going to be decentralized. >>So, you know, Justin, you guys last, geez, I think it was about a year ago, had a session on, on data mesh. It was a great program. You invited Jamma, Dani, of course, she's the creator of the data mesh. And her one of our fundamental premises is that you've got this hyper specialized team that you've gotta go through. And if you want anything, but at the same time, these, these individuals actually become a bottleneck, even though they're some of the most talented people in the organization. So I guess question for you, Richard, how do you deal with that? Do you, do you organize so that there are a few sort of rock stars that, that, you know, build cubes and, and the like, and, and, and, or have you had any success in sort of decentralizing with, you know, your, your constituencies, that data model? >>Yeah. So, so we absolutely have got rockstar, data scientists and data guardians. If you like people who understand what it means to use this data, particularly as the data that we use at emos is very private it's healthcare information. And some of the, the rules and regulations around using the data are very complex and, and strict. So we have to have people who understand the usage of the data, then people who understand how to build models, how to process the data effectively. And you can think of them like consultants to the wider business, because a pharmacist might not understand how to structure a SQL query, but they do understand how they want to process medication information to improve patient lives. And so that becomes a, a consulting type experience from a, a set of rock stars to help a, a more decentralized business who needs to, to understand the data and to generate some valuable output. >>Justin, what do you say to a, to a customer or prospect that says, look, Justin, I'm gonna, I got a centralized team and that's the most cost effective way to serve the business. Otherwise I got, I got duplication. What do you say to that? >>Well, I, I would argue it's probably not the most cost effective and, and the reason being really twofold. I think, first of all, when you are deploying a enterprise data warehouse model, the, the data warehouse itself is very expensive, generally speaking. And so you're putting all of your most valuable data in the hands of one vendor who now has tremendous leverage over you, you know, for many, many years to come. I think that's the story at Oracle or Terra data or other proprietary database systems. But the other aspect I think is that the reality is those central data warehouse teams is as much as they are experts in the technology. They don't necessarily understand the data itself. And this is one of the core tenants of data mash that that jam writes about is this idea of the domain owners actually know the data the best. >>And so by, you know, not only acknowledging that data is generally decentralized and to your earlier point about SAR, brain Oxley, maybe saving the data warehouse, I would argue maybe GDPR and data sovereignty will destroy it because data has to be decentralized for, for those laws to be compliant. But I think the reality is, you know, the data mesh model basically says, data's decentralized, and we're gonna turn that into an asset rather than a liability. And we're gonna turn that into an asset by empowering the people that know the data, the best to participate in the process of, you know, curating and creating data products for, for consumption. So I think when you think about it, that way, you're going to get higher quality data and faster time to insight, which is ultimately going to drive more revenue for your business and reduce costs. So I think that that's the way I see the two, the two models comparing and contrasting. >>So do you think the demise of the data warehouse is inevitable? I mean, I mean, you know, there Theresa you work with a lot of clients, they're not just gonna rip and replace their existing infrastructure. Maybe they're gonna build on top of it, but what does that mean? Does that mean the E D w just becomes, you know, less and less valuable over time, or it's maybe just isolated to specific use cases. What's your take on that? >>Listen, I still would love all my data within a data warehouse would love it. Mastered would love it owned by essential team. Right? I think that's still what I would love to have. That's just not the reality, right? The investment to actually migrate and keep that up to date. I would say it's a losing battle. Like we've been trying to do it for a long time. Nobody has the budgets and then data changes, right? There's gonna be a new technology. That's gonna emerge that we're gonna wanna tap into. There's going to be not enough investment to bring all the legacy, but still very useful systems into that centralized view. So you keep the data warehouse. I think it's a very, very valuable, very high performance tool for what it's there for, but you could have this, you know, new mesh layer that still takes advantage of the things. I mentioned, the data products in the systems that are meaningful today and the data products that actually might span a number of systems, maybe either those that either source systems for the domains that know it best, or the consumer based systems and products that need to be packaged in a way that be really meaningful for that end user, right? Each of those are useful for a different part of the business and making sure that the mesh actually allows you to use all of them. >>So, Richard, let me ask you, you take, take Gemma's principles back to those. You got to, you know, domain ownership and, and, and data as product. Okay, great. Sounds good. But it creates what I would argue are two, you know, challenges, self-serve infrastructure let's park that for a second. And then in your industry, the one of the high, most regulated, most sensitive computational governance, how do you automate and ensure federated governance in that mesh model that Theresa was just talking about? >>Well, it absolutely depends on some of the tooling and processes that you put in place around those tools to be, to centralize the security and the governance of the data. And I think, although a data warehouse makes that very simple, cause it's a single tool, it's not impossible with some of the data mesh technologies that are available. And so what we've done at emus is we have a single security layer that sits on top of our data match, which means that no matter which user is accessing, which data source, we go through a well audited well understood security layer. That means that we know exactly who's got access to which data field, which data tables. And then everything that they do is, is audited in a very kind of standard way, regardless of the underlying data storage technology. So for me, although storing the data in one place might not be possible understanding where your source of truth is and securing that in a common way is still a valuable approach and you can do it without having to bring all that data into a single bucket so that it's all in one place. And, and so having done that and investing quite heavily in making that possible has paid dividends in terms of giving wider access to the platform and ensuring that only data that's available under GDPR and other regulations is being used by, by the data users. >>Yeah. So Justin, I mean, Democrat, we always talk about data democratization and you know, up until recently, they really haven't been line of sight as to how to get there. But do you have anything to add to this because you're essentially taking, you know, do an analytic queries and with data that's all dispersed all over the, how are you seeing your customers handle this, this challenge? >>Yeah. I mean, I think data products is a really interesting aspect of the answer to that. It allows you to, again, leverage the data domain owners, people know the data, the best to, to create, you know, data as a product ultimately to be consumed. And we try to represent that in our product as effectively a almost eCommerce like experience where you go and discover and look for the data products that have been created in your organization. And then you can start to consume them as, as you'd like. And so really trying to build on that notion of, you know, data democratization and self-service, and making it very easy to discover and, and start to use with whatever BI tool you, you may like, or even just running, you know, SQL queries yourself, >>Okay. G guys grab a sip of water. After this short break, we'll be back to debate whether proprietary or open platforms are the best path to the future of data excellence, keep it right there. >>Your company has more data than ever, and more people trying to understand it, but there's a problem. Your data is stored across multiple systems. It's hard to access and that delays analytics and ultimately decisions. The old method of moving all of your data into a single source of truth is slow and definitely not built for the volume of data we have today or where we are headed while your data engineers spent over half their time, moving data, your analysts and data scientists are left, waiting, feeling frustrated, unproductive, and unable to move the needle for your business. But what if you could spend less time moving or copying data? What if your data consumers could analyze all your data quickly? >>Starburst helps your teams run fast queries on any data source. We help you create a single point of access to your data, no matter where it's stored. And we support high concurrency, we solve for speed and scale, whether it's fast, SQL queries on your data lake or faster queries across multiple data sets, Starburst helps your teams run analytics anywhere you can't afford to wait for data to be available. Your team has questions that need answers. Now with Starburst, the wait is over. You'll have faster access to data with enterprise level security, easy connectivity, and 24 7 support from experts, organizations like Zolando Comcast and FINRA rely on Starburst to move their businesses forward. Contact our Trino experts to get started. >>We're back with Jess Borgman of Starburst and Richard Jarvis of EVAs health. Okay, we're gonna get to lie. Number two, and that is this an open source based platform cannot give you the performance and control that you can get with a proprietary system. Is that a lie? Justin, the enterprise data warehouse has been pretty dominant and has evolved and matured. Its stack has mature over the years. Why is it not the default platform for data? >>Yeah, well, I think that's become a lie over time. So I, I think, you know, if we go back 10 or 12 years ago with the advent of the first data lake really around Hudu, that probably was true that you couldn't get the performance that you needed to run fast, interactive, SQL queries in a data lake. Now a lot's changed in 10 or 12 years. I remember in the very early days, people would say, you you'll never get performance because you need to be column there. You need to store data in a column format. And then, you know, column formats we're introduced to, to data apes, you have Parque ORC file in aro that were created to ultimately deliver performance out of that. So, okay. We got, you know, largely over the performance hurdle, you know, more recently people will say, well, you don't have the ability to do updates and deletes like a traditional data warehouse. >>And now we've got the creation of new data formats, again like iceberg and Delta and Hodi that do allow for updates and delete. So I think the data lake has continued to mature. And I remember a, a quote from, you know, Kurt Monash many years ago where he said, you know, know it takes six or seven years to build a functional database. I think that's that's right. And now we've had almost a decade go by. So, you know, these technologies have matured to really deliver very, very close to the same level performance and functionality of, of cloud data warehouses. So I think the, the reality is that's become a line and now we have large giant hyperscale internet companies that, you know, don't have the traditional data warehouse at all. They do all of their analytics in a data lake. So I think we've, we've proven that it's very much possible today. >>Thank you for that. And so Richard, talk about your perspective as a practitioner in terms of what open brings you versus, I mean, look closed is it's open as a moving target. I remember Unix used to be open systems and so it's, it is an evolving, you know, spectrum, but, but from your perspective, what does open give you that you can't get from a proprietary system where you are fearful of in a proprietary system? >>I, I suppose for me open buys us the ability to be unsure about the future, because one thing that's always true about technology is it evolves in a, a direction, slightly different to what people expect. And what you don't want to end up is done is backed itself into a corner that then prevents it from innovating. So if you have chosen a technology and you've stored trillions of records in that technology and suddenly a new way of processing or machine learning comes out, you wanna be able to take advantage and your competitive edge might depend upon it. And so I suppose for us, we acknowledge that we don't have perfect vision of what the future might be. And so by backing open storage technologies, we can apply a number of different technologies to the processing of that data. And that gives us the ability to remain relevant, innovate on our data storage. And we have bought our way out of the, any performance concerns because we can use cloud scale infrastructure to scale up and scale down as we need. And so we don't have the concerns that we don't have enough hardware today to process what we want to do, want to achieve. We can just scale up when we need it and scale back down. So open source has really allowed us to maintain the being at the cutting edge. >>So Jess, let me play devil's advocate here a little bit, and I've talked to Shaak about this and you know, obviously her vision is there's an open source that, that the data meshes open source, an open source tooling, and it's not a proprietary, you know, you're not gonna buy a data mesh. You're gonna build it with, with open source toolings and, and vendors like you are gonna support it, but to come back to sort of today, you can get to market with a proprietary solution faster. I'm gonna make that statement. You tell me if it's a lie and then you can say, okay, we support Apache iceberg. We're gonna support open source tooling, take a company like VMware, not really in the data business, but how, the way they embraced Kubernetes and, and you know, every new open source thing that comes along, they say, we do that too. Why can't proprietary systems do that and be as effective? >>Yeah, well, I think at least with the, within the data landscape saying that you can access open data formats like iceberg or, or others is, is a bit dis disingenuous because really what you're selling to your customer is a certain degree of performance, a certain SLA, and you know, those cloud data warehouses that can reach beyond their own proprietary storage drop all the performance that they were able to provide. So it is, it reminds me kind of, of, again, going back 10 or 12 years ago when everybody had a connector to Haddo and that they thought that was the solution, right? But the reality was, you know, a connector was not the same as running workloads in Haddo back then. And I think similarly, you know, being able to connect to an external table that lives in an open data format, you know, you're, you're not going to give it the performance that your customers are accustomed to. And at the end of the day, they're always going to be predisposed. They're always going to be incentivized to get that data ingested into the data warehouse, cuz that's where they have control. And you know, the bottom line is the database industry has really been built around vendor lockin. I mean, from the start, how, how many people love Oracle today, but our customers, nonetheless, I think, you know, lockin is, is, is part of this industry. And I think that's really what we're trying to change with open data formats. >>Well, that's interesting reminded when I, you know, I see the, the gas price, the tees or gas price I, I drive up and then I say, oh, that's the cash price credit card. I gotta pay 20 cents more, but okay. But so the, the argument then, so let me, let me come back to you, Justin. So what's wrong with saying, Hey, we support open data formats, but yeah, you're gonna get better performance if you, if you keep it into our closed system, are you saying that long term that's gonna come back and bite you cuz you're gonna end up, you mentioned Oracle, you mentioned Teradata. Yeah. That's by, by implication, you're saying that's where snowflake customers are headed. >>Yeah, absolutely. I think this is a movie that, you know, we've all seen before. At least those of us who've been in the industry long enough to, to see this movie play over a couple times. So I do think that's the future. And I think, you know, I loved what Richard said. I actually wrote it down. Cause I thought it was an amazing quote. He said, it buys us the ability to be unsure of the future. Th that that pretty much says it all the, the future is unknowable and the reality is using open data formats. You remain interoperable with any technology you want to utilize. If you want to use spark to train a machine learning model and you want to use Starbust to query via sequel, that's totally cool. They can both work off the same exact, you know, data, data sets by contrast, if you're, you know, focused on a proprietary model, then you're kind of locked in again to that model. I think the same applies to data, sharing to data products, to a wide variety of, of aspects of the data landscape that a proprietary approach kind of closes you in and locks you in. >>So I, I would say this Richard, I'd love to get your thoughts on it. Cause I talked to a lot of Oracle customers, not as many te data customers, but, but a lot of Oracle customers and they, you know, they'll admit, yeah, you know, they're jamming us on price and the license cost they give, but we do get value out of it. And so my question to you, Richard, is, is do the, let's call it data warehouse systems or the proprietary systems. Are they gonna deliver a greater ROI sooner? And is that in allure of, of that customers, you know, are attracted to, or can open platforms deliver as fast in ROI? >>I think the answer to that is it can depend a bit. It depends on your businesses skillset. So we are lucky that we have a number of proprietary teams that work in databases that provide our operational data capability. And we have teams of analytics and big data experts who can work with open data sets and open data formats. And so for those different teams, they can get to an ROI more quickly with different technologies for the business though, we can't do better for our operational data stores than proprietary databases. Today we can back off very tight SLAs to them. We can demonstrate reliability from millions of hours of those databases being run at enterprise scale, but for an analytics workload where increasing our business is growing in that direction, we can't do better than open data formats with cloud-based data mesh type technologies. And so it's not a simple answer. That one will always be the right answer for our business. We definitely have times when proprietary databases provide a capability that we couldn't easily represent or replicate with open technologies. >>Yeah. Richard, stay with you. You mentioned, you know, you know, some things before that, that strike me, you know, the data brick snowflake, you know, thing is, oh, is a lot of fun for analysts like me. You've got data bricks coming at it. Richard, you mentioned you have a lot of rockstar, data engineers, data bricks coming at it from a data engineering heritage. You get snowflake coming at it from an analytics heritage. Those two worlds are, are colliding people like PJI Mohan said, you know what? I think it's actually harder to play in the data engineering. So I E it's easier to for data engineering world to go into the analytics world versus the reverse, but thinking about up and coming engineers and developers preparing for this future of data engineering and data analytics, how, how should they be thinking about the future? What, what's your advice to those young people? >>So I think I'd probably fall back on general programming skill sets. So the advice that I saw years ago was if you have open source technologies, the pythons and Javas on your CV, you commander 20% pay, hike over people who can only do proprietary programming languages. And I think that's true of data technologies as well. And from a business point of view, that makes sense. I'd rather spend the money that I save on proprietary licenses on better engineers, because they can provide more value to the business that can innovate us beyond our competitors. So I think I would my advice to people who are starting here or trying to build teams to capitalize on data assets is begin with open license, free capabilities, because they're very cheap to experiment with. And they generate a lot of interest from people who want to join you as a business. And you can make them very successful early, early doors with, with your analytics journey. >>It's interesting. Again, analysts like myself, we do a lot of TCO work and have over the last 20 plus years. And in world of Oracle, you know, normally it's the staff, that's the biggest nut in total cost of ownership, not an Oracle. It's the it's the license cost is by far the biggest component in the, in the blame pie. All right, Justin, help us close out this segment. We've been talking about this sort of data mesh open, closed snowflake data bricks. Where does Starburst sort of as this engine for the data lake data lake house, the data warehouse fit in this, in this world? >>Yeah. So our view on how the future ultimately unfolds is we think that data lakes will be a natural center of gravity for a lot of the reasons that we described open data formats, lowest total cost of ownership, because you get to choose the cheapest storage available to you. Maybe that's S3 or Azure data lake storage, or Google cloud storage, or maybe it's on-prem object storage that you bought at a, at a really good price. So ultimately storing a lot of data in a deal lake makes a lot of sense, but I think what makes our perspective unique is we still don't think you're gonna get everything there either. We think that basically centralization of all your data assets is just an impossible endeavor. And so you wanna be able to access data that lives outside of the lake as well. So we kind of think of the lake as maybe the biggest place by volume in terms of how much data you have, but to, to have comprehensive analytics and to truly understand your business and understand it holistically, you need to be able to go access other data sources as well. And so that's the role that we wanna play is to be a single point of access for our customers, provide the right level of fine grained access controls so that the right people have access to the right data and ultimately make it easy to discover and consume via, you know, the creation of data products as well. >>Great. Okay. Thanks guys. Right after this quick break, we're gonna be back to debate whether the cloud data model that we see emerging and the so-called modern data stack is really modern, or is it the same wine new bottle? When it comes to data architectures, you're watching the cube, the leader in enterprise and emerging tech coverage. >>Your data is capable of producing incredible results, but data consumers are often left in the dark without fast access to the data they need. Starers makes your data visible from wherever it lives. Your company is acquiring more data in more places, more rapidly than ever to rely solely on a data centralization strategy. Whether it's in a lake or a warehouse is unrealistic. A single source of truth approach is no longer viable, but disconnected data silos are often left untapped. We need a new approach. One that embraces distributed data. One that enables fast and secure access to any of your data from anywhere with Starburst, you'll have the fastest query engine for the data lake that allows you to connect and analyze your disparate data sources no matter where they live Starburst provides the foundational technology required for you to build towards the vision of a decentralized data mesh Starburst enterprise and Starburst galaxy offer enterprise ready, connectivity, interoperability, and security features for multiple regions, multiple clouds and everchanging global regulatory requirements. The data is yours. And with Starburst, you can perform analytics anywhere in light of your world. >>Okay. We're back with Justin Boardman. CEO of Starbust Richard Jarvis is the CTO of EMI health and Theresa tongue is the cloud first technologist from Accenture. We're on July number three. And that is the claim that today's modern data stack is actually modern. So I guess that's the lie it's it is it's is that it's not modern. Justin, what do you say? >>Yeah. I mean, I think new isn't modern, right? I think it's the, it's the new data stack. It's the cloud data stack, but that doesn't necessarily mean it's modern. I think a lot of the components actually are exactly the same as what we've had for 40 years, rather than Terra data. You have snowflake rather than Informatica you have five trend. So it's the same general stack, just, you know, a cloud version of it. And I think a lot of the challenges that it plagued us for 40 years still maintain. >>So lemme come back to you just, but okay. But, but there are differences, right? I mean, you can scale, you can throw resources at the problem. You can separate compute from storage. You really, you know, there's a lot of money being thrown at that by venture capitalists and snowflake, you mentioned it's competitors. So that's different. Is it not, is that not at least an aspect of, of modern dial it up, dial it down. So what, what do you say to that? >>Well, it, it is, it's certainly taking, you know, what the cloud offers and taking advantage of that, but it's important to note that the cloud data warehouses out there are really just separating their compute from their storage. So it's allowing them to scale up and down, but your data still stored in a proprietary format. You're still locked in. You still have to ingest the data to get it even prepared for analysis. So a lot of the same sort of structural constraints that exist with the old enterprise data warehouse model OnPrem still exist just yes, a little bit more elastic now because the cloud offers that. >>So Theresa, let me go to you cuz you have cloud first in your, in your, your title. So what's what say you to this conversation? >>Well, even the cloud providers are looking towards more of a cloud continuum, right? So the centralized cloud, as we know it, maybe data lake data warehouse in the central place, that's not even how the cloud providers are looking at it. They have news query services. Every provider has one that really expands those queries to be beyond a single location. And if we look at a lot of where our, the future goes, right, that that's gonna very much fall the same thing. There was gonna be more edge. There's gonna be more on premise because of data sovereignty, data gravity, because you're working with different parts of the business that have already made major cloud investments in different cloud providers. Right? So there's a lot of reasons why the modern, I guess, the next modern generation of the data staff needs to be much more federated. >>Okay. So Richard, how do you deal with this? You you've obviously got, you know, the technical debt, the existing infrastructure it's on the books. You don't wanna just throw it out. A lot of, lot of conversation about modernizing applications, which a lot of times is a, you know, a microservices layer on top of leg legacy apps. How do you think about the modern data stack? >>Well, I think probably the first thing to say is that the stack really has to include the processes and people around the data as well is all well and good changing the technology. But if you don't modernize how people use that technology, then you're not going to be able to, to scale because just cuz you can scale CPU and storage doesn't mean you can get more people to use your data, to generate you more, more value for the business. And so what we've been looking at is really changing in very much aligned to data products and, and data mesh. How do you enable more people to consume the service and have the stack respond in a way that keeps costs low? Because that's important for our customers consuming this data, but also allows people to occasionally run enormous queries and then tick along with smaller ones when required. And it's a good job we did because during COVID all of a sudden we had enormous pressures on our data platform to answer really important life threatening queries. And if we couldn't scale both our data stack and our teams, we wouldn't have been able to answer those as quickly as we had. So I think the stack needs to support a scalable business, not just the technology itself. >>Well thank you for that. So Justin let's, let's try to break down what the critical aspects are of the modern data stack. So you think about the past, you know, five, seven years cloud obviously has given a different pricing model. De-risked experimentation, you know that we talked about the ability to scale up scale down, but it's, I'm, I'm taking away that that's not enough based on what Richard just said. The modern data stack has to serve the business and enable the business to build data products. I, I buy that. I'm a big fan of the data mesh concepts, even though we're early days. So what are the critical aspects if you had to think about, you know, paying, maybe putting some guardrails and definitions around the modern data stack, what does that look like? What are some of the attributes and, and principles there >>Of, of how it should look like or, or how >>It's yeah. What it should be. >>Yeah. Yeah. Well, I think, you know, in, in Theresa mentioned this in, in a previous segment about the data warehouse is not necessarily going to disappear. It just becomes one node, one element of the overall data mesh. And I, I certainly agree with that. So by no means, are we suggesting that, you know, snowflake or Redshift or whatever cloud data warehouse you may be using is going to disappear, but it's, it's not going to become the end all be all. It's not the, the central single source of truth. And I think that's the paradigm shift that needs to occur. And I think it's also worth noting that those who were the early adopters of the modern data stack were primarily digital, native born in the cloud young companies who had the benefit of, of idealism. They had the benefit of it was starting with a clean slate that does not reflect the vast majority of enterprises. >>And even those companies, as they grow up mature out of that ideal state, they go buy a business. Now they've got something on another cloud provider that has a different data stack and they have to deal with that heterogeneity that is just change and change is a part of life. And so I think there is an element here that is almost philosophical. It's like, do you believe in an absolute ideal where I can just fit everything into one place or do I believe in reality? And I think the far more pragmatic approach is really what data mesh represents. So to answer your question directly, I think it's adding, you know, the ability to access data that lives outside of the data warehouse, maybe living in open data formats in a data lake or accessing operational systems as well. Maybe you want to directly access data that lives in an Oracle database or a Mongo database or, or what have you. So creating that flexibility to really Futureproof yourself from the inevitable change that you will, you won't encounter over time. >>So thank you. So there, based on what Justin just said, I, my takeaway there is it's inclusive, whether it's a data Mar data hub, data lake data warehouse, it's a, just a node on the mesh. Okay. I get that. Does that include there on Preem data? O obviously it has to, what are you seeing in terms of the ability to, to take that data mesh concept on Preem? I mean, most implementations I've seen in data mesh, frankly really aren't, you know, adhering to the philosophy. They're maybe, maybe it's data lake and maybe it's using glue. You look at what JPMC is doing. Hello, fresh, a lot of stuff happening on the AWS cloud in that, you know, closed stack, if you will. What's the answer to that Theresa? >>I mean, I, I think it's a killer case for data. Me, the fact that you have valuable data sources, OnPrem, and then yet you still wanna modernize and take the best of cloud cloud is still, like we mentioned, there's a lot of great reasons for it around the economics and the way ability to tap into the innovation that the cloud providers are giving around data and AI architecture. It's an easy button. So the mesh allows you to have the best of both worlds. You can start using the data products on-prem or in the existing systems that are working already. It's meaningful for the business. At the same time, you can modernize the ones that make business sense because it needs better performance. It needs, you know, something that is, is cheaper or, or maybe just tap into better analytics to get better insights, right? So you're gonna be able to stretch and really have the best of both worlds. That, again, going back to Richard's point, that is meaningful by the business. Not everything has to have that one size fits all set a tool. >>Okay. Thank you. So Richard, you know, talking about data as product, wonder if we could give us your perspectives here, what are the advantages of treating data as a product? What, what role do data products have in the modern data stack? We talk about monetizing data. What are your thoughts on data products? >>So for us, one of the most important data products that we've been creating is taking data that is healthcare data across a wide variety of different settings. So information about patients' demographics about their, their treatment, about their medications and so on, and taking that into a standards format that can be utilized by a wide variety of different researchers because misinterpreting that data or having the data not presented in the way that the user is expecting means that you generate the wrong insight. And in any business, that's clearly not a desirable outcome, but when that insight is so critical, as it might be in healthcare or some security settings, you really have to have gone to the trouble of understanding the data, presenting it in a format that everyone can clearly agree on. And then letting people consume in a very structured, managed way, even if that data comes from a variety of different sources in, in, in the first place. And so our data product journey has really begun by standardizing data across a number of different silos through the data mesh. So we can present out both internally and through the right governance externally to, to researchers. >>So that data product through whatever APIs is, is accessible, it's discoverable, but it's obviously gotta be governed as well. You mentioned you, you appropriately provided to internally. Yeah. But also, you know, external folks as well. So the, so you've, you've architected that capability today >>We have, and because the data is standard, it can generate value much more quickly and we can be sure of the security and, and, and value that that's providing because the data product isn't just about formatting the data into the correct tables, it's understanding what it means to redact the data or to remove certain rows from it or to interpret what a date actually means. Is it the start of the contract or the start of the treatment or the date of birth of a patient? These things can be lost in the data storage without having the proper product management around the data to say in a very clear business context, what does this data mean? And what does it mean to process this data for a particular use case? >>Yeah, it makes sense. It's got the context. If the, if the domains own the data, you, you gotta cut through a lot of the, the, the centralized teams, the technical teams that, that data agnostic, they don't really have that context. All right. Let's send Justin, how does Starburst fit into this modern data stack? Bring us home. >>Yeah. So I think for us, it's really providing our customers with, you know, the flexibility to operate and analyze data that lives in a wide variety of different systems. Ultimately giving them that optionality, you know, and optionality provides the ability to reduce costs, store more in a data lake rather than data warehouse. It provides the ability for the fastest time to insight to access the data directly where it lives. And ultimately with this concept of data products that we've now, you know, incorporated into our offering as well, you can really create and, and curate, you know, data as a product to be shared and consumed. So we're trying to help enable the data mesh, you know, model and make that an appropriate compliment to, you know, the, the, the modern data stack that people have today. >>Excellent. Hey, I wanna thank Justin Theresa and Richard for joining us today. You guys are great. I big believers in the, in the data mesh concept, and I think, you know, we're seeing the future of data architecture. So thank you. Now, remember, all these conversations are gonna be available on the cube.net for on-demand viewing. You can also go to starburst.io. They have some great content on the website and they host some really thought provoking interviews and, and, and they have awesome resources, lots of data mesh conversations over there, and really good stuff in, in the resource section. So check that out. Thanks for watching the data doesn't lie or does it made possible by Starburst data? This is Dave Valante for the cube, and we'll see you next time. >>The explosion of data sources has forced organizations to modernize their systems and architecture and come to terms with one size does not fit all for data management today. Your teams are constantly moving and copying data, which requires time management. And in some cases, double paying for compute resources. Instead, what if you could access all your data anywhere using the BI tools and SQL skills your users already have. And what if this also included enterprise security and fast performance with Starburst enterprise, you can provide your data consumers with a single point of secure access to all of your data, no matter where it lives with features like strict, fine grained, access control, end to end data encryption and data masking Starburst meets the security standards of the largest companies. Starburst enterprise can easily be deployed anywhere and managed with insights where data teams holistically view their clusters operation and query execution. So they can reach meaningful business decisions faster, all this with the support of the largest team of Trino experts in the world, delivering fully tested stable releases and available to support you 24 7 to unlock the value in all of your data. You need a solution that easily fits with what you have today and can adapt to your architecture. Tomorrow. Starbust enterprise gives you the fastest path from big data to better decisions, cuz your team can't afford to wait. Trino was created to empower analytics anywhere and Starburst enterprise was created to give you the enterprise grade performance, connectivity, security management, and support your company needs organizations like Zolando Comcast and FINRA rely on Starburst to move their businesses forward. Contact us to get started.

Published Date : Aug 20 2022

SUMMARY :

famously said the best minds of my generation are thinking about how to get people to the data warehouse ever have featured parody with the data lake or vice versa is So, you know, despite being the industry leader for 40 years, not one of their customers truly had So Richard, from a practitioner's point of view, you know, what, what are your thoughts? although if you were starting from a Greenfield site and you were building something brand new, Y you know, Theresa, I feel like Sarbanes Oxley kinda saved the data warehouse, I, I think you gotta have centralized governance, right? So, you know, Justin, you guys last, geez, I think it was about a year ago, had a session on, And you can think of them Justin, what do you say to a, to a customer or prospect that says, look, Justin, I'm gonna, you know, for many, many years to come. But I think the reality is, you know, the data mesh model basically says, I mean, you know, there Theresa you work with a lot of clients, they're not just gonna rip and replace their existing that the mesh actually allows you to use all of them. But it creates what I would argue are two, you know, Well, it absolutely depends on some of the tooling and processes that you put in place around those do an analytic queries and with data that's all dispersed all over the, how are you seeing your the best to, to create, you know, data as a product ultimately to be consumed. open platforms are the best path to the future of data But what if you could spend less you create a single point of access to your data, no matter where it's stored. give you the performance and control that you can get with a proprietary system. I remember in the very early days, people would say, you you'll never get performance because And I remember a, a quote from, you know, Kurt Monash many years ago where he said, you know, know it takes six or seven it is an evolving, you know, spectrum, but, but from your perspective, And what you don't want to end up So Jess, let me play devil's advocate here a little bit, and I've talked to Shaak about this and you know, And I think similarly, you know, being able to connect to an external table that lives in an open data format, Well, that's interesting reminded when I, you know, I see the, the gas price, And I think, you know, I loved what Richard said. not as many te data customers, but, but a lot of Oracle customers and they, you know, And so for those different teams, they can get to an ROI more quickly with different technologies that strike me, you know, the data brick snowflake, you know, thing is, oh, is a lot of fun for analysts So the advice that I saw years ago was if you have open source technologies, And in world of Oracle, you know, normally it's the staff, easy to discover and consume via, you know, the creation of data products as well. really modern, or is it the same wine new bottle? And with Starburst, you can perform analytics anywhere in light of your world. And that is the claim that today's So it's the same general stack, just, you know, a cloud version of it. So lemme come back to you just, but okay. So a lot of the same sort of structural constraints that exist with So Theresa, let me go to you cuz you have cloud first in your, in your, the data staff needs to be much more federated. you know, a microservices layer on top of leg legacy apps. So I think the stack needs to support a scalable So you think about the past, you know, five, seven years cloud obviously has given What it should be. And I think that's the paradigm shift that needs to occur. data that lives outside of the data warehouse, maybe living in open data formats in a data lake seen in data mesh, frankly really aren't, you know, adhering to So the mesh allows you to have the best of both worlds. So Richard, you know, talking about data as product, wonder if we could give us your perspectives is expecting means that you generate the wrong insight. But also, you know, around the data to say in a very clear business context, It's got the context. And ultimately with this concept of data products that we've now, you know, incorporated into our offering as well, This is Dave Valante for the cube, and we'll see you next time. You need a solution that easily fits with what you have today and can adapt

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RichardPERSON

0.99+

Dave LantaPERSON

0.99+

Jess BorgmanPERSON

0.99+

JustinPERSON

0.99+

TheresaPERSON

0.99+

Justin BorgmanPERSON

0.99+

TeresaPERSON

0.99+

Jeff OckerPERSON

0.99+

Richard JarvisPERSON

0.99+

Dave ValantePERSON

0.99+

Justin BoardmanPERSON

0.99+

sixQUANTITY

0.99+

DaniPERSON

0.99+

MassachusettsLOCATION

0.99+

20 centsQUANTITY

0.99+

TeradataORGANIZATION

0.99+

OracleORGANIZATION

0.99+

JammaPERSON

0.99+

UKLOCATION

0.99+

FINRAORGANIZATION

0.99+

40 yearsQUANTITY

0.99+

Kurt MonashPERSON

0.99+

20%QUANTITY

0.99+

twoQUANTITY

0.99+

fiveQUANTITY

0.99+

JessPERSON

0.99+

2011DATE

0.99+

StarburstORGANIZATION

0.99+

10QUANTITY

0.99+

AccentureORGANIZATION

0.99+

seven yearsQUANTITY

0.99+

thousandsQUANTITY

0.99+

pythonsTITLE

0.99+

BostonLOCATION

0.99+

GDPRTITLE

0.99+

TodayDATE

0.99+

two modelsQUANTITY

0.99+

Zolando ComcastORGANIZATION

0.99+

GemmaPERSON

0.99+

StarbustORGANIZATION

0.99+

JPMCORGANIZATION

0.99+

FacebookORGANIZATION

0.99+

JavasTITLE

0.99+

todayDATE

0.99+

AWSORGANIZATION

0.99+

millionsQUANTITY

0.99+

first lieQUANTITY

0.99+

10DATE

0.99+

12 yearsQUANTITY

0.99+

one placeQUANTITY

0.99+

TomorrowDATE

0.99+

Starburst panel Q3


 

>>Okay. We're back with Justin Boorman CEO of Starburst. Richard Jarvis is the CTO of EMI health and Teresa tongue is the cloud first technologist from Accenture. We're on July number three. And that is the claim that today's modern data stack is actually modern. So I guess that's the lie or it's it is it's is that it's not modern, Justin, what do you say? >>Yeah, I mean, I think new isn't modern, right? I think it's, the's the new data stack. It's the cloud data stack, but that doesn't necessarily mean it's modern. I think a lot of the components actually are exactly the same as what we've had for 40 years, rather than Terra data. You have snowflake rather than Informatica you have five trend. So it's the same general stack, just, you know, a cloud version of it. And I think a lot of the challenges that it plagued us for 40 years still maintain. >>So lemme come back to you just this, but okay. But, but there are differences, right? I mean, you can scale, you can throw resources at the problem. You can separate compute from storage. You really, you know, there's a lot of money being thrown at that by venture capitalists and snowflake, you mentioned it's competitors. So that's different. Is it not, is that not at least an aspect of, of modern dial it up, dial it down. So what, what do you say to that? >>Well, it, it is, it's certainly taking, you know, what the cloud offers and taking advantage of that, but it's important to note that the cloud data warehouses out there are really just separating their compute from their storage. So it's allowing them to scale up and down, but your data's still stored in a proprietary format. You're still locked in. You still have to ingest the data to get it even prepared for analysis. So a lot of the same sort of structural constraints that exist with the old enterprise data warehouse model OnPrem still exists just, yes, a little bit more elastic now because the cloud offers that. >>So Theresa, let me go to you cuz you have cloud first in your, in your, your title. So what's what say you to this conversation? >>Well, even the cloud providers are looking towards more of a cloud continuum, right? So the centralized cloud, as we know it, maybe data lake data warehouse in the central place, that's not even how the cloud providers are looking at it. They have news query services. Every provider has one that really expands those queries to be beyond a single location. And if we look at a lot of where our, the future goes, right, that that's gonna very much fall the same thing. There was gonna be more edge. There's gonna be more on premise because of data sovereignty, data gravity, because you're working with different parts of the business that have already made major cloud investments in different cloud providers. Right? So there's a lot of reasons why the modern, I guess the next modern generation of the data staff needs to be much more federated. >>Okay. So Richard, how do you deal with this? You you've obviously got, you know, the technical debt, the existing infrastructure it's on the books. You don't wanna just throw it out. A lot of, lot of conversation about modernizing applications, which a lot of times is a, you know, of microservices layer on top of leg legacy apps. Ho how do you think about the modern data stack? >>Well, I think probably the first thing to say is that the stack really has to include the processes and people around the data as well is all well and good changing the technology. But if you don't modernize how people use that technology, then you're not going to be able to, to scale because just cuz you can scale CPU and storage doesn't mean you can get more people to use your data, to generate you more value for the business. And so what we've been looking at is really changing in very much aligned to data products and, and data mesh. How do you enable more people to consume the service and have the stack respond in a way that keeps costs low? Because that's important for our customers consuming this data, but also allows people to occasionally run enormous queries and then tick along with smaller ones when required. And it's a good job we did because during COVID all of a sudden we had enormous pressures on our data platform to answer really important life threatening queries. And if we couldn't scale both our data stack and our teams, we wouldn't have been able to answer those as quickly as we had. So I think the stack needs to support a scalable business, not just the technology itself. >>Oh thank you for that. So Justin let's, let's try to break down what the critical aspects are of the modern data stack. So you think about the past, you know, five, seven years cloud obviously has given a different pricing model. Drisk experimentation, you know that we talked about the ability to scale up scale down, but it's, I'm, I'm taking away that that's not enough based on what Richard just said. The modern data stack has to serve the business and enable the business to build data products. I, I buy that I'm, you know, a big fan of the data mesh concepts, even though we're early days. So what are the critical aspects if you had to think about, you know, the paying, maybe putting some guardrails and definitions around the modern data stack, what does that look like? What are some of the attributes and principles there >>Of, of how it should look like or, or how >>Yeah. What it should be? >>Yeah. Yeah. Well, I think, you know, in Theresa mentioned this in, in a previous segment about the data warehouse is not necessarily going to disappear. It just becomes one node, one element of the overall data mesh. And I, I certainly agree with that. So by no means, are we suggesting that, you know, snowflake or Redshift or whatever cloud data warehouse you may be using is going to disappear, but it's, it's not going to become the end all be all. It's not the, the central single source of truth. And I think that's the paradigm shift that needs to occur. And I think it's also worth noting that those who were the early adopters of the modern data stack were primarily digital, native born in the cloud young companies who had the benefit of, of idealism. They had the benefit of starting with a clean slate that does not reflect the vast majority of enterprises. >>And even those companies, as they grow up mature out of that ideal state, they go by a business. Now they've got something on another cloud provider that has a different data stack and they have to deal with that heterogeneity that is just change and change is a part of life. And so I think there is an element here that is almost philosophical. It's like, do you believe in an absolute ideal where I can just fit everything into one place or do I believe in reality? And I think the far more pragmatic approach is really what data mesh represents. So to answer your question directly, I think it's adding, you know, the ability to access data that lives outside of the data warehouse, maybe living in open data formats in a data lake or accessing operational systems as well. Maybe you want to directly access data that lives in an Oracle database or a Mongo database or, or what have you. So creating that flexibility to really Futureproof yourself from the inevitable change that you will, you won't encounter over time. >>So thank you. So there, based on what Justin just said, I, I might take away there is it's inclusive, whether it's a data Mart, data hub, data lake data warehouse, it's a, just a node on the mesh. Okay. I get that. Does that include Theresa on, on Preem data? Obviously it has to, what are you seeing in terms of the ability to, to take that data mesh concept on pre I mean most implementations I've seen and data mesh, frankly really aren't, you know, adhering to the philosophy there. Maybe, maybe it's data lake and maybe it's using glue. You look at what JPMC is doing. Hello, fresh, a lot of stuff happening on the AWS cloud in that, you know, closed stack, if you will. What's the answer to that Theresa? >>I mean, I, I think it's a killer case for data mesh. The fact that you have valuable data sources, OnPrem, and then yet you still wanna modernize and take the best of cloud cloud is still, like we mentioned, there's a lot of great reasons for it around the economics and the way ability to tap into the innovation that the cloud providers are giving around data and AI architecture. It's an easy button. So the mesh allows you to have the best of both world. You can start using the data products on-prem or in the existing systems that are working already. It's meaningful for the business. At the same time, you can modernize the ones that make business sense because it needs better performance. It needs, you know, something that is, is cheaper or, or maybe just tap into better analytics to get better insights, right? So you're gonna be able to stretch and really have the best of both worlds that, again, going back to Richard's point, that is needful by the business. Not everything has to have that one size fits all set a tool. >>Okay. Thank you. So Richard, you know, you're talking about data as product. Wonder if we could give us your perspectives here, what are the advantages of treating data as a product? What, what role do data products have in the modern data stack? We talk about monetizing data. What are your thoughts on data products? >>So for us, one of the most important data products that we've been creating is taking data that is healthcare data across a wide variety of different settings. So information about patients' demographics about their, their treatment, about their medications and so on, and taking that into a standards format that can be utilized by a wide variety of different researchers because misinterpreting that data or having the data not presented in the way that the user is expecting means that you generate the wrong insight and in any business, that's clearly not a desirable outcome, but when that insight is so critical, as it might be in healthcare or some security settings, you really have to have gone to the trouble of understanding the data, presenting it in a format that everyone can clearly agree on. And then letting people consume in a very structured and managed way, even if that data comes from a variety of different sources in, in, in the first place. And so our data product journey has really begun by standardizing data across a number of different silos through the data mesh. So we can present out both internally and through the right governance externally to, to research is >>So that data product through whatever APIs is, is accessible, it's discoverable, but it's obviously gotta be governed as well. You mentioned appropriately provided to internally. Yeah. But also, you know, external folks as well. So the, so you've, you've architected that capability today >>We have and because the data is standard, it can generate value much more quickly and we can be sure of the security and, and, and value that that's providing because the data product isn't just about formatting the data into the right, correct tables, it's understanding what it means to redact the data or to remove certain rows from it or to interpret what a date actually means. Is it the start of the contract or the start of the treatment or the date of birth of a patient? These things can be lost in the data storage without having the proper product management around the data to say in a very clear business context, what does this data mean? And what does it mean to process this data for a particular use >>Case? Yeah, it makes sense. It's got the context. If the, if the domains on the data, you, you gotta cut through a lot of the, the, the centralized teams, the technical teams that, that data agnostic, they don't really have that context. All right. Let's end, Justin, how does Starburst fit into this modern data stack? Bring us home. >>Yeah. So I think for us, it's really providing our customers with, you know, the flexibility to operate and analyze data that lives in a wide variety of different systems. Ultimately giving them that optionality, you know, and optionality provides the ability to reduce costs, store more in a data lake rather than data warehouse. It provides the ability for the fastest time to insight to access the data directly where it lives. And ultimately with this concept of data products that we've now, you know, incorporated into our offering as well, you can really create and, and curate, you know, data as a product to be shared and consumed. So we're trying to help enable the data mesh, you know, model and make that an appropriate compliment to, you know, the, the, the modern data stack that people have today. >>Excellent. Hey, I wanna thank Justin Teresa and Richard for joining us today. You guys are great. I big believers in the, in the data mesh concept, and I think, you know, we're seeing the future of data architecture. So thank you. Now, remember, all these conversations are gonna be available on the cube.net for on-demand viewing. You can also go to starburst.io. They have some great content on the website and they host some really thought provoking interviews and, and, and they have awesome resources, lots of data mesh conversations over there, and really good stuff in, in the resource section. So check that out. Thanks for watching the data doesn't lie or does it made possible by Starburst data? This is Dave ante for the, and we'll see you next time.

Published Date : Aug 2 2022

SUMMARY :

And that is the claim that today's So it's the same general stack, So lemme come back to you just this, but okay. So a lot of the same sort of structural So Theresa, let me go to you cuz you have cloud first in your, in your, So the centralized cloud, as we know it, maybe data lake data warehouse in the central place, a, you know, of microservices layer on top of leg legacy apps. you can get more people to use your data, to generate you more value for the business. So you think about the past, you know, five, seven years cloud obviously has given And I think that's the paradigm shift that needs to occur. from the inevitable change that you will, you won't encounter over time. and data mesh, frankly really aren't, you know, adhering to So the mesh allows you to have the best of both world. So Richard, you know, you're talking about data as product. that data or having the data not presented in the way that the user But also, you know, external folks as well. the proper product management around the data to say in a very clear business It's got the context. So we're trying to help enable the data mesh, you know, I big believers in the, in the data mesh concept, and I think, you know,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RichardPERSON

0.99+

TheresaPERSON

0.99+

Richard JarvisPERSON

0.99+

JustinPERSON

0.99+

Justin BoormanPERSON

0.99+

DavePERSON

0.99+

AWSORGANIZATION

0.99+

fiveQUANTITY

0.99+

40 yearsQUANTITY

0.99+

StarburstORGANIZATION

0.99+

AccentureORGANIZATION

0.99+

40 yearsQUANTITY

0.99+

JPMCORGANIZATION

0.99+

bothQUANTITY

0.99+

Justin TeresaPERSON

0.99+

both worldsQUANTITY

0.99+

todayDATE

0.98+

first thingQUANTITY

0.98+

TeresaPERSON

0.98+

first technologistQUANTITY

0.98+

OracleORGANIZATION

0.98+

firstQUANTITY

0.98+

one elementQUANTITY

0.97+

InformaticaORGANIZATION

0.97+

cube.netOTHER

0.97+

MongoORGANIZATION

0.97+

starburst.ioOTHER

0.96+

seven yearsQUANTITY

0.95+

oneQUANTITY

0.95+

data MartORGANIZATION

0.91+

one placeQUANTITY

0.88+

both worldQUANTITY

0.85+

COVIDTITLE

0.83+

single locationQUANTITY

0.8+

OnPremORGANIZATION

0.8+

TerraORGANIZATION

0.77+

single sourceQUANTITY

0.74+

one sizeQUANTITY

0.73+

EMI healthORGANIZATION

0.73+

July numberDATE

0.7+

dataORGANIZATION

0.64+

five trendQUANTITY

0.63+

moneyQUANTITY

0.51+

threeQUANTITY

0.37+

Starburst Panel Q2


 

>>We're back with Jess Borgman of Starburst and Richard Jarvis of emus health. Okay. We're gonna get into lie. Number two, and that is this an open source based platform cannot give you the performance and control that you can get with a proprietary system. Is that a lie? Justin, the enterprise data warehouse has been pretty dominant and has evolved and matured. Its stack has mature over the years. Why is it not the default platform for data? >>Yeah, well, I think that's become a lie over time. So I, I think, you know, if we go back 10 or 12 years ago with the advent of the first data lake really around Hudu, that probably was true that you couldn't get the performance that you needed to run fast, interactive, SQL queries in a data lake. Now a lot's changed in 10 or 12 years. I remember in the very early days, people would say, you'll, you'll never get performance because you need to be column. You need to store data in a column format. And then, you know, column formats were introduced to, to data lakes. You have Parque ORC file in aro that were created to ultimately deliver performance out of that. So, okay. We got, you know, largely over the performance hurdle, you know, more recently people will say, well, you don't have the ability to do updates and deletes like a traditional data warehouse. >>And now we've got the creation of new data formats, again like iceberg and Delta and DY that do allow for updates and delete. So I think the data lake has continued to mature. And I remember a, a quote from, you know, Kurt Monash many years ago where he said, you know, it takes six or seven years to build a functional database. I think that's that's right. And now we've had almost a decade go by. So, you know, these technologies have matured to really deliver very, very close to the same level performance and functionality of, of cloud data warehouses. So I think the, the reality is that's become a lie and now we have large giant hyperscale internet companies that, you know, don't have the traditional data warehouse at all. They do all of their analytics in a data lake. So I think we've, we've proven that it's very much possible today. >>Thank you for that. And so Richard, talk about your perspective as a practitioner in terms of what open brings you versus, I mean, the closed is it's open as a moving target. I remember Unix used to be open systems and so it's, it is an evolving, you know, spectrum, but, but from your perspective, what does open give you that you can't get from a proprietary system where you are fearful of in a proprietary system? >>I, I suppose for me open buys us the ability to be unsure about the future, because one thing that's always true about technology is it evolves in a, a direction, slightly different to what people expect. And what you don't want to end up is done is backed itself into a corner that then prevents it from innovating. So if you have chosen the technology and you've stored trillions of records in that technology and suddenly a new way of processing or machine learning comes out, you wanna be able to take advantage and your competitive edge might depend upon it. And so I suppose for us, we acknowledge that we don't have perfect vision of what the future might be. And so by backing open storage technologies, we can apply a number of different technologies to the processing of that data. And that gives us the ability to remain relevant, innovate on our data storage. And we have bought our way out of the, any performance concerns because we can use cloud scale infrastructure to scale up and scale down as we need. And so we don't have the concerns that we don't have enough hardware today to process what we want to do, but want to achieve. We can just scale up when we need it and scale back down. So open source has really allowed us to maintain the being at the cutting edge. >>So Justin, let me play devil's advocate here a little bit, and I've talked to JAK about this and you know, obviously her vision is there's an open source that, that data mesh is open source, an open source tooling, and it's not a proprietary, you know, you're not gonna buy a data mesh. You're gonna build it with, with open source toolings and, and vendors like you are gonna support it, but come back to sort of today, you can get to market with a proprietary solution faster. I'm gonna make that statement. You tell me if it's a lie and then you can say, okay, we support Apache iceberg. We're gonna support open source tooling, take a company like VMware, not really in the data business, but how, the way they embraced Kubernetes and, and you know, every new open source thing that comes along, they say, we do that too. Why can't proprietary systems do that and be as effective? >>Yeah, well, I think at least with the, within the data landscape saying that you can access open data formats like iceberg or, or others is, is a bit dis disingenuous because really what you're selling to your customer is a certain degree of performance, a certain SLA, and you know, those cloud data warehouses that can reach beyond their own proprietary storage drop all the performance that they were able to provide. So it is, it reminds me kind of, of, again, going back 10 or 12 years ago when everybody had a connector to hit and that they thought that was the solution, right? But the reality was, you know, a connector was not the same as running workloads in had back then. And I think, think similarly, you know, being able to connect to an external table that lives in an open data format, you know, you're, you're not going to give it the performance that your customers are accustomed to. And at the end of the day, they're always going to be predisposed. They're always going to be incentivized to get that data ingested into the data warehouse, cuz that's where they have control. And you know, the bottom line is the database industry has really been built around vendor lockin. I mean, from the start, how, how many people love Oracle today, but our customers, nonetheless, I think, you know, lockin is, is, is part of this industry. And I think that's really what we're trying to change with open data formats. >>Well, it's interesting reminded when I, you know, I see the, the gas price, the TSR gas price I, I drive up and then I say, oh, that's the cash price credit card. I gotta pay 20 cents more, but okay. But so the, the argument then, so let me, let me come back to you, Justin. So what's wrong with saying, Hey, we support open data formats, but yeah, you're gonna get better performance if you, if you, you keep it into our closed system, are you saying that long term that's gonna come back and bite you cuz you're gonna end up. You mentioned Oracle, you mentioned Teradata. Yeah. That's by, by implication, you're saying that's where snowflake customers are headed. >>Yeah, absolutely. I think this is a movie that, you know, we've all seen before. At least those of us who've been in the industry long enough to, to see this movie play over a couple times. So I do think that's the future. And I think, you know, I loved what Richard said. I actually wrote it down cause I thought it was amazing quote. He said, it buys us the ability to be unsure of the future. That that pretty much says it all the, the future is unknowable and the reality is using open data formats. You remain interoperable with any technology you want to utilize. If you want to use smart to train a machine learning model and you wanna use Starbust to query be a sequel, that's totally cool. They can both work off the same exact, you know, data, data sets by contrast, if you're, you know, focused on a proprietary model, then you're kind of locked in again to that model. I think the same applies to data, sharing to data products, to a wide variety of, of aspects of the data landscape that a proprietary approach kind of closes you and, and locks you in. >>So I would say this Richard, I'd love to get your thoughts on it. Cause I talked to a lot of Oracle customers, not as many te data customers, but, but a lot of Oracle customers and they, you know, they'll admit yeah, you know, they Jimin some price and the license cost they give, but we do get value out of it. And so my question to you, Richard, is, is do the, let's call it data warehouse systems or the proprietary systems. Are they gonna deliver a greater ROI sooner? And is that in allure of, of that customers, you know, are attracted to, or can open platforms deliver as fast an ROI? >>I think the answer to that is it can depend a bit. It depends on your business's skillset. So we are lucky that we have a number of proprietary teams that work in databases that provide our operational data capability. And we have teams of analytics and big data experts who can work with open data sets and open data formats. And so for those different teams, they can get to an ROI more quickly with different technologies for the business though, we can't do better for our operational data stores than proprietary databases. Today we can back off very tight SLAs to them. We can demonstrate reliability from millions of hours of those databases being run enterprise scale, but for an analytics workload where increasing our business is growing in that direction, we can't do better than open data formats with cloud based data mesh type technologies. And so it's not a simple answer. That one will always be the right answer for our business. We definitely have times when proprietary databases provide a capability that we couldn't easily represent or replicate with open technologies. >>Yeah. Richard, stay with you. You mentioned, you know, you know, some things before that, that strike me, you know, the data brick snowflake, you know, thing is a lot of fun for analysts like me. You've got data bricks coming at it. Richard, you mentioned you have a lot of rockstar, data engineers, data bricks coming at it from a data engineering heritage. You get snowflake coming at it from an analytics heritage. Those two worlds are, are colliding people like P Sanji Mohan said, you know what? I think it's actually harder to play in the data engineering. So I E it's easier to for data engineering world to go into the analytics world versus the reverse, but thinking about up and coming engineers and developers preparing for this future of data engineering and data analytics, how, how should they be thinking about the future? What, what's your advice to those young people? >>So I think I'd probably fall back on general programming skill sets. So the advice that I saw years ago was if you have open source technologies, the pythons and Javas on your CV, you command a 20% pay, hike over people who can only do proprietary programming languages. And I think that's true of data technologies as well. And from a business point of view, that makes sense. I'd rather spend the money that I save on proprietary licenses on better engineers, because they can provide more value to the business that can innovate us beyond our competitors. So I think I would my advice to people who are starting here or trying to build teams to capitalize on data assets is begin with open license, free capabilities, because they're very cheap to experiment with. And they generate a lot of interest from people who want to join you as a business. And you can make them very successful early, early doors with, with your analytics journey. >>It's interesting. Again, analysts like myself, we do a lot of TCO work and have over the last 20 plus years and in the world of Oracle, you know, normally it's the staff, that's the biggest nut in total cost of ownership, not an Oracle. It's the it's the license cost is by far the biggest component in the, in the blame pie. All right, Justin, help us close out this segment. We've been talking about this sort of data mesh open, closed snowflake data bricks. Where does Starburst sort of as this engine for the data lake data lake house, the data warehouse, it fit in this, in this world. >>Yeah. So our view on how the future ultimately unfolds is we think that data lakes will be a natural center of gravity for a lot of the reasons that we described open data formats, lowest total cost of ownership, because you get to choose the cheapest storage available to you. Maybe that's S3 or Azure data lake storage, or Google cloud storage, or maybe it's on-prem object storage that you bought at a, at a really good price. So ultimately storing a lot of data in a data lake makes a lot of sense, but I think what makes our perspective unique is we still don't think you're gonna get everything there either. We think that basically centralization of all your data assets is just an impossible endeavor. And so you wanna be able to access data that lives outside of the lake as well. So we kind of think of the lake as maybe the biggest place by volume in terms of how much data you have, but to, to have comprehensive analytics and to truly understand your business and understand it holistically, you need to be able to go access other data sources as well. And so that's the role that we wanna play is to be a single point of access for our customers, provide the right level of fine grained access control so that the right people have access to the right data and ultimately make it easy to discover and consume via, you know, the creation of data products as well. >>Great. Okay. Thanks guys. Right after this quick break, we're gonna be back to debate whether the cloud data model that we see emerging and the so-called modern data stack is really modern, or is it the same wine new bottle when it comes to data architectures, you're watching the cube, the leader in enterprise and emerging tech coverage.

Published Date : Aug 2 2022

SUMMARY :

cannot give you the performance and control that you can get with We got, you know, largely over the performance hurdle, you know, more recently people will say, And I remember a, a quote from, you know, Kurt Monash many years ago where he said, you know, open systems and so it's, it is an evolving, you know, spectrum, And what you don't want to end up So Justin, let me play devil's advocate here a little bit, and I've talked to JAK about this and you know, And I think, think similarly, you know, being able to connect to an external table that lives in an open data Well, it's interesting reminded when I, you know, I see the, the gas price, And I think, you know, I loved what Richard said. not as many te data customers, but, but a lot of Oracle customers and they, you know, I think the answer to that is it can depend a bit. that strike me, you know, the data brick snowflake, you know, thing is a lot of fun for analysts So the advice that I saw years ago was if you have open source technologies, years and in the world of Oracle, you know, normally it's the staff, it easy to discover and consume via, you know, the creation of data products as well. data model that we see emerging and the so-called modern data stack

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RichardPERSON

0.99+

Jess BorgmanPERSON

0.99+

JustinPERSON

0.99+

sixQUANTITY

0.99+

OracleORGANIZATION

0.99+

Richard JarvisPERSON

0.99+

20 centsQUANTITY

0.99+

20%QUANTITY

0.99+

Kurt MonashPERSON

0.99+

P Sanji MohanPERSON

0.99+

TodayDATE

0.99+

seven yearsQUANTITY

0.99+

pythonsTITLE

0.99+

TeradataORGANIZATION

0.99+

JAKPERSON

0.99+

JavasTITLE

0.99+

10DATE

0.99+

todayDATE

0.98+

StarbustTITLE

0.98+

StarburstORGANIZATION

0.97+

VMwareORGANIZATION

0.97+

bothQUANTITY

0.97+

12 years agoDATE

0.96+

single pointQUANTITY

0.96+

millions of hoursQUANTITY

0.95+

10QUANTITY

0.93+

UnixTITLE

0.92+

12 yearsQUANTITY

0.92+

GoogleORGANIZATION

0.9+

two worldsQUANTITY

0.9+

DYORGANIZATION

0.87+

first data lakeQUANTITY

0.86+

HuduLOCATION

0.85+

trillionsQUANTITY

0.85+

one thingQUANTITY

0.83+

many years agoDATE

0.79+

Apache icebergORGANIZATION

0.79+

over a couple timesQUANTITY

0.77+

emus healthORGANIZATION

0.75+

JiminPERSON

0.73+

StarburstTITLE

0.73+

years agoDATE

0.72+

AzureTITLE

0.7+

KubernetesORGANIZATION

0.67+

TCOORGANIZATION

0.64+

S3TITLE

0.62+

DeltaORGANIZATION

0.6+

plus yearsDATE

0.59+

Number twoQUANTITY

0.58+

a decadeQUANTITY

0.56+

icebergTITLE

0.47+

ParqueORGANIZATION

0.47+

lastDATE

0.47+

20QUANTITY

0.46+

Q2QUANTITY

0.31+

ORCORGANIZATION

0.27+

Starburst Panel Q1


 

>>In 2011, early Facebook employee and Cloudera co-founder Jeff Ocker famously said the best minds of my generation are thinking about how to get people to click on ads. And that sucks. Let's face it more than a decade later organizations continue to be frustrated with how difficult it is to get value from data and build a truly agile data driven enterprise. What does that even mean? You ask? Well, it means that everyone in the organization has the data they need when they need it. In a context that's relevant to advance the mission of an organization. Now that could mean cutting costs could mean increasing profits, driving productivity, saving lives, accelerating drug discovery, making better diagnoses, solving, supply chain problems, predicting weather disasters, simplifying processes, and thousands of other examples where data can completely transform people's lives beyond manipulating internet users to behave a certain way. We've heard the prognostications about the possibilities of data before and in fairness we've made progress, but the hard truth is the original promises of master data management, enterprise data, warehouses, data, Mars, data hubs, and yes, even data lakes were broken and left us wanting for more welcome to the data doesn't lie, or does it a series of conversations produced by the cube and made possible by Starburst data. >>I'm your host, Dave Lanta and joining me today are three industry experts. Justin Borgman is this co-founder and CEO of Starburst. Richard Jarvis is the CTO at EMI health and Theresa tongue is cloud first technologist at Accenture. Today we're gonna have a candid discussion that will expose the unfulfilled and yes, broken promises of a data past we'll expose data lies, big lies, little lies, white lies, and hidden truths. And we'll challenge, age old data conventions and bust some data myths. We're debating questions like is the demise of a single source of truth. Inevitable will the data warehouse ever have feature parody with the data lake or vice versa is the so-called modern data stack simply centralization in the cloud, AKA the old guards model in new cloud close. How can organizations rethink their data architectures and regimes to realize the true promises of data can and will and open ecosystem deliver on these promises in our lifetimes, we're spanning much of the Western world today. Richard is in the UK. Teresa is on the west coast and Justin is in Massachusetts with me. I'm in the cube studios about 30 miles outside of Boston folks. Welcome to the program. Thanks for coming on. Thanks for having us. Let's get right into it. You're very welcome. Now here's the first lie. The most effective data architecture is one that is centralized with a team of data specialists serving various lines of business. What do you think Justin? >>Yeah, definitely a lie. My first startup was a company called hit adapt, which was an early SQL engine for IDU that was acquired by Teradata. And when I got to Teradata, of course, Terada is the pioneer of that central enterprise data warehouse model. One of the things that I found fascinating was that not one of their customers had actually lived up to that vision of centralizing all of their data into one place. They all had data silos. They all had data in different systems. They had data on-prem data in the cloud. You know, those companies were acquiring other companies and inheriting their data architecture. So, you know, despite being the industry leader for 40 years, not one of their customers truly had everything in one place. So I think definitely history has proven that to be a lie. >>So Richard, from a practitioner's point of view, you know, what, what are your thoughts? I mean, there, there's a lot of pressure to cut cost, keep things centralized, you know, serve the business as best as possible from that standpoint. What, what is your experience, Joe? >>Yeah, I mean, I think I would echo Justin's experience really that we, as a business have grown up through acquisition, through storing data in different places sometimes to do information governance in different ways to store data in, in a platform that's close to data experts, people who really understand healthcare data from pharmacies or from, from doctors. And so, although if you were starting from a Greenfield site and you were building something brand new, you might be able to centralize all the data and all of the tooling and teams in one place. The reality is that that businesses just don't grow up like that. And, and it's just really impossible to get that academic perfection of, of storing everything in one place. >>Y you know, Theresa, I feel like Sarbanes Oxley kinda saved the data warehouse, you know? Right. But you actually did have to have a single version of the truth for certain financial data, but really for those, some of those other use cases, I, I mentioned, I, I do feel like the industry has kinda let us down. What's your take on this? Where does it make sense to have that sort of centralized approach versus where does it make sense to maybe decentralized? >>I, I think you gotta have centralized governance, right? So from the central team, for things like swans Oxley, for things like security, for certain very core data sets, having a centralized set of roles, responsibilities to really QA, right. To serve as a design authority for your entire data estate, just like you might with security, but how it's implemented has to be distributed. Otherwise you're not gonna be able to scale. Right? So being able to have different parts of the business really make the right data investments for their needs. And then ultimately you're gonna collaborate with your partners. So partners that are not within the company, right. External partners, we're gonna see a lot more data sharing and model creation. And so you're definitely going to be decentralized. >>So, you know, Justin, you guys last, geez, I think it was about a year ago, had a session on, on data mesh. It was a great program. You invited JAK, Dani, of course, she's the creator of the data mesh. And her one of our fundamental premises is that you've got this hyper specialized team that you've gotta go through. And if you want anything, but at the same time, these, these individuals actually become a bottleneck, even though they're some of the most talented people in the organization. So I guess question for you, Richard, how do you deal with that? Do you, do you organize so that there are a few sort of rock stars that, that, you know, build cubes and, and the like, and, and, and, or have you had any success in sort of decentralizing with, you know, your, your constituencies, that data model? >>Yeah. So, so we absolutely have got rockstar, data scientists and data guardians. If you like people who understand what it means to use this data, particularly as the data that we use at emos is very private it's healthcare information. And some of the, the rules and regulations around using the data are very complex and, and strict. So we have to have people who understand the usage of the data, then people who understand how to build models, how to process the data effectively. And you can think of them like consultants to the wider business, because a pharmacist might not understand how to structure a SQL query, but they do understand how they want to process medication information to improve patient lives. And so that becomes a, a consulting type experience from a, a set of rock stars to help a, a more decentralized business who needs to, to understand the data and to generate some valuable output. >>Justin, what do you say to a, to a customer or prospect that says, look, Justin, I'm gonna, I got a centralized team and that's the most cost effective way to serve the business. Otherwise I got, I got duplication. What do you say to that? >>Well, I, I would argue it's probably not the most cost effective and, and the reason being really twofold. I think, first of all, when you are deploying a enterprise data warehouse model, the, the data warehouse itself is very expensive, generally speaking. And so you're putting all of your most valuable data in the hands of one vendor who now has tremendous leverage over you, you know, for many, many years to come, I think that's the story of Oracle or Terra data or other proprietary database systems. But the other aspect I think is that the reality is those central data warehouse teams is as much as they are experts in the technology. They don't necessarily understand the data itself. And this is one of the core tenets of data mash that that jam writes about is this idea of the domain owners actually know the data the best. >>And so by, you know, not only acknowledging that data is generally decentralized and to your earlier point about, so Oxley, maybe saving the data warehouse, I would argue maybe GDPR and data sovereignty will destroy it because data has to be decentralized for, for those laws to be compliant. But I think the reality is, you know, the data mesh model basically says, data's decentralized, and we're gonna turn that into an asset rather than a liability. And we're gonna turn that into an asset by empowering the people that know the data, the best to participate in the process of, you know, curating and creating data products for, for consumption. So I think when you think about it, that way, you're going to get higher quality data and faster time to insight, which is ultimately going to drive more revenue for your business and reduce costs. So I think that that's the way I see the two, the two models comparing and con contrasting. >>So do you think the demise of the data warehouse is inevitable? I mean, I mean, you know, there Theresa you work with a lot of clients, they're not just gonna rip and replace their existing infrastructure. Maybe they're gonna build on top of it, but the, what does that mean? Does that mean the ed w just becomes, you know, less and less valuable over time, or it's maybe just isolated to specific use cases. What's your take on that? >>Listen, I still would love all my data within a data warehouse would love it. Mastered would love it owned by essential team. Right? I think that's still what I would love to have. That's just not the reality, right? The investment to actually migrate and keep that up to date. I would say it's a losing battle. Like we've been trying to do it for a long time. Nobody has the budgets and then data changes, right? There's gonna be a new technology. That's gonna emerge that we're gonna wanna tap into. There's gonna be not enough investment to bring all the legacy, but still very useful systems into that centralized view. So you keep the data warehouse. I think it's a very, very valuable, very high performance tool for what it's there for, but you could have this, you know, new mesh layer that still takes advantage of the things. I mentioned, the data products in the systems that are meaningful today and the data products that actually might span a number of systems. Maybe either those that either source systems, the domains that know it best, or the consumer based systems and products that need to be packaged in a way that be really meaningful for that end user, right? Each of those are useful for a different part of the business and making sure that the mesh actually allows you to lose all of them. >>So, Richard, let me ask you, you take, take Gemma's principles back to those. You got, you know, the domain ownership and, and, and data as product. Okay, great. Sounds good. But it creates what I would argue or two, you know, challenges self-serve infrastructure let's park that for a second. And then in your industry, one of the high, most regulated, most sensitive computational governance, how do you automate and ensure federated governance in that mesh model that Theresa was just talking about? >>Well, it absolutely depends on some of the tooling and processes that you put in place around those tools to be, to centralize the security and the governance of the data. And, and I think, although a data warehouse makes that very simple, cause it's a single tool, it's not impossible with some of the data mesh technologies that are available. And so what we've done at EMI is we have a single security layer that sits on top of our data mesh, which means that no matter which user is accessing, which data source, we go through a well audited well understood security layer. That means that we know exactly who's got access to which data field, which data tables. And then everything that they do is, is audited in a very kind of standard way, regardless of the underlying data storage technology. So for me, although storing the data in one place might not be possible understanding where your source of truth is and securing that in a common way is still a valuable approach and you can do it without having to bring all that data into a single bucket so that it's all in one place. >>And, and so having done that and investing quite heavily in making that possible has paid dividends in terms of giving wider access to the platform and ensuring that only data that's available under GDPR and other regulations is being used by, by the data users. >>Yeah. So Justin mean Democrat, we always talk about data democratization and you know, up until recently, they really haven't been line of sight as to how to get there. But do you have anything to add to this because you're essentially taking, you know, doing analytic queries and with data, that's all dispersed all over the, how are you seeing your customers handle this, this challenge? >>Yeah, I mean, I think data products is a really interesting aspect of the answer to that. It allows you to, again, leverage the data domain owners, people know the data, the best to, to create, you know, data as a product ultimately to be consumed. And we try to represent that in our product as effectively, almost eCommerce, like experience where you go and discover and look for the data products that have been created in your organization. And then you can start to consume them as, as you'd like. And so really trying to build on that notion of, you know, data democratization and self-service, and making it very easy to discover and, and start to use with whatever BI tool you, you may like, or even just running, you know, SQL queries yourself. >>Okay. G guys grab a sip of water. After the short break, we'll be back to debate whether proprietary or open platforms are the best path to the future of data excellence. Keep it right there.

Published Date : Aug 2 2022

SUMMARY :

famously said the best minds of my generation are thinking about how to get people to Teresa is on the west coast and Justin is in Massachusetts with me. So, you know, despite being the industry leader for 40 years, not one of their customers truly had So Richard, from a practitioner's point of view, you know, what, what are your thoughts? you might be able to centralize all the data and all of the tooling and teams in one place. Y you know, Theresa, I feel like Sarbanes Oxley kinda saved the data warehouse, I, I think you gotta have centralized governance, right? of rock stars that, that, you know, build cubes and, and the like, And you can think of them like consultants Justin, what do you say to a, to a customer or prospect that says, look, Justin, I'm gonna, you know, for many, many years to come, I think that's the story of Oracle or Terra data or other proprietary But I think the reality is, you know, the data mesh model basically says, I mean, you know, there Theresa you work with a lot of clients, they're not just gonna rip and replace their existing you know, new mesh layer that still takes advantage of the things. But it creates what I would argue or two, you know, Well, it absolutely depends on some of the tooling and processes that you put in place around And, and so having done that and investing quite heavily in making that possible But do you have anything to add to this because you're essentially taking, you know, the best to, to create, you know, data as a product ultimately to be consumed. open platforms are the best path to the future of

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave LantaPERSON

0.99+

DaniPERSON

0.99+

RichardPERSON

0.99+

Justin BorgmanPERSON

0.99+

JustinPERSON

0.99+

Jeff OckerPERSON

0.99+

TheresaPERSON

0.99+

Richard JarvisPERSON

0.99+

TeresaPERSON

0.99+

MassachusettsLOCATION

0.99+

TeradataORGANIZATION

0.99+

40 yearsQUANTITY

0.99+

OracleORGANIZATION

0.99+

UKLOCATION

0.99+

twoQUANTITY

0.99+

JoePERSON

0.99+

GDPRTITLE

0.99+

JAKPERSON

0.99+

2011DATE

0.99+

StarburstORGANIZATION

0.99+

BostonLOCATION

0.99+

thousandsQUANTITY

0.99+

two modelsQUANTITY

0.99+

EMIORGANIZATION

0.99+

FacebookORGANIZATION

0.99+

GemmaPERSON

0.99+

TeradaORGANIZATION

0.99+

AccentureORGANIZATION

0.99+

EachQUANTITY

0.99+

first lieQUANTITY

0.99+

todayDATE

0.99+

first startupQUANTITY

0.98+

ClouderaORGANIZATION

0.98+

TodayDATE

0.98+

SQLTITLE

0.98+

first technologistQUANTITY

0.97+

one placeQUANTITY

0.97+

DemocratORGANIZATION

0.97+

singleQUANTITY

0.97+

about 30 milesQUANTITY

0.97+

oneQUANTITY

0.96+

three industry expertsQUANTITY

0.95+

more than a decade laterDATE

0.94+

OneQUANTITY

0.94+

hit adaptORGANIZATION

0.94+

Terra dataORGANIZATION

0.93+

GreenfieldLOCATION

0.92+

single sourceQUANTITY

0.91+

single toolQUANTITY

0.91+

OxleyPERSON

0.91+

one vendorQUANTITY

0.9+

single bucketQUANTITY

0.9+

single versionQUANTITY

0.88+

about a year agoDATE

0.85+

Theresa tonguePERSON

0.83+

emosORGANIZATION

0.82+

MarsORGANIZATION

0.8+

swans OxleyPERSON

0.77+

IDUTITLE

0.69+

firstQUANTITY

0.59+

a secondQUANTITY

0.55+

Sarbanes OxleyORGANIZATION

0.53+

MasteredPERSON

0.45+

Q1QUANTITY

0.37+

Justin Borgman, Starburst and Teresa Tung, Accenture | AWS re:Invent 2021


 

>>Hey, welcome back to the cubes. Continuing coverage of AWS reinvent 2021. I'm your host, Lisa Martin. This is day two, our first full day of coverage. But day two, we have two life sets here with AWS and its ecosystem partners to remote sets over a hundred guests on the program. We're going to be talking about the next decade of cloud innovation, and I'm pleased to welcome back to cube alumni to the program. Justin Borkman is here, the co-founder and CEO of Starburst and Teresa Tung, the cloud first chief technologist at Accenture guys. Welcome back to the queue. Thank you. Thank you for having me. Good to have you back. So, so Teresa, I was doing some research on you and I see you are the most prolific prolific inventor at Accenture with over 220 patents and patent applications. That's huge. Congratulations. Thank you. Thank you. And I love your title. I think it's intriguing. I'd like to learn a little bit more about your role cloud-first chief technologist. Tell me about, >>Well, I get to think about the future of cloud and if you think about clouded powers, everything experiences in our everyday lives and our homes and our car in our stores. So pretty much I get to be cute, right? The rest of Accenture's James Bond >>And your queue. I like that. Wow. What a great analogy. Just to talk to me a little bit, I know service has been on the program before, but give me a little bit of an overview of the company, what you guys do. What were some of the gaps in the markets that you saw a few years ago and said, we have an idea to solve this? Sure. >>So Starburst offers a distributed query engine, which essentially means we're able to run SQL queries on data anywhere, uh, could be in traditional relational databases, data lakes in the cloud on-prem. And I think that was the gap that we saw was basically that people had data everywhere and really had a challenge with how they analyze that data. And, uh, my co-founders are the creators of an open source project originally called Presto now called Trino. And it's how Facebook and Netflix and Airbnb and, and a number of the internet companies run their analytics. And so our idea was basically to take that, commercialize that and make it enterprise grade for the thousands of other companies that are struggling with data management, data analytics problems. >>And that's one of the things we've seen explode during the last 22 months, among many other things is data, right? In every company. These days has to be a data company. If they're not, there's a competitor in the rear view rear view mirror, ready to come and take that place. We're going to talk about the data mesh Teresa, we're going to start with you. This is not a new car. This is a new concept. Talk to us about what a data mesh is and why organizations need to embrace this >>Approach. So there's a canonical definition about data mesh with four attributes and any data geek or data architect really resonates with them. So number one, it's really routed decentralized domain ownership. So data is not within a single line of business within a single entity within a single partner has to be across different domains. Second is publishing data as products. And so instead of these really, you know, technology solutions, data sets, data tables, really thinking about the product and who's going to use it. The third one is really around self-service infrastructure. So you want everybody to be able to use those products. And finally, number four, it's really about federated and global governance. So even though their products, you really need to make sure that you're doing the right things, but what's data money. >>We're not talking about a single tool here, right? This is more of a, an approach, a solution. >>It is a data strategy first and foremost, right? So companies, they are multi-cloud, they have many projects going on, they are on premise. So what do you do about it? And so that's the reality of the situation today, and it's first and foremost, a business strategy and framework to think about the data. And then there's a new architecture that underlines and supports that >>Just didn't talk to me about when you're having customer conversations. Obviously organizations need to have a core data strategy that runs the business. They need to be able to, to democratize really truly democratized data access across all business units. What are some of the, what are some of your customer conversations like are customers really embracing the data strategy, vision and approach? >>Yeah, well, I think as you alluded to, you know, every business is data-driven today and the pandemic, if anything has accelerated digital transformation in that move to become data-driven. So it's imperative that every business of every shape and size really put the power of data in the hands of everyone within their organization. And I think part of what's making data mesh resonates so well, is that decentralization concept that Teresa spoke about? Like, I think companies acknowledge that data is inherently decentralized. They have a lot of different database systems, different teams and data mesh is a framework for thinking about that. Then not only acknowledges that reality, but also braces it and basically says there's actually advantages to this decentralized approach. And so I think that's, what's driving the interest level in the data mesh, uh, paradigm. And it's been exciting to work with customers as they think about that strategy. And I think that, you know, essentially every company in the space is, is in transition, whether they're moving from on cloud to the prem, uh, to, uh, sorry, from on-prem to the cloud or from one cloud to another cloud or undergoing that digital transformation, they have left behind data everywhere. And so they're, they're trying to wrestle with how to grasp that. >>And there's, we know that there's so much value in data. The, the need is to be able to get it, to be able to analyze it quickly in real time. I think another thing we learned in the pandemic is it real-time is no longer a nice to have. It is essential for businesses in every organization. So Theresa let's talk about how Accenture and servers are working together to take the data mesh from a concept of framework and put this into production into execution. >>Yeah. I mean, many clients are already doing some aspect of the data mesh as I listed those four attributes. I'm sure everybody thought like I'm already doing some of this. And so a lot of that is reviewing your existing data projects and looking at it from a data product landscape we're at Amazon, right? Amazon famous for being customer obsessed. So in data, we're not always customer obsessed. We put up tables, we put up data sets, feature stores. Who's actually going to use this data. What's the value from it. And I think that's a big change. And so a lot of what we're doing is helping apply that product lens, a literal product lens and thinking about the customer. >>So what are some w you know, we often talk about outcomes, everything being outcomes focused and customers, vendors wanting to help customers deliver big outcomes, you know, cost reduction, et cetera, things like that. How, what are some of the key outcomes Theresa that the data mesh framework unlocks for organizations in any industry to be able to leverage? >>Yeah. I mean, it really depends on the product. Some of it is organizational efficiency and data-driven decisions. So just by the able to see the data, see what's happening now, that's great. But then you have so beyond the, now what the, so what the analytics, right. Both predictive prescriptive analytics. So what, so now I have all this data I can analyze and drive and predict. And then finally, the, what if, if I have this data and my partners have this data in this mesh, and I can use it, I can ask a lot of what if and, and kind of game out scenarios about what if I did things differently, all of this in a very virtualized data-driven fashion, >>Right? Well, we've been talking about being data-driven for years and years and years, but it's one thing to say that it's a whole other thing to actually be able to put that into practice and to use it, to develop new products and services, delight customers, right. And, and really achieve the competitive advantage that businesses want to have. Just so talk to me about how your customer conversations have changed in the last 22 months, as we've seen this massive acceleration of digital transformation companies initially, really trying to survive and figure out how to pivot, not once, but multiple times. How are those customer conversations changing now is as that data strategy becomes core to the survival of every business and its ability to thrive. >>Yeah. I mean, I think it's accelerated everything and, and that's been obviously good for companies like us and like Accenture, cause there's a lot of work to be done out there. Um, but I think it's a transition from a storage centric mindset to more of an analytics centric mindset. You know, I think traditionally data warehousing has been all about moving data into one central place. And, and once you get it there, then you can analyze it. But I think companies don't have the time to wait for that anymore. Right there, there's no time to build all the ETL pipelines and maintain them and get all of that data together. We need to shorten that time to insight. And that's really what we, what we've been focusing on with our, with our customers, >>Shorten that time to insight to get that value out of the data faster. Exactly. Like I said, you know, the time is no longer a nice to have. It's an absolute differentiator for folks in every business. And as, as in our consumer lives, we have this expectation that we can get whatever we want on our phone, on any device, 24 by seven. And of course now in our business lives, we're having the same expectation, but you have to be able to unlock that access to that data, to be able to do the analytics, to make the decisions based on what the data say. Are you, are you finding our total? Let's talk about a little bit about the go to market strategy. You guys go in together. Talk to me about how you're working with AWS, Theresa, we'll start with you. And then Justin we'll head over to you. Okay. >>Well, a lot of this is powered by the cloud, right? So being able to imagine a new data business to run the analytics on it and then push it out, all of that is often cloud-based. But then the great thing about data mesh it's it gives you a framework to look at and tap into multi-cloud on-prem edge data, right? Data that can't be moved because it is a private and secure has to be at the edge and on-prem so you need to have that's their data reality. And the cloud really makes this easier to do. And then with data virtualization, especially coming from the digital natives, we know it scales >>Just to talk to me about it from your perspective that the GTL. >>Yeah. So, I mean, I think, uh, data mesh is really about people process and technology. I think Theresa alluded to it as a strategy. It's, it's more than just technology. Obviously we bring some of that technology to bear by allowing customers to query the data where it lives. But the people in process side is just as important training people to kind of think about how they do data management, data analytics differently is essential thinking about how to create data as a product. That's one of the core principles that Theresa mentioned, you know, that's where I think, um, you know, folks like Accenture can be really instrumental in helping people drive that transformational change within their organization. And that's >>Hard. Transformational change is hard with, you know, the last 22 months. I've been hard on everyone for every reason. How are you facilitating? I'm curious, like to get Theresa, we'll start with you, your perspectives on how our together as servers and Accenture, with the power of AWS, helping to drive that cultural change within organizations. Because like we talked about Justin there, nobody has extra time to waste on anything these days. >>The good news is there's that imperative, right? Every business is a digital business. We found that our technology leaders, right, the top 10% investors in digital, they are outperforming are the laggards. So before pandemic, it's times to post pep devek times five, so there's a need to change. And so data is really the heart of the company. That's how you unlock your technical debt into technical wealth. And so really using cloud and technologies like Starburst and data virtualization is how we can actually do that. >>And so how do you, Justin, how does Starburst help organizations transfer that technical debt or reduce it? How does the D how does the data much help facilitate that? Because we talk about technical debt and it can, it can really add up. >>Yeah, well, a lot of people use us, uh, or think about us as an abstraction layer above the different data sources that they have. So they may have legacy data sources today. Um, then maybe they want to move off of over time, um, could be classical data, warehouses, other classical, uh, relational databases, perhaps they're moving to the cloud. And by leveraging Starburst as this abstraction, they can query the data that they have today, while in the background, moving data into the cloud or moving it into the new data stores that they want to utilize. And it sort of hides that complexity. It decouples the end user experience, the business analyst, the data scientists from where the data lives. And I think that gives people a lot of freedom and a lot of optionality. And I think, you know, the only constant is change. Um, and so creating an architecture that can stand the test of time, I think is really, really important. >>Absolutely. Speaking of change, I just saw the announcement about Starburst galaxy fully managed SAS platform now available in all three major clouds. Of course, here we are at AWS. This is a, is this a big directional shift for servers? >>It is, you know, uh, I think there's great precedent within open source enterprise software companies like Mongo DB or confluent who started with a self managed product, much the way that we did, and then moved in the direction of creating a SAS product, a cloud hosted, fully managed product that really I think, expands the market. And that's really essentially what we're doing with galaxy galaxy is designed to be as easy as possible. Um, you know, Starburst was already powerful. This makes it powerful and easy. And, uh, and, and in our view, can, can hopefully expand the market to thousands of potential customers that can now leverage this technology in a, in a faster, easier way, >>Just in sticking with you for a minute. Talk to me about kind of where you're going in, where services heading in terms of support for the data mesh architecture across industries. >>Yeah. So a couple of things that we've, we've done recently, and whether we're doing, uh, as we speak, one is, uh, we introduced a new capability. We call star gate. Now star gate is a connector between Starburst clusters. So you're going to have a Starbucks cluster, and let's say Azure service cluster in AWS, a Starbucks cluster, maybe an AWS west and AWS east. And this basically pushes the processing to where the data lives. So again, living within this construct of, uh, of decentralized data that a data mesh is all about, this allows you to do that at an even greater level of abstraction. So it doesn't even matter what cloud region the data lives in or what cloud entirely it lives in. And there are a lot of important applications for this, not only latency in terms of giving you fast, uh, ability to join across those different clouds, but also, uh, data sovereignty constraints, right? >>Um, increasingly important, especially in Europe, but increasingly everywhere. And, you know, if your data isn't Switzerland, it needs to stay in Switzerland. So starting date as a way of pushing the processing to Switzerland. So you're minimizing the data that you need to pull back to complete your analysis. And, uh, and so we think that's a big deal about, you know, kind of enabling a data mash on a, on a global scale. Um, another thing we're working on back to the point of data products is how do customers curate and create these data products and share them within their organization. And so we're investing heavily in our product to make that easier as well, because I think back to one of the things, uh, Theresa said, it's, it's really all about, uh, making this practical and finding quick wins that customers can deploy, deploy in their data mess journey, right? >>This quick wins are key. So Theresa, last question to you, where should companies go to get started today? Obviously everybody has gotten, we're still in this work from anywhere environment. Companies have tons of data, tons of sources of data, did it, infrastructure's already in place. How did they go and get started with data? >>I think they should start looking at their data projects and thinking about the best data products. I think just that mindset shift about thinking about who's this for what's the business value. And then underneath that architecture and support comes to bear. And then thinking about who are the products that your product could work better with just like any other practice partnerships, like what we have with AWS, right? Like that's a stronger together sort of thing, >>Right? So there's that kind of that cultural component that really strategic shift in thinking and on the architecture. Awesome guys, thank you so much for joining me on the program, coming back on the cube at re-invent talking about data mesh really help. You can help organizations and industry put that together and what's going on at service. We appreciate your time. Thanks again. All right. For my guests, I'm Lisa Martin, you're watching the cubes coverage of AWS reinvent 2021. The cube is the leader in global live tech coverage. We'll be right back.

Published Date : Nov 30 2021

SUMMARY :

Good to have you back. Well, I get to think about the future of cloud and if you think about clouded powers, I know service has been on the program before, but give me a little bit of an overview of the company, what you guys do. And it's how Facebook and Netflix and Airbnb and, and a number of the internet And that's one of the things we've seen explode during the last 22 months, among many other things is data, So even though their products, you really need to make sure that you're doing the right things, but what's data money. This is more of a, an approach, And so that's the reality of the situation today, and it's first and foremost, Just didn't talk to me about when you're having customer conversations. And I think that, you know, essentially every company in the space is, The, the need is to be able to get it, And so a lot of that is reviewing your existing data projects So what are some w you know, we often talk about outcomes, So just by the able to see the data, see what's happening now, that's great. Just so talk to me about how your customer conversations have changed in the last 22 But I think companies don't have the time to wait for that anymore. Let's talk about a little bit about the go to market strategy. And the cloud really makes this easier to do. That's one of the core principles that Theresa mentioned, you know, that's where I think, I'm curious, like to get Theresa, we'll start with you, your perspectives on how And so data is really the heart of the company. And so how do you, Justin, how does Starburst help organizations transfer that technical And I think, you know, the only constant is change. This is a, is this a big directional can, can hopefully expand the market to thousands of potential customers that can now leverage Talk to me about kind of where you're going in, where services heading in the processing to where the data lives. And, uh, and so we think that's a big deal about, you know, kind of enabling a data mash So Theresa, last question to you, where should companies go to get started today? And then thinking about who are the products that your product could work better with just like any other The cube is the leader in global live tech coverage.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

TheresaPERSON

0.99+

AWSORGANIZATION

0.99+

Teresa TungPERSON

0.99+

Justin BorkmanPERSON

0.99+

Justin BorgmanPERSON

0.99+

TeresaPERSON

0.99+

AmazonORGANIZATION

0.99+

JustinPERSON

0.99+

EuropeLOCATION

0.99+

SwitzerlandLOCATION

0.99+

StarburstORGANIZATION

0.99+

AccentureORGANIZATION

0.99+

SecondQUANTITY

0.99+

thousandsQUANTITY

0.99+

NetflixORGANIZATION

0.99+

FacebookORGANIZATION

0.99+

third oneQUANTITY

0.99+

pandemicEVENT

0.98+

four attributesQUANTITY

0.98+

BothQUANTITY

0.98+

todayDATE

0.98+

24QUANTITY

0.98+

firstQUANTITY

0.98+

AirbnbORGANIZATION

0.98+

over 220 patentsQUANTITY

0.97+

over a hundred guestsQUANTITY

0.97+

2021DATE

0.97+

oneQUANTITY

0.96+

StarbucksORGANIZATION

0.96+

single partnerQUANTITY

0.96+

PrestoORGANIZATION

0.96+

single lineQUANTITY

0.96+

sevenQUANTITY

0.95+

confluentORGANIZATION

0.95+

10%QUANTITY

0.94+

one central placeQUANTITY

0.94+

one thingQUANTITY

0.93+

single toolQUANTITY

0.92+

day twoQUANTITY

0.92+

next decadeDATE

0.92+

single entityQUANTITY

0.92+

star gateTITLE

0.92+

Mongo DBORGANIZATION

0.91+

last 22 monthsDATE

0.91+

two lifeQUANTITY

0.91+

StarburstTITLE

0.88+

last 22 monthsDATE

0.87+

Breaking Analysis: What we hope to learn at Supercloud22


 

>> From theCUBE studios in Palo Alto in Boston bringing you data driven insights from theCUBE and ETR. This is breaking analysis with Dave Vellante. >> The term Supercloud is somewhat new, but the concepts behind it have been bubbling for years, early last decade when NIST put forth a definition of cloud computing it said services had to be accessible over a public network essentially cutting the on-prem crowd out of the cloud conversation. Now a guy named Chuck Hollis, who was a field CTO at EMC at the time and a prolific blogger objected to that criterion and laid out his vision for what he termed a private cloud. Now, in that post, he showed a workload running both on premises and in a public cloud sharing the underlying resources in an automated and seamless manner. What later became known more broadly as hybrid cloud that vision as we now know, really never materialized, and we were left with multi-cloud sets of largely incompatible and disconnected cloud services running in separate silos. The point is what Hollis laid out, IE the ability to abstract underlying infrastructure complexity and run workloads across multiple heterogeneous estates with an identical experience is what super cloud is all about. Hello and welcome to this week's Wikibon cube insights powered by ETR and this breaking analysis. We share what we hope to learn from super cloud 22 next week, next Tuesday at 9:00 AM Pacific. The community is gathering for Supercloud 22 an inclusive pilot symposium hosted by theCUBE and made possible by VMware and other founding partners. It's a one day single track event with more than 25 speakers digging into the architectural, the technical, structural and business aspects of Supercloud. This is a hybrid event with a live program in the morning running out of our Palo Alto studio and pre-recorded content in the afternoon featuring industry leaders, technologists, analysts and investors up and down the technology stack. Now, as I said up front the seeds of super cloud were sewn early last decade. After the very first reinvent we published our Amazon gorilla post, that scene in the upper right corner here. And we talked about how to differentiate from Amazon and form ecosystems around industries and data and how the cloud would change IT permanently. And then up in the upper left we put up a post on the old Wikibon Wiki. Yeah, it used to be a Wiki. Check out my hair by the way way no gray, that's how long ago this was. And we talked about in that post how to compete in the Amazon economy. And we showed a graph of how IT economics were changing. And cloud services had marginal economics that looked more like software than hardware at scale. And this would reset, we said opportunities for both technology sellers and buyers for the next 20 years. And this came into sharper focus in the ensuing years culminating in a milestone post by Greylock's Jerry Chen called Castles in the Cloud. It was an inspiration and catalyst for us using the term Supercloud in John Furrier's post prior to reinvent 2021. So we started to flesh out this idea of Supercloud where companies of all types build services on top of hyperscale infrastructure and across multiple clouds, going beyond multicloud 1.0, if you will, which was really a symptom, as we said, many times of multi-vendor at least that's what we argued. And despite its fuzzy definition, it resonated with people because they knew something was brewing, Keith Townsend the CTO advisor, even though he frankly, wasn't a big fan of the buzzy nature of the term Supercloud posted this awesome Blackboard on Twitter take a listen to how he framed it. Please play the clip. >> Is VMware the right company to make the super cloud work, term that Wikibon came up with to describe the taking of discreet services. So it says RDS from AWS, cloud compute engines from GCP and authentication from Azure to build SaaS applications or enterprise applications that connect back to your data center, is VMware's cross cloud vision 'cause it is just a vision today, the right approach. Or should you be looking towards companies like HashiCorp to provide this overall capability that we all agree, or maybe you don't that we need in an enterprise comment below your thoughts. >> So I really like that Keith has deep practitioner knowledge and lays out a couple of options. I especially like the examples he uses of cloud services. He recognizes the need for cross cloud services and he notes this capability is aspirational today. Remember this was eight or nine months ago and he brings HashiCorp into the conversation as they're one of the speakers at Supercloud 22 and he asks the community, what they think, the thing is we're trying to really test out this concept and people like Keith are instrumental as collaborators. Now I'm sure you're not surprised to hear that mot everyone is on board with the Supercloud meme, in particular Charles Fitzgerald has been a wonderful collaborator just by his hilarious criticisms of the concept. After a couple of super cloud posts, Charles put up his second rendition of "Supercloudifragilisticexpialidoucious". I mean, it's just beautiful, but to boot, he put up this picture of Baghdad Bob asking us to just stop, Bob's real name is Mohamed Said al-Sahaf. He was the minister of propaganda for Sadam Husein during the 2003 invasion of Iraq. And he made these outrageous claims of, you know US troops running in fear and putting down their arms and so forth. So anyway, Charles laid out several frankly very helpful critiques of Supercloud which has led us to really advance the definition and catalyze the community's thinking on the topic. Now, one of his issues and there are many is we said a prerequisite of super cloud was a super PaaS layer. Gartner's Lydia Leong chimed in saying there were many examples of successful PaaS vendors built on top of a hyperscaler some having the option to run in more than one cloud provider. But the key point we're trying to explore is the degree to which that PaaS layer is purpose built for a specific super cloud function. And not only runs in more than one cloud provider, Lydia but runs across multiple clouds simultaneously creating an identical developer experience irrespective of a state. Now, maybe that's what Lydia meant. It's hard to say from just a tweet and she's a sharp lady, so, and knows more about that market, that PaaS market, than I do. But to the former point at Supercloud 22, we have several examples. We're going to test. One is Oracle and Microsoft's recent announcement to run database services on OCI and Azure, making them appear as one rather than use an off the shelf platform. Oracle claims to have developed a capability for developers specifically built to ensure high performance low latency, and a common experience for developers across clouds. Another example we're going to test is Snowflake. I'll be interviewing Benoit Dageville co-founder of Snowflake to understand the degree to which Snowflake's recent announcement of an application development platform is perfect built, purpose built for the Snowflake data cloud. Is it just a plain old pass, big whoop as Lydia claims or is it something new and innovative, by the way we invited Charles Fitz to participate in Supercloud 22 and he decline saying in addition to a few other somewhat insulting things there's definitely interesting new stuff brewing that isn't traditional cloud or SaaS but branding at all super cloud doesn't help either. Well, indeed, we agree with part of that and we'll see if it helps advanced thinking and helps customers really plan for the future. And that's why Supercloud 22 has going to feature some of the best analysts in the business in The Great Supercloud Debate. In addition to Keith Townsend and Maribel Lopez of Lopez research and Sanjeev Mohan from former Gartner analyst and principal at SanjMo participated in this session. Now we don't want to mislead you. We don't want to imply that these analysts are hopping on the super cloud bandwagon but they're more than willing to go through the thought experiment and mental exercise. And, we had a great conversation that you don't want to miss. Maribel Lopez had what I thought was a really excellent way to think about this. She used TCP/IP as an historical example, listen to what she said. >> And Sanjeev Mohan has some excellent thoughts on the feasibility of an open versus de facto standard getting us to the vision of Supercloud, what's possible and what's likely now, again, I don't want to imply that these analysts are out banging the Supercloud drum. They're not necessarily doing that, but they do I think it's fair to say believe that something new is bubbling and whether it's called Supercloud or multicloud 2.0 or cross cloud services or whatever name you choose it's not multicloud of the 2010s and we chose Supercloud. So our goal here is to advance the discussion on what's next in cloud and Supercloud is meant to be a term to describe that future of cloud and specifically the cloud opportunities that can be built on top of hyperscale, compute, storage, networking machine learning, and other services at scale. And that is why we posted this piece on Answering the top 10 questions about Supercloud. Many of which were floated by Charles Fitzgerald and others in the community. Why does the industry need another term what's really new and different? And what is hype? What specific problems does Supercloud solve? What are the salient characteristics of Supercloud? What's different beyond multicloud? What is a super pass? Is it necessary to have a Supercloud? How will applications evolve on superclouds? What workloads will run? All these questions will be addressed in detail as a way to advance the discussion and help practitioners and business people understand what's real today. And what's possible with cloud in the near future. And one other question we'll address is who will build super clouds? And what new entrance we can expect. This is an ETR graphic that we showed in a previous episode of breaking analysis, and it lays out some of the companies we think are building super clouds or in a position to do so, by the way the Y axis shows net score or spending velocity and the X axis depicts presence in the ETR survey of more than 1200 respondents. But the key callouts to this slide in addition to some of the smaller firms that aren't yet showing up in the ETR data like Chaossearch and Starburst and Aviatrix and Clumio but the really interesting additions are industry players Walmart with Azure, Capital one and Goldman Sachs with AWS, Oracle, with Cerner. These we think are early examples, bubbling up of industry clouds that will eventually become super clouds. So we'll explore these and other trends to get the community's input on how this will all play out. These are the things we hope you'll take away from Supercloud 22. And we have an amazing lineup of experts to answer your question. Technologists like Kit Colbert, Adrian Cockcroft, Mariana Tessel, Chris Hoff, Will DeForest, Ali Ghodsi, Benoit Dageville, Muddu Sudhakar and many other tech athletes, investors like Jerry Chen and In Sik Rhee the analyst we featured earlier, Paula Hansen talking about go to market in a multi-cloud world Gee Rittenhouse talking about cloud security, David McJannet, Bhaskar Gorti of Platform9 and many, many more. And of course you, so please go to theCUBE.net and register for Supercloud 22, really lightweight reg. We're not doing this for lead gen. We're doing it for collaboration. If you sign in you can get the chat and ask questions in real time. So don't miss this inaugural event Supercloud 22 on August 9th at 9:00 AM Pacific. We'll see you there. Okay. That's it for today. Thanks for watching. Thank you to Alex Myerson who's on production and manages the podcast. Kristen Martin and Cheryl Knight. They help get the word out on social media and in our newsletters. And Rob Hof is our editor in chief over at SiliconANGLE. Does some really wonderful editing. Thank you to all. Remember these episodes are all available as podcasts wherever you listen, just search breaking analysis podcast. I publish each week on wikibon.com and Siliconangle.com. And you can email me at David.Vellantesiliconangle.com or DM me at Dvellante, comment on my LinkedIn post. Please do check out ETR.AI for the best survey data in the enterprise tech business. This is Dave Vellante for theCUBE insights powered by ETR. Thanks for watching. And we'll see you next week in Palo Alto at Supercloud 22 or next time on breaking analysis. (calm music)

Published Date : Aug 5 2022

SUMMARY :

This is breaking analysis and buyers for the next 20 years. Is VMware the right company is the degree to which that PaaS layer and specifically the cloud opportunities

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Alex MyersonPERSON

0.99+

Dave VellantePERSON

0.99+

David McJannetPERSON

0.99+

Cheryl KnightPERSON

0.99+

Paula HansenPERSON

0.99+

Jerry ChenPERSON

0.99+

Adrian CockcroftPERSON

0.99+

Maribel LopezPERSON

0.99+

Keith TownsendPERSON

0.99+

Kristen MartinPERSON

0.99+

Chuck HollisPERSON

0.99+

Charles FitzPERSON

0.99+

CharlesPERSON

0.99+

Chris HoffPERSON

0.99+

KeithPERSON

0.99+

Mariana TesselPERSON

0.99+

AWSORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

Ali GhodsiPERSON

0.99+

OracleORGANIZATION

0.99+

Charles FitzgeraldPERSON

0.99+

Mohamed Said al-SahafPERSON

0.99+

Kit ColbertPERSON

0.99+

WalmartORGANIZATION

0.99+

Rob HofPERSON

0.99+

ClumioORGANIZATION

0.99+

Goldman SachsORGANIZATION

0.99+

Gee RittenhousePERSON

0.99+

AviatrixORGANIZATION

0.99+

ChaossearchORGANIZATION

0.99+

Benoit DagevillePERSON

0.99+

AmazonORGANIZATION

0.99+

Palo AltoLOCATION

0.99+

NISTORGANIZATION

0.99+

Lydia LeongPERSON

0.99+

Muddu SudhakarPERSON

0.99+

BobPERSON

0.99+

CernerORGANIZATION

0.99+

John FurrierPERSON

0.99+

Sanjeev MohanPERSON

0.99+

Capital oneORGANIZATION

0.99+

David.Vellantesiliconangle.comOTHER

0.99+

StarburstORGANIZATION

0.99+

EMCORGANIZATION

0.99+

2010sDATE

0.99+

Will DeForestPERSON

0.99+

more than 1200 respondentsQUANTITY

0.99+

one dayQUANTITY

0.99+

VMwareORGANIZATION

0.99+

GartnerORGANIZATION

0.99+

2021DATE

0.99+

next weekDATE

0.99+

Supercloud 22EVENT

0.99+

theCUBE.netOTHER

0.99+

Bhaskar GortiPERSON

0.99+

SupercloudORGANIZATION

0.98+

each weekQUANTITY

0.98+

eightDATE

0.98+

SanjMoORGANIZATION

0.98+

LydiaPERSON

0.98+

theCUBEORGANIZATION

0.98+

PaaSTITLE

0.98+

more than 25 speakersQUANTITY

0.98+

SnowflakeORGANIZATION

0.98+

Platform9ORGANIZATION

0.97+

firstQUANTITY

0.97+

oneQUANTITY

0.97+

todayDATE

0.97+

HollisPERSON

0.97+

Sadam HuseinPERSON

0.97+

second renditionQUANTITY

0.97+

BostonLOCATION

0.97+

SiliconANGLEORGANIZATION

0.96+

more than one cloud providerQUANTITY

0.96+

bothQUANTITY

0.95+

super cloud 22EVENT

0.95+

Breaking Analysis: Answering the top 10 questions about SuperCloud


 

>> From the theCUBE studios in Palo Alto in Boston, bringing you data driven insights from theCUBE and ETR. This is "Breaking Analysis" with Dave Vellante. >> Welcome to this week's Wikibon, theCUBE's insights powered by ETR. As we exited the isolation economy last year, supercloud is a term that we introduced to describe something new that was happening in the world of cloud. In this Breaking Analysis, we address the 10 most frequently asked questions we get around supercloud. Okay, let's review these frequently asked questions on supercloud that we're going to try to answer today. Look at an industry that's full of hype and buzzwords. Why the hell does anyone need a new term? Aren't hyperscalers building out superclouds? We'll try to answer why the term supercloud connotes something different from hyperscale clouds. And we'll talk about the problems that superclouds solve specifically. And we'll further define the critical aspects of a supercloud architecture. We often get asked, isn't this just multi-cloud? Well, we don't think so, and we'll explain why in this Breaking Analysis. Now in an earlier episode, we introduced the notion of super PaaS. Well, isn't a plain vanilla PaaS already a super PaaS? Again, we don't think so, and we'll explain why. Who will actually build and who are the players currently building superclouds? What workloads and services will run on superclouds? And 8-A or number nine, what are some examples that we can share of supercloud? And finally, we'll answer what you can expect next from us on supercloud? Okay, let's get started. Why do we need another buzzword? Well, late last year, ahead of re:Invent, we were inspired by a post from Jerry Chen called "Castles in the Cloud." Now in that blog post, he introduced the idea that there were sub-markets emerging in cloud that presented opportunities for investors and entrepreneurs that the cloud wasn't going to suck the hyperscalers. Weren't going to suck all the value out of the industry. And so we introduced this notion of supercloud to describe what we saw as a value layer emerging above the hyperscalers CAPEX gift, we sometimes call it. Now it turns out, that we weren't the only ones using the term as both Cornell and MIT have used the phrase in somewhat similar, but different contexts. The point is something new was happening in the AWS and other ecosystems. It was more than IaaS and PaaS, and wasn't just SaaS running in the cloud. It was a new architecture that integrates infrastructure, platform and software as services to solve new problems that the cloud vendors in our view, weren't addressing by themselves. It seemed to us that the ecosystem was pursuing opportunities across clouds that went beyond conventional implementations of multi-cloud. And we felt there was a structural change going on at the industry level, the supercloud, metaphorically was highlighting. So that's the background on why we felt a new catch phrase was warranted, love it or hate it. It's memorable and it's what we chose. Now to that last point about structural industry transformation. Andy Rappaport is sometimes and often credited with identifying the shift from the vertically integrated IBM mainframe era to the fragmented PC microprocesor-based era in his HBR article in 1991. In fact, it was David Moschella, who at the time was an IDC Analyst who first introduced the concept in 1987, four years before Rappaport's article was published. Moschella saw that it was clear that Intel, Microsoft, Seagate and others would replace the system vendors, and put that forth in a graphic that looked similar to the first two on this chart. We don't have to review the shift from IBM as the center of the industry to Wintel, that's well understood. What isn't as well known or accepted is what Moschella put out in his 2018 book called "Seeing Digital" which introduced the idea of "The Matrix" that's shown on the right hand side of this chart. Moschella posited that new services were emerging built on top of the internet and hyperscale clouds that would integrate other innovations and would define the next era of computing. He used the term Matrix because the conceptual depiction included not only horizontal technology rose like the cloud and the internet, but for the first time included connected industry verticals, the columns in this chart. Moschella pointed out that whereas historically, industry verticals had a closed value chain or stack and ecosystem of R&D, and production, and manufacturing, and distribution. And if you were in that industry, the expertise within that vertical generally stayed within that vertical and was critical to success. But because of digital and data, for the first time, companies were able to traverse industries, jump across industries and compete because data enabled them to do that. Examples, Amazon and content, payments, groceries, Apple, and payments, and content, and so forth. There are many examples. Data was now this unifying enabler and this marked a change in the structure of the technology landscape. And supercloud is meant to imply more than running in hyperscale clouds, rather it's the combination of multiple technologies enabled by CloudScale with new industry participants from those verticals, financial services and healthcare, manufacturing, energy, media, and virtually all in any industry. Kind of an extension of every company is a software company. Basically, every company now has the opportunity to build their own cloud or supercloud. And we'll come back to that. Let's first address what's different about superclouds relative to hyperscale clouds? You know, this one's pretty straightforward and obvious, I think. Hyperscale clouds, they're walled gardens where they want your data in their cloud and they want to keep you there. Sure, every cloud player realizes that not all data will go to their particular cloud so they're meeting customers where their data lives with initiatives like Amazon Outposts and Azure Arc, and Google Anthos. But at the end of the day, the more homogeneous they can make their environments, the better control, security, cost, and performance they can deliver. The more complex the environment, the more difficult it is to deliver on their brand promises. And of course, the lesser margin that's left for them to capture. Will the hyperscalers get more serious about cross-cloud services? Maybe, but they have plenty of work to do within their own clouds and within enabling their own ecosystems. They had a long way to go a lot of runway. So let's talk about specifically, what problems superclouds solve? We've all seen the stats from IDC or Gartner, or whomever the customers on average use more than one cloud. You know, two clouds, three clouds, five clouds, 20 clouds. And we know these clouds operate in disconnected silos for the most part. And that's a problem because each cloud requires different skills because the development environment is different as is the operating environment. They have different APIs, different primitives, and different management tools that are optimized for each respective hyperscale cloud. Their functions and value props don't extend to their competitors' clouds for the most part. Why would they? As a result, there's friction when moving between different clouds. It's hard to share data, it's hard to move work. It's hard to secure and govern data. It's hard to enforce organizational edicts and policies across these clouds, and on-prem. Supercloud is an architecture designed to create a single environment that enables management of workloads and data across clouds in an effort to take out complexity, accelerate application development, streamline operations and share data safely, irrespective of location. It's pretty straightforward, but non-trivial, which is why I always ask a company's CEO and executives if stock buybacks and dividends will yield as much return as building out superclouds that solve really specific and hard problems, and create differential value. Okay, let's dig a bit more into the architectural aspects of supercloud. In other words, what are the salient attributes of supercloud? So first and foremost, a supercloud runs a set of specific services designed to solve a unique problem and it can do so in more than one cloud. Superclouds leverage the underlying cloud native tooling of a hyperscale cloud, but they're optimized for a specific objective that aligns with the problem that they're trying to solve. For example, supercloud might be optimized for lowest cost or lowest latency, or sharing data, or governing, or securing that data, or higher performance for networking, for example. But the point is, the collection of services that is being delivered is focused on a unique value proposition that is not being delivered by the hyperscalers across clouds. A supercloud abstracts the underlying and siloed primitives of the native PaaS layer from the hyperscale cloud and then using its own specific platform as a service tooling, creates a common experience across clouds for developers and users. And it does so in a most efficient manner, meaning it has the metadata knowledge and management capabilities that can optimize for latency, bandwidth, or recovery, or data sovereignty, or whatever unique value that supercloud is delivering for the specific use case in their domain. And a supercloud comprises a super PaaS capability that allows ecosystem partners through APIs to add incremental value on top of the supercloud platform to fill gaps, accelerate features, and of course innovate. The services can be infrastructure-related, they could be application services, they could be data services, security services, user services, et cetera, designed and packaged to bring unique value to customers. Again, that hyperscalers are not delivering across clouds or on-premises. Okay, so another common question we get is, isn't that just multi-cloud? And what we'd say to that is yes, but no. You can call it multi-cloud 2.0, if you want, if you want to use it, it's kind of a commonly used rubric. But as Dell's Chuck Whitten proclaimed at Dell Technologies World this year, multi-cloud by design, is different than multi-cloud by default. Meaning to date, multi-cloud has largely been a symptom of what we've called multi-vendor or of M&A, you buy a company and they happen to use Google Cloud, and so you bring it in. And when you look at most so-called, multi-cloud implementations, you see things like an on-prem stack, which is wrapped in a container and hosted on a specific cloud or increasingly a technology vendor has done the work of building a cloud native version of their stack and running it on a specific cloud. But historically, it's been a unique experience within each cloud with virtually no connection between the cloud silos. Supercloud sets out to build incremental value across clouds and above hyperscale CAPEX that goes beyond cloud compatibility within each cloud. So if you want to call it multi-cloud 2.0, that's fine, but we chose to call it supercloud. Okay, so at this point you may be asking, well isn't PaaS already a version of supercloud? And again, we would say no, that supercloud and its corresponding superPaaS layer which is a prerequisite, gives the freedom to store, process and manage, and secure, and connect islands of data across a continuum with a common experience across clouds. And the services offered are specific to that supercloud and will vary by each offering. Your OpenShift, for example, can be used to construct a superPaaS, but in and of itself, isn't a superPaaS, it's generic. A superPaaS might be developed to support, for instance, ultra low latency database work. It would unlikely again, taking the OpenShift example, it's unlikely that off-the-shelf OpenShift would be used to develop such a low latency superPaaS layer for ultra low latency database work. The point is supercloud and its inherent superPaaS will be optimized to solve specific problems like that low latency example for distributed databases or fast backup and recovery for data protection, and ransomware, or data sharing, or data governance. Highly specific use cases that the supercloud is designed to solve for. Okay, another question we often get is who has a supercloud today and who's building a supercloud, and who are the contenders? Well, most companies that consider themselves cloud players will, we believe, be building or are building superclouds. Here's a common ETR graphic that we like to show with Net Score or spending momentum on the Y axis and overlap or pervasiveness in the ETR surveys on the X axis. And we've randomly chosen a number of players that we think are in the supercloud mix, and we've included the hyperscalers because they are enablers. Now remember, this is a spectrum of maturity it's a maturity model and we've added some of those industry players that we see building superclouds like CapitalOne, Goldman Sachs, Walmart. This is in deference to Moschella's observation around The Matrix and the industry structural changes that are going on. This goes back to every company, being a software company and rather than pattern match an outdated SaaS model, we see new industry structures emerging where software and data, and tools, specific to an industry will lead the next wave of innovation and bring in new value that traditional technology companies aren't going to solve, and the hyperscalers aren't going to solve. You know, we've talked a lot about Snowflake's data cloud as an example of supercloud. After being at Snowflake Summit, we're more convinced than ever that they're headed in this direction. VMware is clearly going after cross-cloud services you know, perhaps creating a new category. Basically, every large company we see either pursuing supercloud initiatives or thinking about it. Dell showed project Alpine at Dell Tech World, that's a supercloud. Snowflake introducing a new application development capability based on their superPaaS, our term of course, they don't use the phrase. Mongo, Couchbase, Nutanix, Pure Storage, Veeam, CrowdStrike, Okta, Zscaler. Yeah, all of those guys. Yes, Cisco and HPE. Even though on theCUBE at HPE Discover, Fidelma Russo said on theCUBE, she wasn't a fan of cloaking mechanisms, but then we talked to HPE's Head of Storage Services, Omer Asad is clearly headed in the direction that we would consider supercloud. Again, those cross-cloud services, of course, their emphasis is connecting as well on-prem. That single experience, which traditionally has not existed with multi-cloud or hybrid. And we're seeing the emergence of companies, smaller companies like Aviatrix and Starburst, and Clumio and others that are building versions of superclouds that solve for a specific problem for their customers. Even ISVs like Adobe, ADP, we've talked to UiPath. They seem to be looking at new ways to go beyond the SaaS model and add value within their cloud ecosystem specifically, around data as part of their and their customers digital transformations. So yeah, pretty much every tech vendor with any size or momentum and new industry players are coming out of hiding, and competing. Building superclouds that look a lot like Moschella's Matrix, with machine intelligence and blockchains, and virtual realities, and gaming, all enabled by the internet and hyperscale cloud CAPEX. So it's moving fast and it's the future in our opinion. So don't get too caught up in the past or you'll be left behind. Okay, what about examples? We've given a number in the past, but let's try to be a little bit more specific. Here are a few we've selected and we're going to answer the two questions in one section here. What workloads and services will run in superclouds and what are some examples? Let's start with analytics. Our favorite example is Snowflake, it's one of the furthest along with its data cloud, in our view. It's a supercloud optimized for data sharing and governance, query performance, and security, and ecosystem enablement. When you do things inside of that data cloud, what we call a super data cloud. Again, our term, not theirs. You can do things that you could not do in a single cloud. You can't do this with Redshift, You can't do this with SQL server and they're bringing new data types now with merging analytics or at least accommodate analytics and transaction type data, and bringing open source tooling with things like Apache Iceberg. And so it ticks the boxes we laid out earlier. I would say that a company like Databricks is also in that mix doing it, coming at it from a data science perspective, trying to create that consistent experience for data scientists and data engineering across clouds. Converge databases, running transaction and analytic workloads is another example. Take a look at what Couchbase is doing with Capella and how it's enabling stretching the cloud to the edge with ARM-based platforms and optimizing for low latency across clouds, and even out to the edge. Document database workloads, look at MongoDB, a very developer-friendly platform that with the Atlas is moving toward a supercloud model running document databases very, very efficiently. How about general purpose workloads? This is where VMware comes into to play. Very clearly, there's a need to create a common operating environment across clouds and on-prem, and out to the edge. And I say VMware is hard at work on that. Managing and moving workloads, and balancing workloads, and being able to recover very quickly across clouds for everyday applications. Network routing, take a look at what Aviatrix is doing across clouds, industry workloads. We see CapitalOne, it announced its cost optimization platform for Snowflake, piggybacking on Snowflake supercloud or super data cloud. And in our view, it's very clearly going to go after other markets is going to test it out with Snowflake, running, optimizing on AWS and it's going to expand to other clouds as Snowflake's business and those other clouds grows. Walmart working with Microsoft to create an on-premed Azure experience that's seamless. Yes, that counts, on-prem counts. If you can create that seamless and continuous experience, identical experience from on-prem to a hyperscale cloud, we would include that as a supercloud. You know, we've written about what Goldman is doing. Again, connecting its on-prem data and software tooling, and other capabilities to AWS for scale. And we can bet dollars to donuts that Oracle will be building a supercloud in healthcare with its Cerner acquisition. Supercloud is everywhere you look. So I'm sorry, naysayers it's happening all around us. So what's next? Well, with all the industry buzz and debate about the future, John Furrier and I, have decided to host an event in Palo Alto, we're motivated and inspired to further this conversation. And we welcome all points of view, positive, negative, multi-cloud, supercloud, hypercloud, all welcome. So theCUBE on Supercloud is coming on August 9th, out of our Palo Alto studios, we'll be running a live program on the topic. We've reached out to a number of industry participants, VMware, Snowflake, Confluent, Sky High Security, Gee Rittenhouse's new company, HashiCorp, CloudFlare. We've hit up Red Hat and we expect many of these folks will be in our studios on August 9th. And we've invited a number of industry participants as well that we're excited to have on. From industry, from financial services, from healthcare, from retail, we're inviting analysts, thought leaders, investors. We're going to have more detail in the coming weeks, but for now, if you're interested, please reach out to me or John with how you think you can advance the discussion and we'll see if we can fit you in. So mark your calendars, stay tuned for more information. Okay, that's it for today. Thanks to Alex Myerson who handles production and manages the podcast for Breaking Analysis. And I want to thank Kristen Martin and Cheryl Knight, they help get the word out on social and in our newsletters. And Rob Hof is our editor in chief over at SiliconANGLE, who does a lot of editing and appreciate you posting on SiliconANGLE, Rob. Thanks to all of you. Remember, all these episodes are available as podcasts wherever you listen. All you got to do is search Breaking Analysis podcast. It publish each week on wikibon.com and siliconangle.com. You can email me directly at david.vellante@siliconangle.com or DM me @DVellante, or comment on my LinkedIn post. And please do check out ETR.ai for the best survey data. And the enterprise tech business will be at AWS NYC Summit next Tuesday, July 12th. So if you're there, please do stop by and say hello to theCUBE, it's at the Javits Center. This is Dave Vellante for theCUBE insights powered by ETR. Thanks for watching. And we'll see you next time on "Breaking Analysis." (bright music)

Published Date : Jul 9 2022

SUMMARY :

From the theCUBE studios and how it's enabling stretching the cloud

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Alex MyersonPERSON

0.99+

SeagateORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

Dave VellantePERSON

0.99+

1987DATE

0.99+

Andy RappaportPERSON

0.99+

David MoschellaPERSON

0.99+

WalmartORGANIZATION

0.99+

Jerry ChenPERSON

0.99+

IntelORGANIZATION

0.99+

Chuck WhittenPERSON

0.99+

Cheryl KnightPERSON

0.99+

Rob HofPERSON

0.99+

1991DATE

0.99+

August 9thDATE

0.99+

AmazonORGANIZATION

0.99+

HPEORGANIZATION

0.99+

Palo AltoLOCATION

0.99+

JohnPERSON

0.99+

MoschellaPERSON

0.99+

OracleORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

IBMORGANIZATION

0.99+

20 cloudsQUANTITY

0.99+

StarburstORGANIZATION

0.99+

Goldman SachsORGANIZATION

0.99+

DellORGANIZATION

0.99+

Fidelma RussoPERSON

0.99+

2018DATE

0.99+

two questionsQUANTITY

0.99+

AppleORGANIZATION

0.99+

AWSORGANIZATION

0.99+

AviatrixORGANIZATION

0.99+

Omer AsadPERSON

0.99+

Sky High SecurityORGANIZATION

0.99+

DatabricksORGANIZATION

0.99+

ConfluentORGANIZATION

0.99+

WintelORGANIZATION

0.99+

NutanixORGANIZATION

0.99+

CapitalOneORGANIZATION

0.99+

CouchbaseORGANIZATION

0.99+

HashiCorpORGANIZATION

0.99+

five cloudsQUANTITY

0.99+

Kristen MartinPERSON

0.99+

last yearDATE

0.99+

david.vellante@siliconangle.comOTHER

0.99+

two cloudsQUANTITY

0.99+

RobPERSON

0.99+

SnowflakeORGANIZATION

0.99+

MongoORGANIZATION

0.99+

Pure StorageORGANIZATION

0.99+

each cloudQUANTITY

0.99+

VeeamORGANIZATION

0.99+

John FurrierPERSON

0.99+

GartnerORGANIZATION

0.99+

VMwareORGANIZATION

0.99+

first twoQUANTITY

0.99+

ClumioORGANIZATION

0.99+

CrowdStrikeORGANIZATION

0.99+

OktaORGANIZATION

0.99+

three cloudsQUANTITY

0.99+

MITORGANIZATION

0.99+

Javits CenterLOCATION

0.99+

first timeQUANTITY

0.99+

ZscalerORGANIZATION

0.99+

RappaportPERSON

0.99+

MoschellaORGANIZATION

0.99+

each weekQUANTITY

0.99+

late last yearDATE

0.99+

UiPathORGANIZATION

0.99+

10 most frequently asked questionsQUANTITY

0.99+

CloudFlareORGANIZATION

0.99+

IDCORGANIZATION

0.99+

one sectionQUANTITY

0.99+

SiliconANGLEORGANIZATION

0.98+

Seeing DigitalTITLE

0.98+

eachQUANTITY

0.98+

firstQUANTITY

0.98+

bothQUANTITY

0.98+

AdobeORGANIZATION

0.98+

more than one cloudQUANTITY

0.98+

each offeringQUANTITY

0.98+

Breaking Analysis: Answering the top 10 questions about supercloud


 

>> From theCUBE Studios in Palo Alto and Boston, bringing you data-driven insights from theCUBE and ETR. This is "Breaking Analysis" with Dave Vallante. >> Welcome to this week's Wikibon CUBE Insights powered by ETR. As we exited the isolation economy last year, Supercloud is a term that we introduced to describe something new that was happening in the world of cloud. In this "Breaking Analysis," we address the 10 most frequently asked questions we get around Supercloud. Okay, let's review these frequently asked questions on Supercloud that we're going to try to answer today. Look at an industry that's full of hype and buzzwords. Why the hell does anyone need a new term? Aren't hyperscalers building out Superclouds? We'll try to answer why the term Supercloud connotes something different from hyperscale clouds. And we'll talk about the problems that Superclouds solve specifically, and we'll further define the critical aspects of a Supercloud architecture. We often get asked, "Isn't this just multi-cloud?" Well, we don't think so, and we'll explain why in this "Breaking Analysis." Now, in an earlier episode, we introduced the notion of super PaaS. Well, isn't a plain vanilla PaaS already a super PaaS? Again, we don't think so, and we'll explain why. Who will actually build and who are the players currently building Superclouds? What workloads and services will run on Superclouds? And eight A or number nine, what are some examples that we can share of Supercloud? And finally, we'll answer what you can expect next from us on Supercloud. Okay, let's get started. Why do we need another buzzword? Well, late last year ahead of re:Invent, we were inspired by a post from Jerry Chen called castles in the cloud. Now, in that blog post, he introduced the idea that there were submarkets emerging in cloud that presented opportunities for investors and entrepreneurs. That the cloud wasn't going to suck the hyperscalers, weren't going to suck all the value out of the industry. And so we introduced this notion of Supercloud to describe what we saw as a value layer emerging above the hyperscalers CAPEX gift, we sometimes call it. Now, it turns out that we weren't the only ones using the term, as both Cornell and MIT, have used the phrase in somewhat similar, but different contexts. The point is, something new was happening in the AWS and other ecosystems. It was more than IS and PaaS, and wasn't just SaaS running in the cloud. It was a new architecture that integrates infrastructure, platform and software as services, to solve new problems that the cloud vendors, in our view, weren't addressing by themselves. It seemed to us that the ecosystem was pursuing opportunities across clouds that went beyond conventional implementations of multi-cloud. And we felt there was a structural change going on at the industry level. The Supercloud metaphorically was highlighting. So that's the background on why we felt a new catch phrase was warranted. Love it or hate it, it's memorable and it's what we chose. Now, to that last point about structural industry transformation. Andy Rapaport is sometimes and often credited with identifying the shift from the vertically integrated IBM mainframe era to the fragmented PC microprocesor based era in his HBR article in 1991. In fact, it was David Moschella, who at the time was an IDC analyst who first introduced the concept in 1987, four years before Rapaport's article was published. Moschella saw that it was clear that Intel, Microsoft, Seagate and others would replace the system vendors and put that forth in a graphic that looked similar to the first two on this chart. We don't have to review the shift from IBM as the center of the industry to Wintel. That's well understood. What isn't as well known or accepted is what Moschella put out in his 2018 book called "Seeing Digital" which introduced the idea of the matrix that's shown on the right hand side of this chart. Moschella posited that new services were emerging, built on top of the internet and hyperscale clouds that would integrate other innovations and would define the next era of computing. He used the term matrix, because the conceptual depiction included, not only horizontal technology rows, like the cloud and the internet, but for the first time included connected industry verticals, the columns in this chart. Moschella pointed out that, whereas historically, industry verticals had a closed value chain or stack and ecosystem of R&D and production and manufacturing and distribution. And if you were in that industry, the expertise within that vertical generally stayed within that vertical and was critical to success. But because of digital and data, for the first time, companies were able to traverse industries jump across industries and compete because data enabled them to do that. Examples, Amazon and content, payments, groceries, Apple and payments, and content and so forth. There are many examples. Data was now this unifying enabler and this marked a change in the structure of the technology landscape. And Supercloud is meant to imply more than running in hyperscale clouds. Rather, it's the combination of multiple technologies, enabled by cloud scale with new industry participants from those verticals; financial services, and healthcare, and manufacturing, energy, media, and virtually all and any industry. Kind of an extension of every company is a software company. Basically, every company now has the opportunity to build their own cloud or Supercloud. And we'll come back to that. Let's first address what's different about Superclouds relative to hyperscale clouds. Now, this one's pretty straightforward and obvious, I think. Hyperscale clouds, they're walled gardens where they want your data in their cloud and they want to keep you there. Sure, every cloud player realizes that not all data will go to their particular cloud. So they're meeting customers where their data lives with initiatives like Amazon Outposts and Azure Arc and Google Antos. But at the end of the day, the more homogeneous they can make their environments, the better control, security, costs, and performance they can deliver. The more complex the environment, the more difficult it is to deliver on their brand promises. And, of course, the less margin that's left for them to capture. Will the hyperscalers get more serious about cross cloud services? Maybe, but they have plenty of work to do within their own clouds and within enabling their own ecosystems. They have a long way to go, a lot of runway. So let's talk about specifically, what problems Superclouds solve. We've all seen the stats from IDC or Gartner or whomever, that customers on average use more than one cloud, two clouds, three clouds, five clouds, 20 clouds. And we know these clouds operate in disconnected silos for the most part. And that's a problem, because each cloud requires different skills, because the development environment is different as is the operating environment. They have different APIs, different primitives, and different management tools that are optimized for each respective hyperscale cloud. Their functions and value props don't extend to their competitors' clouds for the most part. Why would they? As a result, there's friction when moving between different clouds. It's hard to share data. It's hard to move work. It's hard to secure and govern data. It's hard to enforce organizational edicts and policies across these clouds and on-prem. Supercloud is an architecture designed to create a single environment that enables management of workloads and data across clouds in an effort to take out complexity, accelerate application development, streamline operations, and share data safely, irrespective of location. It's pretty straightforward, but non-trivial, which is why I always ask a company's CEO and executives if stock buybacks and dividends will yield as much return as building out Superclouds that solve really specific and hard problems and create differential value. Okay, let's dig a bit more into the architectural aspects of Supercloud. In other words, what are the salient attributes of Supercloud? So, first and foremost, a Supercloud runs a set of specific services designed to solve a unique problem, and it can do so in more than one cloud. Superclouds leverage the underlying cloud native tooling of a hyperscale cloud, but they're optimized for a specific objective that aligns with the problem that they're trying to solve. For example, Supercloud might be optimized for lowest cost or lowest latency or sharing data or governing or securing that data or higher performance for networking, for example. But the point is, the collection of services that is being delivered is focused on a unique value proposition that is not being delivered by the hyperscalers across clouds. A Supercloud abstracts the underlying and siloed primitives of the native PaaS layer from the hyperscale cloud, and then using its own specific platform as a service tooling, creates a common experience across clouds for developers and users. And it does so in the most efficient manner, meaning it has the metadata knowledge and management capabilities that can optimize for latency, bandwidth, or recovery or data sovereignty, or whatever unique value that Supercloud is delivering for the specific use case in their domain. And a Supercloud comprises a super PaaS capability that allows ecosystem partners through APIs to add incremental value on top of the Supercloud platform to fill gaps, accelerate features, and of course, innovate. The services can be infrastructure related, they could be application services, they could be data services, security services, user services, et cetera, designed and packaged to bring unique value to customers. Again, that hyperscalers are not delivering across clouds or on premises. Okay, so another common question we get is, "Isn't that just multi-cloud?" And what we'd say to that is yeah, "Yes, but no." You can call it multi-cloud 2.0, if you want. If you want to use, it's kind of a commonly used rubric. But as Dell's Chuck Whitten proclaimed at Dell Technologies World this year, multi-cloud, by design, is different than multi-cloud by default. Meaning, to date, multi-cloud has largely been a symptom of what we've called multi-vendor or of M&A. You buy a company and they happen to use Google cloud. And so you bring it in. And when you look at most so-called multi-cloud implementations, you see things like an on-prem stack, which is wrapped in a container and hosted on a specific cloud. Or increasingly, a technology vendor has done the work of building a cloud native version of their stack and running it on a specific cloud. But historically, it's been a unique experience within each cloud, with virtually no connection between the cloud silos. Supercloud sets out to build incremental value across clouds and above hyperscale CAPEX that goes beyond cloud compatibility within each cloud. So, if you want to call it multi-cloud 2.0, that's fine, but we chose to call it Supercloud. Okay, so at this point you may be asking, "Well isn't PaaS already a version of Supercloud?" And again, we would say, "No." That Supercloud and its corresponding super PaaS layer, which is a prerequisite, gives the freedom to store, process, and manage and secure and connect islands of data across a continuum with a common experience across clouds. And the services offered are specific to that Supercloud and will vary by each offering. OpenShift, for example, can be used to construct a super PaaS, but in and of itself, isn't a super PaaS, it's generic. A super PaaS might be developed to support, for instance, ultra low latency database work. It would unlikely, again, taking the OpenShift example, it's unlikely that off the shelf OpenShift would be used to develop such a low latency, super PaaS layer for ultra low latency database work. The point is, Supercloud and its inherent super PaaS will be optimized to solve specific problems like that low latency example for distributed databases or fast backup in recovery for data protection and ransomware, or data sharing or data governance. Highly specific use cases that the Supercloud is designed to solve for. Okay, another question we often get is, "Who has a Supercloud today and who's building a Supercloud and who are the contenders?" Well, most companies that consider themselves cloud players will, we believe, be building or are building Superclouds. Here's a common ETR graphic that we like to show with net score or spending momentum on the Y axis, and overlap or pervasiveness in the ETR surveys on the X axis. And we've randomly chosen a number of players that we think are in the Supercloud mix. And we've included the hyperscalers because they are enablers. Now, remember, this is a spectrum of maturity. It's a maturity model. And we've added some of those industry players that we see building Superclouds like Capital One, Goldman Sachs, Walmart. This is in deference to Moschella's observation around the matrix and the industry structural changes that are going on. This goes back to every company being a software company. And rather than pattern match and outdated SaaS model, we see new industry structures emerging where software and data and tools specific to an industry will lead the next wave of innovation and bring in new value that traditional technology companies aren't going to solve. And the hyperscalers aren't going to solve. We've talked a lot about Snowflake's data cloud as an example of Supercloud. After being at Snowflake Summit, we're more convinced than ever that they're headed in this direction. VMware is clearly going after cross cloud services, perhaps creating a new category. Basically, every large company we see either pursuing Supercloud initiatives or thinking about it. Dell showed Project Alpine at Dell Tech World. That's a Supercloud. Snowflake introducing a new application development capability based on their super PaaS, our term, of course. They don't use the phrase. Mongo, Couchbase, Nutanix, Pure Storage, Veeam, CrowdStrike, Okta, Zscaler. Yeah, all of those guys. Yes, Cisco and HPE. Even though on theCUBE at HPE Discover, Fidelma Russo said on theCUBE, she wasn't a fan of cloaking mechanisms. (Dave laughing) But then we talked to HPE's head of storage services, Omer Asad, and he's clearly headed in the direction that we would consider Supercloud. Again, those cross cloud services, of course, their emphasis is connecting as well on-prem. That single experience, which traditionally has not existed with multi-cloud or hybrid. And we're seeing the emergence of smaller companies like Aviatrix and Starburst and Clumio and others that are building versions of Superclouds that solve for a specific problem for their customers. Even ISVs like Adobe, ADP, we've talked to UiPath. They seem to be looking at new ways to go beyond the SaaS model and add value within their cloud ecosystem, specifically around data as part of their and their customer's digital transformations. So yeah, pretty much every tech vendor with any size or momentum, and new industry players are coming out of hiding and competing, building Superclouds that look a lot like Moschella's matrix, with machine intelligence and blockchains and virtual realities and gaming, all enabled by the internet and hyperscale cloud CAPEX. So it's moving fast and it's the future in our opinion. So don't get too caught up in the past or you'll be left behind. Okay, what about examples? We've given a number in the past but let's try to be a little bit more specific. Here are a few we've selected and we're going to answer the two questions in one section here. What workloads and services will run in Superclouds and what are some examples? Let's start with analytics. Our favorite example of Snowflake. It's one of the furthest along with its data cloud, in our view. It's a Supercloud optimized for data sharing and governance, and query performance, and security, and ecosystem enablement. When you do things inside of that data cloud, what we call a super data cloud. Again, our term, not theirs. You can do things that you could not do in a single cloud. You can't do this with Redshift. You can't do this with SQL server. And they're bringing new data types now with merging analytics or at least accommodate analytics and transaction type data and bringing open source tooling with things like Apache Iceberg. And so, it ticks the boxes we laid out earlier. I would say that a company like Databricks is also in that mix, doing it, coming at it from a data science perspective trying to create that consistent experience for data scientists and data engineering across clouds. Converge databases, running transaction and analytic workloads is another example. Take a look at what Couchbase is doing with Capella and how it's enabling stretching the cloud to the edge with arm based platforms and optimizing for low latency across clouds, and even out to the edge. Document database workloads, look at Mongo DB. A very developer friendly platform that where the Atlas is moving toward a Supercloud model, running document databases very, very efficiently. How about general purpose workloads? This is where VMware comes into play. Very clearly, there's a need to create a common operating environment across clouds and on-prem and out to the edge. And I say, VMware is hard at work on that, managing and moving workloads and balancing workloads, and being able to recover very quickly across clouds for everyday applications. Network routing, take a look at what Aviatrix is doing across clouds. Industry workloads, we see Capital One. It announced its cost optimization platform for Snowflake, piggybacking on Snowflake's Supercloud or super data cloud. And in our view, it's very clearly going to go after other markets. It's going to test it out with Snowflake, optimizing on AWS, and it's going to expand to other clouds as Snowflake's business and those other clouds grows. Walmart working with Microsoft to create an on-premed Azure experience that's seamless. Yes, that counts, on-prem counts. If you can create that seamless and continuous experience, identical experience from on-prem to a hyperscale cloud, we would include that as a Supercloud. We've written about what Goldman is doing. Again, connecting its on-prem data and software tooling, and other capabilities to AWS for scale. And you can bet dollars to donuts that Oracle will be building a Supercloud in healthcare with its Cerner acquisition. Supercloud is everywhere you look. So I'm sorry, naysayers, it's happening all around us. So what's next? Well, with all the industry buzz and debate about the future, John Furrier and I have decided to host an event in Palo Alto. We're motivated and inspired to further this conversation. And we welcome all points of view, positive, negative, multi-cloud, Supercloud, HyperCloud, all welcome. So theCUBE on Supercloud is coming on August 9th out of our Palo Alto studios. We'll be running a live program on the topic. We've reached out to a number of industry participants; VMware, Snowflake, Confluent, Skyhigh Security, G. Written House's new company, HashiCorp, CloudFlare. We've hit up Red Hat and we expect many of these folks will be in our studios on August 9th. And we've invited a number of industry participants as well that we're excited to have on. From industry, from financial services, from healthcare, from retail, we're inviting analysts, thought leaders, investors. We're going to have more detail in the coming weeks, but for now, if you're interested, please reach out to me or John with how you think you can advance the discussion, and we'll see if we can fit you in. So mark your calendars, stay tuned for more information. Okay, that's it for today. Thanks to Alex Myerson who handles production and manages the podcast for "Breaking Analysis." And I want to thank Kristen Martin and Cheryl Knight. They help get the word out on social and in our newsletters. And Rob Hof is our editor in chief over at SiliconANGLE, who does a lot of editing and appreciate you posting on SiliconANGLE, Rob. Thanks to all of you. Remember, all these episodes are available as podcasts wherever you listen. All you got to do is search, breaking analysis podcast. I publish each week on wikibon.com and siliconangle.com. Or you can email me directly at david.vellante@siliconangle.com. Or DM me @DVallante, or comment on my LinkedIn post. And please, do check out etr.ai for the best survey data in the enterprise tech business. We'll be at AWS NYC summit next Tuesday, July 12th. So if you're there, please do stop by and say hello to theCUBE. It's at the Javits Center. This is Dave Vallante for theCUBE Insights, powered by ETR. Thanks for watching. And we'll see you next time on "Breaking Analysis." (slow music)

Published Date : Jul 8 2022

SUMMARY :

This is "Breaking Analysis" stretching the cloud to the edge

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Alex MyersonPERSON

0.99+

SeagateORGANIZATION

0.99+

1987DATE

0.99+

Dave VallantePERSON

0.99+

MicrosoftORGANIZATION

0.99+

WalmartORGANIZATION

0.99+

1991DATE

0.99+

Andy RapaportPERSON

0.99+

Jerry ChenPERSON

0.99+

MoschellaPERSON

0.99+

OracleORGANIZATION

0.99+

Cheryl KnightPERSON

0.99+

David MoschellaPERSON

0.99+

Rob HofPERSON

0.99+

Palo AltoLOCATION

0.99+

August 9thDATE

0.99+

IntelORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

HPEORGANIZATION

0.99+

Chuck WhittenPERSON

0.99+

IBMORGANIZATION

0.99+

Goldman SachsORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

Fidelma RussoPERSON

0.99+

20 cloudsQUANTITY

0.99+

AWSORGANIZATION

0.99+

WintelORGANIZATION

0.99+

DatabricksORGANIZATION

0.99+

two questionsQUANTITY

0.99+

DellORGANIZATION

0.99+

John FurrierPERSON

0.99+

2018DATE

0.99+

AppleORGANIZATION

0.99+

JohnPERSON

0.99+

BostonLOCATION

0.99+

AviatrixORGANIZATION

0.99+

StarburstORGANIZATION

0.99+

ConfluentORGANIZATION

0.99+

five cloudsQUANTITY

0.99+

ClumioORGANIZATION

0.99+

CouchbaseORGANIZATION

0.99+

first timeQUANTITY

0.99+

NutanixORGANIZATION

0.99+

MoschellaORGANIZATION

0.99+

Skyhigh SecurityORGANIZATION

0.99+

MITORGANIZATION

0.99+

HashiCorpORGANIZATION

0.99+

last yearDATE

0.99+

RobPERSON

0.99+

two cloudsQUANTITY

0.99+

three cloudsQUANTITY

0.99+

david.vellante@siliconangle.comOTHER

0.99+

first twoQUANTITY

0.99+

Kristen MartinPERSON

0.99+

MongoORGANIZATION

0.99+

GartnerORGANIZATION

0.99+

CrowdStrikeORGANIZATION

0.99+

OktaORGANIZATION

0.99+

Pure StorageORGANIZATION

0.99+

Omer AsadPERSON

0.99+

Capital OneORGANIZATION

0.99+

each cloudQUANTITY

0.99+

SnowflakeORGANIZATION

0.99+

VeeamORGANIZATION

0.99+

OpenShiftTITLE

0.99+

10 most frequently asked questionsQUANTITY

0.99+

RapaportPERSON

0.99+

SiliconANGLEORGANIZATION

0.99+

CloudFlareORGANIZATION

0.99+

one sectionQUANTITY

0.99+

Seeing DigitalTITLE

0.99+

VMwareORGANIZATION

0.99+

IDCORGANIZATION

0.99+

ZscalerORGANIZATION

0.99+

each weekQUANTITY

0.99+

Javits CenterLOCATION

0.99+

late last yearDATE

0.98+

firstQUANTITY

0.98+

AdobeORGANIZATION

0.98+

more than one cloudQUANTITY

0.98+

each offeringQUANTITY

0.98+

Day 2 Wrap Up | HPE Discover 2022


 

>>The cube presents HPE discover 2022 brought to you by HPE. >>Welcome back to the Cube's coverage. We're wrapping up day two, John furrier and Dave ante. We got some friends and colleagues, longtime friends, Crawford Del Pret is the president of IDC. Matt Eastwood is the senior vice president of infrastructure and cloud guys. Thanks for coming on spending time. Great to you guys. >>That's fun to do it. Awesome. >>Cravin I want to ask you, I, I think this correct me if I'm wrong, but this was your first physical directions as, as president. Is that true or did you do one in 2019? >>Uh, no, we did one in 20. We did, we did one in 20. I was president at the time and then, and then everything started, >>Well, how was directions this year? You must have been stoked to get back together. Yeah, >>It was great. I mean, it was actually pretty emotional, you know, it's, it's a community, right? I mean, we have a lot of customers that have been coming to that event for a long, long time and to stand up on the stage and look out and see people, you know, getting a little bit emotional and a lot of hugs and a lot of bringing people together. And this year in Boston, we were the first event really of any size that kind of came back. And when I kind of didn't see that coming in terms of how people, how ready people were to be together. Cause >>When did you did it April >>In Boston? Yeah, we did it March in March. Yeah, it was, it was, it was, it was a game day decision. I mean, we were, we had negotiated it, we were going back and forth and then I kind of made the call at the last minute, say, let's go and do it. And in Santa Clara, I felt like we were kind of opening up the crypt at the convention center. I mean, all the production people said, you know what? You guys were really the first event to be back. And attendance was really strong. You know, we, we, we got over a thousand. It was, it was really good. >>Good. It's always a fun when I was there. It was, it's a big deal. You guys prepare for it. Yeah. Some new faces up on the stage. Yeah. So, so Matt, um, you've been doing the circuit. I take it like, like all top analysts, super busy. Right. This is kind of end of the spring. I mean, I know it's summer, right. That's right. But, um, how do you look at, at discover relative some, some of the other events you've been at? >>So I think if you go back to what Crawford was just talking about our event in March, I mean, March was sort of the, the reopening and there was, I think people just felt so happy to be, to be back out there. You still get a little bit at, at these events. I mean, cuz for each, each company it's their first time back at it, but I think we're starting to get down what these events are gonna feel like going forward. Um, and it, I mean, there's good energy here. There's been a good attendance. I think the, the interest in getting back live and having face to face meetings is clearly strong. >>Yeah. I mean, this definitely shows that hybrids, the steady state, both events cloud. Yeah. Virtualization remotes. So what are you guys seeing with that hybrid mode? Just from a workforce, certainly people excited to get back together, but it's gonna continue. You're starting to see that digital piece. How is that impacting some of the, some of the customers you're tracking, who's winning and who's losing, coming out of the pandemic. What's the big picture look like? >>Yeah. I mean, if you, if you take a look at hybrid work, um, people are testing many, many, many different models. And I think as we move from a pandemic to an em, we're gonna have just waves and waves and waves of people needing that flexibility for a lot of different reasons, whether they have, uh, you know, preexisting conditions, whether they're just not comfortable, whether they have people who can't be vaccinated at home. So I think we're gonna be in this hybrid work for a long, long time. I do think though that we are gonna transition back into some kind of a normal, um, and I, and I think the big difference is that I think leaders back in the day, a long time ago, when people weren't coming into work, it was kind of like, oh, I know nothing's going on there. People aren't getting worked. And I think we're over that stage. Yeah. I think we're now into a stage where we know people can be productive. We know people can effectively work from home and now we're into the reason to be in the office. And the reason to be in the office is that collaboration, it's that mentoring it's that, you know, think about your 25 year old self. Do you wanna be staring at a windshield all day long and not kind of building those relationships? People want face to face, it's difficult. They want face >>To face and I would, and you guys had a great culture and it's a young culture. How are you handling it as an executive in terms of, is there a policy for hybrid or >>Yeah, so, so, so at IDC, what we did is we're in a pilot period and we've kind of said that the summertime is gonna be a pilot period and we've asked people, we're actually serving shocker, we're >>Serving, >>But we're, but we're, but, but we're actually asking people to work with their manager on what works for them. And then we'll come up with, you know, whether you are in, out of the office worker, which will be less than two days a hybrid worker, which will be three days or, uh, in, in the office, which is more than three days a week. And you know, we all know there's, there's, there's limitation, there's, there's, there's variability in that, but that's kind of what we're shooting for. And we'd like to be able to have that in place in the fall. >>Are you pretty much there? >>Yeah, I am. I, I am there three days a week. I I, Mondays and Fridays, unless, >>Because you got the CEO radius, right? Yeah. >><laugh>, <laugh> >>The same way I'm in the office, the smaller, smaller office. But so, uh, let's talk a little bit about the, the numbers we were chatting earlier, trying to squint through you guys are, you know, obviously the gold standard for what the market does, what happened in, you know, during the pandemic, what happened in 2021 and what do you expect to happen in, in 2022 in terms of it spending growth? >>Yeah. So this is, this is a crazy time, right? We've never seen this. You and I have a long history of, uh, of tracking this. So we saw in, in, in, in 2020, the market decelerated dramatically, um, the GDP went down to a negative like it always does in these cases, it was, you know, probably negative six in that, in that, in that kind of range for the first time, since I've been tracking it, which goes back over 30 years, tech didn't go negative tech went to about just under 3%. And then as we went to 2021, we saw, you know, everything kind of snap back, we saw tech go up to about 11% growth. And then of course we saw, you know, GDP come back to about a 4%, you know, ki kind of range growth. Now what's I think the story there is that companies and you saw this anecdotally everywhere companies leaned into tech, uh, company. >>You know, I think, you know, Matt, you have a great statistic that, you know, 80% of companies used COVID as their point to pivot into digital transformation, right. And to invest in a different way. And so what we saw now is that tech is now where I think companies need to focus. They need to invest in tech. They need to make people more productive with tech and it played out in the numbers now. So this year what's fascinating is we're looking at two Fastly different markets. We've got gasoline at $7 a gallon. We've got that affecting food prices. Uh, interesting fun fact recently it now costs over $1,000 to fill an 18 Wheeler. All right. Based on, I mean this just kind of can't continue. So you think about it, don't put the boat >>In the wall. Yeah. Yeah. >>Good, good, good, good luck. It's good. Yeah, exactly. <laugh> so a family has kind of this bag of money, right? And that bag of money goes up by maybe three, 4% every year, depending upon earnings. So that is sort of sloshing around. So if food and fuel and rent is taking up more gadgets and consumer tech are not, you know, you're gonna use that iPhone a little longer. You're gonna use that Android phone a little longer. You're gonna use that TV a little longer. So consumer tech is getting crushed, you know, really it's very, very, and you saw it immediately and ad spending, you've seen it in meta. You've seen it in Facebook. Consumer tech is doing very, very it's tough enterprise tech. We haven't been in the office for two and a half years. We haven't upgraded whether that be campus wifi, whether that be, uh, servers, whether that be, uh, commercial PCs, as much as we would have. So enterprise tech, we're seeing double digit order rates. We're seeing strong, strong demand. Um, we have combined that with a component shortage and you're seeing some enterprise companies with a quarter of backlog. I mean, that's, you know, really unheard at higher >>Prices, which >>Also, and therefore that drives that >>Drives. It shouldn't be that way. If there's a shortage of chips, it shouldn't be that way, >>But it is, but it is, but it is. And then you look at software and we saw this, you know, we've seen this in previous cycles, but we really saw it in the COVID downturn where, uh, in software, the stickiness of SaaS means that you just, you're not gonna take that stuff out. So the, the second half of last year we saw double digit rates in software surprise. We're seeing high single digit revenue growth in software now, so that we think is gonna sustain, which means that overall it demand. We expect to be between five and 6% this year. Okay, fine. We have a war going on. We have, you know, potentially, uh, a recession. We think if we do, it'll be with a lower case, R maybe you see a banded down to maybe 4% growth, but it's gonna grow this. >>Is it, is it both the structural change of the disruption of COVID plus the digital transformation yeah. Together? Or is it, >>I, I think you make a great point. Um, I, I, I think that we are entering a new era for tech. I think that, you know, Andrew's famous wall street journal oped 10 years ago, software is even world was absolutely correct. And now we're finding that software is, is eing into every nook and cranny people have to invest. They, they know disruptors are coming around every single corner. And if I'm not leaning into digital transformation, I'm dead. So >>The number of players in tech is, is growing, >>Cuz there's well, the number of players in tech number >>Industry's coming >>In. Yeah. The industry's coming in. So I think the interesting dynamic you're gonna see there is now we have high interest rates. Yeah. Which means that the price of funding these companies and buying them and putting data on is gonna get higher and higher, which means that I think you could, you could see another wave of consolidation. Mm-hmm <affirmative> because tech large install based tech companies are saying, oh, you know what? I like that now >>4 0 9 S are being reset too. That's another point. >>Yeah. I mean, so if you think about this, this transformation, right. So it's all about apps, absent data and differentiating and absent data. What the, the big winner the last couple years was cloud. And I would just say that if this is the first potential recession that we're talking about, where the cloud service providers. So I think a cloud as an operating model, not necessarily a destination, but for these cloud service providers, they've actually never experienced a slowdown. So how, and, and if you think about the numbers, 30% of, of the typical it budget is now quote, unquote cloud and 30% of all expenditures are it related. So there's a lot of exposure there. And I think you're gonna see a lot of, a lot of focus on how we can rationalize some of those investments. >>Well, that's a great point. I want to just double click on that. So yeah, the cloud did well during the pandemic. We saw that with SAS, have you guys tracked like the Tams of what got pulled forward? So the bit, a big discussion about something that pulled forward because of the pandemic, um, like zoom, for instance, obviously everyone's using zoom. Yeah, yeah, yeah. Was there fake Tams? There was one, uh, couple analysts who were pointing out that some companies were hot during the pandemic will go away that that Tam doesn't really exist, but there's some that got pulled forward early. That's where the growth is. So is there a, is there a line between the, I call fake Tam or pulled forward TA that was only for the pandemic situationally, um, devices might be like virtual event, virtual event. Software was one, I know Hoppin got laid a lot of layoffs. And so that was kind of gone coming, coming and going. And you got SAS which got pulled forward. Yep. And it's not going away, but it's >>Sustaining. Yeah. Yeah. But it's, but, but it's sustaining, um, you know, I definitely think there was a, there was a lot of spending that absolutely got pulled forward. And I think it's really about CEO's ability to control expectations and to kind of message what it, what it looks like. Um, you know, I think I look, I, I, I think virtual event platforms probably have a role. I think you can, you can definitely, you know, raise your margins in the event, business, significantly using those platforms. There's a role for them. But if you were out there thinking that this thing was gonna continue, then you know, that that was unrealistic, you know, Dave, to, to your point on devices, I'm not necessarily, you know. Sure. I think, I think we definitely got ahead of our expectations and things like consumer PCs, those things will go back to historical growth >>Rates. Yeah. I mean, you got the install base is pretty young right now, but I think the one way to look at it too, is there was some technical debt brought in because people didn't necessarily expect that we'd be moving to a permanent hybrid state two years ago. So now we have to actually invest on both. We have to make, create a little bit more permanency around the hybrid world. And then also like Crawford's talking about the permanency of, of having an office and having people work in, in multiple modes. Yeah. It actually requires investment in both the office. And >>Also, so you're saying operationally, you gotta run the company and do the digital transformation to level up the hybrid. >>Yeah. Yeah. Just the way people work. Right. So, so, you know, you basically have to, I mean, even for like us internally, Crawford was saying, we're experimenting with what works for us. My team before the pandemic was like one third virtual. Now it's two third virtual, which means that all of our internal meetings are gonna be on, on teams or zoom. Right. Yeah. They're not gonna necessarily be, Hey, just coming to the office today, cuz two thirds of people aren't in the Boston area. >>Right. Matt, you said if you see cloud as an operating model, not necessarily a place. I remember when you were out, I was in the, on the, on the, on the zoom when, when first met Adam Celski yeah. Um, he said, you were asking him about, you know, the, the on-prem guys and he's like, nah, it's not cloud. And he kind of was very dismissive of it. Yeah. Yeah. I wanna get your take on, you know, what we're seeing with as Azure service GreenLake, apex, Cisco's got their version. IBM. Fewer is doing it. Is that cloud. >>I think if it's, I, I don't think all of it is by default. I think it is. If I actually think what HPE is doing is cloud, because it's really about how you present the services and how you allow customers to engage with the platform. So they're actually creating a cloud model. I think a lot of people get lost in the transition from, you know, CapEx to OPEX and the financing element of this. But the reality is what HPE is doing and they're sort of setting the standard. I think for the industry here is actually setting up what I would consider a cloud model. >>Well, in the early days of, of GreenLake, for sure it was more of a financial, you >>Know, it was kind of bespoke, right. But now you've got 70 services. And so you can, you can build that out. But >>You know, we were talking to Keith Townsend right after the keynote and we were sort of UN unpacking it a little bit. And I, I asked the question, you know, if you, if you had to pin this in terms of AWS's maturity, where are we? And the consensus was 2014 console filling, is that fair or unfair? >>Oh, that's a good question. I mean, um, I think it's, well, clouds come a long way, right? So it'd be, I, I, I think 20, fourteen's probably a little bit too far back because >>You have more modern tools I Kubernetes is. Yeah. >>And, but you also have, I would say the market still getting to a point of, of, of readiness and in terms of buying this way. So if you think about the HP's kind of strategy around edge, the core platform as a, as a service, you know, we're all big believers in edge and the apps follow the data and the data's being created in new locations and you gotta put the infrastructure there. And for an end user, there's a lot of risk there because they don't know how to actually plan for capacity at the edge. So they're gonna look to offload that, but this is a long term play to actually, uh, build out and deploy at the edge. It's not gonna happen tomorrow. It's a five, 10 year play. >>Yeah. I mean, I like the operating model. I'd agree with you, Matt, that if it's, if it's cloud operations, DevSecOps and all that, all that jazz it's cloud it's cloud operating and, and, and public cloud is a public cloud hyperscaler on premise. And the storage folks were presented. That's a single pane of glass. That's old school concepts, but cloud based. Yep. Shipping hardwares, auto figures. Yeah. That's the kind of consumption they're going for now. I like it. Then I, then they got the partner led thing is the partner piece. How do you guys see that? Because if I'm a partner, there's two things, wait a minute, am I at bottleneck to the direct self-service? Or is that an enabler to get more cash, to make more money? If I'm a partner. Cause you see what Essentia's doing with what they do with Amazon and Deloitte and et C. Yeah. You know, it's interesting, right? Like they've a channel partner, I'm making more cash. >>Yeah. I mean, well, and those channel partners are all in transition too. They're trying to yeah. Right. Figure out. Right, right. Are they, you know, what are their managed services gonna look like? You know, what kind of applications are they gonna stand up? They're they're not gonna just be >>Reselling, bought a big house in a boat. The box is not selling. I wanna ask you guys about growth because you know, the big three cloud, big four growing pick a number, I dunno, 30, 35% revenue big. And like you said, it's 30% of the business now. I think Dell's growing double digits. I don't know how much of that is sustainable. A lot of that is PCs, but still strong growth. Yep. I think Cisco has promised 9% >>In, in that. Right, right. >>About that. Something like that. I think IBM Arvin is at 6%. Yep. And I think HPE has said, Hey, we're gonna do three to 4%. Right. Which is so really sort of lagging and which I think a lot of people in wall street is like, okay, well that's not necessarily so compelling. Right. What does HPE have to do to double that growth? Or even triple that growth. >>Yeah. So they're gonna need, so, so obviously you're right. I mean, being able to show growth is Tanem out to this company getting, you know, more attention, more heat from, from investors. I think that they're rightly pointing to the triple digit growth that they've seen on green lake. I think if you look at the trailing, you know, 12 month bookings, you got over, you know, 7 billion, which means that in a year, you're gonna have a significant portion of the company is as a service. And you're gonna see that revenue that's rat being, you know, recognized over a series of months. So I think that this is sort of the classic SAS trough that we've seen applied to an infrastructure company where you're basically have to kind of be in the desert for a long time. But if they can, I think the most important number for HPE right now is that GreenLake booking snow. >>And if you look at that number and you see that number, you know, rapidly come down, which it hasn't, I mean off a very large number, you're still in triple digits. They will ultimately start to show revenue growth, um, in the business. And I think the one thing people are missing about HPE is there aren't, there are a lot of companies that want to build a platform, but they're small and nobody cares. And nobody let's say they throw a party and nobody comes. HP has such a significant installed base that if they do build a platform, they can attract partners to that platform. What I mean by that is partners that deliver services on GreenLake that they're not delivering. They have the girth to really start to change an industry and change the way stuff is being built. And that's the be they're making. And frankly, they are showing progress in that direction. >>So I buy that. But the one thing that concerns me is they kind of hide the ball on services. Right. And I, and I worry about that is like, is this a services kind of just, you know, same wine, new bottle or, >>Or, yeah. So, so I, I, I would argue that it's not about hiding the ball. It's about eliminating confusion of the marketplace. This is the company that bought EDS only to spin it off <laugh>. Okay. And so you don't wanna have a situation where you're getting back into services. >>Yeah. They're the only one >>They're product, not the only ones who does, I mean, look at the way IBM used to count and still >>I get it. I get it. But I think it's, it's really about clarity of mission. Well, I point next they are in the Ts business, absolutely. Point of it. It's important prop >>Drive for them at the top. Right. The global 50 say there's still a lot of uniqueness in what they want to buy. So there's definitely a lot of bespoke kind of delivery. That's still happening there. The real promise here is when you get into the global 2000 and yeah. And can start them to getting them to consume very standardized offers. And then the margins are, are healthy >>And they got they're what? Below 30, 33, 30 3%. I think 34% last quarter gross margin. Yeah. That that's solid. Just compare that with Dell is, I don't know. They're happy with 20, 21% of correct. You get that, which is, you know, I I'll come back. Go ahead. I want, I wanna ask >>Guys. No, I wanna, I wanna just, he said one thing I like, which was, I think he nailed it. They have such, um, big install base. They have a great channel. They know how to use it. Right. That's a real asset. Yeah. And Microsoft, I remember when their stock was trading at 26 when Baltimore was CEO. Yep. What they did with no, they had office and windows, so a little bit different. Yep. But similar strategy, leverage our install base, bring something up to them. That's what you're kind of connecting the >>Absolutely. You have this velocity, uh, machine with a significant girth that you can now move to a new model. They move that to a new model. To Matt's point. They lead the industry, they change the way large swath the customers buy and you will see it in steady revenue growth over time. Okay. So I just in that, well, >>So your point is the focus and there the right it's the right focus. And I would agree what's >>What's the other move. What's their other move, >>The problem. Triple digit booking growth off a number that gets bigger >>Inspired. Okay. >>Whats what's the scoreboard. Okay. Now they're go at the growth. That's the scoreboard. What are the signals? Are you looking at on the scoreboard Crawford and Matt in terms of success? What are the benchmarks? Is it ecosystem growth, number of services, triple growth. Yeah. What's the, what are some of the metrics that you guys are gonna be watching and we should be watching? >>Yeah. I mean, I dunno if >>You wanna jump in, I mean, I think ecosystem's really critical. Yeah. You want to, you want to have well and, and you need to sell both ways like HPE needs to be selling their technology on other cloud providers and vice versa. You need to have the VMs of the world on, you know, offering services on your platform and, and kind of capturing some, some motion off that. I think that's pretty critical. The channel definitely. I mean, you have to help and what you're gonna see happen there is there will be channel partners that succeed in transforming and succeeding and there'll be a lot that go away and that some, some of that's, uh, generational there'll be people that just kind of age outta the system and, and just go home. >>Yeah. Yeah. So I would argue it's, it's, it's, it's gonna be, uh, bookings growth rate. It's gonna be retention rate of the, of, of, of the customers, uh, that they have. And then it's gonna be that, that, um, you know, ultimately you're gonna see revenue, um, growth, and which is that revenue growth is gonna have to be correlated to the booking's growth for green lake cross. >>What's the Achilles heel on, on HPE. If you had to do the SWAT, what's the, what's the w for HPE that they really need to pay >>Attention to. I mean, they, they need to continue their relentless focus on cost, particularly in the, in the core compute, you know, segment they need to be, they need to be able to be as cost effective as possible while the higher profit dollars associated with GreenLake and other services come in and then increase the overall operating margin and gross margin >>Picture for the, I mean, I think the biggest thing is they just have, they have to continue the motion that they've been on. Right. And they've been consistent about that. Mm-hmm, <affirmative> what you see where others have, have kind of slipped up is when you go to, to customers and you present the, the OPEX as a service and the traditional CapEx side by side, and the customers put in this position of trying to detangle what's in that OPEX service, you don't wanna do that obviously. And, and HP has not done that, but we've seen others kind of slip up. And, but >>A lot of companies still wanna buy CapEx. Right. Absolutely liquid. And, and I think, >>But you shouldn't do a, you shouldn't do that bake off by putting those two offers out. You should basically ascertain what they want to do. >>What's kind of what Dell does. Right. Hey, how, what do you want? We got this, we got >>This on one hand, we got this, the, we got that, right. Uh, the two hand sales rep, no, this CapEx. Thing's interesting. And if you're Amazon and Azure and, and GCP, what are they thinking right now? Cause remember what, four years ago outpost was launched, which essentially hardware. Yeah. This is cloud operating model. Yep. Yeah. They're essentially bringing outpost. This is what they got basically is Amazon and Azure, like, is this ABL on the radar for them? How would you, what, what are they thinking in your mind if we're on, if we're in their office, in their brain trust, are they laughing? Are they like saying, oh, they're scared. Is this real threat >>Opportunity? I, I, I mean, I wouldn't say they're laughing at all. I, I would say they're probably discounting a little bit and saying, okay, fine. You know, that's a strategy that a traditional hardware company is moving to. But I think if you look underneath the covers, you know, two years ago it was, you know, pretty basic stuff they were offering. But now when you start getting into some, you know, HPC is a service, you start getting into data fabric, you start getting into some of the more, um, sophisticated services that they're offering. And, and I think what's interesting about HP. What my, my take is that they're not gonna go after the 250 services the Amazon's offering, they're gonna basically have a portfolio of services that really focus on the core use cases of their infrastructure set. And, and I think one of the danger things, one, one of the, one of the red flags would be, if they start going way up the stack and wanting to offer the entire application stack, that would be like a big flashing warning sign, cuz it's not their sweet spot. It's not, not what they have. >>So machine learning, machine learning and quantum, okay. One you can argue might be up the stack machine learning quantum should be in their wheelhouse. >>I would argue machine learning is not up the stack because what they would focus on is inference. They'd focus on learning. If they came out and said, machine learning all the way up to the, you know, what a, what, what a drug discovery company needs to do. >>So they're bringing it down. >>Yeah. Yeah. Well, no, I think they're focusing on that middle layer, right? That, that, that data layer. And I think that helping companies manage their data make more sense outta their data structure, their data that's core to what they wanna do. >>I, I feel as though what they're doing now is table stakes. Honestly, I do. I do feel like, okay, Hey finally, you know, I say the same thing about apex, you >>Know, we finally got, >>It's like, okay guys, the >>Party. Great. Welcome to the, >>But the one thing I would just say about, about AWS and the other big clouds is whether they might be a little dismissive of what's truly gonna happen at the edge. I think the traditional OEMs that are transforming are really betting on that edge, being a huge play and a huge differentiator for them where the public cloud obviously have their own bets there. But I think they were pretty dismissive initially about how big that went. >>I don't, and I don't think anybody's really figured out the edge yet. >>Well, that's an, it's a battleground. That's what he's saying. I think you're >>Saying, but on the ecosystem, I wanna say up the stack, I think it's the ecosystem. That's gotta fill that out. You gotta see more governance tools and catalogs and AI tools and, and >>It immediately goes more, it goes more vertical when you go edge, you're gonna have different conversations and >>They're >>Lacking. Yeah. And they, but they're in there though. They're in the verticals. HP's in the, yeah, >>For sure. But they gotta build out an ego. Like you walk around here, the data, the number of data companies here. I mean, Starburst is here. I'm actually impressed that Starburst is here. Cause I think they're a forward thinking company. I wanna see that times a hundred. Right. I mean, that's >>You see HP's in all the verticals. That's I think the point here, >>So they should be able to attract that ecosystem and build that, that flywheel that's the, that's the hallmark of a cloud that marketplace. >>Yeah, it is. But I think there's a, again, I go back to, they really gotta stay focused on that infrastructure and data management. Yeah. >>But they'll be focused on that, but, but their ecosystem, >>Their ecosystem will then take it up from there. And I think that's the next stage >>And that ecosystem's gotta include OT players and communications technologies players as well. Right. Because that stuff gets kind of sucked up in that, in that edge play. Do >>You feel like HPE has a, has a leg up on that or like a little, a little bit of a lead or is it pretty much, you know, even raced right now? >>I think they've, I think the big infrastructure companies have all had OEM businesses and they've all played there. It's it's, it's also helping those OT players actually convert their own needs into more of a software play and, and not so much of >>Physical. You've been, you've been following and you guys both have been following HP and HPE for years. They've been on the edge for a long time. I've been focused on this edge. Yeah. Now they might not have the product traction that's right. Or they might not develop as fast, but industrial OT and IOT they've been talking about it, focused on it. I think Amazon was mostly like, okay, we gotta get to the edge and like the enterprise. And, and I think HP's got a leg up in my opinion on that. Well, I question is can they execute? >>Yeah. I mean, PTC was here years ago on stage talking >>About, but I mean, you think about, if you think about the edge, right. I mean, I would argue one of the best acquisitions this company ever did was Aruba. Right. I mean, it basically changed the whole conversation of the edge changed the whole conversation. >>If >>Became GreenLake, it was GreenLake. >>Well, it became a big department. They gave a big, but, but, but I mean, you know, I mean they, they, they went after going selling edge line servers and frankly it's very difficult to gain traction there. Yeah. Aruba, huge area. And I think the March announcement was when they brought Aruba management into. Yeah. Yeah. >>Totally. >>Last question. Love >>That. >>What are you guys saying about the, the Broadcom VMware acquisition? What's the, what are the implications for the ecosystem for companies like HPE and just generally for the it business? >>Yeah. So >>You start. Yeah, sure. I'll start, I'll start there. So look, you know, we've, you know, spent some time, uh, going through it spent some time, you know, speaking, uh, to the, to the, to the folks involved and, and, and I gotta tell you, I think this is a really interesting moment for Broadcom. This is Broadcom's opportunity to basically build a different kind of a conversation with developers to, uh, try to invest in. I mean, just for perspective, right? These numbers may not be exact. And I know a dollar is not a dollar, but in 2001, anybody, remember what HP paid for? Compact >>8,000,000,020, >>So 25 billion, 25 billion. Wow. VMware just got sold for 61 billion. Wow. Okay. Unbill dollars. Okay. That gives you a perspective. No, again, I know a dollar is not a dollar 2000. >>It's still big numbers, >>2022. So having said that, if you just did it to, to, to basically build your DCF model and say, okay, over this amount of time, I'll pay you this. And I'll take the money out of this period of time, which is what people have criticized them for. I think that's a little shortsighted. I, yeah, I think this is Broadcom's opportunity to invest in that product and really try to figure out how to get a seat at the table in software and pivot their company to enterprise software in a different way. They have to prove that they're willing to do that. And then frankly, that they can develop the skills to do that over time. But I do believe this is a, a different, this is a pivot point. This is not >>CA this is not CA >>It's not CA >>In my, in my mind, it can't be CA they would, they would destroy too much. Now you and I, Dave had some, had some conversations on Twitter. I, I don't think it's the step up to them sort of thinking differently about semiconductor, dying, doing some custom semi I, I don't think that's. Yeah. I agree with that. Yeah. I think I, I think this is really about, I got two aspiration for them pivoting the company. They could >>Justify the >>Price to the, getting a seat at the adults table in software is, >>Well, if, if Broadcom has been squeezing their supplies, we all hear the scutle butt. Yeah. If they're squeezing, they can use VMware to justify the prices. Yeah. Maybe use that hostage. And that installed base. That's kind of Mike conspiracy. >>I think they've told us what they're gonna do. >><laugh> I do. >>Maybe it's not like C what's your conspiracy theory like Symantec, but what >>Do you think? Well, I mean, there's still, I mean, so VMware there's really nobody that can do all the things that VMware does say. So really impossible for an enterprise to just rip 'em out. But obviously you can, you can sour people's taste and you can very much influence the direction they head in with the collection of, of providers. One thing, interesting thing here is, was the 37% of VMware's revenues sold through Dell. So there's, there's lots of dependencies. It's not, it's not as simple as I think John, you you're right. You can't just pull the CA playbook out and rerun it here. This is a lot more complex. Yeah. It's a lot more volume of, of, of distribution, but a fair amount of VMware's install >>Base Dell's influence is still there basically >>Is in the mid-market. It's not, it's not something that they're gonna touch directly. >>You think about what VMware did. I mean, they kept adding new businesses, buying new businesses. I mean, is security business gonna stay >>Networking security, I think are interesting. >>Same >>Customers >>Over and over. Haven't done anything. VMware has the same customers. What new >>Customers. So imagine simplifying VMware. Right, right. Becomes a different equation. It's really interesting. And to your point, yeah. I mean, I think Broadcom is, I mean, Tom Crouse knows how to run a business. >>Yeah. He knows how to run a business. He's gonna, I, I think it's gonna be, you know, it's gonna be an efficient business. It's gonna be a well run business, but I think it's a pivot point for >>Broadcom. It's amazing to me, Broadcom sells to HPE. They sell it to Dell and they've got a market cap. That's 10 X, you know? Yes. Yeah. All we gotta go guys. Awesome. Great conversation guys. >>A lot. Thanks for having us on. >>Okay. Listen, uh, day two is a, is a wrap. We'll be here tomorrow, all day. Dave ante, John furrier, Lisa Martin, Lisa. Hope you're feeling okay. We'll see you tomorrow. Thanks for watching the cube, your leader in enterprise tech, live coverage.

Published Date : Jun 30 2022

SUMMARY :

Great to you guys. That's fun to do it. Is that true or did you do one in 2019? I was president at the time and then, You must have been stoked to get back together. I mean, it was actually pretty emotional, you know, it's, it's a community, right? I mean, all the production people said, you know what? But, um, how do you look at, at discover relative some, So I think if you go back to what Crawford was just talking about our event in March, I mean, March was sort of the, So what are you guys seeing with that hybrid mode? And I think as we move from a pandemic to an em, To face and I would, and you guys had a great culture and it's a young culture. And then we'll come up with, you know, whether you are in, out of the office worker, which will be less than two days a I I, Mondays and Fridays, Because you got the CEO radius, right? you know, during the pandemic, what happened in 2021 and what do you expect to happen in, in 2022 And then of course we saw, you know, GDP come back to about a 4%, you know, ki kind of range growth. You know, I think, you know, Matt, you have a great statistic that, you know, 80% of companies used COVID as their point to pivot In the wall. I mean, that's, you know, really unheard at higher It shouldn't be that way. And then you look at software and we saw this, you know, Is it, is it both the structural change of the disruption of COVID plus I think that, you know, Andrew's famous wall street journal oped 10 years ago, software is even world was absolutely on is gonna get higher and higher, which means that I think you could, you could see another That's another point. And I think you're gonna see a lot of, a lot of focus on how we can rationalize some of those investments. We saw that with SAS, have you guys tracked like the Tams of what got pulled forward? I think you can, you can definitely, create a little bit more permanency around the hybrid world. the hybrid. So, so, you know, you basically have to, I remember when you were the transition from, you know, CapEx to OPEX and the financing element of this. And so you can, you can build that out. And I, I asked the question, you know, if you, if you had to pin this in terms of AWS's maturity, I mean, um, I think it's, well, clouds come a long way, right? Yeah. the core platform as a, as a service, you know, we're all big believers in edge and the apps follow And the storage folks were presented. Are they, you know, what are their managed services gonna look like? I wanna ask you guys about growth because In, in that. And I think HPE has said, I think if you look at the trailing, you know, 12 month bookings, you got over, you know, 7 billion, which means that in a And I think the one thing people are missing about HPE is there aren't, there are a lot of companies that want And I, and I worry about that is like, is this a services kind of just, you know, And so you don't wanna have a situation where you're But I think it's, it's really about clarity of mission. The real promise here is when you get into the global 2000 and yeah. You get that, which is, you know, I I'll come back. They know how to use it. You have this velocity, uh, machine with a significant girth that you can now move And I would agree what's What's the other move. Triple digit booking growth off a number that gets bigger Okay. What's the, what are some of the metrics that you guys are gonna be watching I mean, you have to help and what you're gonna see And then it's gonna be that, that, um, you know, ultimately you're gonna see revenue, If you had to do the SWAT, what's the, what's the w for HPE that I mean, they, they need to continue their relentless focus on cost, Mm-hmm, <affirmative> what you see where others have, have kind of slipped up is when you go A lot of companies still wanna buy CapEx. But you shouldn't do a, you shouldn't do that bake off by putting those two offers out. Hey, how, what do you want? And if you're Amazon and Azure and, and GCP, But I think if you look underneath the covers, you know, two years ago it was, One you can argue might be up the stack machine learning quantum should If they came out and said, machine learning all the way up to the, you know, what a, what, what a drug discovery company needs to do. And I think that helping companies manage their data make more sense outta their data structure, their data that's core to okay, Hey finally, you know, I say the same thing about apex, you Welcome to the, But I think they were pretty dismissive initially about how big that went. I think you're Saying, but on the ecosystem, I wanna say up the stack, I think it's the ecosystem. They're in the verticals. Cause I think they're a forward thinking company. You see HP's in all the verticals. So they should be able to attract that ecosystem and build that, that flywheel that's the, But I think there's a, again, I go back to, they really gotta stay focused And I think that's the next stage And that ecosystem's gotta include OT players and communications technologies players as well. I think they've, I think the big infrastructure companies have all had OEM businesses and they've all played there. I think Amazon was mostly like, okay, we gotta get to the edge and like the enterprise. I mean, it basically changed the whole conversation of the edge changed the whole conversation. And I think the March announcement was when they brought So look, you know, we've, you know, spent some time, uh, going through it spent some time, That gives you a perspective. And I'll take the money out of this period of time, which is what people have criticized them for. I think I, I think this is really about, I got two aspiration for them pivoting the company. And that installed base. think John, you you're right. Is in the mid-market. I mean, they kept adding new businesses, buying new businesses. VMware has the same customers. I mean, I think Broadcom is, I mean, Tom Crouse knows how to run a business. He's gonna, I, I think it's gonna be, you know, it's gonna be an efficient business. That's 10 X, you know? Thanks for having us on. We'll see you tomorrow.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavePERSON

0.99+

Lisa MartinPERSON

0.99+

AmazonORGANIZATION

0.99+

DeloitteORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

Matt EastwoodPERSON

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

25 yearQUANTITY

0.99+

JohnPERSON

0.99+

MattPERSON

0.99+

LisaPERSON

0.99+

Tom CrousePERSON

0.99+

Adam CelskiPERSON

0.99+

three daysQUANTITY

0.99+

BostonLOCATION

0.99+

BroadcomORGANIZATION

0.99+

2020DATE

0.99+

2014DATE

0.99+

AWSORGANIZATION

0.99+

DellORGANIZATION

0.99+

MarchDATE

0.99+

Keith TownsendPERSON

0.99+

34%QUANTITY

0.99+

HPORGANIZATION

0.99+

Santa ClaraLOCATION

0.99+

20QUANTITY

0.99+

John furrierPERSON

0.99+

61 billionQUANTITY

0.99+

tomorrowDATE

0.99+

2022DATE

0.99+

30QUANTITY

0.99+

2001DATE

0.99+

HPEORGANIZATION

0.99+

12 monthQUANTITY

0.99+

2021DATE

0.99+

30%QUANTITY

0.99+

80%QUANTITY

0.99+

AndrewPERSON

0.99+

VMwareORGANIZATION

0.99+

25 billionQUANTITY

0.99+

iPhoneCOMMERCIAL_ITEM

0.99+

fiveQUANTITY

0.99+

SymantecORGANIZATION

0.99+

33QUANTITY

0.99+

Crawford Del PretPERSON

0.99+

2019DATE

0.99+

9%QUANTITY

0.99+

GreenLakeORGANIZATION

0.99+

CrawfordPERSON

0.99+

37%QUANTITY

0.99+

IDCORGANIZATION

0.99+

70 servicesQUANTITY

0.99+

7 billionQUANTITY

0.99+

twoQUANTITY

0.99+

less than two daysQUANTITY

0.99+

OPEXORGANIZATION

0.99+

250 servicesQUANTITY

0.99+

Heiko Meyer & Paul Hunter, HPE | HPE Discover 2022


 

>>The cube presents HPE discover 2022 brought to you by HPE. >>Welcome back to HPE. Discover 2022. You're watching the cubes, uh, coverage where day two here at Dave ante with John furrier, HaCo Myers here. He's the executive vice president and chief sales officer, newly minted, relatively newly minted chief sales officer at HPE at HPE and Paul Hunter. Who's the senior vice president and managing director of north America for Hewlett Packard enterprise gentlemen. Welcome to the cube. >>Thank you. Thanks having us. >>Hi coach. This is the first time back in Vegas in a while three years. I think it's been. And your first as the chief sales officer. Yep. What's the vibe like how how's, how's >>It feel? I can tell you. It's so cool. Is it you, you walk down the hallway, everybody's smiling and you see people from, you have seen three years ago or in this format on your screen the last three years, I think, uh, what is amazing. We had exactly three years ago, we had this event and Antonio mentioned, Hey, and by end of 20, 22, you will see everything being available as a service. Yeah. And nobody thought about that. We will not meet in person until 2022 at that point in time. Yeah, indeed. And that's what happened that I can tell you, what's the best decision to make an in person event here in vigor with so many people, uh, because it's about, Hey, the change in the market, the demand, the transition. And, uh, so I think it, I, couldn't be more happy to see the last two days and looking for, for the, to the rest of the event, >>Paul, you have a, a, a background in the, the channel, um, and now you're heading north America. What are you seeing in the ecosystem? Is it, is there a difference as HaCo was saying from 2019, is there a different, you know, feeling different conversations? What are you seeing? Yeah. >>Well, the good thing is like, because we haven't been here for three years, you've got a really marked moment of comparison. So you cast your mind back. What were the conversations like? I think three years ago we were talking about cloud services and partners were nodding their heads and thinking, yeah, but the world is gonna continue as normal and we fast forward three years and, uh, the partners are really talking about, uh, proactively how do they build up that cloud services? And, uh, they're also talking about customer experiences as well. We've landed and won new customers. So, uh, that's really sort of thrilling to hear that they're really excited about the journey on with us. >>You know, I'd like to get your perspective on the, what happened during the pandemic, because we saw, um, first of all, you know, zoom and video com saved the internet, uh, had meetings, but the partner, the partners delivered a lot of value. Um, customers had to pivot, or if they had a tailwind, they had, they took advantage of it. Some had headwinds with the pandemic everyone's working at home. So a lot of disruptions for all the companies, but a lot of the partners had success during the pandemic. And because they have that solution. What was the, uh, uh, the learnings that you guys saw during the pandemic, because now with cloud cloud scale hybrid, mainstream, and now steady state people lived it and partners delivered a lot of solutions in hybrid mode. Yeah. In virtual mode. What was the learnings for you guys out, coming out of that with customers and partners? >>I think first of all, we, we all learned during the pandemic that, uh, you can business, uh, do business in a different way, but as well, you learned, uh, how to pivot faster in the digital transformation. This makes a difference. And this creates value. And I think together with our partner ecosystem, we were able to develop faster solutions there while we developed everything as a service and came up with more and more cloud services. The good thing is it resonates. And our history with the partners is I donors, as long as I can, uh, think back in my career. And you only can do that together with the channel partners. And I think they appreciate that. We learn from each other. We do the same enablement from my guys, like from the partner guys and this close relation, I think made a difference, >>You know, in 2019 GreenLake, as a service was really a financial vehicle right now that's, that's evolved. And now, you know, two years on three years on it's actually a cloud service. Absolutely. And so what's the resonance been with customers because I mean, every everybody says they want that cloud experience. They may not all want OPEX. Yeah. But so what have you hearing from customers? So >>First of all, what I hear is, um, not the, if, so the strategy is clear, the customer they'll love it. They like it. They have, they want to have the cloud-like experience and guess what? We have 70 cloud services now. Yeah. And we have announced a lot of new one the last couple of days, but it's not so much that if they should do this, it's more the question, how can we help me to scale faster? Yeah. And, uh, that's the, the, the, the feedback I got the last couple of days, and for us, it's a motivation. We are on the right track. There is, this is a moment where you have a demand from the market and a strategy that fits, and this is so strong and you can do this with the partner through the partners and you see the, the customers, they love it. I have never seen an event where I got so many requests the last two days where I say, I thought that, can you help me to get there faster? It's perfect. >>Yeah. I think, I think it was also a landmark moment when we presented the cloud platform as part of the Antonio's keynote, I've had a lot of partners say this was sort of really marked the moment where we felt there was there's real substance to the offering now. And, uh, I had one of the sales guys relate to me a story where they have a, a, a client in the audience. And, uh, they're thinking about how they might, um, have a relationship with us and through seeing the kind of significance of it for us, we're able to close deals. So that's also, you know, a really exciting thing. We're actually know we're closing deals and, and winning new >>Customers, Hey, being agile and closing deals fast is a good thing. Right? I mean, that's what you guys like. Yeah. >>I mean, that's >>What it, so I, so I love the channel conversation partners because one of the things that I've observed and, and, and, and, and knowing the HP channels so strong, they're obviously they want make money. Gross margin is all about the profit, the profit motive, but the enablement that you guys have, how is that translated into this, this, this shift everyone's aligned behind GreenLake and as a service, cuz this seems to be a good fit for partner. Cause they're gonna go to the customer, the ultimate end customer and bolt on services. >>Yeah. >>How is that going? Cuz this is, to me, seems like a dream scenario for services, which we all know is high gross margin. >>Yeah. Yeah. I think it's a journey. What's a journey for our sales organization. Like it is for the partners, but it's a journey worth to do that. And um, so what, what is our, our strategy to have this together with our channel partners in mind, uh, to, to combine their strengths and they can, we, we have a kind of modular approach so that they can plug in their strengths, their IP, or as well, their services, which makes them sticky and, uh, relevant to the customer. And it drives profitability. And I think that's the, the, the secret behind the model, working with the channel, not, uh, separate to the channel. And I think this resonates this story, it's, it's a journey. And, uh, we learned a lot the last three years how to sell it. We, in the past we were selling, uh, transactional hardware. Yeah. Now we are selling services, cloud services, like you mentioned different game and this is an enablement. And we, we, um, we offer the same trainings we are doing with our folks to our channel partners because we are together in this journey. >>Yeah. It, you, you described it really well. And uh, so did, Hico essentially, this is, it requires a lot of persistence because you, you're not gonna get it right the first time. And so we now seen partners try and fail several times, but now try fail and succeed. So that's exciting. Um, and also I think what we're also seeing is partners is doing quite a good job of building services that integrate into the cloud services. So they're right into the APIs. I was, I was with a meeting with a partner called CBTS and they talked about the whole of their services portfolio now is embedded in, in GreenLake. So that certainly was not the case three years ago. >>Yeah. And the other, the big tailwind too, as you got the open source software movement, you're seeing, you know, the ability for partners and ultimately the channel being software enabled they're adding services, not just professional services, but cloud services where they have the domain expertise. Yeah. They're close to the customer. Yeah. And they could really be, um, customizing solutions. Um, and that's gonna always be great for the customer. The question I have for you guys is do you see that domain specialism with machine learning and with software, do you see partners start to get vertically focused and like, and start, get more targeted towards save verticals? >>Yeah. >>You go, no, go first. Yeah. >>Well, again, I was, uh, it's funny, your questions are completely resonating with the conversations we've been having all day. Like I was with our partner called connection and they're talking about how do they build practices in four areas? And they're, I'm quite closely allowed to aligned to our areas of edge cloud and data. Um, they have another one which is also workplace transformation. So, and they're thinking, how do we add expertise? How do we hire, recruit and retain the best talent? And, uh, that again, that wasn't a conversation we were having two, three years ago. So where partners really add value to us is through their services and their expertise and progressive partners are hiring and doing that. >>Yeah. And this transformation I mentioned earlier, it's selling outcomes, business outcomes for the end customer. And, uh, I think selling outcomes means you need to be specialized in something, be it on a domain area or be on a vertical. And I think, uh, when you focus on that, uh, that's the best way you can add value to a customer. This creates this trust, this trust relationship. Yeah. >>So edge cloud and data, obviously, I, I think edge, you guys, you got sending stuff and outta space, that's the ultimate edge. So you got some proof points there. Deep edge, I think. Deep edge. Yeah. >><laugh> I is very good. >>I think cloud, you showed the console Alma. It was very, had a very clear and strong platform message say, okay, now go build the data piece to me is the least mature when I walk around. Although I did see Starburst. Yeah, yeah. Out there. I think Starbursts a very advanced leading edge thinker. So that was a good sign. What do you see as having to happen to really build out that data ecosystem now? >>So I think what is important, this, this is all connected to each other edge cloud and data. And at the end, it's about, uh, how we can create insights out of the data, uh, and uh, where they, they live, where they come up, the data, how we structure them, how we get insights out of the data. So I think this is an area we see much more. It's not only about AI, but it's about having a data strategy as a customer. This is one of those areas. We have customer advisory bots that tell us, Hey, help us. We want to create our data strategy. And this is something where I think we can play together with our partners to really create value, get these insights out of the data. >>Are you hearing conversations where cus customers or partners are saying, okay, I wanna get insights out of the data, but I actually want to build a data business. I wanna build data products on, on, on GreenLake. Are you hearing that yet? >>Yeah, we are. Um, particularly the sort of, we, we think of them as sort of information, um, modern companies, um, they're building out new service lines. I mean, you, you, we see it in a lot of industries. Now you can see like how car manufacturers are increasingly thinking about how do they monetize their data, they're getting from it. And, uh, so there are new businesses being established, like in lots of different verticals. Pharmaceuticals will be another one where, um, traditional players are really being challenged and there are big businesses growing very rapidly based off data. So we're seeing it quite extensively. And, and we have to think about how do we access those new customers? How do we intersect them? And it's not just the people that we've been dealing with for 10, 20 years. They're very new companies, >>Which, which announcement's got the most buzz in your conversations with customers, partners. >><laugh>, it's funny. So I, I had, uh, when, when, when I started my conversation at a couple of, uh, meetings now, the last two days always started and said, what, what resonates? Yeah. And first of all, the funny thing is everybody told me the clarity of the message, the strategy second, uh, the consistency that we do, what we promise to do. Um, a couple of them, I know that down ISS, private cloud enterprise, uh, it's a great solution here. And, uh, then, uh, what, what I hear as well with our clarity and the strategy we are leapfrogging the competition. That's what I get out of these meetings. And I think that's the best compliment we can get for the two days. Yeah. >>Yeah. And I think the platform and the conversations around machine learning, AI, we even had an HP executive talk about quantum. Yeah. So you guys are already starting to think about what's around the corner. And I think if the platform works, the test will be, and the results will be enablement ecosystem will be flourishing, and we're gonna watch that. So I wanna get your, your take on the early, um, shift. Cause I think this year with GreenLake and the platform it's, it's maturing enough to the right. No doubt about it. We see the momentum, but there's still a lot more to do than go. So how do you guys envision the ecosystem developing? Because that'll be the true test, the flourishing, cuz if you enable people will get value out of it and it's gotta be a step function, not incremental value. >>Yeah. I think we, we, we, we always talk about, Hey, we landed and then we expand from there. That's the beauty of the model. And the good thing is there's no window looking for the customer. So they are free. That's a modular system. And what we see it's uh, really first of all, to understand the customer digital journey, where are the journey and they're all in a different place. And we have this digital, uh, uh, next advisor workshops when we have this anchor point mm-hmm, <affirmative>, you start there, you really can grow. And then you add workloads based on where the customer sits, what are the partnerships we have to bring to that? So it's really a model which starts and is, uh, designed for the future. >>The field must love it. The folks in the field, we love it. Yeah. You guys love that. Absolutely. Yeah. Yeah. Give it the customer plan future and >>I can tell you the partners love it the, well, yeah, >>Got it. When I talk to CIOs, I, I, and I ask them, you know, what's changed, you know, with Ukraine and supply chain and inflation and rising interest rate, what's changed in terms of your assumption from the beginning of the year, you know, let's say, you know, in, in terms of it spend and they're saying, well, not a lot, actually we're gonna continue to spend, we are reprioritizing. You know, we got, we're taking Robb a little bit from over here to put it into security. Yeah. Okay. But generally speaking, it's, it's the same as we expected, let's call it six, 7% growth, which is pretty good on top of last year. Um, and, and maybe there's some dry powder there, depending on how business goes. It also seems like there's, there's a lot of headwinds at the macro and B to C you know, some of the consumer companies, but B2B is booming. >>So I think >>That what do you guys are seeing? >>Absolutely. I, I completely agree that the demand will continue. Mm-hmm <affirmative> for different reasons. It could be a little bit shift within the demand as you described, but, uh, they know exactly they're on a journey in the digital transformation. If they stop now, they have a competitive disadvantage. So they are wisely in investing. So I think that the, the demand will stay here. Yes. Everybody talks about macroeconomics recession. Uh, we are confident we will see in our B2B >>Part continued demand and they're well capitalized as are a lot of the ecosystem partners. >>Yeah. And it's not a nice to have. It's a must have, I mean, I dunno of any customers that are deinvest in technology, deinvest in the life blood of their business >>Business. Exactly. Guys, thanks so much for coming on the cube. Great. Great to see you. Yeah. Congratulations on being here and, and, and best of luck with all the follow up from the show. I'm sure that lot we're gonna update next year. You see how it turned out? Yeah. >><laugh> numbers >><laugh> thanks for having us. Thank you for watching this segment. This is Dave ante for John furrier, the cubes coverage of HPE discover 22 from Las Vegas. We'll be right back.

Published Date : Jun 29 2022

SUMMARY :

He's the executive vice president and chief sales Thank you. This is the first time back in Vegas in a while three years. Hey, and by end of 20, 22, you will see everything being available as is there a difference as HaCo was saying from 2019, is there a different, you know, Well, the good thing is like, because we haven't been here for three years, you've got a really marked moment of comparison. So a lot of disruptions for all the companies, but a lot of the partners had success during the pandemic. And I think together with our partner ecosystem, And now, you know, and this is so strong and you can do this with the partner through the partners and you see the, So that's also, you know, a really exciting thing. I mean, that's what you guys like. but the enablement that you guys have, how is that translated into this, this, Cuz this is, to me, seems like a dream scenario for services, And I think this resonates this story, it's, it's a journey. job of building services that integrate into the cloud services. with software, do you see partners start to get vertically focused and like, and start, get more targeted towards Yeah. And, uh, that again, that wasn't a conversation we were having two, three years ago. And I think, uh, when you focus on that, uh, So edge cloud and data, obviously, I, I think edge, you guys, you got sending stuff I think cloud, you showed the console Alma. And at the end, it's about, uh, how we can create insights out of the data, uh, Are you hearing that yet? And it's not just the people that we've been dealing with for 10, Which, which announcement's got the most buzz in your conversations with customers, And I think that's the best compliment we can get for the two Because that'll be the true test, the flourishing, cuz if you enable people And the good thing is there's no window looking for the customer. The folks in the field, we love it. at the macro and B to C you know, some of the consumer companies, but B2B is booming. So I think that the, the demand will stay here. technology, deinvest in the life blood of their business Guys, thanks so much for coming on the cube. This is Dave ante for John furrier, the cubes coverage of HPE discover 22

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AntonioPERSON

0.99+

VegasLOCATION

0.99+

Paul HunterPERSON

0.99+

2019DATE

0.99+

HPEORGANIZATION

0.99+

Las VegasLOCATION

0.99+

Hewlett PackardORGANIZATION

0.99+

two yearsQUANTITY

0.99+

PaulPERSON

0.99+

three yearsQUANTITY

0.99+

70 cloud servicesQUANTITY

0.99+

six, 7%QUANTITY

0.99+

next yearDATE

0.99+

firstQUANTITY

0.99+

north AmericaLOCATION

0.99+

last yearDATE

0.99+

CBTSORGANIZATION

0.99+

three years agoDATE

0.99+

two daysQUANTITY

0.99+

first timeQUANTITY

0.99+

HPORGANIZATION

0.99+

OPEXORGANIZATION

0.99+

John furrierPERSON

0.98+

three years agoDATE

0.98+

2022DATE

0.97+

twoDATE

0.97+

oneQUANTITY

0.97+

Heiko MeyerPERSON

0.95+

HaCoORGANIZATION

0.94+

GreenLakeORGANIZATION

0.93+

Dave antePERSON

0.93+

FirstQUANTITY

0.92+

HaCo MyersORGANIZATION

0.9+

this yearDATE

0.89+

22DATE

0.87+

day twoQUANTITY

0.87+

StarburstORGANIZATION

0.86+

pandemicEVENT

0.85+

UkraineLOCATION

0.85+

last three yearsDATE

0.83+

20DATE

0.81+

Dave anteORGANIZATION

0.81+

endDATE

0.79+

10, 20 yearsQUANTITY

0.77+

last couple of daysDATE

0.75+

secondQUANTITY

0.75+

StarburstsORGANIZATION

0.73+

daysDATE

0.66+

HicoORGANIZATION

0.65+

last two daysDATE

0.61+

lastDATE

0.53+

22TITLE

0.51+

twoQUANTITY

0.49+

AlmaORGANIZATION

0.33+

Data Power Panel V3


 

(upbeat music) >> The stampede to cloud and massive VC investments has led to the emergence of a new generation of object store based data lakes. And with them two important trends, actually three important trends. First, a new category that combines data lakes and data warehouses aka the lakehouse is emerged as a leading contender to be the data platform of the future. And this novelty touts the ability to address data engineering, data science, and data warehouse workloads on a single shared data platform. The other major trend we've seen is query engines and broader data fabric virtualization platforms have embraced NextGen data lakes as platforms for SQL centric business intelligence workloads, reducing, or somebody even claim eliminating the need for separate data warehouses. Pretty bold. However, cloud data warehouses have added complimentary technologies to bridge the gaps with lakehouses. And the third is many, if not most customers that are embracing the so-called data fabric or data mesh architectures. They're looking at data lakes as a fundamental component of their strategies, and they're trying to evolve them to be more capable, hence the interest in lakehouse, but at the same time, they don't want to, or can't abandon their data warehouse estate. As such we see a battle royale is brewing between cloud data warehouses and cloud lakehouses. Is it possible to do it all with one cloud center analytical data platform? Well, we're going to find out. My name is Dave Vellante and welcome to the data platform's power panel on theCUBE. Our next episode in a series where we gather some of the industry's top analysts to talk about one of our favorite topics, data. In today's session, we'll discuss trends, emerging options, and the trade offs of various approaches and we'll name names. Joining us today are Sanjeev Mohan, who's the principal at SanjMo, Tony Baers, principal at dbInsight. And Doug Henschen is the vice president and principal analyst at Constellation Research. Guys, welcome back to theCUBE. Great to see you again. >> Thank guys. Thank you. >> Thank you. >> So it's early June and we're gearing up with two major conferences, there's several database conferences, but two in particular that were very interested in, Snowflake Summit and Databricks Data and AI Summit. Doug let's start off with you and then Tony and Sanjeev, if you could kindly weigh in. Where did this all start, Doug? The notion of lakehouse. And let's talk about what exactly we mean by lakehouse. Go ahead. >> Yeah, well you nailed it in your intro. One platform to address BI data science, data engineering, fewer platforms, less cost, less complexity, very compelling. You can credit Databricks for coining the term lakehouse back in 2020, but it's really a much older idea. You can go back to Cloudera introducing their Impala database in 2012. That was a database on top of Hadoop. And indeed in that last decade, by the middle of that last decade, there were several SQL on Hadoop products, open standards like Apache Drill. And at the same time, the database vendors were trying to respond to this interest in machine learning and the data science. So they were adding SQL extensions, the likes Hudi and Vertical we're adding SQL extensions to support the data science. But then later in that decade with the shift to cloud and object storage, you saw the vendor shift to this whole cloud, and object storage idea. So you have in the database camp Snowflake introduce Snowpark to try to address the data science needs. They introduced that in 2020 and last year they announced support for Python. You also had Oracle, SAP jumped on this lakehouse idea last year, supporting both the lake and warehouse single vendor, not necessarily quite single platform. Google very recently also jumped on the bandwagon. And then you also mentioned, the SQL engine camp, the Dremios, the Ahanas, the Starbursts, really doing two things, a fabric for distributed access to many data sources, but also very firmly planning that idea that you can just have the lake and we'll help you do the BI workloads on that. And then of course, the data lake camp with the Databricks and Clouderas providing a warehouse style deployments on top of their lake platforms. >> Okay, thanks, Doug. I'd be remiss those of you who me know that I typically write my own intros. This time my colleagues fed me a lot of that material. So thank you. You guys make it easy. But Tony, give us your thoughts on this intro. >> Right. Well, I very much agree with both of you, which may not make for the most exciting television in terms of that it has been an evolution just like Doug said. I mean, for instance, just to give an example when Teradata bought AfterData was initially seen as a hardware platform play. In the end, it was basically, it was all those after functions that made a lot of sort of big data analytics accessible to SQL. (clears throat) And so what I really see just in a more simpler definition or functional definition, the data lakehouse is really an attempt by the data lake folks to make the data lake friendlier territory to the SQL folks, and also to get into friendly territory, to all the data stewards, who are basically concerned about the sprawl and the lack of control in governance in the data lake. So it's really kind of a continuing of an ongoing trend that being said, there's no action without counter action. And of course, at the other end of the spectrum, we also see a lot of the data warehouses starting to edit things like in database machine learning. So they're certainly not surrendering without a fight. Again, as Doug was mentioning, this has been part of a continual blending of platforms that we've seen over the years that we first saw in the Hadoop years with SQL on Hadoop and data warehouses starting to reach out to cloud storage or should say the HDFS and then with the cloud then going cloud native and therefore trying to break the silos down even further. >> Now, thank you. And Sanjeev, data lakes, when we first heard about them, there were such a compelling name, and then we realized all the problems associated with them. So pick it up from there. What would you add to Doug and Tony? >> I would say, these are excellent points that Doug and Tony have brought to light. The concept of lakehouse was going on to your point, Dave, a long time ago, long before the tone was invented. For example, in Uber, Uber was trying to do a mix of Hadoop and Vertical because what they really needed were transactional capabilities that Hadoop did not have. So they weren't calling it the lakehouse, they were using multiple technologies, but now they're able to collapse it into a single data store that we call lakehouse. Data lakes, excellent at batch processing large volumes of data, but they don't have the real time capabilities such as change data capture, doing inserts and updates. So this is why lakehouse has become so important because they give us these transactional capabilities. >> Great. So I'm interested, the name is great, lakehouse. The concept is powerful, but I get concerned that it's a lot of marketing hype behind it. So I want to examine that a bit deeper. How mature is the concept of lakehouse? Are there practical examples that really exist in the real world that are driving business results for practitioners? Tony, maybe you could kick that off. >> Well, put it this way. I think what's interesting is that both data lakes and data warehouse that each had to extend themselves. To believe the Databricks hype it's that this was just a natural extension of the data lake. In point of fact, Databricks had to go outside its core technology of Spark to make the lakehouse possible. And it's a very similar type of thing on the part with data warehouse folks, in terms of that they've had to go beyond SQL, In the case of Databricks. There have been a number of incremental improvements to Delta lake, to basically make the table format more performative, for instance. But the other thing, I think the most dramatic change in all that is in their SQL engine and they had to essentially pretty much abandon Spark SQL because it really, in off itself Spark SQL is essentially stop gap solution. And if they wanted to really address that crowd, they had to totally reinvent SQL or at least their SQL engine. And so Databricks SQL is not Spark SQL, it is not Spark, it's basically SQL that it's adapted to run in a Spark environment, but the underlying engine is C++, it's not scale or anything like that. So Databricks had to take a major detour outside of its core platform to do this. So to answer your question, this is not mature because these are all basically kind of, even though the idea of blending platforms has been going on for well over a decade, I would say that the current iteration is still fairly immature. And in the cloud, I could see a further evolution of this because if you think through cloud native architecture where you're essentially abstracting compute from data, there is no reason why, if let's say you are dealing with say, the same basically data targets say cloud storage, cloud object storage that you might not apportion the task to different compute engines. And so therefore you could have, for instance, let's say you're Google, you could have BigQuery, perform basically the types of the analytics, the SQL analytics that would be associated with the data warehouse and you could have BigQuery ML that does some in database machine learning, but at the same time for another part of the query, which might involve, let's say some deep learning, just for example, you might go out to let's say the serverless spark service or the data proc. And there's no reason why Google could not blend all those into a coherent offering that's basically all triggered through microservices. And I just gave Google as an example, if you could generalize that with all the other cloud or all the other third party vendors. So I think we're still very early in the game in terms of maturity of data lakehouses. >> Thanks, Tony. So Sanjeev, is this all hype? What are your thoughts? >> It's not hype, but completely agree. It's not mature yet. Lakehouses have still a lot of work to do, so what I'm now starting to see is that the world is dividing into two camps. On one hand, there are people who don't want to deal with the operational aspects of vast amounts of data. They are the ones who are going for BigQuery, Redshift, Snowflake, Synapse, and so on because they want the platform to handle all the data modeling, access control, performance enhancements, but these are trade off. If you go with these platforms, then you are giving up on vendor neutrality. On the other side are those who have engineering skills. They want the independence. In other words, they don't want vendor lock in. They want to transform their data into any number of use cases, especially data science, machine learning use case. What they want is agility via open file formats using any compute engine. So why do I say lakehouses are not mature? Well, cloud data warehouses they provide you an excellent user experience. That is the main reason why Snowflake took off. If you have thousands of cables, it takes minutes to get them started, uploaded into your warehouse and start experimentation. Table formats are far more resonating with the community than file formats. But once the cost goes up of cloud data warehouse, then the organization start exploring lakehouses. But the problem is lakehouses still need to do a lot of work on metadata. Apache Hive was a fantastic first attempt at it. Even today Apache Hive is still very strong, but it's all technical metadata and it has so many different restrictions. That's why we see Databricks is investing into something called Unity Catalog. Hopefully we'll hear more about Unity Catalog at the end of the month. But there's a second problem. I just want to mention, and that is lack of standards. All these open source vendors, they're running, what I call ego projects. You see on LinkedIn, they're constantly battling with each other, but end user doesn't care. End user wants a problem to be solved. They want to use Trino, Dremio, Spark from EMR, Databricks, Ahana, DaaS, Frink, Athena. But the problem is that we don't have common standards. >> Right. Thanks. So Doug, I worry sometimes. I mean, I look at the space, we've debated for years, best of breed versus the full suite. You see AWS with whatever, 12 different plus data stores and different APIs and primitives. You got Oracle putting everything into its database. It's actually done some interesting things with MySQL HeatWave, so maybe there's proof points there, but Snowflake really good at data warehouse, simplifying data warehouse. Databricks, really good at making lakehouses actually more functional. Can one platform do it all? >> Well in a word, I can't be best at breed at all things. I think the upshot of and cogen analysis from Sanjeev there, the database, the vendors coming out of the database tradition, they excel at the SQL. They're extending it into data science, but when it comes to unstructured data, data science, ML AI often a compromise, the data lake crowd, the Databricks and such. They've struggled to completely displace the data warehouse when it really gets to the tough SLAs, they acknowledge that there's still a role for the warehouse. Maybe you can size down the warehouse and offload some of the BI workloads and maybe and some of these SQL engines, good for ad hoc, minimize data movement. But really when you get to the deep service level, a requirement, the high concurrency, the high query workloads, you end up creating something that's warehouse like. >> Where do you guys think this market is headed? What's going to take hold? Which projects are going to fade away? You got some things in Apache projects like Hudi and Iceberg, where do they fit Sanjeev? Do you have any thoughts on that? >> So thank you, Dave. So I feel that table formats are starting to mature. There is a lot of work that's being done. We will not have a single product or single platform. We'll have a mixture. So I see a lot of Apache Iceberg in the news. Apache Iceberg is really innovating. Their focus is on a table format, but then Delta and Apache Hudi are doing a lot of deep engineering work. For example, how do you handle high concurrency when there are multiple rights going on? Do you version your Parquet files or how do you do your upcerts basically? So different focus, at the end of the day, the end user will decide what is the right platform, but we are going to have multiple formats living with us for a long time. >> Doug is Iceberg in your view, something that's going to address some of those gaps in standards that Sanjeev was talking about earlier? >> Yeah, Delta lake, Hudi, Iceberg, they all address this need for consistency and scalability, Delta lake open technically, but open for access. I don't hear about Delta lakes in any worlds, but Databricks, hearing a lot of buzz about Apache Iceberg. End users want an open performance standard. And most recently Google embraced Iceberg for its recent a big lake, their stab at having supporting both lakes and warehouses on one conjoined platform. >> And Tony, of course, you remember the early days of the sort of big data movement you had MapR was the most closed. You had Horton works the most open. You had Cloudera in between. There was always this kind of contest as to who's the most open. Does that matter? Are we going to see a repeat of that here? >> I think it's spheres of influence, I think, and Doug very much was kind of referring to this. I would call it kind of like the MongoDB syndrome, which is that you have... and I'm talking about MongoDB before they changed their license, open source project, but very much associated with MongoDB, which basically, pretty much controlled most of the contributions made decisions. And I think Databricks has the same iron cloud hold on Delta lake, but still the market is pretty much associated Delta lake as the Databricks, open source project. I mean, Iceberg is probably further advanced than Hudi in terms of mind share. And so what I see that's breaking down to is essentially, basically the Databricks open source versus the everything else open source, the community open source. So I see it's a very similar type of breakdown that I see repeating itself here. >> So by the way, Mongo has a conference next week, another data platform is kind of not really relevant to this discussion totally. But in the sense it is because there's a lot of discussion on earnings calls these last couple of weeks about consumption and who's exposed, obviously people are concerned about Snowflake's consumption model. Mongo is maybe less exposed because Atlas is prominent in the portfolio, blah, blah, blah. But I wanted to bring up the little bit of controversy that we saw come out of the Snowflake earnings call, where the ever core analyst asked Frank Klutman about discretionary spend. And Frank basically said, look, we're not discretionary. We are deeply operationalized. Whereas he kind of poo-pooed the lakehouse or the data lake, et cetera, saying, oh yeah, data scientists will pull files out and play with them. That's really not our business. Do any of you have comments on that? Help us swing through that controversy. Who wants to take that one? >> Let's put it this way. The SQL folks are from Venus and the data scientists are from Mars. So it means it really comes down to it, sort that type of perception. The fact is, is that, traditionally with analytics, it was very SQL oriented and that basically the quants were kind of off in their corner, where they're using SaaS or where they're using Teradata. It's really a great leveler today, which is that, I mean basic Python it's become arguably one of the most popular programming languages, depending on what month you're looking at, at the title index. And of course, obviously SQL is, as I tell the MongoDB folks, SQL is not going away. You have a large skills base out there. And so basically I see this breaking down to essentially, you're going to have each group that's going to have its own natural preferences for its home turf. And the fact that basically, let's say the Python and scale of folks are using Databricks does not make them any less operational or machine critical than the SQL folks. >> Anybody else want to chime in on that one? >> Yeah, I totally agree with that. Python support in Snowflake is very nascent with all of Snowpark, all of the things outside of SQL, they're very much relying on partners too and make things possible and make data science possible. And it's very early days. I think the bottom line, what we're going to see is each of these camps is going to keep working on doing better at the thing that they don't do today, or they're new to, but they're not going to nail it. They're not going to be best of breed on both sides. So the SQL centric companies and shops are going to do more data science on their database centric platform. That data science driven companies might be doing more BI on their leagues with those vendors and the companies that have highly distributed data, they're going to add fabrics, and maybe offload more of their BI onto those engines, like Dremio and Starburst. >> So I've asked you this before, but I'll ask you Sanjeev. 'Cause Snowflake and Databricks are such great examples 'cause you have the data engineering crowd trying to go into data warehousing and you have the data warehousing guys trying to go into the lake territory. Snowflake has $5 billion in the balance sheet and I've asked you before, I ask you again, doesn't there has to be a semantic layer between these two worlds? Does Snowflake go out and do M&A and maybe buy ad scale or a data mirror? Or is that just sort of a bandaid? What are your thoughts on that Sanjeev? >> I think semantic layer is the metadata. The business metadata is extremely important. At the end of the day, the business folks, they'd rather go to the business metadata than have to figure out, for example, like let's say, I want to update somebody's email address and we have a lot of overhead with data residency laws and all that. I want my platform to give me the business metadata so I can write my business logic without having to worry about which database, which location. So having that semantic layer is extremely important. In fact, now we are taking it to the next level. Now we are saying that it's not just a semantic layer, it's all my KPIs, all my calculations. So how can I make those calculations independent of the compute engine, independent of the BI tool and make them fungible. So more disaggregation of the stack, but it gives us more best of breed products that the customers have to worry about. >> So I want to ask you about the stack, the modern data stack, if you will. And we always talk about injecting machine intelligence, AI into applications, making them more data driven. But when you look at the application development stack, it's separate, the database is tends to be separate from the data and analytics stack. Do those two worlds have to come together in the modern data world? And what does that look like organizationally? >> So organizationally even technically I think it is starting to happen. Microservices architecture was a first attempt to bring the application and the data world together, but they are fundamentally different things. For example, if an application crashes, that's horrible, but Kubernetes will self heal and it'll bring the application back up. But if a database crashes and corrupts your data, we have a huge problem. So that's why they have traditionally been two different stacks. They are starting to come together, especially with data ops, for instance, versioning of the way we write business logic. It used to be, a business logic was highly embedded into our database of choice, but now we are disaggregating that using GitHub, CICD the whole DevOps tool chain. So data is catching up to the way applications are. >> We also have databases, that trans analytical databases that's a little bit of what the story is with MongoDB next week with adding more analytical capabilities. But I think companies that talk about that are always careful to couch it as operational analytics, not the warehouse level workloads. So we're making progress, but I think there's always going to be, or there will long be a separate analytical data platform. >> Until data mesh takes over. (all laughing) Not opening a can of worms. >> Well, but wait, I know it's out of scope here, but wouldn't data mesh say, hey, do take your best of breed to Doug's earlier point. You can't be best of breed at everything, wouldn't data mesh advocate, data lakes do your data lake thing, data warehouse, do your data lake, then you're just a node on the mesh. (Tony laughs) Now you need separate data stores and you need separate teams. >> To my point. >> I think, I mean, put it this way. (laughs) Data mesh itself is a logical view of the world. The data mesh is not necessarily on the lake or on the warehouse. I think for me, the fear there is more in terms of, the silos of governance that could happen and the silo views of the world, how we redefine. And that's why and I want to go back to something what Sanjeev said, which is that it's going to be raising the importance of the semantic layer. Now does Snowflake that opens a couple of Pandora's boxes here, which is one, does Snowflake dare go into that space or do they risk basically alienating basically their partner ecosystem, which is a key part of their whole appeal, which is best of breed. They're kind of the same situation that Informatica was where in the early 2000s, when Informatica briefly flirted with analytic applications and realized that was not a good idea, need to redouble down on their core, which was data integration. The other thing though, that raises the importance of and this is where the best of breed comes in, is the data fabric. My contention is that and whether you use employee data mesh practice or not, if you do employee data mesh, you need data fabric. If you deploy data fabric, you don't necessarily need to practice data mesh. But data fabric at its core and admittedly it's a category that's still very poorly defined and evolving, but at its core, we're talking about a common meta data back plane, something that we used to talk about with master data management, this would be something that would be more what I would say basically, mutable, that would be more evolving, basically using, let's say, machine learning to kind of, so that we don't have to predefine rules or predefine what the world looks like. But so I think in the long run, what this really means is that whichever way we implement on whichever physical platform we implement, we need to all be speaking the same metadata language. And I think at the end of the day, regardless of whether it's a lake, warehouse or a lakehouse, we need common metadata. >> Doug, can I come back to something you pointed out? That those talking about bringing analytic and transaction databases together, you had talked about operationalizing those and the caution there. Educate me on MySQL HeatWave. I was surprised when Oracle put so much effort in that, and you may or may not be familiar with it, but a lot of folks have talked about that. Now it's got nowhere in the market, that no market share, but a lot of we've seen these benchmarks from Oracle. How real is that bringing together those two worlds and eliminating ETL? >> Yeah, I have to defer on that one. That's my colleague, Holger Mueller. He wrote the report on that. He's way deep on it and I'm not going to mock him. >> I wonder if that is something, how real that is or if it's just Oracle marketing, anybody have any thoughts on that? >> I'm pretty familiar with HeatWave. It's essentially Oracle doing what, I mean, there's kind of a parallel with what Google's doing with AlloyDB. It's an operational database that will have some embedded analytics. And it's also something which I expect to start seeing with MongoDB. And I think basically, Doug and Sanjeev were kind of referring to this before about basically kind of like the operational analytics, that are basically embedded within an operational database. The idea here is that the last thing you want to do with an operational database is slow it down. So you're not going to be doing very complex deep learning or anything like that, but you might be doing things like classification, you might be doing some predictives. In other words, we've just concluded a transaction with this customer, but was it less than what we were expecting? What does that mean in terms of, is this customer likely to turn? I think we're going to be seeing a lot of that. And I think that's what a lot of what MySQL HeatWave is all about. Whether Oracle has any presence in the market now it's still a pretty new announcement, but the other thing that kind of goes against Oracle, (laughs) that they had to battle against is that even though they own MySQL and run the open source project, everybody else, in terms of the actual commercial implementation it's associated with everybody else. And the popular perception has been that MySQL has been basically kind of like a sidelight for Oracle. And so it's on Oracles shoulders to prove that they're damn serious about it. >> There's no coincidence that MariaDB was launched the day that Oracle acquired Sun. Sanjeev, I wonder if we could come back to a topic that we discussed earlier, which is this notion of consumption, obviously Wall Street's very concerned about it. Snowflake dropped prices last week. I've always felt like, hey, the consumption model is the right model. I can dial it down in when I need to, of course, the street freaks out. What are your thoughts on just pricing, the consumption model? What's the right model for companies, for customers? >> Consumption model is here to stay. What I would like to see, and I think is an ideal situation and actually plays into the lakehouse concept is that, I have my data in some open format, maybe it's Parquet or CSV or JSON, Avro, and I can bring whatever engine is the best engine for my workloads, bring it on, pay for consumption, and then shut it down. And by the way, that could be Cloudera. We don't talk about Cloudera very much, but it could be one business unit wants to use Athena. Another business unit wants to use some other Trino let's say or Dremio. So every business unit is working on the same data set, see that's critical, but that data set is maybe in their VPC and they bring any compute engine, you pay for the use, shut it down. That then you're getting value and you're only paying for consumption. It's not like, I left a cluster running by mistake, so there have to be guardrails. The reason FinOps is so big is because it's very easy for me to run a Cartesian joint in the cloud and get a $10,000 bill. >> This looks like it's been a sort of a victim of its own success in some ways, they made it so easy to spin up single note instances, multi note instances. And back in the day when compute was scarce and costly, those database engines optimized every last bit so they could get as much workload as possible out of every instance. Today, it's really easy to spin up a new node, a new multi node cluster. So that freedom has meant many more nodes that aren't necessarily getting that utilization. So Snowflake has been doing a lot to add reporting, monitoring, dashboards around the utilization of all the nodes and multi node instances that have spun up. And meanwhile, we're seeing some of the traditional on-prem databases that are moving into the cloud, trying to offer that freedom. And I think they're going to have that same discovery that the cost surprises are going to follow as they make it easy to spin up new instances. >> Yeah, a lot of money went into this market over the last decade, separating compute from storage, moving to the cloud. I'm glad you mentioned Cloudera Sanjeev, 'cause they got it all started, the kind of big data movement. We don't talk about them that much. Sometimes I wonder if it's because when they merged Hortonworks and Cloudera, they dead ended both platforms, but then they did invest in a more modern platform. But what's the future of Cloudera? What are you seeing out there? >> Cloudera has a good product. I have to say the problem in our space is that there're way too many companies, there's way too much noise. We are expecting the end users to parse it out or we expecting analyst firms to boil it down. So I think marketing becomes a big problem. As far as technology is concerned, I think Cloudera did turn their selves around and Tony, I know you, you talked to them quite frequently. I think they have quite a comprehensive offering for a long time actually. They've created Kudu, so they got operational, they have Hadoop, they have an operational data warehouse, they're migrated to the cloud. They are in hybrid multi-cloud environment. Lot of cloud data warehouses are not hybrid. They're only in the cloud. >> Right. I think what Cloudera has done the most successful has been in the transition to the cloud and the fact that they're giving their customers more OnRamps to it, more hybrid OnRamps. So I give them a lot of credit there. They're also have been trying to position themselves as being the most price friendly in terms of that we will put more guardrails and governors on it. I mean, part of that could be spin. But on the other hand, they don't have the same vested interest in compute cycles as say, AWS would have with EMR. That being said, yes, Cloudera does it, I think its most powerful appeal so of that, it almost sounds in a way, I don't want to cast them as a legacy system. But the fact is they do have a huge landed legacy on-prem and still significant potential to land and expand that to the cloud. That being said, even though Cloudera is multifunction, I think it certainly has its strengths and weaknesses. And the fact this is that yes, Cloudera has an operational database or an operational data store with a kind of like the outgrowth of age base, but Cloudera is still based, primarily known for the deep analytics, the operational database nobody's going to buy Cloudera or Cloudera data platform strictly for the operational database. They may use it as an add-on, just in the same way that a lot of customers have used let's say Teradata basically to do some machine learning or let's say, Snowflake to parse through JSON. Again, it's not an indictment or anything like that, but the fact is obviously they do have their strengths and their weaknesses. I think their greatest opportunity is with their existing base because that base has a lot invested and vested. And the fact is they do have a hybrid path that a lot of the others lack. >> And of course being on the quarterly shock clock was not a good place to be under the microscope for Cloudera and now they at least can refactor the business accordingly. I'm glad you mentioned hybrid too. We saw Snowflake last month, did a deal with Dell whereby non-native Snowflake data could access on-prem object store from Dell. They announced a similar thing with pure storage. What do you guys make of that? Is that just... How significant will that be? Will customers actually do that? I think they're using either materialized views or extended tables. >> There are data rated and residency requirements. There are desires to have these platforms in your own data center. And finally they capitulated, I mean, Frank Klutman is famous for saying to be very focused and earlier, not many months ago, they called the going on-prem as a distraction, but clearly there's enough demand and certainly government contracts any company that has data residency requirements, it's a real need. So they finally addressed it. >> Yeah, I'll bet dollars to donuts, there was an EBC session and some big customer said, if you don't do this, we ain't doing business with you. And that was like, okay, we'll do it. >> So Dave, I have to say, earlier on you had brought this point, how Frank Klutman was poo-pooing data science workloads. On your show, about a year or so ago, he said, we are never going to on-prem. He burnt that bridge. (Tony laughs) That was on your show. >> I remember exactly the statement because it was interesting. He said, we're never going to do the halfway house. And I think what he meant is we're not going to bring the Snowflake architecture to run on-prem because it defeats the elasticity of the cloud. So this was kind of a capitulation in a way. But I think it still preserves his original intent sort of, I don't know. >> The point here is that every vendor will poo-poo whatever they don't have until they do have it. >> Yes. >> And then it'd be like, oh, we are all in, we've always been doing this. We have always supported this and now we are doing it better than others. >> Look, it was the same type of shock wave that we felt basically when AWS at the last moment at one of their reinvents, oh, by the way, we're going to introduce outposts. And the analyst group is typically pre briefed about a week or two ahead under NDA and that was not part of it. And when they dropped, they just casually dropped that in the analyst session. It's like, you could have heard the sound of lots of analysts changing their diapers at that point. >> (laughs) I remember that. And a props to Andy Jassy who once, many times actually told us, never say never when it comes to AWS. So guys, I know we got to run. We got some hard stops. Maybe you could each give us your final thoughts, Doug start us off and then-- >> Sure. Well, we've got the Snowflake Summit coming up. I'll be looking for customers that are really doing data science, that are really employing Python through Snowflake, through Snowpark. And then a couple weeks later, we've got Databricks with their Data and AI Summit in San Francisco. I'll be looking for customers that are really doing considerable BI workloads. Last year I did a market overview of this analytical data platform space, 14 vendors, eight of them claim to support lakehouse, both sides of the camp, Databricks customer had 32, their top customer that they could site was unnamed. It had 32 concurrent users doing 15,000 queries per hour. That's good but it's not up to the most demanding BI SQL workloads. And they acknowledged that and said, they need to keep working that. Snowflake asked for their biggest data science customer, they cited Kabura, 400 terabytes, 8,500 users, 400,000 data engineering jobs per day. I took the data engineering job to be probably SQL centric, ETL style transformation work. So I want to see the real use of the Python, how much Snowpark has grown as a way to support data science. >> Great. Tony. >> Actually of all things. And certainly, I'll also be looking for similar things in what Doug is saying, but I think sort of like, kind of out of left field, I'm interested to see what MongoDB is going to start to say about operational analytics, 'cause I mean, they're into this conquer the world strategy. We can be all things to all people. Okay, if that's the case, what's going to be a case with basically, putting in some inline analytics, what are you going to be doing with your query engine? So that's actually kind of an interesting thing we're looking for next week. >> Great. Sanjeev. >> So I'll be at MongoDB world, Snowflake and Databricks and very interested in seeing, but since Tony brought up MongoDB, I see that even the databases are shifting tremendously. They are addressing both the hashtag use case online, transactional and analytical. I'm also seeing that these databases started in, let's say in case of MySQL HeatWave, as relational or in MongoDB as document, but now they've added graph, they've added time series, they've added geospatial and they just keep adding more and more data structures and really making these databases multifunctional. So very interesting. >> It gets back to our discussion of best of breed, versus all in one. And it's likely Mongo's path or part of their strategy of course, is through developers. They're very developer focused. So we'll be looking for that. And guys, I'll be there as well. I'm hoping that we maybe have some extra time on theCUBE, so please stop by and we can maybe chat a little bit. Guys as always, fantastic. Thank you so much, Doug, Tony, Sanjeev, and let's do this again. >> It's been a pleasure. >> All right and thank you for watching. This is Dave Vellante for theCUBE and the excellent analyst. We'll see you next time. (upbeat music)

Published Date : Jun 2 2022

SUMMARY :

And Doug Henschen is the vice president Thank you. Doug let's start off with you And at the same time, me a lot of that material. And of course, at the and then we realized all the and Tony have brought to light. So I'm interested, the And in the cloud, So Sanjeev, is this all hype? But the problem is that we I mean, I look at the space, and offload some of the So different focus, at the end of the day, and warehouses on one conjoined platform. of the sort of big data movement most of the contributions made decisions. Whereas he kind of poo-pooed the lakehouse and the data scientists are from Mars. and the companies that have in the balance sheet that the customers have to worry about. the modern data stack, if you will. and the data world together, the story is with MongoDB Until data mesh takes over. and you need separate teams. that raises the importance of and the caution there. Yeah, I have to defer on that one. The idea here is that the of course, the street freaks out. and actually plays into the And back in the day when the kind of big data movement. We are expecting the end And the fact is they do have a hybrid path refactor the business accordingly. saying to be very focused And that was like, okay, we'll do it. So Dave, I have to say, the Snowflake architecture to run on-prem The point here is that and now we are doing that in the analyst session. And a props to Andy Jassy and said, they need to keep working that. Great. Okay, if that's the case, Great. I see that even the databases I'm hoping that we maybe have and the excellent analyst.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DougPERSON

0.99+

Dave VellantePERSON

0.99+

DavePERSON

0.99+

TonyPERSON

0.99+

UberORGANIZATION

0.99+

FrankPERSON

0.99+

Frank KlutmanPERSON

0.99+

Tony BaersPERSON

0.99+

MarsLOCATION

0.99+

Doug HenschenPERSON

0.99+

2020DATE

0.99+

AWSORGANIZATION

0.99+

VenusLOCATION

0.99+

OracleORGANIZATION

0.99+

2012DATE

0.99+

DatabricksORGANIZATION

0.99+

DellORGANIZATION

0.99+

HortonworksORGANIZATION

0.99+

Holger MuellerPERSON

0.99+

Andy JassyPERSON

0.99+

last yearDATE

0.99+

$5 billionQUANTITY

0.99+

$10,000QUANTITY

0.99+

14 vendorsQUANTITY

0.99+

Last yearDATE

0.99+

last weekDATE

0.99+

San FranciscoLOCATION

0.99+

SanjMoORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

8,500 usersQUANTITY

0.99+

SanjeevPERSON

0.99+

InformaticaORGANIZATION

0.99+

32 concurrent usersQUANTITY

0.99+

twoQUANTITY

0.99+

Constellation ResearchORGANIZATION

0.99+

MongoORGANIZATION

0.99+

Sanjeev MohanPERSON

0.99+

AhanaORGANIZATION

0.99+

DaaSORGANIZATION

0.99+

EMRORGANIZATION

0.99+

32QUANTITY

0.99+

AtlasORGANIZATION

0.99+

DeltaORGANIZATION

0.99+

SnowflakeORGANIZATION

0.99+

PythonTITLE

0.99+

eachQUANTITY

0.99+

AthenaORGANIZATION

0.99+

next weekDATE

0.99+

Breaking Analysis: Technology & Architectural Considerations for Data Mesh


 

>> From theCUBE Studios in Palo Alto and Boston, bringing you data driven insights from theCUBE in ETR, this is Breaking Analysis with Dave Vellante. >> The introduction in socialization of data mesh has caused practitioners, business technology executives, and technologists to pause, and ask some probing questions about the organization of their data teams, their data strategies, future investments, and their current architectural approaches. Some in the technology community have embraced the concept, others have twisted the definition, while still others remain oblivious to the momentum building around data mesh. Here we are in the early days of data mesh adoption. Organizations that have taken the plunge will tell you that aligning stakeholders is a non-trivial effort, but necessary to break through the limitations that monolithic data architectures and highly specialized teams have imposed over frustrated business and domain leaders. However, practical data mesh examples often lie in the eyes of the implementer, and may not strictly adhere to the principles of data mesh. Now, part of the problem is lack of open technologies and standards that can accelerate adoption and reduce friction, and that's what we're going to talk about today. Some of the key technology and architecture questions around data mesh. Hello, and welcome to this week's Wikibon CUBE Insights powered by ETR, and in this Breaking Analysis, we welcome back the founder of data mesh and director of Emerging Technologies at Thoughtworks, Zhamak Dehghani. Hello, Zhamak. Thanks for being here today. >> Hi Dave, thank you for having me back. It's always a delight to connect and have a conversation. Thank you. >> Great, looking forward to it. Okay, so before we get into it in the technology details, I just want to quickly share some data from our friends at ETR. You know, despite the importance of data initiative since the pandemic, CIOs and IT organizations have had to juggle of course, a few other priorities, this is why in the survey data, cyber and cloud computing are rated as two most important priorities. Analytics and machine learning, and AI, which are kind of data topics, still make the top of the list, well ahead of many other categories. And look, a sound data architecture and strategy is fundamental to digital transformations, and much of the past two years, as we've often said, has been like a forced march into digital. So while organizations are moving forward, they really have to think hard about the data architecture decisions that they make, because it's going to impact them, Zhamak, for years to come, isn't it? >> Yes, absolutely. I mean, we are moving really from, slowly moving from reason based logical algorithmic to model based computation and decision making, where we exploit the patterns and signals within the data. So data becomes a very important ingredient, of not only decision making, and analytics and discovering trends, but also the features and applications that we build for the future. So we can't really ignore it, and as we see, some of the existing challenges around getting value from data is not necessarily that no longer is access to computation, is actually access to trustworthy, reliable data at scale. >> Yeah, and you see these domains coming together with the cloud and obviously it has to be secure and trusted, and that's why we're here today talking about data mesh. So let's get into it. Zhamak, first, your new book is out, 'Data Mesh: Delivering Data-Driven Value at Scale' just recently published, so congratulations on getting that done, awesome. Now in a recent presentation, you pulled excerpts from the book and we're going to talk through some of the technology and architectural considerations. Just quickly for the audience, four principles of data mesh. Domain driven ownership, data as product, self-served data platform and federated computational governance. So I want to start with self-serve platform and some of the data that you shared recently. You say that, "Data mesh serves autonomous domain oriented teams versus existing platforms, which serve a centralized team." Can you elaborate? >> Sure. I mean the role of the platform is to lower the cognitive load for domain teams, for people who are focusing on the business outcomes, the technologies that are building the applications, to really lower the cognitive load for them, to be able to work with data. Whether they are building analytics, automated decision making, intelligent modeling. They need to be able to get access to data and use it. So the role of the platform, I guess, just stepping back for a moment is to empower and enable these teams. Data mesh by definition is a scale out model. It's a decentralized model that wants to give autonomy to cross-functional teams. So it is core requires a set of tools that work really well in that decentralized model. When we look at the existing platforms, they try to achieve this similar outcome, right? Lower the cognitive load, give the tools to data practitioners, to manage data at scale because today centralized teams, really their job, the centralized data teams, their job isn't really directly aligned with a one or two or different, you know, business units and business outcomes in terms of getting value from data. Their job is manage the data and make the data available for then those cross-functional teams or business units to use the data. So the platforms they've been given are really centralized around or tuned to work with this structure as a team, structure of centralized team. Although on the surface, it seems that why not? Why can't I use my, you know, cloud storage or computation or data warehouse in a decentralized way? You should be able to, but some changes need to happen to those online platforms. As an example, some cloud providers simply have hard limits on the number of like account storage, storage accounts that you can have. Because they never envisaged you have hundreds of lakes. They envisage one or two, maybe 10 lakes, right. They envisage really centralizing data, not decentralizing data. So I think we see a shift in thinking about enabling autonomous independent teams versus a centralized team. >> So just a follow up if I may, we could be here for a while. But so this assumes that you've sorted out the organizational considerations? That you've defined all the, what a data product is and a sub product. And people will say, of course we use the term monolithic as a pejorative, let's face it. But the data warehouse crowd will say, "Well, that's what data march did. So we got that covered." But Europe... The primest of data mesh, if I understand it is whether it's a data march or a data mart or a data warehouse, or a data lake or whatever, a snowflake warehouse, it's a node on the mesh. Okay. So don't build your organization around the technology, let the technology serve the organization is that-- >> That's a perfect way of putting it, exactly. I mean, for a very long time, when we look at decomposition of complexity, we've looked at decomposition of complexity around technology, right? So we have technology and that's maybe a good segue to actually the next item on that list that we looked at. Oh, I need to decompose based on whether I want to have access to raw data and put it on the lake. Whether I want to have access to model data and put it on the warehouse. You know I need to have a team in the middle to move the data around. And then try to figure organization into that model. So data mesh really inverses that, and as you said, is look at the organizational structure first. Then scale boundaries around which your organization and operation can scale. And then the second layer look at the technology and how you decompose it. >> Okay. So let's go to that next point and talk about how you serve and manage autonomous interoperable data products. Where code, data policy you say is treated as one unit. Whereas your contention is existing platforms of course have independent management and dashboards for catalogs or storage, et cetera. Maybe we double click on that a bit. >> Yeah. So if you think about that functional, or technical decomposition, right? Of concerns, that's one way, that's a very valid way of decomposing, complexity and concerns. And then build solutions, independent solutions to address them. That's what we see in the technology landscape today. We will see technologies that are taking care of your management of data, bring your data under some sort of a control and modeling. You'll see technology that moves that data around, will perform various transformations and computations on it. And then you see technology that tries to overlay some level of meaning. Metadata, understandability, discovery was the end policy, right? So that's where your data processing kind of pipeline technologies versus data warehouse, storage, lake technologies, and then the governance come to play. And over time, we decomposed and we compose, right? Deconstruct and reconstruct back this together. But, right now that's where we stand. I think for data mesh really to become a reality, as in independent sources of data and teams can responsibly share data in a way that can be understood right then and there can impose policies, right then when the data gets accessed in that source and in a resilient manner, like in a way that data changes structure of the data or changes to the scheme of the data, doesn't have those downstream down times. We've got to think about this new nucleus or new units of data sharing. And we need to really bring back transformation and governing data and the data itself together around these decentralized nodes on the mesh. So that's another, I guess, deconstruction and reconstruction that needs to happen around the technology to formulate ourselves around the domains. And again the data and the logic of the data itself, the meaning of the data itself. >> Great. Got it. And we're going to talk more about the importance of data sharing and the implications. But the third point deals with how operational, analytical technologies are constructed. You've got an app DevStack, you've got a data stack. You've made the point many times actually that we've contextualized our operational systems, but not our data systems, they remain separate. Maybe you could elaborate on this point. >> Yes. I think this is, again, has a historical background and beginning. For a really long time, applications have dealt with features and the logic of running the business and encapsulating the data and the state that they need to run that feature or run that business function. And then we had for anything analytical driven, which required access data across these applications and across the longer dimension of time around different subjects within the organization. This analytical data, we had made a decision that, "Okay, let's leave those applications aside. Let's leave those databases aside. We'll extract the data out and we'll load it, or we'll transform it and put it under the analytical kind of a data stack and then downstream from it, we will have analytical data users, the data analysts, the data sciences and the, you know, the portfolio of users that are growing use that data stack. And that led to this really separation of dual stack with point to point integration. So applications went down the path of transactional databases or urban document store, but using APIs for communicating and then we've gone to, you know, lake storage or data warehouse on the other side. If we are moving and that again, enforces the silo of data versus app, right? So if we are moving to the world that our missions that are ambitions around making applications, more intelligent. Making them data driven. These two worlds need to come closer. As in ML Analytics gets embedded into those app applications themselves. And the data sharing, as a very essential ingredient of that, gets embedded and gets closer, becomes closer to those applications. So, if you are looking at this now cross-functional, app data, based team, right? Business team, then the technology stacks can't be so segregated, right? There has to be a continuum of experience from app delivery, to sharing of the data, to using that data, to embed models back into those applications. And that continuum of experience requires well integrated technologies. I'll give you an example, which actually in some sense, we are somewhat moving to that direction. But if we are talking about data sharing or data modeling and applications use one set of APIs, you know, HTTP compliant, GraQL or RAC APIs. And on the other hand, you have proprietary SQL, like connect to my database and run SQL. Like those are very two different models of representing and accessing data. So we kind of have to harmonize or integrate those two worlds a bit more closely to achieve that domain oriented cross-functional teams. >> Yeah. We are going to talk about some of the gaps later and actually you look at them as opportunities, more than barriers. But they are barriers, but they're opportunities for more innovation. Let's go on to the fourth one. The next point, it deals with the roles that the platform serves. Data mesh proposes that domain experts own the data and take responsibility for it end to end and are served by the technology. Kind of, we referenced that before. Whereas your contention is that today, data systems are really designed for specialists. I think you use the term hyper specialists a lot. I love that term. And the generalist are kind of passive bystanders waiting in line for the technical teams to serve them. >> Yes. I mean, if you think about the, again, the intention behind data mesh was creating a responsible data sharing model that scales out. And I challenge any organization that has a scaled ambitions around data or usage of data that relies on small pockets of very expensive specialists resources, right? So we have no choice, but upscaling cross-scaling. The majority population of our technologists, we often call them generalists, right? That's a short hand for people that can really move from one technology to another technology. Sometimes we call them pandric people sometimes we call them T-shaped people. But regardless, like we need to have ability to really mobilize our generalists. And we had to do that at Thoughtworks. We serve a lot of our clients and like many other organizations, we are also challenged with hiring specialists. So we have tested the model of having a few specialists, really conveying and translating the knowledge to generalists and bring them forward. And of course, platform is a big enabler of that. Like what is the language of using the technology? What are the APIs that delight that generalist experience? This doesn't mean no code, low code. We have to throw away in to good engineering practices. And I think good software engineering practices remain to exist. Of course, they get adopted to the world of data to build resilient you know, sustainable solutions, but specialty, especially around kind of proprietary technology is going to be a hard one to scale. >> Okay. I'm definitely going to come back and pick your brain on that one. And, you know, your point about scale out in the examples, the practical examples of companies that have implemented data mesh that I've talked to. I think in all cases, you know, there's only a handful that I've really gone deep with, but it was their hadoop instances, their clusters wouldn't scale, they couldn't scale the business and around it. So that's really a key point of a common pattern that we've seen now. I think in all cases, they went to like the data lake model and AWS. And so that maybe has some violation of the principles, but we'll come back to that. But so let me go on to the next one. Of course, data mesh leans heavily, toward this concept of decentralization, to support domain ownership over the centralized approaches. And we certainly see this, the public cloud players, database companies as key actors here with very large install bases, pushing a centralized approach. So I guess my question is, how realistic is this next point where you have decentralized technologies ruling the roost? >> I think if you look at the history of places, in our industry where decentralization has succeeded, they heavily relied on standardization of connectivity with, you know, across different components of technology. And I think right now you are right. The way we get value from data relies on collection. At the end of the day, collection of data. Whether you have a deep learning machinery model that you're training, or you have, you know, reports to generate. Regardless, the model is bring your data to a place that you can collect it, so that we can use it. And that leads to a naturally set of technologies that try to operate as a full stack integrated proprietary with no intention of, you know, opening, data for sharing. Now, conversely, if you think about internet itself, web itself, microservices, even at the enterprise level, not at the planetary level, they succeeded as decentralized technologies to a large degree because of their emphasis on open net and openness and sharing, right. API sharing. We don't talk about, in the API worlds, like we don't say, you know, "I will build a platform to manage your logical applications." Maybe to a degree but we actually moved away from that. We say, "I'll build a platform that opens around applications to manage your APIs, manage your interfaces." Right? Give you access to API. So I think the shift needs to... That definition of decentralized there means really composable, open pieces of the technology that can play nicely with each other, rather than a full stack, all have control of your data yet being somewhat decentralized within the boundary of my platform. That's just simply not going to scale if data needs to come from different platforms, different locations, different geographical locations, it needs to rethink. >> Okay, thank you. And then the final point is, is data mesh favors technologies that are domain agnostic versus those that are domain aware. And I wonder if you could help me square the circle cause it's nuanced and I'm kind of a 100 level student of your work. But you have said for example, that the data teams lack context of the domain and so help us understand what you mean here in this case. >> Sure. Absolutely. So as you said, we want to take... Data mesh tries to give autonomy and decision making power and responsibility to people that have the context of those domains, right? The people that are really familiar with different business domains and naturally the data that that domain needs, or that naturally the data that domains shares. So if the intention of the platform is really to give the power to people with most relevant and timely context, the platform itself naturally becomes as a shared component, becomes domain agnostic to a large degree. Of course those domains can still... The platform is a (chuckles) fairly overloaded world. As in, if you think about it as a set of technology that abstracts complexity and allows building the next level solutions on top, those domains may have their own set of platforms that are very much doing agnostic. But as a generalized shareable set of technologies or tools that allows us share data. So that piece of technology needs to relinquish the knowledge of the context to the domain teams and actually becomes domain agnostic. >> Got it. Okay. Makes sense. All right. Let's shift gears here. Talk about some of the gaps and some of the standards that are needed. You and I have talked about this a little bit before, but this digs deeper. What types of standards are needed? Maybe you could walk us through this graphic, please. >> Sure. So what I'm trying to depict here is that if we imagine a world that data can be shared from many different locations, for a variety of analytical use cases, naturally the boundary of what we call a node on the mesh will encapsulates internally a fair few pieces. It's not just the boundary of that, not on the mesh, is the data itself that it's controlling and updating and maintaining. It's of course a computation and the code that's responsible for that data. And then the policies that continue to govern that data as long as that data exists. So if that's the boundary, then if we shift that focus from implementation details, that we can leave that for later, what becomes really important is the scene or the APIs and interfaces that this node exposes. And I think that's where the work that needs to be done and the standards that are missing. And we want the scene and those interfaces be open because that allows, you know, different organizations with different boundaries of trust to share data. Not only to share data to kind of move that data to yes, another location, to share the data in a way that distributed workloads, distributed analytics, distributed machine learning model can happen on the data where it is. So if you follow that line of thinking around the centralization and connection of data versus collection of data, I think the very, very important piece of it that needs really deep thinking, and I don't claim that I have done that, is how do we share data responsibly and sustainably, right? That is not brittle. If you think about it today, the ways we share data, one of the very common ways is around, I'll give you a JDC endpoint, or I give you an endpoint to your, you know, database of choice. And now as technology, whereas a user actually, you can now have access to the schema of the underlying data and then run various queries or SQL queries on it. That's very simple and easy to get started with. That's why SQL is an evergreen, you know, standard or semi standard, pseudo standard that we all use. But it's also very brittle, because we are dependent on a underlying schema and formatting of the data that's been designed to tell the computer how to store and manage the data. So I think that the data sharing APIs of the future really need to think about removing this brittle dependencies, think about sharing, not only the data, but what we call metadata, I suppose. Additional set of characteristics that is always shared along with data to make the data usage, I suppose ethical and also friendly for the users and also, I think we have to... That data sharing API, the other element of it, is to allow kind of computation to run where the data exists. So if you think about SQL again, as a simple primitive example of computation, when we select and when we filter and when we join, the computation is happening on that data. So maybe there is a next level of articulating, distributed computational data that simply trains models, right? Your language primitives change in a way to allow sophisticated analytical workloads run on the data more responsibly with policies and access control and force. So I think that output port that I mentioned simply is about next generation data sharing, responsible data sharing APIs. Suitable for decentralized analytical workloads. >> So I'm not trying to bait you here, but I have a follow up as well. So you schema, for all its good creates constraints. No schema on right, that didn't work, cause it was just a free for all and it created the data swamps. But now you have technology companies trying to solve that problem. Take Snowflake for example, you know, enabling, data sharing. But it is within its proprietary environment. Certainly Databricks doing something, you know, trying to come at it from its angle, bringing some of the best to data warehouse, with the data science. Is your contention that those remain sort of proprietary and defacto standards? And then what we need is more open standards? Maybe you could comment. >> Sure. I think the two points one is, as you mentioned. Open standards that allow... Actually make the underlying platform invisible. I mean my litmus test for a technology provider to say, "I'm a data mesh," (laughs) kind of compliant is, "Is your platform invisible?" As in, can I replace it with another and yet get the similar data sharing experience that I need? So part of it is that. Part of it is open standards, they're not really proprietary. The other angle for kind of sharing data across different platforms so that you know, we don't get stuck with one technology or another is around APIs. It is around code that is protecting that internal schema. So where we are on the curve of evolution of technology, right now we are exposing the internal structure of the data. That is designed to optimize certain modes of access. We're exposing that to the end client and application APIs, right? So the APIs that use the data today are very much aware that this database was optimized for machine learning workloads. Hence you will deal with a columnar storage of the file versus this other API is optimized for a very different, report type access, relational access and is optimized around roles. I think that should become irrelevant in the API sharing of the future. Because as a user, I shouldn't care how this data is internally optimized, right? The language primitive that I'm using should be really agnostic to the machine optimization underneath that. And if we did that, perhaps this war between warehouse or lake or the other will become actually irrelevant. So we're optimizing for that human best human experience, as opposed to the best machine experience. We still have to do that but we have to make that invisible. Make that an implementation concern. So that's another angle of what should... If we daydream together, the best experience and resilient experience in terms of data usage than these APIs with diagnostics to the internal storage structure. >> Great, thank you for that. We've wrapped our ankles now on the controversy, so we might as well wade all the way in, I can't let you go without addressing some of this. Which you've catalyzed, which I, by the way, I see as a sign of progress. So this gentleman, Paul Andrew is an architect and he gave a presentation I think last night. And he teased it as quote, "The theory from Zhamak Dehghani versus the practical experience of a technical architect, AKA me," meaning him. And Zhamak, you were quick to shoot back that data mesh is not theory, it's based on practice. And some practices are experimental. Some are more baked and data mesh really avoids by design, the specificity of vendor or technology. Perhaps you intend to frame your post as a technology or vendor specific, specific implementation. So touche, that was excellent. (Zhamak laughs) Now you don't need me to defend you, but I will anyway. You spent 14 plus years as a software engineer and the better part of a decade consulting with some of the most technically advanced companies in the world. But I'm going to push you a little bit here and say, some of this tension is of your own making because you purposefully don't talk about technologies and vendors. Sometimes doing so it's instructive for us neophytes. So, why don't you ever like use specific examples of technology for frames of reference? >> Yes. My role is pushes to the next level. So, you know everybody picks their fights, pick their battles. My role in this battle is to push us to think beyond what's available today. Of course, that's my public persona. On a day to day basis, actually I work with clients and existing technology and I think at Thoughtworks we have given the talk we gave a case study talk with a colleague of mine and I intentionally got him to talk about (indistinct) I want to talk about the technology that we use to implement data mesh. And the reason I haven't really embraced, in my conversations, the specific technology. One is, I feel the technology solutions we're using today are still not ready for the vision. I mean, we have to be in this transitional step, no matter what we have to be pragmatic, of course, and practical, I suppose. And use the existing vendors that exist and I wholeheartedly embrace that, but that's just not my role, to show that. I've gone through this transformation once before in my life. When microservices happened, we were building microservices like architectures with technology that wasn't ready for it. Big application, web application servers that were designed to run these giant monolithic applications. And now we're trying to run little microservices onto them. And the tail was riding the dock, the environmental complexity of running these services was consuming so much of our effort that we couldn't really pay attention to that business logic, the business value. And that's where we are today. The complexity of integrating existing technologies is really overwhelmingly, capturing a lot of our attention and cost and effort, money and effort as opposed to really focusing on the data product themselves. So it's just that's the role I have, but it doesn't mean that, you know, we have to rebuild the world. We've got to do with what we have in this transitional phase until the new generation, I guess, technologies come around and reshape our landscape of tools. >> Well, impressive public discipline. Your point about microservice is interesting because a lot of those early microservices, weren't so micro and for the naysayers look past this, not prologue, but Thoughtworks was really early on in the whole concept of microservices. So be very excited to see how this plays out. But now there was some other good comments. There was one from a gentleman who said the most interesting aspects of data mesh are organizational. And that's how my colleague Sanji Mohan frames data mesh versus data fabric. You know, I'm not sure, I think we've sort of scratched the surface today that data today, data mesh is more. And I still think data fabric is what NetApp defined as software defined storage infrastructure that can serve on-prem and public cloud workloads back whatever, 2016. But the point you make in the thread that we're showing you here is that you're warning, and you referenced this earlier, that the segregating different modes of access will lead to fragmentation. And we don't want to repeat the mistakes of the past. >> Yes, there are comments around. Again going back to that original conversation that we have got this at a macro level. We've got this tendency to decompose complexity based on technical solutions. And, you know, the conversation could be, "Oh, I do batch or you do a stream and we are different."' They create these bifurcations in our decisions based on the technology where I do events and you do tables, right? So that sort of segregation of modes of access causes accidental complexity that we keep dealing with. Because every time in this tree, you create a new branch, you create new kind of new set of tools and then somehow need to be point to point integrated. You create new specialization around that. So the least number of branches that we have, and think about really about the continuum of experiences that we need to create and technologies that simplify, that continuum experience. So one of the things, for example, give you a past experience. I was really excited around the papers and the work that came around on Apache Beam, and generally flow based programming and stream processing. Because basically they were saying whether you are doing batch or whether you're doing streaming, it's all one stream. And sometimes the window of time, narrows and sometimes the window of time over which you're computing, widens and at the end of today, is you are just getting... Doing the stream processing. So it is those sort of notions that simplify and create continuum of experience. I think resonate with me personally, more than creating these tribal fights of this type versus that mode of access. So that's why data mesh naturally selects kind of this multimodal access to support end users, right? The persona of end users. >> Okay. So the last topic I want to hit, this whole discussion, the topic of data mesh it's highly nuanced, it's new, and people are going to shoehorn data mesh into their respective views of the world. And we talked about lake houses and there's three buckets. And of course, the gentleman from LinkedIn with Azure, Microsoft has a data mesh community. See you're going to have to enlist some serious army of enforcers to adjudicate. And I wrote some of the stuff down. I mean, it's interesting. Monte Carlo has a data mesh calculator. Starburst is leaning in, chaos. Search sees themselves as an enabler. Oracle and Snowflake both use the term data mesh. And then of course you've got big practitioners J-P-M-C, we've talked to Intuit, Orlando, HelloFresh has been on, Netflix has this event based sort of streaming implementation. So my question is, how realistic is it that the clarity of your vision can be implemented and not polluted by really rich technology companies and others? (Zhamak laughs) >> Is it even possible, right? Is it even possible? That's a yes. That's why I practice then. This is why I should practice things. Cause I think, it's going to be hard. What I'm hopeful, is that the socio-technical, Leveling Data mentioned that this is a socio-technical concern or solution, not just a technology solution. Hopefully always brings us back to, you know, the reality that vendors try to sell you safe oil that solves all of your problems. (chuckles) All of your data mesh problems. It's just going to cause more problem down the track. So we'll see, time will tell Dave and I count on you as one of those members of, (laughs) you know, folks that will continue to share their platform. To go back to the roots, as why in the first place? I mean, I dedicated a whole part of the book to 'Why?' Because we get, as you said, we get carried away with vendors and technology solution try to ride a wave. And in that story, we forget the reason for which we even making this change and we are going to spend all of this resources. So hopefully we can always come back to that. >> Yeah. And I think we can. I think you have really given this some deep thought and as we pointed out, this was based on practical knowledge and experience. And look, we've been trying to solve this data problem for a long, long time. You've not only articulated it well, but you've come up with solutions. So Zhamak, thank you so much. We're going to leave it there and I'd love to have you back. >> Thank you for the conversation. I really enjoyed it. And thank you for sharing your platform to talk about data mesh. >> Yeah, you bet. All right. And I want to thank my colleague, Stephanie Chan, who helps research topics for us. Alex Myerson is on production and Kristen Martin, Cheryl Knight and Rob Hoff on editorial. Remember all these episodes are available as podcasts, wherever you listen. And all you got to do is search Breaking Analysis Podcast. Check out ETR's website at etr.ai for all the data. And we publish a full report every week on wikibon.com, siliconangle.com. You can reach me by email david.vellante@siliconangle.com or DM me @dvellante. Hit us up on our LinkedIn post. This is Dave Vellante for theCUBE Insights powered by ETR. Have a great week, stay safe, be well. And we'll see you next time. (bright music)

Published Date : Apr 20 2022

SUMMARY :

bringing you data driven insights Organizations that have taken the plunge and have a conversation. and much of the past two years, and as we see, and some of the data and make the data available But the data warehouse crowd will say, in the middle to move the data around. and talk about how you serve and the data itself together and the implications. and the logic of running the business and are served by the technology. to build resilient you I think in all cases, you know, And that leads to a that the data teams lack and naturally the data and some of the standards that are needed. and formatting of the data and it created the data swamps. We're exposing that to the end client and the better part of a decade So it's just that's the role I have, and for the naysayers look and at the end of today, And of course, the gentleman part of the book to 'Why?' and I'd love to have you back. And thank you for sharing your platform etr.ai for all the data.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Kristen MartinPERSON

0.99+

Rob HoffPERSON

0.99+

Cheryl KnightPERSON

0.99+

Stephanie ChanPERSON

0.99+

Alex MyersonPERSON

0.99+

DavePERSON

0.99+

ZhamakPERSON

0.99+

oneQUANTITY

0.99+

Dave VellantePERSON

0.99+

AWSORGANIZATION

0.99+

10 lakesQUANTITY

0.99+

Sanji MohanPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Paul AndrewPERSON

0.99+

twoQUANTITY

0.99+

NetflixORGANIZATION

0.99+

Zhamak DehghaniPERSON

0.99+

Data Mesh: Delivering Data-Driven Value at ScaleTITLE

0.99+

BostonLOCATION

0.99+

OracleORGANIZATION

0.99+

14 plus yearsQUANTITY

0.99+

Palo AltoLOCATION

0.99+

two pointsQUANTITY

0.99+

siliconangle.comOTHER

0.99+

second layerQUANTITY

0.99+

2016DATE

0.99+

LinkedInORGANIZATION

0.99+

todayDATE

0.99+

SnowflakeORGANIZATION

0.99+

hundreds of lakesQUANTITY

0.99+

theCUBEORGANIZATION

0.99+

david.vellante@siliconangle.comOTHER

0.99+

theCUBE StudiosORGANIZATION

0.98+

SQLTITLE

0.98+

one unitQUANTITY

0.98+

firstQUANTITY

0.98+

100 levelQUANTITY

0.98+

third pointQUANTITY

0.98+

DatabricksORGANIZATION

0.98+

EuropeLOCATION

0.98+

three bucketsQUANTITY

0.98+

ETRORGANIZATION

0.98+

DevStackTITLE

0.97+

OneQUANTITY

0.97+

wikibon.comOTHER

0.97+

bothQUANTITY

0.97+

ThoughtworksORGANIZATION

0.96+

one setQUANTITY

0.96+

one streamQUANTITY

0.96+

IntuitORGANIZATION

0.95+

one wayQUANTITY

0.93+

two worldsQUANTITY

0.93+

HelloFreshORGANIZATION

0.93+

this weekDATE

0.93+

last nightDATE

0.91+

fourth oneQUANTITY

0.91+

SnowflakeTITLE

0.91+

two different modelsQUANTITY

0.91+

ML AnalyticsTITLE

0.91+

Breaking AnalysisTITLE

0.87+

two worldsQUANTITY

0.84+

AIOps Virtual Forum 2020


 

>>From around the globe. It's the cube with digital coverage of an AI ops virtual forum brought to you by Broadcom. >>Welcome to the AI ops virtual forum. Finally, some Artan extended to be talking with rich lane now, senior analyst, serving infrastructure and operations professionals at Forrester. Rich. It's great to have you today. >>Thank you for having me. I think it's going to be a really fun conversation to have today. >>It is. We're going to be setting the stage for, with Richard, for the it operations challenges and the need for AI ops. That's kind of our objective here in the next 15 minutes. So rich talk to us about some of the problems that enterprise it operations are facing now in this year, that is 2020 that are going to be continuing into the next year. >>Yeah, I mean, I think we've been on this path for a while, but certainly the last eight months has, uh, has accelerated, uh, this problem and, and brought a lot of things to light that, that people were, you know, they were going through the day to day firefighting as their goal way of life. Uh, it's just not sustainable anymore. You a highly distributed environment or in the need for digital services. And, you know, one of them has been building for a while really is in the digital age, you know, we're providing so many, uh, uh, the, the interactions with customers online. Um, we've, we've added these layers of complexity, um, to applications, to infrastructure, you know, or we're in the, in the cloud or a hybrid or multi-cloud, or do you know you name it using cloud native technologies? We're using legacy stuff. We still have mainframe out there. >>Uh, you know, the, just the, the vast amount of things we have to keep track of now and process and look at the data and signals from, it's just, it's a really untenable for, for humans to do that in silos now, uh, in, in, you know, when you add to that, you know, when companies are so heavily invested in gone on the digital transformation path, and it's accelerated so much in the last, uh, year or so that, you know, we're getting so much of our business in revenue derived from these services that they become core to the business. They're not afterthoughts anymore. It's not just about having a website presence. It's, it's about deriving core business value from the services you're providing to your, through your customers. And a lot of cases, customers you're never going to meet or see at that. So it's even more important to be vigilant. >>And on top of the quality of that service that you're giving them. And then when you think about just the staffing issues we have, there's just not enough bodies to go around it in operations anymore. Um, you know, we're not going to be able to hire, you know, like we did 10 years ago, even. Uh, so that's where we need the systems to be able to bring those operational efficiencies to bear. When we say operational efficiencies, we don't mean, you know, uh, lessening head count because we can't do that. That'd be foolish. What we mean is getting the head count. We have back to burping on and higher level things, you know, working on, uh, technology refreshes and project work that that brings better digital services to customers and get them out of doing these sort of, uh, low, uh, complexity, high volume tasks that they're spending at least 20%, if not more on our third day, each day. So I think that the more we can bring intelligence to bear and automation to take those things out of their hands, the better off we are going forward. >>And I'm sure those workers are wanting to be able to have the time to deliver more value, more strategic value to the organization, to their role. And as you're saying, you know, was the demand for digital services is spiking. It's not going to go down and as consumers, if w if we have another option and we're not satisfied, we're going to go somewhere else. So, so it's really about not just surviving this time right now, it's about how do I become a business that's going to thrive going forward and exceeding expectations that are now just growing and growing. So let's talk about AI ops as a facilitator of collaboration, across business folks, it folks developers, operations, how can it facilitate collaboration, which is even more important these days? >>Yeah. So one of the great things about it is now, you know, years ago, have I gone years, as they say, uh, we would buy a tool to fit each situation. And, you know, someone that worked in network and others who will somebody worked in infrastructure from a, you know, Linux standpoint, have their tool, somebody who's from storage would have their tool. And what we found was we would have an incident, a very high impact incident occur. Everybody would get on the phone, 24 people all be looking at their siloed tool, they're siloed pieces of data. And then we'd still have to try to like link point a to B to C together, you know, just to institutional knowledge. And, uh, there was just ended up being a lot of gaps there because we couldn't understand that a certain thing happening over here was related to an advantage over here. >>Um, now when we bring all that data into one umbrella, one data Lake, whatever we want to call it, a lot of smart analytics to that data, uh, and normalize that data in a way we can contextualize it from, you know, point a to point B all the way through the application infrastructure stack. Now, the conversation changes now, the conversation changes to here is the problem, how are we going to fix it? And we're getting there immediately versus three, four or five hours of, uh, you know, hunting and pecking and looking at things and trying to try to extrapolate what we're seeing across disparate systems. Um, and that's really valuable. And in what that does is now we can change the conversation for measuring things. And in server up time and data center, performance metrics as to how are we performing as a business? How are we overall in, in real time, how are businesses being impacted by service disruption? >>We know how much money losing per minute hour, or what have you, uh, and what that translate lights into brand damage and things along those lines, that people are very interested in that. And, you know, what is the effect of making decisions either brief from a product change side? You know, if we're, we're, we're always changing the mobile apps and we're always changing the website, but do we understand what value that brings us or what negative impact that has? We can measure that now and also sales, marketing, um, they run a campaign here's your, you know, coupon for 12% off today only, uh, what does that drive to us with user engagement? We can measure that now in real time, we don't have to wait for those answers anymore. And I think, you know, having all those data and understanding the cause and effect of things increases, it enhances these feedback loops of we're making decisions as a business, as a whole to make, bring better value to our customers. >>You know, how does that tie into ops and dev initiatives? How does everything that we do if I make a change to the underlying architectures that help move the needle forward, does that hinder things, uh, all these things factor into it. In fact, there into the customer experience, which is what we're trying to do at the end of the day, w w whether operations people like it or not, we are all in the customer experience business now. And we have to realize that and work closer than ever with our business and dev partners to make sure we're delivering the highest level of customer experience we can. >>Uh, customer experience is absolutely critical for a number of reasons. I always kind of think it's inextricably linked with employee experience, but let's talk about long-term value because as organizations and every industry has pivoted multiple times this year and will probably continue to do so for the foreseeable future, for them to be able to get immediate value that let's, let's not just stop the bleeding, but let's allow them to get a competitive advantage and be really become resilient. What are some of the, uh, applications that AI ops can deliver with respect to long-term value for an organization? >>Yeah, and I think that it's, you know, you touched upon this a very important point that there is a set of short term goals you want to achieve, but they're really going to be looking towards 12, 18 months down the road. What is it going to have done for you? And I think this helps framing out for you what's most important because it'd be different for every enterprise. Um, and it also shows the ROI of doing this because there is some, you know, change is going to be involved with things you're gonna have to do. But when you look at the, the, the longer time horizon of what it brings to your business as a whole, uh it's to me, at least it all seems, it seems like a no brainer to not do it. Um, you know, thinking about the basic things, like, you know, faster remediation of, of, uh, client impacting incidents, or maybe, maybe even predictive of sort of detection of these incidents that will affect clients. >>So now you're getting, you know, at scale, you know, it's very hard to do when you have hundreds of thousands of optics of the management that relate to each other, but now you're having letting the machines and the intelligence layer find out where that problem is. You know, it's not the red thing, it's the yellow thing. Go look at that. Um, it's reducing the amount of finger pointing and what have you like resolved between teams now, everybody's looking at the same data, the same sort of, uh, symptoms and like, Oh yeah, okay. This is telling us, you know, here's the root cause you should investigate this huge, huge thing. Um, and, and it's something we never thought we'd get to where, uh, this, this is where we smart enough to tell us these things, but this, again, this is the power of having all the data under one umbrella >>And the smart analytics. >>Um, and I think really, you know, it's a boat. Uh, if you look at where infrastructure and operations people are today, and especially, you know, eight months, nine months, whatever it is into the pandemic, uh, a lot of them are getting really burnt out with doing the same repetitive tasks over and over again. Um, just trying to keep the lights on, you know, we need, we need to extract those things for those people, uh, just because it just makes no sense to do something over and over again, the same remediation step, just we should automate those things. So getting that sort of, uh, you know, drudgery off their hands, if you will, and, and get them into, into all their important things they should be doing, you know, they're really hard to solve problems. That's where the human shine, um, and that's where, you know, having a, you know, really high level engineers, that's what they should be doing, you know, and just being able to do things I >>Think in a much faster, >>In a more efficient manner, when you think about an incident occurring, right. In, in a level, one technician picks that up and he goes and triaged that maybe run some tests. He has a script, >>Uh, or she, uh, and, >>You know, uh, they open a ticket and they enrich the ticket. They call it some log files. They can look up for the servers on it. You're in an hour and a half into an incident before anyone's even looked at it. If we could automate all of that, >>Why wouldn't we, that makes it easier for everyone. Um, >>Yeah. And I really think that's where the future is, is, is, is bringing this intelligent automation to bear, to take, knock down all the little things that consume the really, the most amount of time. When you think about it, if you aggregate it over the course of a quarter or a year, a great deal of your time is spent just doing that minutiae again, why don't we automate that? And we should. So I really think that's, that's where you get to look long-term. I think also the sense of we're going to be able to measure everything in the sense of business KPIs versus just IT-centric KPIs. That's really where we going to get to in the digital age. And I think we waited too long to do that. I think our operations models were all voted. I think, uh, you know, a lot of, a lot of the KPIs we look at today are completely outmoded. They don't really change if you think about it. When we look at the monthly reports over the course of a year, uh, so let's do something different. And now having all this data and the smart analytics, we can do something different. Absolutely. I'm glad >>That you brought up kind of looking at the impact that AI ops can make on, on minutiae and burnout. That's a really huge problem that so many of us are facing in any industry. And we know that there's some amount of this that's going to continue for a while longer. So let's get our let's leverage intelligent automation to your point, because we can to be able to allow our people to not just be more efficient, but to be making a bigger impact. And there's that mental component there that I think is absolutely critical. I do want to ask you what are some of these? So for those folks going, all right, we've got to do this. It makes sense. We see some short-term things that we need. We need short-term value. We need long-term value as you've just walked us through. What are some of the obstacles that you'd say, Hey, be on the lookout for this to wipe it out of the way. >>Yeah. I, I think there's, you know, when you think about the obstacles, I think people don't think about what are big changes for their organization, right? You know, they're, they're going to change process. They're going to change the way teams interact. They're they're going to change a lot of things, but they're all for the better. So what we're traditionally really bad in infrastructure and operations is communication, marketing, a new initiative, right? We don't go out and get our peers agreement to it where the product owner is, you know, and say, okay, this is what it gets you. This is where it changes. People just hear I'm losing something, I'm losing control over something. You're going to get rid of the tools that I have, but I love I've spent years building out perfecting, um, and that's threatening to people and understandably so because people think if I start losing tools, I start losing head count. >>And then, whereas my department at that point, um, but that's not what this is all about. Uh, this, this isn't a replacement for people. This isn't a replacement for teams. This isn't augmentation. This is getting them back to doing the things they should be doing and less of the stuff they shouldn't be doing. And frankly, it's, it's about providing better services. So when in the end, it's counterintuitive to be against it because it's gonna make it operations look better. It's gonna make us show us that we are the thought leaders in delivering digital services that we can, um, constantly be perfecting the way we're doing it. And Oh, by the way, we can help the business be better. Also at the same time. Uh, I think some of the mistakes people really don't make, uh, really do make, uh, is not looking at their processes today, trying to figure out what they're gonna look like tomorrow when we bring in advanced automation and intelligence, uh, but also being prepared for what the future state is, you know, in talking to one company, they were like, yeah, we're so excited for this. >>Uh, we, we got rid of our old 15 year old laundering system and the same day we stepped a new system. Uh, one problem we had though, was we weren't ready for the amount of incidents that had generated on day one. And it wasn't because we did anything wrong or the system was wrong or what have you. It did the right thing actually, almost too. Well, what it did is it uncovered a lot of really small incidents through advanced correlations. We didn't know we had, so there were things lying out there that were always like, huh, that's weird. That system acts strange sometimes, but we can never pin it down. We found all of those things, which is good. It goes, but it kind of made us all kind of sit back and think, and then our readership are these guys doing their job. Right? >>And then we had to go through an evolution of, you know, just explaining we were 15 years behind from a visibility standpoint to our environment, but technologies that we deployed in applications had moved ahead and modernized. So this is like a cautionary tale of falling too far behind from a sort of a monitoring and intelligence and automation standpoint. Um, so I thought that was a really good story for something like, think about as Eagle would deploy these modern systems. But I think if he really, you know, the marketing to people, so they're not threatened, I think thinking about your process and then what's, what's your day one and then look like, and then what's your six and 12 months after that looks like, I think settling all that stuff upfront just sets you up for success. >>All right. Rich, take us home here. Let's summarize. How can clients build a business case for AI ops? What do you recommend? >>Yeah. You know, I actually get that question a lot. It's usually, uh, almost always the number one, uh, question in, in, um, you know, webinars like this and conversations that, that the audience puts in. So I wouldn't be surprised, but if that was true, uh, going forward from this one, um, yeah, people are like, you know, Hey, we're all in. We want to do this. We know this is the way forward, but the guy who writes the checks, the CIO, the VP of ops is like, you know, I I've signed lots of checks over the years for tools wise is different. Um, and when I guide people to do is to sit back and, and start doing some hard math, right. Uh, one of the things that resonates with the leadership is dollars and cents. It's not percentages. So saying, you know, it's, it brings us a 63% reduction and MTTR is not going to resonate. >>Uh, Oh, even though it's a really good number, you know, uh, I think what it is, you have to put it in terms of avoid, if we could avoid that 63%. Right. You know, um, what does that mean for our, our digital services as far as revenue, right. We know that every hour system down, I think, uh, you know, typically in the market, you see is about $500,000 an hour for enterprise. We'll add that up over the course of the year. What are you losing in revenue? Add to that brand damage loss of customers, you know, uh, Forrester puts out a really big, uh, casino, um, uh, customer experience index every year that measures that if you're delivering good Udall services, bad digital services, if you could raise that up, what does that return to you in revenue? And that's a key thing. And then you just look at the, the, uh, hours of lost productivity. >>I call it, I might call it something else, but I think it's a catchy name. Meaning if a core internal system is down say, and you know, you have a customer service desk of a thousand customer service people, and they can't do that look up or fix that problem for clients for an hour. How much money does that lose you? And you multiply it out. You know, average customer service desk person makes X amount an hour times this much time. This many times it happens. Then you start seeing the real, sort of a power of AI ops for this incident avoidance, or at least lowering the impact of these incidents. And people have put out in graphs and spreadsheets and all this, and then I'm doing some research around this actually to, to, to put out something that people can use to say, the project funds itself in six to 12 months, it's paid for itself. And then after that it's returning money to the business. Why would you not do that? And when you start framing the conversation, that way, the little light bulb turn on for the people that sign the checks. For sure. >>That's great advice for folks to be thinking about. I loved how you talked about the 63% reduction in something. I think that's great. What does it impact? How does it impact the revenue for the organization? If we're avoiding costs here, how do we drive up revenue? So having that laser focus on revenue is great advice for folks in any industry, looking to build a business case for AI ops. I think you set the stage for that rich beautifully, and you were right. This was a fun conversation. Thank you for your time. Thank you. And thanks for watching >>From around the globe with digital coverage. >>Welcome back to the Broadcom AI ops, virtual forum, Lisa Martin here talking with Eastman Nasir global product management at Verizon. We spent welcome back. >>Hi. Hello. Uh, what a pleasure. >>So 2020 the year of that needs no explanation, right? The year of massive challenges and wanting to get your take on the challenges that organizations are facing this year as the demand to deliver digital products and services has never been higher. >>Yeah. So I think this is something it's so close to all the far far, right? It's, uh, it's something that's impacted the whole world equally. And I think regardless of which industry you rent, you have been impacted by this in one form or the other, and the ICT industry, the information and communication technology industry, you know, Verizon being really massive player in that whole arena. It has just been sort of struck with this massive consummation we have talked about for a long time, we have talked about these remote surgery capabilities whereby you've got patients in Kenya who are being treated by an expert sitting in London or New York, and also this whole consciousness about, you know, our carbon footprint and being environmentally conscious. This pandemic has taught us all of that and brought us to the forefront of organization priorities, right? The demand. I think that's, that's a very natural consequence of everybody sitting at home. >>And the only thing that can keep things still going is this data communication, right? But I wouldn't just say that that is, what's kind of at the heart of all of this. Just imagine if we are to realize any of these targets of the world is what leadership is setting for themselves. Hey, we have to be carbon neutral by X year as a country, as a geography, et cetera, et cetera. You know, all of these things require you to have this remote working capabilities, this remote interaction, not just between humans, but machine to machine interactions. And this there's a unique value chain, which is now getting created that you've got people who are communicating with other people or communicating with other machines, but the communication is much more. I wouldn't even use the term real time because we've used real time for voice and video, et cetera. >>We're talking low latency, microsecond decision-making that can either cut somebody's, you know, um, our trees or that could actually go and remove the tumor, that kind of stuff. So that has become a reality. Everybody's asking for it, remote learning, being an extremely massive requirement where, you know, we've had to enable these, uh, these virtual classrooms ensuring the type of connectivity, ensuring the type of type of privacy, which is just so, so critical. You can't just have everybody in a go on the internet and access a data source. You have to be concerned about the integrity and security of that data as the foremost. So I think all of these things, yes, we have not been caught off guard. We were pretty forward-looking in our plans and our evolution, but yes, it's fast track the journey that we would probably believe we would have taken in three years. It has brought that down to two quarters where we've had to execute them. >>Right. Massive acceleration. All right. So you articulated the challenges really well. And a lot of the realities that many of our viewers are facing. Let's talk now about motivations, AI ops as a tool, as a catalyst for helping organizations overcome those challenges. >>So yeah. Now on that I said, you can imagine, you know, it requires microsecond decision-making which human being on this planet can do microsecond decision-making on complex network infrastructure, which is impacting end user applications, which have multitudes of effect. You know, in real life, I use the example of a remote surgeon. Just imagine that, you know, even because of you just use your signal on the quality of that communication for that microsecond, it could be the difference between killing somebody in saving somebody's life. And it's not predictable. We talk about autonomous vehicles. Uh, we talk about this transition to electric vehicles, smart motorways, et cetera, et cetera, in federal environment, how is all of that going to work? You have so many different components coming in. You don't just have a network and security anymore. You have software defined networking. That's coming, becoming a part of that. >>You have mobile edge computing that is rented for the technologies. 5g enables we're talking augmented reality. We're talking virtual reality. All of these things require that resources and why being carbon conscious. We told them we just want to build a billion data centers on this planet, right? We, we have to make sure that resources are given on demand and the best way of resources can be given on demand and could be most efficient is that the thing is being made at million microsecond and those resources are accordingly being distributed, right? If you're relying on people, sipping their coffees, having teas, talking to somebody else, you know, just being away on holiday. I don't think we're going to be able to handle that one that we have already stepped into. Verizon's 5g has already started businesses on that transformational journey where they're talking about end user experience personalization. >>You're going to have events where people are going to go, and it's going to be three-dimensional experiences that are purely customized for you. How, how does that all happen without this intelligence sitting there and a network with all of these multiple layers? So spectrum, it doesn't just need to be intuitive. Hey, this is my private IP traffic. This is public traffic. You know, it has to not be in two, or this is an application that I have to prioritize over another task to be intuitive to the criticality and the context of those transactions. Again, that's surgeons. So be it's much more important than postman setting and playing a video game. >>I'm glad that you think that that's excellent. Let's go into some specific use cases. What are some of the examples that you gave? Let's kind of dig deeper into some of the, what you think are the lowest hanging fruit for organizations kind of pan industry to go after. >>Excellent. Brian, and I think this, this like different ways to look at the lowest hanging fruit, like for somebody like revising who is a managed services provider, you know, very comprehensive medicines, but we obviously have food timing, much lower potentially for some of our customers who want to go on that journey. Right? So for them to just go and try and harness the power of the foods might be a bit higher hanging, but for somebody like us, the immediate ones would be to reduce the number of alarms that are being generated by these overlay services. You've got your basic network, then you've got your whole software defined networking on top of that, you have your hybrid clouds, you have your edge computing coming on top of that. You know? So all of that means if there's an outage on one device on the network, I want to make this very real for everybody, right? >>It's like device and network does not stop all of those multiple applications or monitoring tools from raising and raising thousands of alarm and everyone, one capacity. If people are attending to those thousands of alarms, it's like you having a police force and there's a burglary in one time and the alarm goes off and 50 bags. How, how are you kind of make the best use of your police force? You're going to go investigate 50 bags or do you want to investigate where the problem is? So it's as real as that, I think that's the first wins where people can save so much cost, which is coming from being wasted and resources running around, trying to figure stuff out immediately. I'm tied this with network and security network and security is something which has you did even the most, you know, I mean single screens in our engineering, well, we took it to have network experts, separate people, security experts, separate people to look for different things, but there are security events that can impact the performance of a network. >>And then just drop the case on the side of et cetera, which could be falsely attributed to the metric. And then if you've got multiple parties, which are then the chapter clear stakeholders, you can imagine the blame game that goes on finding fingers, taking names, not taking responsibility that don't has all this happened. This is the only way to bring it all together to say, okay, this is what takes priority. If there's an event that has happened, what is its correlation to the other downstream systems, devices, components, and these are applications. And then subsequently, you know, like isolating it to the right cost where you can most effectively resolve that problem. Thirdly, I would say on demand, virtualized resource, virtualized resources, the heart and soul, the spirit of status that you can have them on demand. So you can automate the allocation of these resources based on customer's consumption their peaks, their cramps, all of that comes in. >>You see, Hey, typically on a Wednesday, the traffic was up significantly for this particular application, you know, going to this particular data center, you could have this automated system, uh, which is just providing those resources, you know, on demand. And so it is to have a much better commercial engagement with customers and just a much better service assurance model. And then one more thing on top of that, which is very critical is that as I was saying, giving that intelligence to the networks to start having context of the criticality of a transaction, that doesn't make sense to them. You can't have that because for that, you need to have this, you know, monkey their data. You need to have multi-cam system, which are monitoring and controlling different aspects of your overall end user application value chain to be communicating with each other. And, you know, that's the only way to sort of achieve that goal. And that only happens with AI. It's not possible >>So it was when you clearly articulated some obvious, low hanging fruit and use cases that organizations can go after. Let's talk now about some of the considerations, you talked about the importance of a network and AI ops, the approach I assume, needs to be modular support needs to be heterogeneous. Talk to us about some of those key considerations that you would recommend. >>Absolutely. So again, basically starting with the network, because if there's, if the metrics sitting at the middle of all of this is not working, then things can communicate with each other, right? And the cloud doesn't work, nothing metal. That's the hardest part of this, but that's the frequency. When you talk about machine to machine communication or IOT, it's just the biggest transformation of the span of every company is going for IOT now to drive those costs, efficiencies, and had, something's got some experience, the integrity of the topic karma, right? The security, integrity of that. How do you maintain integrity of your data beyond just a secure network components? That is true, right? That's where you're getting to the whole arena blockchain technologies, where you have to use digital signatures or barcodes that machine then, and then an intelligence system is automatically able to validate and verify the integrity of the data and the commands that are being executed by those end-user told them what I need to tell them that. >>So it's IOT machines, right? That is paramount. And if anybody is not keeping that into their equation, that in its own self is any system that is therefore maintaining the integrity of your commands and your hold that sits on those, those machines. Right? Second, you have your network. You need to have any else platform, which is able to restless all the fast network information, et cetera. And coupled with that data integrity piece, because for the management, ultimately they need to have a coherent view of the analytics, et cetera, et cetera. They need to know where the problems are again, right? So let's say if there's a problem with the integrity of the commands that are being executed by the machine, that's a much bigger problem than not being able to communicate with that machine and the best thing, because you'd rather not talk to the machine or have to do anything if it's going to start doing wrong things. >>So I think that's where it is. It's very intuitive. It's not true. You have to have subsequently if you have some kind of faith and let me use that use case self autonomous vehicles. Again, I think we're going to see in the next five years, because he's smart with the rates, et cetera, it won't separate autonomous cars. It's much more efficient, it's much more space, et cetera, et cetera. So within that equation, you're going to have systems which will be specialists in looking at aspects and transactions related to those systems. For example, in autonomous moving vehicles, brakes are much more important than the Vipers, right? So this kind of intelligence, it will be multiple systems who have to sit, N nobody has to, one person has to go in one of these systems. I think these systems should be open source enough that they, if you were able to integrate them, right, if something's sitting in the cloud, you were able to integrate for that with obviously the regard of the security and integrity of your data that has to traverse from one system to the other extremely important. >>So I'm going to borrow that integrity theme for a second, as we go into our last question, and that is this kind of take a macro look at the overall business impact that AI ops can help customers make. I'm thinking of, you know, the integrity of teams aligning business in it, which we probably can't talk about enough. We're helping organizations really effectively measure KPIs that deliver that digital experience that all of us demanding consumers expect. What's the overall impact. What would you say in summary fashion? >>So I think the overall impact is a lot of costs. That's customized and businesses gives the time to the time of enterprises. Defense was inevitable. It's something that for the first time, it will come to life. And it's something that is going to, you know, start driving cost efficiencies and consciousness and awareness within their own business, which is obviously going to have, you know, it domino kind of an effect. So one example being that, you know, you have problem isolation. I talked about network security, this multi-layers architecture, which enables this new world of 5g, um, at the heart of all of it, it has to identify the problem to the source, right? Not be bogged down by 15 different things that are going wrong. What is causing those 15 things to go wrong, right? That speed to isolation in its own sense can make millions and millions of dollars to organizations after we organize it. Next one is obviously overall impacted customer experience. Uh, 5g was given out of your customers, expecting experiences from you, even if you're not expecting to deliver them in 2021, 2022, it would have customers asking for those experience or walking away, if you do not provide those experience. So it's almost like a business can do nothing every year. They don't have to reinvest if they just want to die on the line, businesses want remain relevant. >>Businesses want to adopt the latest and greatest in technology, which enables them to, you know, have that superiority and continue it. So from that perspective that continue it, he will read that they write intelligence systems that tank rationalizing information and making decisions supervised by people, of course were previously making some of those. >>That was a great summary because you're right, you know, with how demanding consumers are. We don't get what we want quickly. We churn, right? We go somewhere else and we could find somebody that can meet those expectations. So it has been thanks for doing a great job of clarifying the impact and the value that AI ops can bring to organizations that sounds really now is we're in this even higher demand for digital products and services, which is not going away. It's probably going to only increase it's table stakes for any organization. Thank you so much for joining me today and giving us your thoughts. >>Pleasure. Thank you. We'll be right back with our next segment. >>Digital applications and services are more critical to a positive customer and employee experience than ever before. But the underlying infrastructure that supports these apps and services has become increasingly complex and expanding use of multiple clouds, mobile and microservices, along with modern and legacy infrastructure can make it difficult to pinpoint the root cause when problems occur, it can be even more difficult to determine the business impact your problems that occur and resolve them efficiently. AI ops from Broadcom can help first by providing 360 degree visibility, whether you have hybrid cloud or a cloud native AI ops from Broadcom provides a clear line of sight, including apt to infrastructure and network visibility across hybrid environments. Second, the solution gives you actionable insights by correlating an aggregating data and applying AI and machine learning to identify root causes and even predict problems before users are impacted. Third AI ops from Broadcom provides intelligent automation that identifies potential solutions when problems occur applied to the best one and learns from the effectiveness to improve response in case the problem occurs. Again, finally, the solution enables organizations to achieve digit with jelly by providing feedback loops across development and operations to allow for continuous improvements and innovation through these four capabilities. AI ops from Broadcom can help you reduce service outages, boost, operational efficiency, and effectiveness and improve customer and employee experience. To learn more about AI ops from Broadcom, go to broadcom.com/ai ops from around the globe. >>It's the cube with digital coverage of AI ops virtual forum brought to you by Broadcom. >>Welcome back to the AI ops virtual forum, Lisa Martin here with Srinivasan, Roger Rajagopal, the head of product and strategy at Broadcom. Raj, welcome here, Lisa. I'm excited for our conversation. So I wanted to dive right into a term that we hear all the time, operational excellence, right? We hear it everywhere in marketing, et cetera, but why is it so important to organizations as they head into 2021? And tell us how AI ops as a platform can help. >>Yeah. Well, thank you. First off. I wanna, uh, I want to welcome our viewers back and, uh, I'm very excited to, uh, to share, um, uh, more info on this topic. You know, uh, here's what we believe as we work with large organizations, we see all our organizations are poised to get out of the, uh, the pandemic and look for a brood for their own business and helping customers get through this tough time. So fiscal year 2021, we believe is going to be a combination of, uh, you know, resiliency and agility at the, at the same time. So operational excellence is critical because the business has become more digital, right? There are going to be three things that are going to be more sticky. Uh, you know, remote work is going to be more sticky, um, cost savings and efficiency is going to be an imperative for organizations and the continued acceleration of digital transformation of enterprises at scale is going to be in reality. So when you put all these three things together as a, as a team that is, uh, you know, that's working behind the scenes to help the businesses succeed, operational excellence is going to be, make or break for organizations, >>Right with that said, if we kind of strip it down to the key capabilities, what are those key capabilities that companies need to be looking for in an AI ops solution? >>Yeah, you know, so first and foremost, AI ops means many things to many, many folks. So let's take a moment to simply define it. The way we define AI ops is it's a system of intelligence, human augmented system that brings together full visibility across app infra and network elements that brings together disparate data sources and provides actionable intelligence and uniquely offers intelligent automation. Now, the, the analogy many folks draw is the self-driving car. I mean, we are in the world of Teslas, uh, but you know, uh, but self-driving data center is it's too far away, right? Autonomous systems are still far away. However, uh, you know, application of AI ML techniques to help deal with volume velocity, veracity of information, uh, is, is critical. So that's how we look at AI ops and some of the key capabilities that we, uh, that we, uh, that we work with our customers to help them on our own for eight years. >>Right? First one is eyes and ears. What we call full stack observability. If you do not know what is happening in your systems, uh, you know, that that serve up your business services. It's going to be pretty hard to do anything, uh, in terms of responsiveness, right? So from stack observability, the second piece is what we call actionable insights. So when you have disparate data sources, tools, sprawls data coming at you from, uh, you know, uh, from a database systems, it systems customer management systems, ticketing systems. How do you find the needle from the haystack? And how do you respond rapidly from a myriad of problems as CEO of red? The third area is what we call intelligent automation. Well, identifying the problem to act on is important, and then acting on automating that and creating, uh, a recommendation system where, uh, you know, you can be proactive about it is even more important. And finally, all of this focuses on efficiency. What about effectiveness? Effectiveness comes when you create a feedback loop, when what happens in production is related to your support systems and your developers so that they can respond rapidly. So we call that continuous feedback. So these are the four key capabilities that, uh, you know, uh, you should look for in an AI ops system. And that's what we offer as well. >>Russia, there's four key capabilities that businesses need to be looking for. I'm wondering how those help to align business. And it it's, again like operational excellence. It's something that we talk about a lot is the alignment of business. And it a lot more challenging, easier said than done, right. But I want you to explain how can AI ops help with that alignment and align it outputs to business outcomes? >>Yeah. So, you know, one of the things, uh, I'm going to say something that is, uh, that is, uh, that is simple, but, but, but this harder, but alignment is not on systems alignment is with people, right? So when people align, when organizations align, when cultures align, uh, dramatic things can happen. So in the context of AI ops VC, when, when SRE is aligned with the DevOps engineers and information architects and, uh, uh, you know, it operators, uh, you know, they enable organizations to reduce the gap between intent and outcome or output and outcome that said, uh, you know, these personas need mechanisms to help them better align, right. Help them better visualize, see the, you know, what we call single source of truth, right? So there are four key things that I want to call out. When we work with large enterprises, we find that customer journey alignment with the, you know, what we call it systems is critical. >>So how do you understand your business imperatives and your customer journey goals, whether it is car to a purchase or whether it is, uh, you know, bill shock scenarios and Swan alignment on customer journey to your it systems is one area that you can reduce the gap. The second area is how do you create a scenario where your teams can find problems before your customers do right outage scenarios and so on. So that's the second area of alignment. The third area of alignment is how can you measure business impact driven services? Right? There are several services that an organization offers versus an it system. Some services are more critical to the business than others, and these change in a dynamic environment. So how do you, how do you understand that? How do you measure that and how, how do you find the gaps there? So that's the third area of alignment that we, that we help and last but not least there are, there are things like NPS scores and others that, that help us understand alignment, but those are more long-term. But in the, in the context of, uh, you know, operating digitally, uh, you want to use customer experience and business, uh, you know, a single business outcome, uh, as a, as a key alignment factor, and then work with your systems of engagement and systems of interaction, along with your key personas to create that alignment. It's a people process technology challenge. >>So, whereas one of the things that you said there is that it's imperative for the business to find a problem before a customer does, and you talked about outages there, that's always a goal for businesses, right. To prevent those outages, how can AI ops help with that? Yeah, >>So, you know, outages, uh, talk, you know, go to resiliency of a system, right? And they also go to, uh, uh, agility of the same system, you know, if you're a customer and if you're whipping up your mobile app and it takes more than three milliseconds, uh, you know, you're probably losing that customer, right. So outages mean different things, you know, and there's an interesting website called down detector.com that actually tracks all the old pages of publicly available services, whether it's your bank or your, uh, you know, tele telecom service or a mobile service and so on and so forth. In fact, the key question around outages for, from, uh, from, uh, you know, executives are the question of, are you ready? Right? Are you ready to respond to the needs of your customers and your business? Are you ready to rapidly resolve an issue that is impacting customer experience and therefore satisfaction? >>Are you creating a digital trust system where customers can be, you know, um, uh, you know, customers can feel that their information is secure when they transact with you, all of these, getting into the notion of resiliency and outages. Now, you know, one of the things that, uh, that I, I often, uh, you know, work with customers around, you know, would that be find as the radius of impact is important when you deal with outages? What I mean by that is problems occur, right? How do you respond? How quickly do you take two seconds, two minutes, 20 minutes, two hours, 20 hours, right? To resolve the problem that radius of impact is important. That's where, you know, you have to bring a gain people, process technology together to solve that. And the key thing is you need a system of intelligence that can aid your teams, you know, look at the same set of parameters so that you can respond faster. That's the key here. >>We look at digital transformation at scale. Raj, how does AI ops help influence that? >>You know, um, I'm going to take a slightly long-winded way to answer this question. See when it comes to digital transformation at scale, the focus on business purpose and business outcome becomes extremely critical. And then the alignment of that to your digital supply chain, right, are the, are the, are the key factors that differentiate winners in the, in their digital transformation game? Really, what we have seen, uh, with, with winners is they operate very differently. Like for example, uh, you know, Nike matures, its digital business outcomes by shoes per second, right? Uh, Apple by I-phones per minute, Tesla by model threes per month, are you getting this, getting it right? I mean, you want to have a clear business outcome, which is a measure of your business, uh, in effect, I mean, ENC, right? Which, which, uh, um, my daughter use and I use very well. >>Right. Uh, you know, uh, they measure by revenue per hour, right? I mean, so these are key measures. And when you have a key business outcome measure like that, you can everything else, because you know what these measures, uh, you know, uh, for a bank, it may be deposits per month, right now, when you move money from checking account to savings account, or when you do direct deposits, those are, you know, banks need liquidity and so on and so forth. But, you know, the, the key thing is that single business outcome has a Starburst effect inside the it organization that touches a single money moment from checking a call to savings account can touch about 75 disparate systems internally. Right? So those think about it, right? I mean, all, all we're doing is moving money from checking account a savings account. Now that goats into a it production system, there are several applications. >>There is a database, there is, there are infrastructures, there are load balancers that are webs. You know, you know, the web server components, which then touches your, your middleware component, which is a queuing system, right. Which then touches your transactional system. Uh, and, uh, you know, which may be on your main frames, what we call mobile to mainframe scenario, right? And we are not done yet. Then you have a security and regulatory compliance system that you have to touch a fraud prevention system that you have to touch, right? A state department regulation that you may have to meet and on and on and on, right? This is the chat that it operations teams face. And when you have millions of customers transacting, right, suddenly this challenge cannot be managed by human beings alone. So therefore you need a system of intelligence that augments human intelligence and acts as your, you know, your, your eyes and ears in a way to, to point pinpoint where problems are. >>Right. So digital transformation at scale really requires a very well thought out AI ops system, a platform, an open extensible platform that, uh, you know, uh, that is heterogeneous in nature because there's tools, products in organizations. There is a lot of databases in systems. There are millions of, uh, uh, you know, customers and hundreds of partners and vendors, you know, making up that digital supply chain. So, you know, AI ops is at the center of an enabling an organization achieve digital op you know, transformation at scale last but not least. You need continuous feedback loop. Continuous feedback loop is the ability for a production system to inform your dev ops teams, your finance teams, your customer experience teams, your cost modeling teams about what is going on so that they can so that they can reduce the intent, come gap. >>All of this need to come together, what we call BizOps. >>That was a great example of how you talked about the Starburst effect. I actually never thought about it in that way, when you give the banking example, but what you should is the magnitude of systems. The fact that people alone really need help with that, and why intelligent automation and AI ops can be transformative and enable that scale. Raj, it's always a pleasure to talk with you. Thanks for joining me today. And we'll be right back with our next segment. Welcome back to the AI ops virtual forum. We've heard from our guests about the value of AI ops and why and how organizations are adopting AI ops platforms. But now let's see AI ops inaction and get a practical view of AI ops to deep Dante. The head of AI ops at Broadcom is now going to take you through a quick demo. >>Hello. So they've gotta head off AI ops and automation here. What I'm going to do today is talk through some of the key capabilities and differentiators of Broadcom's CII ops solution in this solution, which can be delivered on cloud or on-prem. We bring a variety of metric alarm log and applauded data from multiple sources, EPM, NetApps, and infrastructure monitoring tools to provide a single point of observability and control. Let me start where our users mostly stock key enterprises like FSI, telcos retailers, et cetera, do not manage infrastructure or applications without having a business context. At the end of the day, they offer business services governed by SLS service level objectives and SLI service level indicators are service analytics, which can scale to a few thousand services, lets our customers create and monitor the services as per their preference. They can create a hierarchy of services based on their business practice. >>For example, here, the sub services are created based on functional subsistence for certain enterprises. It could be based on location. Users can import these services from their favorite CMDB. What's important to note that not all services are born equal. If you are a modern bank, you may want to prioritize tickets coming from digital banking, for example, and this application lets you rank them as per the KPI of your choice. We can source the availability, not merely from the state of the infrastructure, whether they're running or not. But from the SLS that represent the state of the application, when it comes to triaging issues related to the service, it is important to have a complete view of the topology. The typology can show both east-west elements from mobile to mainframe or not South elements in a network flow. This is particularly relevant for a large enterprise who could be running the systems of engagement on the cloud and system of records on mainframe inside the firewall here, you can see that the issue is related to the mainframe kick server. >>You can expand to see the actual alarm, which is sourced from the mainframe operational intelligence. Similarly, clicking on network will give the hub and spoke view of the network devices, the Cisco switches and routers. I can click on the effected router and see all the details Broadcom's solution stores, the ontological model of the typology in the form of a journal graph where one can not only view the current state of the typology, but the past as well, talking of underlying data sources, the solution uses best of the pre data stores for structured and unstructured data. We have not only leveraged the power of open source, but have actively contributed back to the community. One of the key innovations is evident in our dashboarding framework because we have enhanced the open source Grafana technology to support these diverse data sources here. You can see a single dashboard representing applications to infrastructure, to mainframe again, sourcing a variety of data from these sources. >>When we talk to customers, one of the biggest challenges that they face today is related to alarms because of a proliferation of tools. They are currently drowning in an ocean of hundreds and thousands of alarms. This increases the Elmont support cost to tens of dollars per ticket, and also affects LTO efficiency leading to an average of five to six hours of meantime to resolution here is where we have the state of the art innovation utilizing the power of machine learning and ontology to arrive at the root cause we not only clusterize alarms based on text, but employ the technique of 41st. We look at the topology then at the time window duplicate text based on NLP. And lastly learn from continuous training of the model to deduce what we call situations. This is an example of a situation. As you can see, we provide a time-based evidence of how things unfolded and arrive at a root cause. >>Lastly, the solution provides a three 60 degree closed loop remediation either through a ticketing system or by direct invocation of automation actions instead of firing hard-coded automation runbooks for certain conditions, the tool leverage is machine learning to rank automation actions based on past heuristics. That's why we call it intelligent automation to summarize AI ops from Broadcom helps you achieve operational excellence through full stack observability, coupled with AIML that applies across modern hybrid cloud environments, as well as legacy ones uniquely. It ties these insights with intelligent automation to improve customer experience. Thank you for watching from around the globe. It's the cube with digital coverage of AI ops virtual forum brought to you by Broadcom. >>Welcome to our final segment today. So we've discussed today. The value that AI ops will bring to organizations in 2021, we'll discuss that through three different perspectives. And so now we want to bring those perspectives together and see if we can get a consensus on where AI ops needs to go for folks to be successful with it in the future. So bringing back some folks Richland is back with us. Senior analysts, serving infrastructure and operations professionals at Forrester smartness here is also back in global product management at Verizon and Srinivasan, Reggie Gopaul head of product and strategy at Broadcom guys. Great to have you back. So let's jump in and rich, we're going to, we're going to start with you, but we are going to get all three of you, a chance to answer the questions. So we've talked about why organizations should adopt AI ops, but what happens if they choose not to what challenges would they face? Basically what's the cost of organizations doing nothing >>Good question, because I think in operations for a number of years, we've kind of stand stood, Pat, where we are, where we're afraid change things sometimes, or we just don't think about a tooling as often. The last thing to change because we're spending so much time doing project work and modernization and fighting fires on a daily basis. >>Problem is going to get worse. If we do nothing, >>You know, we're building new architectures like containers and microservices, which means more things to mind and keep running. Um, we're building highly distributed systems. We're moving more and more into this hybrid world, a multi-cloud world, uh, it's become over-complicate and I'll give a short anecdote. I think, eliminate this. Um, when I go to conferences and give speeches, it's all infrastructure operations people. And I say, you know, how many people have three X, five X, you know, uh, things to monitor them. They had, you know, three years ago, two years ago, and everyone's saying how many people have hired more staff in that time period, zero hands go up. That's the gap we have to fill. And we have to fill that through better automation, more intelligent systems. It's the only way we're going to be able to fill back out. >>What's your perspective, uh, if organizations choose not to adopt AI ops. Yeah. So I'll do that. Yeah. So I think it's, I would just relate it to a couple of things that probably everybody >>Tired off lately and everybody can relate to. And this would resonate that we have 5g, which is all set to transform the world. As we know it, I don't have a lot of communication with these smart cities, smart communities, IOT, which is going to make us pivotal to the success of businesses. And as you've seen with this call with, you know, transformation of the world, that there's a, there's a much bigger cost consciousness out there. People are trying to become much more, forward-looking much more sustainable. And I think at the heart of all of this, that the necessity that you have intelligent systems, which are bastardizing more than enough information that previously could've been overlooked because if you don't measure engagement, not going right. People not being on the same page of this using two examples or hundreds of things, you know, that play a part in things, but not coming together in the best possible way. So I think it has an absolute necessity to drive those cost efficiencies rather than, you know, left right and center laying off people who are like 10 Mattel to your business and have a great tribal knowledge of your business. So to speak, you can drive these efficiencies through automating a lot of those tasks that previously were being very manually intensive or resource intensive. And you could allocate those resources towards doing much better things, which let's be very honest going into 20, 21 after what we've seen with 2020, it's going to be mandate treat. >>And so Raj, I saw you shaking your head there when he was mom was sharing his thoughts. What are your thoughts about that sounds like you agree. Yeah. I mean, uh, you know, uh, to put things in perspective, right? I mean we're firmly in the digital economy, right? Digital economy, according to the Bureau of economic analysis is 9% of the U S GDP. Just, you know, think about it in, in, in, in, in the context of the GDP, right? It's only ranked lower, slightly lower than manufacturing, which is at 11.3% GDP and slightly about finance and insurance, which is about seven and a half percent GDP. So the digital economy is firmly in our lives, right. And as Huisman was talking about it, you know, software eats the world and digital, operational excellence is critical for customers, uh, to, uh, you know, to, uh, to drive profitability and growth, uh, in the digital economy. >>It's almost, you know, the key is digital at scale. So when, uh, when rich talks about some of the challenges and when Huseman highlights 5g as an example, those are the things that, that, that come to mind. So to me, what is the cost or perils of doing nothing? You know, uh, it's not an option. I think, you know, more often than not, uh, you know, C-level execs are asking head of it and they are key influencers, a single question, are you ready? Are you ready in the context of addressing spikes in networks because of the pandemic scenario, are you ready in the context of automating away toil? Are you ready to respond rapidly to the needs of the digital business? I think AI ops is critical. >>That's a great point. Roger, where does stick with you? So we got kind of consensus there, as you said, wrapping it up. This is basically a, not an option. This is a must to go forward for organizations to be successful. So let's talk about some quick wins, or as you talked about, you know, organizations and sea levels asking, are you ready? What are some quick wins that that organizations can achieve when they're adopting AI? >>You know, um, immediate value. I think I would start with a question. How often do your customers find problems in your digital experience before you do think about that? Right. You know, if you, if you, you know, there's an interesting web, uh, website, um, uh, you know, down detector.com, right? I think, uh, in, in Europe there is an equal amount of that as well. It ha you know, people post their digital services that are down, whether it's a bank that, uh, you know, customers are trying to move money from checking account, the savings account and the digital services are down and so on and so forth. So some and many times customers tend to find problems before it operations teams do. So a quick win is to be proactive and immediate value is visibility. If you do not know what is happening in your complex systems that make up your digital supply chain, it's going to be hard to be responsive. So I would start there >>Visibility this same question over to you from Verizon's perspective, quick wins. >>Yeah. So I think first of all, there's a need to ingest this multi-care spectrum data, which I don't think is humanly possible. You don't have people having expertise, you know, all the seven layers of the OSI model and then across network and security and at the application level. So I think you need systems which are now able to get that data. It shouldn't just be wasted reports that you're paying for on a monthly basis. It's about time that you started making the most of those in the form of identifying what are the efficiencies within your ecosystem. First of all, what are the things, you know, which could be better utilized subsequently you have the >>Opportunity to reduce the noise of a trouble tickets handling. It sounds pretty trivial, but >>An average you can imagine every trouble tickets has the cost in dollars, right? >>So, and there's so many tickets and there's art >>That get created on a network and across an end user application value, >>We're talking thousands, you know, across and end user >>Application value chain could be million in >>A year. So, and so many of those are not really, >>He, you know, a cause of concern because the problem is something. >>So I think that whole triage is an immediate cost saving and the bigger your network, the bigger >>There's a cost of things, whether you're a provider, whether you're, you know, the end customer at the end of the day, not having to deal with problems, which nobody can resolve, which are not meant to be dealt with. There's so many of those situations, right, where service has just been adopted, >>Which is just coordinate quality, et cetera, et cetera. So many reasons. So those are the, >>So there's some of the immediate cost saving them. They are really, really significant. >>Secondly, I would say Raj mentioned something about, you know, the user, >>Your application value chain, and an understanding of that, especially with this hybrid cloud environment, >>Et cetera, et cetera, right? The time it takes to identify a problem in an end user application value chain across the seven layers that I mentioned with the OSI reference model across network and security and the application environment. It's something that >>In its own self has massive cost to business, >>Right? That could be >>No sale transactions that could be obstructed because of this. There could be, and I'm going to use a really interesting example. >>We talk about IOT. The integrity of the IOT machine is exciting. >>Family is pivotal in this new world that we're stepping into. >>You could be running commands, >>Super efficient. He has, everything is being told to the machine really fast with sending yeah. >>Everything there. What if it's hacked? And if that's okay, >>Robotic arm starts to involve the things you don't want it to do. >>So there's so much of that. That becomes a part of this naturally. And I believe, yes, this is not just like from a cost >>standpoint, but anything going wrong with that code base, et cetera, et cetera. These are massive costs to the business in the form of the revenue. They have lost the perception in the market as a result, the fed, >>You know, all that stuff. So >>These are a couple of very immediate problems, but then you also have the whole player virtualized resources where you can automate the allocation, you know, the quantification of an orchestration of those virtualized resources, rather than a person having to, you know, see something and then say, Oh yeah, I need to increase capacity over here, because then it's going to have this particular application. You have systems doing this stuff and to, you know, Roger's point your customer should not be identifying your problems before you, because this digital is where it's all about perception. >>Absolutely. We definitely don't want the customers finding it before. So rich, let's wrap this particular question up with you from that senior analyst perspective, how can companies use make big impact quickly with AI ops? Yeah, >>Yeah, I think, you know, and it was been really summed up some really great use cases there. I think with the, uh, you know, one of the biggest struggles we've always had in operations is isn't, you know, the mean time to resolve. We're pretty good at resolving the things. We just have to find the thing we have to resolve. That's always been the problem and using these advanced analytics and machine learning algorithms now across all machine and application data, our tendency is humans is to look at the console and say, what's flashing red. That must be what we have to fix, but it could be something that's yellow, somewhere else, six services away. And we have made things so complicated. And I think this is what it was when I was saying that we can't get there anymore on our own. We need help to get there in all of this stuff that the outline. >>So, so well builds up to a higher level thing of what is the customer experience about what is the customer journey? And we've struggled for years in the digital world and measuring that a day-to-day thing. We know an online retail. If you're having a bad experience at one retailer, you just want your thing. You're going to go to another retailer, brand loyalty. Isn't one of like it, wasn't a brick and mortal world where you had a department store near you. So you were loyal to that because it was in your neighborhood, um, online that doesn't exist anymore. So we need to be able to understand the customer from that first moment, they touch a digital service all the way from their, their journey through that digital service, the lowest layer, whether it be a database or the network, what have you, and then back to them again, and we're not understanding, is that a good experience? >>We gave them. How does that compare to last week's experience? What should we be doing to improve that next week? Uh, and I think companies are starting and then the pandemic certainly, you know, push this timeline. If you listened to the, the, the CEO of Microsoft, he's like, you know, 10 years of digital transformation written down. And the first several months of this, um, in banks and in financial institutions, I talked to insurance companies, aren't slowing down. They're trying to speed up. In fact, what they've discovered is that they're, you know, obviously when we were on lockdown or what have you, they use of digital servers is spiked very high. What they've learned is they're never going to go back down. They're never going to return to pretend endemic levels. So now they're stuck with this new reality. Well, how do we service those customers and how do we make sure we keep them loyal to our brand? >>Uh, so, you know, they're looking for modernization opportunities. A lot of that that's things have been exposed. And I think Raj touched upon this very early in the conversation is visibility gaps. Now that we're on the outside, looking in at the data center, we know we architect things in a very way. Uh, we better ways of making these correlations across the Sparrow technologies to understand where the problems lies. We can give better services to our customers. And I think that's really what we're going to see a lot of the innovation and the people really clamoring for these new ways of doing things that starting, you know, now, I mean, I've seen it in customers, but I think really the push through the end of this year to next year when, you know, economy and things like that straightened out a little bit more, I think it really, people are gonna take a hard look of where they are and is, you know, AI ops the way forward for them. And I think they'll find it. The answer is yes, for sure. >>So we've, we've come to a consensus that, of what the parallels are of organizations, basically the cost of doing nothing. You guys have given some great advice on where some of those quick wins are. Let's talk about something Raj touched on earlier is organizations, are they really ready for truly automated AI? Raj, I want to start with you readiness factor. What are your thoughts? >>Uh, you know, uh, I think so, you know, we place our, her lives on automated systems all the time, right? In our, in our day-to-day lives, in the, in the digital world. I think, uh, you know, our, uh, at least the customers that I talk to our customers are, uh, are, uh, you know, uh, have a sophisticated systems. Like for example, advanced automation is a reality. If you look at social media, AI and ML and automation are used to automate away, uh, misinformation, right? If you look at financial institutions, AI and ML are used to automate away a fraud, right? So I want to ask our customers why can't we automate await oil in it, operation systems, right? And that's where our customers are. Then the, you know, uh, I'm a glass half full, uh, cleanup person, right? Uh, this pandemic has been harder on many of our customers, but I think what we have learned from our customers is they've Rose to the occasion. >>They've used digital as a key needs, right? At scale. That's what we see with, you know, when, when Huseman and his team talk about, uh, you know, network operational intelligence, right. That's what it means to us. So I think they are ready, the intersection of customer experience it and OT, operational technology is ripe for automation. Uh, and, uh, you know, I, I wanna, I wanna sort of give a shout out to three key personas in this mix. It's about people, right? One is the SRE persona, you know, site, reliability engineer. The other is the information security persona. And the third one is the it operator automation engineer persona. These folks in organizations are building a system of intelligence that can respond rapidly to the needs of their digital business. We at Broadcom, we are in the business of helping them construct a system of intelligence that will create a human augmented solution for them. Right. So when I see, when I interact with large enterprise customers, I think they, they, you know, they, they want to achieve what I would call advanced automation and AI ML solutions. And that's squarely, very I ops is, you know, is going as it, you know, when I talk to rich and what, everything that rich says, you know, that's where it's going and that's what we want to help our customers to. So, which about your perspective of organizations being ready for truly automated AI? >>I think, you know, the conversation has shifted a lot in the last, in, in pre pandemic. Uh, I'd say at the end of last year, we're, you know, two years ago, people I'd go to conferences and people come up and ask me like, this is all smoke and mirrors, right? These systems can't do this because it is such a leap forward for them, for where they are today. Right. We we've sort of, you know, in software and other systems, we iterate and we move forward slowly. So it's not a big shock. And this is for a lot of organizations that big, big leap forward where they're, they're running their operations teams today. Um, but now they've come around and say, you know what? We want to do this. We want all the automations. We want my staff not doing the low complexity, repetitive tasks over and over again. >>Um, you know, and we have a lot of those kinds of legacy systems. We're not going to rebuild. Um, but they need certain care and feeding. So why are we having operations? People do those tasks? Why aren't we automating those out? I think the other piece is, and I'll, I'll, I'll send this out to any of the operations teams that are thinking about going down this path is that you have to understand that the operations models that we're operating under in, in INO and have been for the last 25 years are super outdated and they're fundamentally broken for the digital age. We have to start thinking about different ways of doing things and how do we do that? Well, it's, it's people, organization, people are going to work together differently in an AI ops world, um, for the better. Um, but you know, there's going to be the, the age of the 40 person bridge call thing. >>Troubleshooting is going away. It's going to be three, four, five focused engineers that need to be there for that particular incident. Um, a lot of process mailer process we have in our level, one level, two engineering. What have you running of tickets, gathering of artifacts, uh, during an incident is going to be automated. That's a good thing. We should be doing those, those things by hand anymore. So I'd say that the, to people's like start thinking about what this means to your organization. Start thinking about the great things we can do by automating things away from people, having to do them over and over again. And what that means for them, getting them matched to what they want to be doing is high level engineering tasks. They want to be doing monitorization, working with new tools and technologies. Um, these are all good things that help the organization perform better as a whole great advice and great kind of some of the thoughts that you shared rich for what the audience needs to be on the lookout. For one, I want to go over to you, give me your thoughts on what the audience that should be on the lookout for, or put on your agendas in the next 12 months. >>So there's like a couple of ways to answer that question. One thing would be in the form of, you know, what are some of the things they have to be concerned about in terms of implementing this solution or harnessing its power. The other one could be, you know, what are the perhaps advantages they should look to see? So if I was to talk about the first one, let's say that, what are some of the things I have to watch out for like possible pitfalls that everybody has data, right? So yeah, there's one strategy we say, okay, you've got the data, let's see what we can do with them. But then there's the exact opposite side, which has to be considered when you're doing that analysis. What are the use cases that you're looking to drive? Right. But then use cases you have to understand, are you taking a reactive use case approach? >>Are you taking active use cases, right? Or, yeah, that's a very, very important concentration. Then you have to be very cognizant of where does this data that you have, where does it reside? What are the systems and where does it need to go to in order for this AI function to happen and subsequently if there needs to be any backward communication with all of that data in a process manner. So I think these are some of the very critical points because you can have an AI solution, which is sitting in a customer data center. It could be in a managed services provider data center, like, right, right. It could be in a cloud data center, like an AWS or something, or you could have hybrid views, et cetera, all of that stuff. So you have to be very mindful of where you're going to get the data from is going to go to what are the use cases you're trying to get out to do a bit of backward forward. >>Okay, we've got this data thing and I think it's a journey. Nobody can come in and say, Hey, you've built this fantastic thing. It's like Terminator two. I think it's a journey where we built starting with the network. My personal focus always comes down to the network and with 5g so much, so much more right with 5g, you're talking low latency communication. That's like the true power of 5g, right? It's low latency, it's ultra high bandwidth, but what's the point of that low latency. If then subsequently the actions that need to be taken to prevent any problems in application, IOT applications, remote surgeries, uh, self driving vehicles, et cetera, et cetera. What if that's where people are sitting and sipping their coffees and trying to take action that needs to be in low latency as well. Right? So these are, I think some of the fundamental things that you have to know your data, your use cases, that location, where it needs to be exchanged, what are the parameters around that for extending that data? >>And I think from that point at one word, it's all about realizing, you know, sense of business outcomes. Unless AI comes in as a digital labor that shows you, I have, I have reduced your this amount of time and that's a result of big problems or identified problems for anything. Or I have saved you this much resource in a month, in a year or whatever timeline that people want to see it. So I think those are some of the initial starting points, and then it all starts coming together. But the key is it's not one system that can do everything. You have to have a way where, you know, you can share data once you've caught all of that data into one system. Maybe you can send it to another system at make more, take more advantage, right? That system might be an AI and IOT system, which is just looking at all of your street and make it sure that Hey parents. So it's still off just to be more carbon neutral and all that great stuff, et cetera, et cetera, >>Stuff for the audience to can cigarette rush, take us time from here. What are some of the takeaways that you think the audience really needs to be laser focused on as we move forward into the next year? You know, one thing that, uh, I think a key takeaway is, um, uh, you know, as we embark on 2021, closing the gap between intent and outcome and outputs and outcome will become critical, is critical. Uh, you know, especially for, uh, you know, uh, digital transformation at scale for organizations context in the, you know, for customer experience becomes even more critical as who Swan Huseman was talking, uh, you know, being network network aware network availability is, is a necessary condition, but not sufficient condition anymore. Right? The what, what, what customers have to go towards is going from network availability to network agility with high security, uh, what we call app aware networks, right? How do you differentiate between a trade, a million dollar trade that's happening between, uh, you know, London and New York, uh, uh, versus a YouTube video training that an employee is going through? Worse is a YouTube video that millions of customers are, are >>Watching, right? Three different context, three different customer scenarios, right? That is going to be critical. And last but not least feedback loop, uh, you know, responsiveness is all about feedback loop. You cannot predict everything, but you can respond to things faster. I think these are sort of the three, three things that, uh, that, uh, you know, customers aren't going to have to have to really think about. And that's also where I believe AI ops, by the way, AI ops and I I'm. Yeah. You know, one of the points that was smart and shout out to what he was saying was heterogeneity is key, right? There is no homogeneous tool in the world that can solve problems. So you want an open extensible system of intelligence that, that can harness data from disparate data sources provide that visualization, the actionable insight and the human augmented recommendation systems that are so needed for, uh, you know, it operators to be successful. I think that's where it's going. >>Amazing. You guys just provided so much content context recommendations for the audience. I think we accomplished our goal on this. I'll call it power panel of not only getting to a consensus of what, where AI ops needs to go in the future, but great recommendations for what businesses in any industry need to be on the lookout for rich Huisman Raj, thank you for joining me today. We want to thank you for watching. This was such a rich session. You probably want to watch it again. Thanks for your time. Thanks so much for attending and participating in the AI OBS virtual forum. We really appreciate your time and we hope you really clearly understand the value that AI ops platforms can deliver to many types of organizations. I'm Lisa Martin, and I want to thank our speakers today for joining. We have rich lane from Forrester who's fund here from Verizon and Raj from Broadcom. Thanks everyone. Stay safe..

Published Date : Dec 2 2020

SUMMARY :

ops virtual forum brought to you by Broadcom. It's great to have you today. I think it's going to be a really fun conversation to have today. that is 2020 that are going to be continuing into the next year. to infrastructure, you know, or we're in the, in the cloud or a hybrid or multi-cloud, in silos now, uh, in, in, you know, when you add to that, we don't mean, you know, uh, lessening head count because we can't do that. It's not going to go down and as consumers, you know, just to institutional knowledge. four or five hours of, uh, you know, hunting and pecking and looking at things and trying to try And I think, you know, having all those data and understanding the cause and effect of things increases, if I make a change to the underlying architectures that help move the needle forward, continue to do so for the foreseeable future, for them to be able and it also shows the ROI of doing this because there is some, you know, you know, here's the root cause you should investigate this huge, huge thing. So getting that sort of, uh, you know, In a more efficient manner, when you think about an incident occurring, You know, uh, they open a ticket and they enrich the ticket. Um, I think, uh, you know, a lot of, a lot of I do want to ask you what are some of these? it where the product owner is, you know, and say, okay, this is what it gets you. you know, in talking to one company, they were like, yeah, we're so excited for this. And it wasn't because we did anything wrong or the system And then we had to go through an evolution of, you know, just explaining we were 15 What do you recommend? the CIO, the VP of ops is like, you know, I I've signed lots of checks over We know that every hour system down, I think, uh, you know, is down say, and you know, you have a customer service desk of a thousand customer I think you set the stage for that rich beautifully, and you were right. Welcome back to the Broadcom AI ops, virtual forum, Lisa Martin here talking with Eastman Nasir Uh, what a pleasure. So 2020 the year of that needs no explanation, right? or New York, and also this whole consciousness about, you know, You know, all of these things require you to have this you know, we've had to enable these, uh, these virtual classrooms ensuring So you articulated the challenges really well. you know, even because of you just use your signal on the quality talking to somebody else, you know, just being away on holiday. So spectrum, it doesn't just need to be intuitive. What are some of the examples that you gave? fruit, like for somebody like revising who is a managed services provider, you know, You're going to go investigate 50 bags or do you want to investigate where And then subsequently, you know, like isolating it to the right cost uh, which is just providing those resources, you know, on demand. So it was when you clearly articulated some obvious, low hanging fruit and use cases that How do you maintain integrity of your you have your network. right, if something's sitting in the cloud, you were able to integrate for that with obviously the I'm thinking of, you know, the integrity of teams aligning business in it, which we probably can't talk So one example being that, you know, you know, have that superiority and continue it. Thank you so much for joining me today and giving us We'll be right back with our next segment. the solution gives you actionable insights by correlating an aggregating data and applying AI brought to you by Broadcom. Welcome back to the AI ops virtual forum, Lisa Martin here with Srinivasan, as a, as a team that is, uh, you know, that's working behind the scenes However, uh, you know, application of AI ML uh, you know, that that serve up your business services. But I want you to explain how can AI ops help with that alignment and align it outcome that said, uh, you know, these personas need mechanisms But in the, in the context of, uh, you know, So, whereas one of the things that you said there is that it's imperative for the business to find a problem before of the same system, you know, if you're a customer and if you're whipping up your mobile app I often, uh, you know, work with customers around, you know, We look at digital transformation at scale. uh, you know, Nike matures, its digital business outcomes by shoes per second, these measures, uh, you know, uh, for a bank, it may be deposits per month, Uh, and, uh, you know, which may be on your main frames, what we call mobile to mainframe scenario, There are millions of, uh, uh, you know, customers and hundreds The head of AI ops at Broadcom is now going to take you through a quick demo. I'm going to do today is talk through some of the key capabilities and differentiators of here, you can see that the issue is related to the mainframe kick server. You can expand to see the actual alarm, which is sourced from the mainframe operational intelligence. This increases the Elmont support cost to tens of dollars per virtual forum brought to you by Broadcom. Great to have you back. The last thing to change because we're spending so much time doing project work and modernization and fighting Problem is going to get worse. And I say, you know, how many people have three X, five X, you know, uh, things to monitor them. So I think it's, I would just relate it to a couple of things So to speak, you can drive these efficiencies through automating a lot of I mean, uh, you know, uh, to put things in perspective, I think, you know, more often than not, uh, you know, So we got kind of consensus there, as you said, uh, website, um, uh, you know, down detector.com, First of all, what are the things, you know, which could be better utilized Opportunity to reduce the noise of a trouble tickets handling. So, and so many of those are not really, not having to deal with problems, which nobody can resolve, which are not meant to be dealt with. So those are the, So there's some of the immediate cost saving them. the seven layers that I mentioned with the OSI reference model across network and security and I'm going to use a really interesting example. The integrity of the IOT machine is He has, everything is being told to the machine really fast with sending yeah. And if that's okay, And I believe, to the business in the form of the revenue. You know, all that stuff. to, you know, Roger's point your customer should not be identifying your problems before up with you from that senior analyst perspective, how can companies use I think with the, uh, you know, one of the biggest struggles we've always had in operations is isn't, So you were loyal to that because it was in your neighborhood, um, online that doesn't exist anymore. Uh, and I think companies are starting and then the pandemic certainly, you know, and is, you know, AI ops the way forward for them. Raj, I want to start with you readiness factor. I think, uh, you know, our, And that's squarely, very I ops is, you know, is going as it, Uh, I'd say at the end of last year, we're, you know, two years ago, people I'd and I'll, I'll, I'll send this out to any of the operations teams that are thinking about going down this path is that you have to understand So I'd say that the, to people's like start thinking about what this means One thing would be in the form of, you know, what are some of the things they have to be concerned So I think these are some of the very critical points because you can have an AI solution, you have to know your data, your use cases, that location, where it needs to be exchanged, You have to have a way where, you know, you can share data once you've uh, you know, uh, digital transformation at scale for organizations context recommendation systems that are so needed for, uh, you know, and we hope you really clearly understand the value that AI ops platforms can deliver to many

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
RichardPERSON

0.99+

VerizonORGANIZATION

0.99+

Lisa MartinPERSON

0.99+

LondonLOCATION

0.99+

two minutesQUANTITY

0.99+

EuropeLOCATION

0.99+

50 bagsQUANTITY

0.99+

BroadcomORGANIZATION

0.99+

two hoursQUANTITY

0.99+

threeQUANTITY

0.99+

KenyaLOCATION

0.99+

RogerPERSON

0.99+

BrianPERSON

0.99+

CiscoORGANIZATION

0.99+

millionsQUANTITY

0.99+

20 minutesQUANTITY

0.99+

Roger RajagopalPERSON

0.99+

sixQUANTITY

0.99+

360 degreeQUANTITY

0.99+

11.3%QUANTITY

0.99+

2021DATE

0.99+

12%QUANTITY

0.99+

RajPERSON

0.99+

20 hoursQUANTITY

0.99+

15 thingsQUANTITY

0.99+

63%QUANTITY

0.99+

Reggie GopaulPERSON

0.99+

SrinivasanPERSON

0.99+

two secondsQUANTITY

0.99+

New YorkLOCATION

0.99+

eight yearsQUANTITY

0.99+

MicrosoftORGANIZATION

0.99+

12QUANTITY

0.99+

second areaQUANTITY

0.99+

10 yearsQUANTITY

0.99+

2020DATE

0.99+

9%QUANTITY

0.99+

AWSORGANIZATION

0.99+

second pieceQUANTITY

0.99+

next weekDATE

0.99+

NikeORGANIZATION

0.99+

2022DATE

0.99+

third areaQUANTITY

0.99+

15 yearsQUANTITY

0.99+

fiveQUANTITY

0.99+

LisaPERSON

0.99+

SecondQUANTITY

0.99+

40 personQUANTITY

0.99+

six hoursQUANTITY

0.99+

thousandsQUANTITY

0.99+

24 peopleQUANTITY

0.99+

next yearDATE

0.99+

HusemanPERSON

0.99+

Swan HusemanPERSON

0.99+

hundredsQUANTITY

0.99+

Bureau of economic analysisORGANIZATION

0.99+

fourQUANTITY

0.99+

last weekDATE

0.99+

YouTubeORGANIZATION

0.99+

TeslaORGANIZATION

0.99+

third dayQUANTITY

0.99+

AppleORGANIZATION

0.99+

six servicesQUANTITY

0.99+

three yearsQUANTITY

0.99+

one systemQUANTITY

0.99+

Sreenivasan Rajagopal, Broadcom | AIOps Virtual Forum 2020


 

>>from around the globe. It's the Cube with digital coverage of AI ops Virtual Forum Brought to you by Broadcom Welcome back to the AI Ops Virtual Forum. Lisa Martin here with Srinivasan Rajagopal, the head of products and strategy at Broadcom Raj Welcome. >>Good to be here, Lisa. >>I'm excited for our conversation, so I wanted to dive right into a term that we hear all the time. Operational excellence, right? We hear it everywhere in marketing, etcetera. But why is it so important to organizations as they head into 2021 tell us how ai ops as a platform can help? >>Yeah. Thank you. First off, I wanna I wanna welcome our viewers back and I'm very excited Toe share more info on this topic. You know, here's what we believe. As we work with large organizations, we see all our organizations are poised toe get out off the pandemic and look for growth for their own business and helping customers get through this tough time. So fiscal year 2021 we believe, is going to be a combination off, you know, resiliency and agility at the at the same time. So operational excellence is critical because the business has become mawr digital, right? There are gonna be three things that are gonna be more sticky. You know, remote work is gonna be more sticky. Um, cost savings and efficiency is going to be an imperative for organizations. And the continued acceleration of digital transformation off enterprises at scale is going to be in reality. So when you put all these three things together as a team, that is, you know, that's working behind the scenes toe help the businesses succeed. Operational excellence is going to be make or break for organizations. >>Russia with that said, if we kind of strip it down to the key capabilities, what are those key capabilities that companies need to be looking for in an AI ops solution? >>Yeah, you know. So first foremost AI ops means many things to many, many folks. So let's take a moment to simply define it. The way we defined AI ops is it's a system off intelligence human augmented system that brings together full visibility across app, infra and network elements that brings together despite of data sources on provides actionable intelligence and uniquely offers intelligent automation. Now the technology many folks draw is the self driving car, right? I mean, we are in the world of Tesla's, but, you know, but self driving data center is is too far away, right? Autonomous systems are still far away. However, you know, application off the I M l techniques toe help deal with volume velocity, veracity of information. Eyes is critical. So that's how we look at AI ops and some of the key capabilities that we that we that we work with our customers to help them around 48 years. Right? First one is eyes and years. What we call full stack, observe ability. If you do not know what is happening in your systems, you know that that serve up your business services, it's gonna be pretty hard to do anything in terms of responsiveness, right? So from stack of their ability, the second piece is what we call actionable insights. So when you have disparaged data sources, tool sprawls, data coming at you from, you know from database systems, I T systems, customer management systems, ticketing systems, how do you find the needle from the haystack? And how do you respond rapidly from a myriad off problems? A sea off read The third area is what we call intelligent automation. Well, Identifying the problem toe Act on is important and then acting on. Automating that and creating a recommendation system where you know you could be proactive about that is even more important. And finally, all of this focuses on efficiency. What about effectiveness? Effectiveness comes when you create a feedback loop when what happens in production is related to your support systems and your developers so that they can respond rapidly. So we call that continuous feedback. So these are the four key capabilities that you know you should look for in an AI ops system. And that's what we offer us. >>Alright, Russia. There's four key capabilities that businesses need to be looking for. I'm wondering how those help to align business and i t. It's again like operational excellence. It's something that we talk about a lot is the alignment of business and I t a lot more challenging. Is your something done right? But I want you to explain how can a iob help with that alignment and align? I t outputs to business outcomes. >>So you know, one of the things I'm going to say something that this, that is that is simple. But it's harder. Alignment is not on systems. Alignment is with people, right? So when people align when organizations aligned, when cultures align, dramatic things can happen. So in the context off AI ops, we see when when saris aligned with the develops engineers and information architects. And, uh, you know, I t operators, you know, they enable organizations to reduce the gap between intent and outcome or output an outcome that said, you know, these personas need mechanisms toe help them better align, right, help them Better visual. I see the you know what we call single source of truth, right? So there are four key things that I wanna call out when we work with large enterprises. We find that customer journey alignment with the you know what we call I T systems is critical. So how do you understand your business imperatives and your customer journey goals? Whether it is card toe purchase or whether it is, you know, Bill shock scenarios and swan alignment on customer journey to your I T systems is one area that you can reduce the gap. The second area is how do you create a scenario where your teams can find problems before your customers do right out. It's scenarios and so on. So that's the second area off alignment. The third area off alignment is how can you measure business impact driven services right? There are several services that an organization off course as the 19 system. Some services are more critical to the business. Well, then, others and thes change in a dynamic environment. So how do you How do you understand that? How do you measure that? And how? How do you find the gaps there? So that that's the 3rd 80 off alignment that we that we help. And last but not least, there are. There are things like NPS scores and others that that help us understand alignment. But those are more long term. But in the in the context off, you know, operating digitally. You want to use customer experience and, you know single business outcome as as a key alignment factor, and then work with your systems of engagement and systems of interaction, along with your key personas to create that alignment. It's a people process technology challenge, actually. >>So where is one of the things that you said there is that it's imperative for the business toe. Find a problem before a customer does. And you talked about outages there. That's always a goal for businesses, right to prevent those outages. How can Ai ops help with that? >>Yeah, so, you know, out they just talk, you know, go to resiliency off a system, right? And they also goto have, you know, agility off the same system. You know, if you are a customer and if you're ripping up your mobile happened, it takes more than you know, three milliseconds. You know, you're probably losing that customer, right? So I would just mean different things, you know? And there's an interesting website called don't detector dot com that actually tracks all the outages of publicly available services, whether it's your bank or your, you know, telecom service or mobile service and so on and so forth. In fact, the key question around outages for from from you know, executives are the question of Are you ready? Right? Are you ready to respond to the needs off your customers and your business? Are you ready toe rapidly to solve an issue that is impacting customer experience and therefore satisfaction. Are you creating a digital trust system where customers can be, You know, you know, customers can feel that their information is secure when they transact with you. All of these getting toe the notion of resiliency and outages. Now, you know, one of the things that I often you know work with customers around, you know, that we find is the radius off. Impact is important when you deal with outages. What I mean by that is problems occur, right? How do you respond? How quickly do you take? Two seconds? Two minutes, 20 minutes. Two hours, 20 hours. Right To resolve that problem. That radius of impact is important. That's where you know you have to bring again. People process technology together to solve that. And the key thing is, you need a system of intelligence that can aid you your teams, you know, look at the same set of parameters so that you can respond faster. That's the key here. >>But as we look at digital transformation at scale, Raj, how does a apps help influence that? >>You know, I'm gonna take a slightly long winded way to answer this question. See, when it comes to digital transformation at scale, the focus on business purpose and business outcome becomes extremely critical. And then the alignment off that to your digital supply chain right are the are the are the key factors that differentiate vintners in the in their digital transformation game. Really? What we have seen with with winners is they operate very differently. Like, for example, you know, 19 assures its digital business outcomes by shoes per second, right apple buy iPhones per per minute. Tesla by model threes per month. Are you getting getting it right? I mean, you wanna have, ah, clear business outcome, which is a measure off your business. In effect, I mean, easy right, which which my daughter use. And I use very well, right? You know, they measured by revenue per hour, right? I mean, so these are key measures, and when you have a key business outcome measure like that, you can align everything else because you know what these measures you know, for a bank, it may be deposits per month. Right now, when you move money from checking account to savings account or when you do direct deposits, those are you know, banks need liquidity and so on and so forth. But, you know, the key thing is that single business outcome has a starburst effect inside the I T. Organization that touches a single money movement from checking account to savings account can touch about 75 disparage systems internally. Right? So those think about right. I mean, all we're doing is moving money from checking accounts savings account. Now that goats in tow, a IittIe production system, there are several applications. There is a database there is there are infrastructures, their load balancers, that our webs, you know, the Web server components, which then touches your your middleware component, which is a queuing system right, which then touches your transactional system on. Do you know which may be on your mainframes what we call mobile toe mainframe scenario, right? And we're not done yet. Then you have a security and regulatory compliance system that you have to touch a fraud prevention system that you have to touch right, a State Department regulation that you may have to meet and on and on and on, right? This is the challenge that I t operation teams phase. And when you have millions of customers transacting right? Certainly this challenge cannot be, you know, managed by, you know, human beings alone. So therefore, you need a system off intelligence that augments human intelligence and acts as you, you know, your your eyes and ears in of a toe point pinpoint. Their problems are right. So digital transformation at scale really requires a well thought out ai ops system a platform and open extensible platform that you know, that is heterogeneous in nature because their stools problems in organizations. There is, uh, you know, a lot of data bases in systems. There are million's off, you know, customers and hundreds off partners and vendors, you know, making up that digital supply chain. So, you know, AI ops is at the center off, enabling an organization achieved digital up, you know, transformation at scale. Last but not least, you need continuous feedback loop. Continuous feedback loop is the ability for a production system toe. Inform your develops teams your finance teams, your customer experience teams your cost Modeling teams about what is going on say that they can so that they can reduce the intent outcome gap. All of this need to come together. What we call biz obs for ideal abs. >>That was a great example of how you talked about the Starburst effect. Actually never thought about it in that way. When you give the banking example but what you should is the magnitude of systems, the fact that people alone really need help with that and why intelligent automation and air ops could be transformative and enable that scale. Raj, it's always a pleasure to talk with you. Thanks for joining me today. Yeah, >>great to be here >>and we'll be right back with our next segment.

Published Date : Nov 23 2020

SUMMARY :

AI ops Virtual Forum Brought to you by Broadcom Welcome the time. that is, you know, that's working behind the scenes toe help the businesses So when you have disparaged data sources, But I want you to explain how can a iob help with that alignment So you know, one of the things I'm going to say something that this, that is that So where is one of the things that you said there is that it's imperative for the business toe. the key question around outages for from from you know, that our webs, you know, the Web server components, which then touches your your middleware component, When you give the banking example but what you should is the magnitude of

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Srinivasan RajagopalPERSON

0.99+

Lisa MartinPERSON

0.99+

Two hoursQUANTITY

0.99+

BroadcomORGANIZATION

0.99+

2021DATE

0.99+

Two secondsQUANTITY

0.99+

20 hoursQUANTITY

0.99+

Sreenivasan RajagopalPERSON

0.99+

Two minutesQUANTITY

0.99+

second areaQUANTITY

0.99+

third areaQUANTITY

0.99+

second pieceQUANTITY

0.99+

TeslaORGANIZATION

0.99+

20 minutesQUANTITY

0.99+

appleORGANIZATION

0.99+

LisaPERSON

0.99+

todayDATE

0.99+

RajPERSON

0.99+

oneQUANTITY

0.98+

fiscal year 2021DATE

0.98+

RussiaLOCATION

0.98+

firstQUANTITY

0.97+

three thingsQUANTITY

0.97+

one areaQUANTITY

0.97+

FirstQUANTITY

0.97+

singleQUANTITY

0.96+

First oneQUANTITY

0.95+

iPhonesCOMMERCIAL_ITEM

0.95+

around 48 yearsQUANTITY

0.95+

about 75 disparage systemsQUANTITY

0.94+

three millisecondsQUANTITY

0.94+

3rd 80QUANTITY

0.93+

pandemicEVENT

0.92+

single sourceQUANTITY

0.92+

2020DATE

0.92+

I T.ORGANIZATION

0.89+

single moneyQUANTITY

0.88+

single businessQUANTITY

0.85+

four key capabilitiesQUANTITY

0.82+

Broadcom RajORGANIZATION

0.81+

millions of customersQUANTITY

0.8+

million'sQUANTITY

0.79+

four key thingsQUANTITY

0.79+

threesQUANTITY

0.77+

StarburstOTHER

0.76+

secondQUANTITY

0.75+

19 systemQUANTITY

0.74+

hundreds offQUANTITY

0.72+

19QUANTITY

0.68+

thingsQUANTITY

0.59+

ForumEVENT

0.57+

State DepartmentORGANIZATION

0.55+

AIOpsEVENT

0.53+

doORGANIZATION

0.51+

opsORGANIZATION

0.42+

AI Ops Virtual ForumEVENT

0.4+

ForumORGANIZATION

0.33+

Abhinav Joshi & Tushar Katarki, Red Hat | KubeCon + CloudNativeCon Europe 2020 – Virtual


 

>> Announcer: From around the globe, it's theCUBE with coverage of KubeCon + CloudNativeCon Europe 2020 Virtual brought to you by Red Hat, the Cloud Native Computing Foundation and Ecosystem partners. >> Welcome back I'm Stu Miniman, this is theCUBE's coverage of KubeCon + CloudNativeCon Europe 2020, the virtual event. Of course, when we talk about Cloud Native we talk about Kubernetes there's a lot that's happening to modernize the infrastructure but a very important thing that we're going to talk about today is also what's happening up the stack, what sits on top of it and some of the new use cases and applications that are enabled by all of this modern environment and for that we're going to talk about artificial intelligence and machine learning or AI and ML as we tend to talk in the industry, so happy to welcome to the program. We have two first time guests joining us from Red Hat. First of all, we have Abhinav Joshi and Tushar Katarki they are both senior managers, part of the OpenShift group. Abhinav is in the product marketing and Tushar is in product management. Abhinav and Tushar thank you so much for joining us. >> Thanks a lot, Stu, we're glad to be here. >> Thanks Stu and glad to be here at KubeCon. >> All right, so Abhinav I mentioned in the intro here, modernization of the infrastructure is awesome but really it's an enabler. We know... I'm an infrastructure person the whole reason we have infrastructure is to be able to drive those applications, interact with my data and the like and of course, AI and ML are exciting a lot going on there but can also be challenging. So, Abhinav if I could start with you bring us inside your customers that you're talking to, what are the challenges, the opportunities? What are they seeing in this space? Maybe what's been holding them back from really unlocking the value that is expected? >> Yup, that's a very good question to kick off the conversation. So what we are seeing as an organization they typically face a lot of challenges when they're trying to build an AI/ML environment, right? And the first one is like a talent shortage. There is a limited amount of the AI, ML expertise in the market and especially the data scientists that are responsible for building out the machine learning and the deep learning models. So yeah, it's hard to find them and to be able to retain them and also other talents like a data engineer or app DevOps folks as well and the lack of talent can actually stall the project. And the second key challenge that we see is the lack of the readily usable data. So the businesses collect a lot of data but they must find the right data and make it ready for the data scientists to be able to build out, to be able to test and train the machine learning models. If you don't have the right kind of data to the predictions that your model is going to do in the real world is only going to be so good. So that becomes a challenge as well, to be able to find and be able to wrangle the right kind of data. And the third key challenge that we see is the lack of the rapid availability of the compute infrastructure, the data and machine learning, and the app dev tools for the various personas like a data scientist or data engineer, the software developers and so on that can also slow down the project, right? Because if all your teams are waiting on the infrastructure and the tooling of their choice to be provisioned on a recurring basis and they don't get it in a timely manner, it can stall the projects. And then the next one is the lack of collaboration. So you have all these kinds of teams that are involved in the AI project, and they have to collaborate with each other because the work one of the team does has a dependency on a different team like say for example, the data scientists are responsible for building the machine learning models and then what they have to do is they have to work with the app dev teams to make sure the models get integrated as part of the app dev processes and ultimately rolled out into the production. So if all these teams are operating in say silos and there is lack of collaboration between the teams, so this can stall the projects as well. And finally, what we see is the data scientists they typically start the machine learning modeling on their individual PCs or laptops and they don't focus on the operational aspects of the solution. So what this means is when the IT teams have to roll all this out into a production kind of deployment, so they get challenged to take all the work that has been done by the individuals and then be able to make sense out of it, be able to make sure that it can be seamlessly brought up in a production environment in a consistent way, be it on-premises, be it in the cloud or be it say at the edge. So these are some of the key challenges that we see that the organizations are facing, as they say try to take the AI projects from pilot to production. >> Well, some of those things seem like repetition of what we've had in the past. Obviously silos have been the bane of IT moving forward and of course, for many years we've been talking about that gap between developers and what's happening in the operation side. So Tushar, help us connect the dots, containers, Kubernetes, the whole DevOps movement. How is this setting us up to actually be successful for solutions like AI and ML? >> Sure Stu I mean, in fact you said it right like in the world of software, in the world of microservices, in the world of app modernization, in the world of DevOps in the past 10, 15 years, but we have seen this evolution revolution happen with containers and Kubernetes driving more DevOps behavior, driving more agile behavior so this in fact is what we are trying to say here can ease up the cable to EIML also. So the various containers, Kubernetes, DevOps and OpenShift for software development is directly applicable for AI projects to make them move agile, to get them into production, to make them more valuable to organization so that they can realize the full potential of AI. We already touched upon a few personas so it's useful to think about who the users are, who the personas are. Abhinav I talked about data scientists these are the people who obviously do the machine learning itself, do the modeling. Then there are data engineers who do the plumbing who provide the essential data. Data is so essential to machine learning and deep learning and so there are data engineers that are app developers who in some ways will then use the output of what the data scientists have produced in terms of models and then incorporate them into services and of course, none of these things are purely cast in stone there's a lot of overlap you could find that data scientists are app developers as well, you'll see some of app developers being data scientist later data engineer. So it's a continuum rather than strict boundaries, but regardless what all of these personas groups of people need or experts need is self service to that preferred tools and compute and storage resources to be productive and then let's not forget the IT, engineering and operations teams that need to make all this happen in an easy, reliable, available manner and something that is really safe and secure. So containers help you, they help you quickly and easily deploy a broad set of machine learning tools, data tools across the cloud, the hybrid cloud from data center to public cloud to the edge in a very consistent way. Teams can therefore alternatively modify, change a shared container images, machine learning models with (indistinct) and track changes. And this could be applicable to both containers as well as to the data by the way and be transparent and transparency helps in collaboration but also it could help with the regulatory reasons later on in the process. And then with containers because of the inherent processes solution, resource control and protection from threat they can also be very secure. Now, Kubernetes takes it to the next level first of all, it forms a cluster of all your compute and data resources, and it helps you to run your containerized tools and whatever you develop on them in a consistent way with access to these shared compute and centralized compute and storage and networking resources from the data center, the edge or the public cloud. They provide things like resource management, workload scheduling, multi-tendency controls so that you can be a proper neighbors if you will, and quota enforcement right? Now that's Kubernetes now if you want to up level it further if you want to enhance what Kubernetes offers then you go into how do you write applications? How do you actually make those models into services? And that's where... and how do you lifecycle them? And that's sort of the power of Helm and for the more Kubernetes operators really comes into the picture and while Helm helps in installing some of this for a complete life cycle experience. A kubernetes operator is the way to go and they simplify the acceleration and deployment and life cycle management from end-to-end of your entire AI, ML tool chain. So all in all organizations therefore you'll see that they need to dial up and define models rapidly just like applications that's how they get ready out of it quickly. There is a lack of collaboration across teams as Abhinav pointed out earlier, as you noticed that has happened still in the world of software also. So we're talking about how do you bring those best practices here to AI, ML. DevOps approaches for machine learning operations or many analysts and others have started calling as MLOps. So how do you kind of bring DevOps to machine learning, and fosters better collaboration between teams, application developers and IT operations and create this feedback loop so that the time to production and the ability to take more machine learning into production and ML-powered applications into production increase is significant. So that's kind of the, where I wanted shine the light on what you were referring to earlier, Stu. >> All right, Abhinav of course one of the good things about OpenShift is you have quite a lot of customers that have deployed the solution over the years, bring us inside some of your customers what are they doing for AI, ML and help us understand really what differentiates OpenShift in the marketplace for this solution set. >> Yeah, absolutely that's a very good question as well and we're seeing a lot of traction in terms of all kinds of industries, right? Be it the financial services like healthcare, automotive, insurance, oil and gas, manufacturing and so on. For a wide variety of use cases and what we are seeing is at the end of the day like all these deployments are focused on helping improve the customer experience, be able to automate the business processes and then be able to help them increase the revenue, serve their customers better, and also be able to save costs. If you go to openshift.com/ai-ml it's got like a lot of customer stories in there but today I will not touch on three of the customers we have in terms of the different industries. The first one is like Royal Bank of Canada. So they are a top global financial institution based out of Canada and they have more than 17 million clients globally. So they recently announced that they build out an AI-powered private cloud platform that was based on OpenShift as well as the NVIDIA DGX AI compute system and this whole solution is actually helping them to transform the customer banking experience by being able to deliver an AI-powered intelligent apps and also at the same time being able to improve the operational efficiency of their organization. And now with this kind of a solution, what they're able to do is they're able to run thousands of simulations and be able to analyze millions of data points in a fraction of time as compared to the solution that they had before. Yeah, so like a lot of great work going on there but now the next one is the ETCA healthcare. So like ETCA is one of the leading healthcare providers in the country and they're based out of the Nashville, Tennessee. And they have more than 184 hospitals as well as more than 2,000 sites of care in the U.S. as well as in the UK. So what they did was they developed a very innovative machine learning power data platform on top of our OpenShift to help save lives. The first use case was to help with the early detection of sepsis like it's a life-threatening condition and then more recently they've been able to use OpenShift in the same kind of stack to be able to roll out the new applications that are powered by machine learning and deep learning let say to help them fight COVID-19. And recently they did a webinar as well that had all the details on the challenges they had like how did they go about it? Like the people, process and technology and then what the outcomes are. And we are proud to be a partner in the solution to help with such a noble cause. And the third example I want to share here is the BMW group and our partner DXC Technology what they've done is they've actually developed a very high performing data-driven data platform, a development platform based on OpenShift to be able to analyze the massive amount of data from the test fleet, the data and the speed of the say to help speed up the autonomous driving initiatives. And what they've also done is they've redesigned the connected drive capability that they have on top of OpenShift that's actually helping them provide various use cases to help improve the customer experience. With the customers and all of the customers are able to leverage a lot of different value-add services directly from within the car, their own cars. And then like last year at the Red Hat Summit they had a keynote as well and then this year at Summit, they were one of the Innovation Award winners. And we have a lot more stories but these are the three that I thought are actually compelling that I should talk about here on theCUBE. >> Yeah Abhinav just a quick follow up for you. One of the things of course we're looking at in 2020 is how has the COVID-19 pandemic, people working from home how has that impacted projects? I have to think that AI and ML are one of those projects that take a little bit longer to deploy, is it something that you see are they accelerating it? Are they putting on pause or are new project kicking off? Anything you can share from customers you're hearing right now as to the impact that they're seeing this year? >> Yeah what we are seeing is that the customers are now even more keen to be able to roll out the digital (indistinct) but we see a lot of customers are now on the accelerated timeline to be able to say complete the AI, ML project. So yeah, it's picking up a lot of momentum and we talk to a lot of analyst as well and they are reporting the same thing as well. But there is the interest that is actually like ramping up on the AI, ML projects like across their customer base. So yeah it's the right time to be looking at the innovation services that it can help improve the customer experience in the new virtual world that we live in now about COVID-19. >> All right, Tushar you mentioned that there's a few projects involved and of course we know at this conference there's a very large ecosystem. Red Hat is a strong contributor to many, many open source projects. Give us a little bit of a view as to in the AI, ML space who's involved, which pieces are important and how Red Hat looks at this entire ecosystem? >> Thank you, Stu so as you know technology partnerships and the power of open is really what is driving the technology world these days in any ways and particularly in the AI ecosystem. And that is mainly because one of the machine learning is in a bootstrap in the past 10 years or so and a lot of that emerging technology to take advantage of the emerging data as well as compute power has been built on the kind of the Linux ecosystem with openness and languages like popular languages like Python, et cetera. And so what you... and of course tons of technology based in Java but the point really here is that the ecosystem plays a big role and open plays a big role and that's kind of Red Hat's best cup of tea, if you will. And that really has plays a leadership role in the open ecosystem so if we take your question and kind of put it into two parts, what is the... what we are doing in the community and then what we are doing in terms of partnerships themselves, commercial partnerships, technology partnerships we'll take it one step at a time. In terms of the community itself, if you step back to the three years, we worked with other vendors and users, including Google and NVIDIA and H2O and other Seldon, et cetera, and both startups and big companies to develop this Kubeflow ecosystem. The Kubeflow is upstream community that is focused on developing MLOps as we talked about earlier end-to-end machine learning on top of Kubernetes. So Kubeflow right now is in 1.0 it happened a few months ago now it's actually at 1.1 you'll see that coupon here and then so that's the Kubeflow community in addition to that we are augmenting that with the Open Data Hub community which is something that extends the capabilities of the Kubeflow community to also add some of the data pipelining stuff and some of the data stuff that I talked about and forms a reference architecture on how to run some of this on top of OpenShift. So the Open Data Hub community also has a great way of including partners from a technology partnership perspective and then tie that with something that I mentioned earlier, which is the idea of Kubernetes operators. Now, if you take a step back as I mentioned earlier, Kubernetes operators help manage the life cycle of the entire application or containerized application including not only the configuration on day one but also day two activities like update and backups, restore et cetera whatever the application needs. Afford proper functioning that a "operator" needs for it to make sure so anyways, the Kubernetes operators ecosystem is also flourishing and we haven't faced that with the OperatorHub.io which is a community marketplace if you will, I don't call it marketplace a community hub because it's just comprised of community operators. So the Open Data Hub actually can take community operators and can show you how to run that on top of OpenShift and manage the life cycle. Now that's the reference architecture. Now, the other aspect of it really is as I mentioned earlier is the commercial aspect of it. It is from a customer point of view, how do I get certified, supported software? And to that extent, what we have is at the top of the... from a user experience point of view, we have certified operators and certified applications from the AI, ML, ISV community in the Red Hat marketplace. And from the Red Hat marketplace is where it becomes easy for end users to easily deploy these ISVs and manage the complete life cycle as I said. Some of the examples of these kinds of ISVs include startups like H2O although H2O is kind of well known in certain sectors PerceptiLabs, Cnvrg, Seldon, Starburst et cetera and then on the other side, we do have other big giants also in this which includes partnerships with NVIDIA, Cloudera et cetera that we have announced, including our also SaaS I got to mention. So anyways these provide... create that rich ecosystem for data scientists to take advantage of. A TEDx Summit back in April, we along with Cloudera, SaaS Anaconda showcased a live demo that shows all these things to working together on top of OpenShift with this operator kind of idea that I talked about. So I welcome people to go and take a look the openshift.com/ai-ml that Abhinav already referenced should have a link to that it take a simple Google search might download if you need some of that, but anyways and the other part of it is really our work with the hardware OEMs right? And so obviously NVIDIA GPUs is obviously hardware, and that accelerations is really important in this world but we are also working with other OEM partners like HP and Dell to produce this accelerated AI platform that turnkey solutions to run your data-- to create this open AI platform for "private cloud" or the data center. The other thing obviously is IBM, IBM Cloud Pak for Data is based on OpenShift that has been around for some time and is seeing very good traction, if you think about a very turnkey solution, IBM Cloud Pak is definitely kind of well ahead in that and then finally Red Hat is about driving innovation in the open-source community. So, as I said earlier, we are doing the Open Data Hub which that reference architecture that showcases a combination of upstream open source projects and all these ISV ecosystems coming together. So I welcome you to take a look at that at opendatahub.io So I think that would be kind of the some total of how we are not only doing open and community building but also doing certifications and providing to our customers that assurance that they can run these tools in production with the help of a rich certified ecosystem. >> And customer is always key to us so that's the other thing that the goal here is to provide our customers with a choice, right? They can go with open source or they can go with a commercial solution as well. So you want to make sure that they get the best in cloud experience on top of our OpenShift and our broader portfolio as well. >> All right great, great note to end on, Abhinav thank you so much and Tushar great to see the maturation in this space, such an important use case. Really appreciate you sharing this with theCUBE and Kubecon community. >> Thank you, Stu. >> Thank you, Stu. >> Okay thank you and thanks a lot and have a great rest of the show. Thanks everyone, stay safe. >> Thanks you and stay with us for a lot more coverage from KubeCon + CloudNativeCon Europe 2020, the virtual edition I'm Stu Miniman and thank you as always for watching theCUBE. (soft upbeat music plays)

Published Date : Aug 18 2020

SUMMARY :

the globe, it's theCUBE and some of the new use Thanks a lot, Stu, to be here at KubeCon. and the like and of course, and make it ready for the data scientists in the operation side. and for the more Kubernetes operators that have deployed the and also at the same time One of the things of course is that the customers and how Red Hat looks at and some of the data that the goal here is great to see the maturation and have a great rest of the show. the virtual edition I'm Stu Miniman

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Brian GilmorePERSON

0.99+

David BrownPERSON

0.99+

Tim YoakumPERSON

0.99+

Lisa MartinPERSON

0.99+

Dave VolantePERSON

0.99+

Dave VellantePERSON

0.99+

BrianPERSON

0.99+

DavePERSON

0.99+

Tim YokumPERSON

0.99+

StuPERSON

0.99+

Herain OberoiPERSON

0.99+

JohnPERSON

0.99+

Dave ValantePERSON

0.99+

Kamile TaoukPERSON

0.99+

John FourierPERSON

0.99+

Rinesh PatelPERSON

0.99+

Dave VellantePERSON

0.99+

Santana DasguptaPERSON

0.99+

EuropeLOCATION

0.99+

CanadaLOCATION

0.99+

BMWORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

ICEORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

Jack BerkowitzPERSON

0.99+

AustraliaLOCATION

0.99+

NVIDIAORGANIZATION

0.99+

TelcoORGANIZATION

0.99+

VenkatPERSON

0.99+

MichaelPERSON

0.99+

CamillePERSON

0.99+

Andy JassyPERSON

0.99+

IBMORGANIZATION

0.99+

Venkat KrishnamachariPERSON

0.99+

DellORGANIZATION

0.99+

Don TapscottPERSON

0.99+

thousandsQUANTITY

0.99+

Palo AltoLOCATION

0.99+

Intercontinental ExchangeORGANIZATION

0.99+

Children's Cancer InstituteORGANIZATION

0.99+

Red HatORGANIZATION

0.99+

telcoORGANIZATION

0.99+

Sabrina YanPERSON

0.99+

TimPERSON

0.99+

SabrinaPERSON

0.99+

John FurrierPERSON

0.99+

GoogleORGANIZATION

0.99+

MontyCloudORGANIZATION

0.99+

AWSORGANIZATION

0.99+

LeoPERSON

0.99+

COVID-19OTHER

0.99+

Santa AnaLOCATION

0.99+

UKLOCATION

0.99+

TusharPERSON

0.99+

Las VegasLOCATION

0.99+

ValentePERSON

0.99+

JL ValentePERSON

0.99+

1,000QUANTITY

0.99+