Image Title

Search Results for Redis:

Manish Gupta, Redis Labs | Spark Summit East 2017


 

>> Announcer: Live from Boston, Massachusetts, it's theCUBE, covering Spark Summit East 2017. Brought to you by Databricks. Now, here are your hosts Dave Vellante and George Gilbert. >> Welcome back to snowy Boston, everybody. This is theCUBE, the leader in live tech coverage. We're here at Spark Summit East, hashtag SparkSummit. Manish Gupta is here, he's the CMO at Redis Labs. Manish, welcome to theCUBE. >> Thank you, good to be here. >> So, you know, 10 years ago you say you're in the database business and everybody would yawn. Now you're the life of the party. >> Yeah, the world has changed. I think the party has lots and lots of players. We are happy to be on the top of that heap. >> It is a crowded space, so how does Redis Labs differentiate? >> Redis Labs is the company behind the massively popular open source Redis, and Redis became popular because of its performance primarily, and then simplicity. Developers could very easily run up an instance of Redis, solve some very hairy problems, and time to market was a big issue for them. Redis Enterprise took that forward and enabled it to be mission critical, ready for the largest workloads, ready for things that the enterprises need in a highly distributed clustered environment. So they have resilience and they benefit from the performance of Redis. >> And your claim to fame, as you say, is that top-gun performance, you guys will talk about some of the benchmarks later. We're talking about use cases like fraud detection, as example. Obviously ad serving would be another one. But add some color to that if you would. >> Redis is whatever you need to make real time real, Redis plays a very important role. It is able to deliver millions of operations per second with sub-millisecond latency, and that's the hallmark. With data structures that comprise Redis, you can solve the problems in a way, and the reason you can get that performance is because the data structures take some very complex issues and simplify the operation. Depending on the use case, you could use one of the data structures, you can mix and match the data structures, so that's the power of a Redis. We're used for ITO, for machine learning, for metering of billing and telecommunications environment, for personalization, for ad serving with companies like Groupon and others, and the list goes on and on. >> Yeah, you've got a big list on your website of all your customers, so you can check that out. Let's get the business model piece out of the way. Everybody's always fascinated. Okay, you got open source, how do you make money? How does Redis make money? >> Yeah, you know, we believe strategically fostering the growth of open source is foundational in our business model, and we invest heavily both R&D and marketing to do that. On top of that, to enable enterprise success and deployment of Redis, we have the mission critical, highly available Redis Enterprise offerings. Our monetization is entirely based on the Redis Enterprise platform, which takes advantage of the data structures and performance of core Redis, but layers on top management and the capabilities that make things like auto-recovery, auto-sorting, management much, much easier for the enterprise. We make that available in four deployment models. The enterprise can select us as Redis cloud, which runs on a public infrastructure on any of the four major platforms. We also allow for the enterprise to select a VPC environment in their own private clouds. They can also get software and self-manage that, or get our software and we can manage it for them. Four deployment options are the modalities in other ways where the enterprise customers help us monetize. >> When you said four major platforms, you meant cloud platforms? >> That's right. AWS, >> So, AWS, Azure >> Azure, Google, and IBM. >> Is IBM software, got there in the fourth, alright. >> That's right, all four. >> Go to the whip IBM. Go ahead, George. >> Along the lines of the business model, and we were sort of starting to talk about this earlier offline, you're just one component in building an application, and there's always this challenge of, well, I can manage my component better than anyone else, but it's got to fit with a bunch of other vendors' components. How do you make that seamless to the customer so that it's not defaulting over to a cloud vendor who has to build all the components themselves to make it work together? >> Certainly, you know, database is an integral part of your stack, of your application stack, but it is a stack, so there are other components. Redis and Redis Labs has a very, very large ecosystem within which we operate. We work closely with others for interfaces, for connectors, for interoperability, and that's a sustained environment that we invest in on a continuous basis. >> How do handle application consistency? A lot of in the no-SQL world, even in the AWS world, you hear about eventual consistency, but in the real-time world, there's a need for more rigorous, what's your philosophy there, how do you approach that? >> I think that's an issue that many no-SQL vendors have not been able to crack. Redis Labs has been at the forefront of that. We are taking an approach, and we are offering what we call tuneable consistency. Depending on the economics and the business model and the use case, the needs of consistency vary. In some cases, you do need immediate consistency. In other cases, you don't ever need consistency. And to give that flexibility to the customer is very important, so we've taken the approach where you can go from loose consistency to what we call strong eventual consistency. That approach is based on a fairly well trusted architecture and approach called CRDT, Conflict-free Replication Data Type. That approach allows us to, regardless of what the cluster magnitude or the distribution looks like geographically, we can deliver strong eventual consistency which meets the needs of majority of the customers. >> What are you seeing in terms of, you know, also in that a discussion about acid properties, and how many workloads really need acid properties. What are seeing now as you get more cloud native workloads and more no-SQL oriented workloads in terms of the requirement for those acid properties? >> First of all, we truly believe and agree that not all environments required acid support. Having said that, to be a truly credible database, you must support acid, and we do. Redis is acid-compli, supports acid, and Redis Labs certainly supports that. >> I remember on a stage once with Curt Monash, I'm sure you know Curt, right? Very famous database person. And he basically had a similar answer. But you would say that increasingly there are workloads that, the growth workloads don't necessarily require that, is that fair statement? >> That's a fair statement I would say. >> Dave: Great, good. >> There's a trade-off, though, when you talked about strong eventual consistency, potentially you have to wait for, presumably, a quorum of the partitions, I'm getting really technical here, but in other words, you've got a copy of the data here-- >> Dave: Good CMO question. (laughing) >> But your value proposition to the customers, we get this stuff done fast, but if you have to wait for a couple other servers to make sure that they've got the update, that can slow things way down. How does that trade-off work? >> I think that's part of the power of our architecture. We have a nothing shared, single proxy architecture where all of the replication, the disaster recovery, and the consistency management of the back end is handled by the proxy, and we ensure that the performance is not degraded when you are working through the consistency challenges, and that's where significant amount of IP is in the development of that proxy. >> I'll take that as a, let's go into it even more offline. >> Manish: Sounds good. >> And I have some other CMO questions, if I may. A lot of young companies like yours, especially in open source world, when they go to get the word out, they rely on their community, their open source community, and that's the core, and that makes a lot of sense, it's their peeps. As you become, grow more into enterprise grade apps and workloads, how do you extend beyond that? What is Redis Labs doing to sort of reach that C-Suite, are you even trying to reach that C-Suite up level to messaging? How do you as a CMO deal with those challenges? >> Maybe I'll begin by talking about our personas that matter to us in the ecosystem. The enterprise level, the architects, the developers, are the primary target, which we try to influence in early part of the decision cycle, it's at the architectural level. The ultimate teams that manage, run, and operate the infrastructure is certainly the DevOps, or the operations teams, and we spend time there. All along for some of the enterprise engagements, CIOs, chief data officers, and CTOs tend to play a very important role in the decisions and the selection process, and so, we do influence and interact with the C-Suite quite heavily. What the power of the open source gives us is that groundswell of love for Redis. Literally you can walk around a developer environment, such as the Spark Summit here, and you'll find people wearing Redis Geek shirts. And we get emails from Kazakhstan and strange, places from all over the world where we don't necessarily have salesforce, and requesting t-shirts, "send us stickers." Because people love Redis, and the word of mouth, that ground level love for the technology enables the decisions to be so much easier and smoother. We're not convincing, it's not a philosophical battle anymore. It's simply about the use case and the solution where Redis Enterprise fits or doesn't fit. >> Okay, so it really is that core developer community that are your advocates, and they're able to internally sell to the C-Suite. A lot of times the C-Suite, not the CTO so much, but certainly the CIO, CDO are like, "Yeah, yeah, they're geekin' out on some new hot thing. "What's the business impact?" Do you get that question a lot, and how do address it? >> I think then you get to some of the very basic tools, ROI calculators and the value proposition. For the C-level, the message is very simple. We are the least risky bet. We are the best long-term proposition, and we are the best cost answer for their implementation. Particularly as the needs are increasingly becoming more real-time in nature, they are not batch processed. Yes, there will always be some of that, but as the workloads are becoming, there is a need for faster processing, there is a need for quick insights, and real-time is not a moniker anymore, right. Real-time truly needs to be delivered today. And so, I think those three propositions for the C-Suite are resonating very well. >> Let's talk about ROI calculators for a second. I love talking about it because it underscores what a company feels as though its core value proposition is. I would think with Redis Labs part of the value proposition is you are enabling new types of workloads and new types of, whether it's sources of revenue or productivity. And these are generally telephone numbers as compared to some of the cost savings head to head to your competition, which of course you want to stress as well because the CFO cares about the cap-backs. What do you emphasize in that, and we don't have to get into the calculator itself, but in the conceptual model, what's the emphasis? Is it on those sort of business value attributes, is it on the sort of cost-savings? How do you translate performance into that business value? A lot of questions there, but if you could summarize, that'd be great. >> Well, I think you can think of it in three dimensions. The very first one is, does the performance support the use case or the solution that is required? That's the very first one. The second piece that fits in it, and that's in our books, that's operations per second and the latency. The second piece is the cost side, and that has two components to it. The first component is, what are the compute requirements? So, what is the infrastructure underneath that has to support it? And the efficiency that Redis and Redis Enterprise has is dramatically superior to the alternatives. And so, the economics show up. To run a million operations per second, we can do that on two nodes as opposed to alternative, which might need 50 nodes or 300 nodes. >> You can utilize your assets on the floor much better than maybe the competition can. >> This is where the data structures come into play quite a bit. That's one part of-- >> Dave: That's one part of the cost. >> Yeah. The other part of the cost is the human cost. >> Dave: People, yeah. >> And because, and this goes back to the open source, because the people available with the talent and the competency and appreciation for Redis, it's easy to procure those people, and your cost of acquisition and deploying goes down quite a bit. So, there's a human cost to it. The third dimension to this whole equation is time to market. And time to market is measured in many ways. Is it lost revenue if it takes you longer to get there? And Redis consistently from multiple analysts' reports gets top ranking for fastest way to get to market because of how simple it is. Beyond performance, simplicity is a second hallmark. >> That's a benefit acceleration, and you can quantify that. >> Absolutely, absolutely. And that's a revenue parameter, right. >> For years, people have been saying this Cambrian explosion of databases is unsustainable, and sort of in response we've gotten a squaring of the Cambrian explosion. The question is, with your sort of very flexible, I don't want to get too geeky, 'cause Dave'll cut me off, but the idea that you can accommodate time series and all these different ways of, all these different types of data, are we approaching a situation where customers can start consolidating their database choices and have fewer vendors, fewer products in their landscape? >> I think not only are we getting there, but we must get there. You've got over 300 databases in the marketplace, and imagine a CIO or an architect trying to have to sort through that to make a decision, it's difficult, and you certainly cannot support it from a trading standpoint or from an investment, cap-backs, and all that standpoint. What we have done with Redis is introduce something called Redis Modules. We released that at the last RedisConf in May in San Francisco. And the Redis Module is a very simple concept but a very powerful concept. It's an API which can be utilized to take an existing development effort, written as CC++, that can be ported onto the Redis data structures. This gives you the flexibility without having to reinvent the wheel every single time to take that investment, port it on top of Redis, and you get the performance, and you can make now Redis becomes a multi-model database. And I'm going to get to your answer of how do you address the multiple needs so you don't need multiple databases. To give you some examples, since the introduction of Redis Modules, we have now over 50 modules that have been published by a variety of places, not just Redis Labs. To indicate how simple and how powerful this model is. We took Lucene and developed the world's fastest full-text search engine as a module. We have very recently introduced Redis machine learning as a module that works with Spark ML and serves as a great serving layer in the machine learning domain. Just two very simple examples, but work that's being done ported over onto Redis data structures and now you have ability to do some very powerful things because of what Redis is. And this is the way future's going to be. I think every database is trying to offer multi-functionality to be multi-model in nature, but instead of doing it one step at a time, this approach gives us the ability to leverage the entire ecosystem. >> Your point being consolidation's inevitable in this business as well. >> Manish: Architectural consolidation. >> Yes, but also you would think, company consolidation, isn't that going to follow? What do you make of the market, and tell me, if you look back on the database market and what Oracle was able to achieve in the face of, maybe not as many players, but you had Sybase and Informix, and certainly DB2's still around, and SQL Server's still around, but Oracle won, and maybe it was SQL standards that. It's great to be lucky and good. Can we learn from that, or is this a whole different world? Are there similarities, and how do you, how do you see that consolidation potentially shaking out, if you agree that there will be consolidation? >> Yeah, there has to be, first and foremost, an architectural approach that solves the OPEX, CAPEX challenge for the enterprise. But beyond that, no industry can sustain the diversity and the fragmentation that exists in database world. I think there will always be new things coming out, of universities particularly. There's great innovation and research happening, and that is required to augment. But at the end of the day, the commercial enterprises cannot be of the fragmented volume that we have today in the database world, so there is going to be some consolidation, and it's not unnatural. I think it's natural, it's expected, time will tell what that looks like. We've seen some of our competitors acquire smaller companies to add graph functionality, to add search functionality. We just don't think that's the level of consolidation that really moves the needle for the industry. It's got to be at a higher level of consolidation. >> I don't want to, don't take this the wrong way, don't hate me for saying it, but is Oracle sort of the enemy, if I can say that. I mean, it's like, no, okay. >> Depends how you define enemy. >> I'm not going to go do many of the workloads that you're talking about on Oracle, despite what Larry tells me at Oracle OpenWorld. And I'm not going to make Oracle my choice for any of the workloads that you guys are working on. I guess in terms, I mean, everybody who's in the database business looks at that and say, "Hey, we can do it cheaper, better, "more productively," but, could you respond to that, and what do you make of Amazon's moves in the database world? Does that concern you? >> We think of Amazon and Oracle as two very different philosophies, if you can use that word. The approach we have taken is really a forward-looking approach and philosophy. We believe that the needs of the market need to be solved in new ways, and new ways should not be encumbered by old approaches. We're not trying to go and replicate what was done in the SQL world or in a relational database world. Our approach is how do you deliver a multi-model database that has the real-time attribute attached to it in a way that requires very limited computer force power and very few resources to manage? You take all of those things as kind of the core philosophy, which is a forward-looking philosophy. We are definitely not trying to replicate what an Oracle used to be. AWS I think is a very different animal. >> Dave: Interesting, though. >> They have defined the cloud, and I think play a very important role. We are a strong partner of theirs, much of our traffic runs on AWS infrastructure, certainly also on other clouds. I think AWS is one to watch in how they evolve. They have database offerings, including Redis offerings. However, we fully recognize, and the industry recognizes that that's not to the same capability as Redis Enterprise. It's open sourced Redis managed by AWS, and that's fine as a cache, but you cannot persist, and you really cannot have a multi-model capability that's a full database in that approach. >> And you're in the marketplace. >> Manish: We are in the marketplace. >> Obviously. >> And actually, we announced earlier, a few weeks ago, that you can buy and get Redis cloud access, which is Redis Enterprise cloud, on AWS through the integrated billing approach on their marketplace. You can have an AWS account and get our service, the true Redis Enterprise service. >> And as a software company, you'd figure, okay, the cloud infrastructures are service, we don't care what infrastructure it runs on. Whatever the customer wants, but you see AWS making these moves up-market, you got to obviously be paying attention to that. >> Manish: Certainly, certainly. >> Go ahead, last question. >> Interesting that you were saying that to solve this problem of proliferation of choice it has to be multi-model with speed and low resource requirement. If I were to interpret that from an old-style database perspective, it would be you're going to get, the multi-model is something you are addressing now, with the extensibility, but the speed means taking out that abstraction layer that was the query optimizer sort of and working almost at the storage layer, or having an option to do that. Would that be a fair way to say? >> No, I don't think that necessarily needs to be the case. For us, speed translates from the simplicity and the power of the data structures. Instead of having to serialize, deserialize before you process data in a Spark context, or instead of having to look for data that is perhaps not put in sorted sets for a use case that you might be doing, running a query on, if the data is already handled through one of the data structures, you now have a much faster query time, you now have the ability to reach the data in the right approach. And again, this is no-SQL, right, so it's a schema lesson write and it sets your scheme as you want it be on read. We marry that with the data structures, and that gives you the ultimate speed. >> We have to leave it there, but Manish, I'll give you the last word. Things we should be paying attention to for Redis Labs this year, events, announcements? >> I think the big thing I would leave the audience with is RedisConf 2017. It's May 31 to June 2 in San Francisco. We are expecting over 1,000 people. The brightest minds around Redis of the database world will be there, and anybody who is considering deploying the next generation database should attend. >> Dave: Where are you doing that? >> It's the Marriott Marquis in San Franciso. >> Great, is that on Howard Street, across from the--? >> It is right across from Moscone. >> Great, awesome location. People know it, easy to get to. Well, congratulations on the success. We'll be lookin' for outputs from that event, and hope to see you again on theCUBE. >> Thank you, enjoyed the conversation. >> Alright, good. Keep it right there, everybody, we'll be back with our next guest. This is theCUBE, we're live from Spark Summit East. Be right back. (upbeat electronic rock music)

Published Date : Feb 9 2017

SUMMARY :

Brought to you by Databricks. Manish Gupta is here, he's the CMO at Redis Labs. So, you know, 10 years ago you say We are happy to be on the top of that heap. Redis Labs is the company behind But add some color to that if you would. and the reason you can get that performance Let's get the business model piece out of the way. We also allow for the enterprise to select a VPC environment That's right. Google, and IBM. Go to the whip IBM. Along the lines of the business model, Certainly, you know, database is an integral part and the use case, the needs of consistency vary. in terms of the requirement for those acid properties? you must support acid, and we do. the growth workloads don't necessarily require that, Dave: Good CMO question. but if you have to wait for a couple other servers and the consistency management of the back end and that's the core, and that makes and the word of mouth, that ground level love but certainly the CIO, CDO are like, For the C-level, the message is very simple. part of the value proposition is you are enabling That's the very first one. much better than maybe the competition can. This is where the data structures of the cost. The other part of the cost is the human cost. and the competency and appreciation for Redis, And that's a revenue parameter, right. but the idea that you can accommodate time series We released that at the last RedisConf in this business as well. and tell me, if you look back on the database market that really moves the needle for the industry. but is Oracle sort of the enemy, if I can say that. for any of the workloads that you guys are working on. We believe that the needs of the market and that's fine as a cache, but you cannot persist, the true Redis Enterprise service. okay, the cloud infrastructures are service, the multi-model is something you are addressing now, and the power of the data structures. but Manish, I'll give you the last word. of the database world will be there, and hope to see you again on theCUBE. This is theCUBE, we're live from Spark Summit East.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmazonORGANIZATION

0.99+

Dave VellantePERSON

0.99+

George GilbertPERSON

0.99+

DavePERSON

0.99+

AWSORGANIZATION

0.99+

GeorgePERSON

0.99+

IBMORGANIZATION

0.99+

OracleORGANIZATION

0.99+

Howard StreetLOCATION

0.99+

CurtPERSON

0.99+

second pieceQUANTITY

0.99+

San FranciscoLOCATION

0.99+

Redis LabsORGANIZATION

0.99+

Manish GuptaPERSON

0.99+

two nodesQUANTITY

0.99+

RedisORGANIZATION

0.99+

two componentsQUANTITY

0.99+

twoQUANTITY

0.99+

San FrancisoLOCATION

0.99+

LarryPERSON

0.99+

ManishPERSON

0.99+

first componentQUANTITY

0.99+

Boston, MassachusettsLOCATION

0.99+

over 50 modulesQUANTITY

0.99+

June 2DATE

0.99+

May 31DATE

0.99+

GoogleORGANIZATION

0.99+

BostonLOCATION

0.99+

Curt MonashPERSON

0.99+

MayDATE

0.99+

millionsQUANTITY

0.99+

third dimensionQUANTITY

0.98+

50 nodesQUANTITY

0.98+

MosconeLOCATION

0.98+

fourthQUANTITY

0.98+

Redis EnterpriseTITLE

0.98+

300 nodesQUANTITY

0.98+

RedisTITLE

0.98+

KazakhstanLOCATION

0.98+

over 1,000 peopleQUANTITY

0.98+

one partQUANTITY

0.98+

bothQUANTITY

0.98+

one stepQUANTITY

0.97+

C-SuiteTITLE

0.97+

Marriott MarquisORGANIZATION

0.97+

second hallmarkQUANTITY

0.97+

10 years agoDATE

0.97+

Spark Summit East 2017EVENT

0.97+

GrouponORGANIZATION

0.97+

first oneQUANTITY

0.97+

CDOTITLE

0.97+

over 300 databasesQUANTITY

0.96+

SQL ServerTITLE

0.96+

Redis Enterprise cloudTITLE

0.96+

Domenic Ravita, SingleStore | AWS re:Invent 2022


 

>>Hey guys and girls, welcome back to The Cube's Live coverage of AWS Reinvent 22 from Sin City. We've been here, this is our third day of coverage. We started Monday night first. Full day of the show was yesterday. Big news yesterday. Big news. Today we're hearing north of 50,000 people, and I'm hearing hundreds of thousands online. We've been having great conversations with AWS folks in the ecosystem, AWS customers, partners, ISVs, you name it. We're pleased to welcome back one of our alumni to the program, talking about partner ecosystem. Dominic Rav Vida joins us, the VP of Developer relations at single store. It's so great to have you on the program. Dominic. Thanks for coming. >>Thanks. Great. Great to see you >>Again. Great to see you too. We go way back. >>We do, yeah. >>So let's talk about reinvent 22. This is the 11th reinvent. Yeah. What are some of the things that you've heard this week that are exciting that are newsworthy from single stores perspective? >>I think in particular what we heard AWS announce on the zero ETL between Aurora and Redshift, I think it's, it's significant in that AWS has provided lots of services for building blocks for applications for a long time. And that's a great amount of flexibility for developers. But there are cases where, you know, it's a common thing to need to move data from transactional systems to analytics systems and making that easy with zero etl, I think it's a significant thing and in general we see in the market and especially in the data management market in the cloud, a unification of different types of workloads. So I think that's a step in the right direction for aws. And I think for the market as a whole, why it's significant for single store is, that's our specialty in particular, is to unify transactions and analytics for realtime applications and analytics. When you've got customer facing analytic applications and you need low latency data from realtime streaming data sources and you've gotta crunch and compute that. Those are diverse types of workloads over document transactional workloads as well as, you know, analytical workloads of various shapes and the data types could be diverse from geospatial time series. And then you've gotta serve that because we're all living in this digital service first world and you need that relevant, consistent, fresh data. And so that unification is what we think is like the big thing in data right >>Now. So validation for single store, >>It does feel like that. I mean, I'd say in the recent like six months, you've seen announcements from Google with Alloy db basically adding the complement to their workload types. You see it with Snowflake adding the complement to their traditional analytical workload site. You see it with Mongo and others. And yeah, we do feel it was validation cuz at single store we completed the functionality for what we call universal storage, which is, is the industry's first third type of storage after row store and column store, single store dbs, universal storage, unifies those. So on a single copy of data you can form these diverse workloads. And that was completed three years ago. So we sort of see like, you know, we're onto something >>Here. Welcome to the game guys. >>That's right. >>What's the value in that universal storage for customers, whether it's a healthcare organization, a financial institution, what's the value in it in those business outcomes that you guys are really helping to fuel? >>I think in short, if there were like a, a bumper sticker for that message, it's like, are you ready for the next interaction? The next interaction with your customer, the next interaction with your supply chain partner, the next interaction with your internal stakeholders, your operational managers being ready for that interaction means you've gotta have the historical data at the ready, accessible, efficiently accessible, and and, and queryable along with the most recent fresh data. And that's the context that's expected and be able to serve that instantaneously. So being ready for that next interaction is what single store helps companies do. >>Talk about single store helping customers. You know, every company these days has to be a data company. I always think, whether it's my grocery store that has all my information and helps keep me fed or a gas station or a car dealer or my bank. And we've also here, one of the things that John Furrier got to do, and he does this every year before aws, he gets to sit down with the CEO and gets really kind of a preview of what's gonna happen at at the show, right? And Adams Lisky said to him some interesting very poignant things. One is that that data, we talk about data democratization, but he says the role of the data analyst is gonna go away. Or that maybe that term in, in that every person within an organization, whether you're marketing, sales, ops, finance, is going to be analyzing data for their jobs to become data driven. Right? How does single store help customers really become data companies, especially powering data intensive apps like I know you do. >>Yeah, that's, there's a lot of talk about that and, and I think there's a lot of work that's been done with companies to make that easier to analyze data in all these different job functions. While we do that, it's not really our starting point because, and our starting point is like operationalizing that analytics as part of the business. So you can think of it in terms of database terms. Like is it batch analysis? Batch analytics after the fact, what happened last week? What happened last month? That's a lot of what those data teams are doing and those analysts are doing. What single store focuses more is in putting those insights into action for the business operations, which typically is more on the application side, it's the API side, you might call it a data product. If you're monetizing your data and you're transacting with that providing as an api, or you're delivering it as software as a service, and you're providing an end-to-end function for, you know, our marketing marketer, then we help power those kinds of real time data applications that have the interactivity and have that customer touchpoint or that partner touchpoint. So you can say we sort of, we put the data in action in that way. >>And that's the most, one of the most important things is putting data in action. If it's, it can be gold, it can be whatever you wanna call it, but if you can't actually put it into action, act on insights in real time, right? The value goes way down or there's liability, >>Right? And I think you have to do that with privacy in mind as well, right? And so you have to take control of that data and use it for your business strategy And the way that you can do that, there's technology like single store makes that possible in ways that weren't possible before. And I'll give you an example. So we have a, a customer named Fathom Analytics. They provide web analytics for marketers, right? So if you're in marketing, you understand this use case. Any demand gen marketer knows that they want to see what the traffic that hits their site is. What are the page views, what are the click streams, what are the sequences? Have these visitors to my website hit certain goals? So the big name in that for years of course has been Google Analytics and that's a free service. And you interact with that and you can see how your website's performing. >>So what Fathom does is a privacy first alternative to Google Analytics. And when you think about, well, how is that possible that they, and as a paid service, it's as software, as a service, how, first of all, how can you keep up with that real time deluge of clickstream data at the rate that Google Analytics can do it? That's the technical problem. But also at the data layer, how could you keep up with Google has, you know, in terms of databases And Fathom's answer to that is to use single store. Their, their prior architecture had four different types of database technologies under the hood. They were using Redis to have fast read time cash. They were using MySEQ database as the application database they were using. They were looking at last search to do full tech search. And they were using DynamoDB as part of a another kind of fast look up fast cash. They replaced all four of those with single store. And, and again, what they're doing is like sort of battling defacto giant in Google Analytics and having a great success at doing that for posting tens of thousands of websites. Some big names that you've heard of as well. >>I can imagine that's a big reduction from four to one, four x reduction in databases. The complexities that go away, the simplification that happens, I can imagine is quite huge for them. >>And we've done a study, an independent study with Giga Home Research. We published this back in June looking at total cost of ownership with benchmarks and the relevant benchmarks for transactions and analytics and databases are tpcc for transactions, TPC H for analytics, TPC DS for analytics. And we did a TCO study using those benchmark datas on a combination of transactional and analytical databases together and saw some pretty big improvements. 60% improvement over Myse Snowflake, for >>Instance. Awesome. Big business outcomes. We only have a few seconds left, so you've already given me a bumper sticker. Yeah. And I know I live in Silicon Valley, I've seen those billboards. I know single store has done some cheeky billboard marketing campaigns. But if you had a new billboard to create from your perspective about single store, what does it say? >>I, I think it's that, are you, are you ready for the next interaction? Because business is won and lost in every moment, in every location, in every digital moment passing by. And if you're not ready to, to interact and transact rather your systems on your behalf, then you're behind the curve. It's easy to be displaced people swipe left and pick your competitor. So I think that's the next bumper sticker. I may, I would say our, my favorite billboard so far of what we've run is cover your SaaS, which is what is how, what is the data layer to, to manage the next level of SaaS applications, the next generation. And we think single store is a big part >>Of that. Cover your SaaS. Love it. Dominic, thank you so much for joining me, giving us an update on single store from your perspective, what's going on there, kind of really where you are in the market. We appreciate that. We'll have to >>Have you back. Thank you. Glad to >>Be here. All right. For Dominic rta, I'm Lisa Martin. You're watching The Cube, the leader in live, emerging and enterprise tech coverage.

Published Date : Dec 1 2022

SUMMARY :

It's so great to have you on the program. Great to see you Great to see you too. What are some of the things that you've heard this week that are exciting that are newsworthy from And so that unification is what we think is like the So on a single copy of data you can form these diverse And that's the context that's expected and be able to serve that instantaneously. one of the things that John Furrier got to do, and he does this every year before aws, he gets to sit down with the CEO So you can think of it in terms of database terms. And that's the most, one of the most important things is putting data in action. And I think you have to do that with privacy in mind as well, right? But also at the data layer, how could you keep up with Google has, you know, The complexities that go away, the simplification that happens, I can imagine is quite huge for them. And we've done a study, an independent study with Giga Home Research. But if you had a new billboard to create from your perspective And if you're not ready to, to interact and transact rather your systems on Dominic, thank you so much for joining me, giving us an update on single store from your Have you back. the leader in live, emerging and enterprise tech coverage.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DominicPERSON

0.99+

Lisa MartinPERSON

0.99+

AWSORGANIZATION

0.99+

60%QUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

Sin CityLOCATION

0.99+

JuneDATE

0.99+

TodayDATE

0.99+

last weekDATE

0.99+

yesterdayDATE

0.99+

Monday nightDATE

0.99+

last monthDATE

0.99+

Giga Home ResearchORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

John FurrierPERSON

0.99+

oneQUANTITY

0.99+

third dayQUANTITY

0.99+

single storeQUANTITY

0.99+

Domenic RavitaPERSON

0.99+

Dominic Rav VidaPERSON

0.99+

Google AnalyticsTITLE

0.99+

Adams LiskyPERSON

0.99+

FathomORGANIZATION

0.98+

three years agoDATE

0.98+

singleQUANTITY

0.98+

Fathom AnalyticsORGANIZATION

0.98+

single copyQUANTITY

0.98+

hundreds of thousandsQUANTITY

0.98+

MongoORGANIZATION

0.97+

OneQUANTITY

0.97+

SingleStoreORGANIZATION

0.97+

this weekDATE

0.97+

The CubeTITLE

0.96+

first third typeQUANTITY

0.94+

fourQUANTITY

0.94+

tens of thousands of websitesQUANTITY

0.92+

MySEQTITLE

0.91+

six monthsQUANTITY

0.91+

RedisTITLE

0.9+

DynamoDBTITLE

0.9+

Dominic rtaPERSON

0.89+

Reinvent 22TITLE

0.89+

north of 50,000 peopleQUANTITY

0.86+

first alternativeQUANTITY

0.81+

firstQUANTITY

0.81+

single storesQUANTITY

0.78+

first worldQUANTITY

0.73+

AlloyTITLE

0.69+

few secondsQUANTITY

0.67+

RedshiftORGANIZATION

0.67+

SnowflakeORGANIZATION

0.67+

awsORGANIZATION

0.67+

TCOORGANIZATION

0.66+

InventEVENT

0.66+

reinvent 22EVENT

0.65+

SnowflakeTITLE

0.65+

2022DATE

0.64+

AuroraORGANIZATION

0.56+

11th reinventEVENT

0.52+

FathomPERSON

0.51+

MyseORGANIZATION

0.49+

CubeORGANIZATION

0.46+

zeroQUANTITY

0.34+

Breaking Analysis: We Have the Data…What Private Tech Companies Don’t Tell you About Their Business


 

>> From The Cube Studios in Palo Alto and Boston, bringing you data driven insights from The Cube at ETR. This is "Breaking Analysis" with Dave Vellante. >> The reverse momentum in tech stocks caused by rising interest rates, less attractive discounted cash flow models, and more tepid forward guidance, can be easily measured by public market valuations. And while there's lots of discussion about the impact on private companies and cash runway and 409A valuations, measuring the performance of non-public companies isn't as easy. IPOs have dried up and public statements by private companies, of course, they accentuate the good and they kind of hide the bad. Real data, unless you're an insider, is hard to find. Hello and welcome to this week's "Wikibon Cube Insights" powered by ETR. In this "Breaking Analysis", we unlock some of the secrets that non-public, emerging tech companies may or may not be sharing. And we do this by introducing you to a capability from ETR that we've not exposed you to over the past couple of years, it's called the Emerging Technologies Survey, and it is packed with sentiment data and performance data based on surveys of more than a thousand CIOs and IT buyers covering more than 400 companies. And we've invited back our colleague, Erik Bradley of ETR to help explain the survey and the data that we're going to cover today. Erik, this survey is something that I've not personally spent much time on, but I'm blown away at the data. It's really unique and detailed. First of all, welcome. Good to see you again. >> Great to see you too, Dave, and I'm really happy to be talking about the ETS or the Emerging Technology Survey. Even our own clients of constituents probably don't spend as much time in here as they should. >> Yeah, because there's so much in the mainstream, but let's pull up a slide to bring out the survey composition. Tell us about the study. How often do you run it? What's the background and the methodology? >> Yeah, you were just spot on the way you were talking about the private tech companies out there. So what we did is we decided to take all the vendors that we track that are not yet public and move 'em over to the ETS. And there isn't a lot of information out there. If you're not in Silicon (indistinct), you're not going to get this stuff. So PitchBook and Tech Crunch are two out there that gives some data on these guys. But what we really wanted to do was go out to our community. We have 6,000, ITDMs in our community. We wanted to ask them, "Are you aware of these companies? And if so, are you allocating any resources to them? Are you planning to evaluate them," and really just kind of figure out what we can do. So this particular survey, as you can see, 1000 plus responses, over 450 vendors that we track. And essentially what we're trying to do here is talk about your evaluation and awareness of these companies and also your utilization. And also if you're not utilizing 'em, then we can also figure out your sales conversion or churn. So this is interesting, not only for the ITDMs themselves to figure out what their peers are evaluating and what they should put in POCs against the big guys when contracts come up. But it's also really interesting for the tech vendors themselves to see how they're performing. >> And you can see 2/3 of the respondents are director level of above. You got 28% is C-suite. There is of course a North America bias, 70, 75% is North America. But these smaller companies, you know, that's when they start doing business. So, okay. We're going to do a couple of things here today. First, we're going to give you the big picture across the sectors that ETR covers within the ETS survey. And then we're going to look at the high and low sentiment for the larger private companies. And then we're going to do the same for the smaller private companies, the ones that don't have as much mindshare. And then I'm going to put those two groups together and we're going to look at two dimensions, actually three dimensions, which companies are being evaluated the most. Second, companies are getting the most usage and adoption of their offerings. And then third, which companies are seeing the highest churn rates, which of course is a silent killer of companies. And then finally, we're going to look at the sentiment and mindshare for two key areas that we like to cover often here on "Breaking Analysis", security and data. And data comprises database, including data warehousing, and then big data analytics is the second part of data. And then machine learning and AI is the third section within data that we're going to look at. Now, one other thing before we get into it, ETR very often will include open source offerings in the mix, even though they're not companies like TensorFlow or Kubernetes, for example. And we'll call that out during this discussion. The reason this is done is for context, because everyone is using open source. It is the heart of innovation and many business models are super glued to an open source offering, like take MariaDB, for example. There's the foundation and then there's with the open source code and then there, of course, the company that sells services around the offering. Okay, so let's first look at the highest and lowest sentiment among these private firms, the ones that have the highest mindshare. So they're naturally going to be somewhat larger. And we do this on two dimensions, sentiment on the vertical axis and mindshare on the horizontal axis and note the open source tool, see Kubernetes, Postgres, Kafka, TensorFlow, Jenkins, Grafana, et cetera. So Erik, please explain what we're looking at here, how it's derived and what the data tells us. >> Certainly, so there is a lot here, so we're going to break it down first of all by explaining just what mindshare and net sentiment is. You explain the axis. We have so many evaluation metrics, but we need to aggregate them into one so that way we can rank against each other. Net sentiment is really the aggregation of all the positive and subtracting out the negative. So the net sentiment is a very quick way of looking at where these companies stand versus their peers in their sectors and sub sectors. Mindshare is basically the awareness of them, which is good for very early stage companies. And you'll see some names on here that are obviously been around for a very long time. And they're clearly be the bigger on the axis on the outside. Kubernetes, for instance, as you mentioned, is open source. This de facto standard for all container orchestration, and it should be that far up into the right, because that's what everyone's using. In fact, the open source leaders are so prevalent in the emerging technology survey that we break them out later in our analysis, 'cause it's really not fair to include them and compare them to the actual companies that are providing the support and the security around that open source technology. But no survey, no analysis, no research would be complete without including these open source tech. So what we're looking at here, if I can just get away from the open source names, we see other things like Databricks and OneTrust . They're repeating as top net sentiment performers here. And then also the design vendors. People don't spend a lot of time on 'em, but Miro and Figma. This is their third survey in a row where they're just dominating that sentiment overall. And Adobe should probably take note of that because they're really coming after them. But Databricks, we all know probably would've been a public company by now if the market hadn't turned, but you can see just how dominant they are in a survey of nothing but private companies. And we'll see that again when we talk about the database later. >> And I'll just add, so you see automation anywhere on there, the big UiPath competitor company that was not able to get to the public markets. They've been trying. Snyk, Peter McKay's company, they've raised a bunch of money, big security player. They're doing some really interesting things in developer security, helping developers secure the data flow, H2O.ai, Dataiku AI company. We saw them at the Snowflake Summit. Redis Labs, Netskope and security. So a lot of names that we know that ultimately we think are probably going to be hitting the public market. Okay, here's the same view for private companies with less mindshare, Erik. Take us through this one. >> On the previous slide too real quickly, I wanted to pull that security scorecard and we'll get back into it. But this is a newcomer, that I couldn't believe how strong their data was, but we'll bring that up in a second. Now, when we go to the ones of lower mindshare, it's interesting to talk about open source, right? Kubernetes was all the way on the top right. Everyone uses containers. Here we see Istio up there. Not everyone is using service mesh as much. And that's why Istio is in the smaller breakout. But still when you talk about net sentiment, it's about the leader, it's the highest one there is. So really interesting to point out. Then we see other names like Collibra in the data side really performing well. And again, as always security, very well represented here. We have Aqua, Wiz, Armis, which is a standout in this survey this time around. They do IoT security. I hadn't even heard of them until I started digging into the data here. And I couldn't believe how well they were doing. And then of course you have AnyScale, which is doing a second best in this and the best name in the survey Hugging Face, which is a machine learning AI tool. Also doing really well on a net sentiment, but they're not as far along on that access of mindshare just yet. So these are again, emerging companies that might not be as well represented in the enterprise as they will be in a couple of years. >> Hugging Face sounds like something you do with your two year old. Like you said, you see high performers, AnyScale do machine learning and you mentioned them. They came out of Berkeley. Collibra Governance, InfluxData is on there. InfluxDB's a time series database. And yeah, of course, Alex, if you bring that back up, you get a big group of red dots, right? That's the bad zone, I guess, which Sisense does vis, Yellowbrick Data is a NPP database. How should we interpret the red dots, Erik? I mean, is it necessarily a bad thing? Could it be misinterpreted? What's your take on that? >> Sure, well, let me just explain the definition of it first from a data science perspective, right? We're a data company first. So the gray dots that you're seeing that aren't named, that's the mean that's the average. So in order for you to be on this chart, you have to be at least one standard deviation above or below that average. So that gray is where we're saying, "Hey, this is where the lump of average comes in. This is where everyone normally stands." So you either have to be an outperformer or an underperformer to even show up in this analysis. So by definition, yes, the red dots are bad. You're at least one standard deviation below the average of your peers. It's not where you want to be. And if you're on the lower left, not only are you not performing well from a utilization or an actual usage rate, but people don't even know who you are. So that's a problem, obviously. And the VCs and the PEs out there that are backing these companies, they're the ones who mostly are interested in this data. >> Yeah. Oh, that's great explanation. Thank you for that. No, nice benchmarking there and yeah, you don't want to be in the red. All right, let's get into the next segment here. Here going to look at evaluation rates, adoption and the all important churn. First new evaluations. Let's bring up that slide. And Erik, take us through this. >> So essentially I just want to explain what evaluation means is that people will cite that they either plan to evaluate the company or they're currently evaluating. So that means we're aware of 'em and we are choosing to do a POC of them. And then we'll see later how that turns into utilization, which is what a company wants to see, awareness, evaluation, and then actually utilizing them. That's sort of the life cycle for these emerging companies. So what we're seeing here, again, with very high evaluation rates. H2O, we mentioned. SecurityScorecard jumped up again. Chargebee, Snyk, Salt Security, Armis. A lot of security names are up here, Aqua, Netskope, which God has been around forever. I still can't believe it's in an Emerging Technology Survey But so many of these names fall in data and security again, which is why we decided to pick those out Dave. And on the lower side, Vena, Acton, those unfortunately took the dubious award of the lowest evaluations in our survey, but I prefer to focus on the positive. So SecurityScorecard, again, real standout in this one, they're in a security assessment space, basically. They'll come in and assess for you how your security hygiene is. And it's an area of a real interest right now amongst our ITDM community. >> Yeah, I mean, I think those, and then Arctic Wolf is up there too. They're doing managed services. You had mentioned Netskope. Yeah, okay. All right, let's look at now adoption. These are the companies whose offerings are being used the most and are above that standard deviation in the green. Take us through this, Erik. >> Sure, yet again, what we're looking at is, okay, we went from awareness, we went to evaluation. Now it's about utilization, which means a survey respondent's going to state "Yes, we evaluated and we plan to utilize it" or "It's already in our enterprise and we're actually allocating further resources to it." Not surprising, again, a lot of open source, the reason why, it's free. So it's really easy to grow your utilization on something that's free. But as you and I both know, as Red Hat proved, there's a lot of money to be made once the open source is adopted, right? You need the governance, you need the security, you need the support wrapped around it. So here we're seeing Kubernetes, Postgres, Apache Kafka, Jenkins, Grafana. These are all open source based names. But if we're looking at names that are non open source, we're going to see Databricks, Automation Anywhere, Rubrik all have the highest mindshare. So these are the names, not surprisingly, all names that probably should have been public by now. Everyone's expecting an IPO imminently. These are the names that have the highest mindshare. If we talk about the highest utilization rates, again, Miro and Figma pop up, and I know they're not household names, but they are just dominant in this survey. These are applications that are meant for design software and, again, they're going after an Autodesk or a CAD or Adobe type of thing. It is just dominant how high the utilization rates are here, which again is something Adobe should be paying attention to. And then you'll see a little bit lower, but also interesting, we see Collibra again, we see Hugging Face again. And these are names that are obviously in the data governance, ML, AI side. So we're seeing a ton of data, a ton of security and Rubrik was interesting in this one, too, high utilization and high mindshare. We know how pervasive they are in the enterprise already. >> Erik, Alex, keep that up for a second, if you would. So yeah, you mentioned Rubrik. Cohesity's not on there. They're sort of the big one. We're going to talk about them in a moment. Puppet is interesting to me because you remember the early days of that sort of space, you had Puppet and Chef and then you had Ansible. Red Hat bought Ansible and then Ansible really took off. So it's interesting to see Puppet on there as well. Okay. So now let's look at the churn because this one is where you don't want to be. It's, of course, all red 'cause churn is bad. Take us through this, Erik. >> Yeah, definitely don't want to be here and I don't love to dwell on the negative. So we won't spend as much time. But to your point, there's one thing I want to point out that think it's important. So you see Rubrik in the same spot, but Rubrik has so many citations in our survey that it actually would make sense that they're both being high utilization and churn just because they're so well represented. They have such a high overall representation in our survey. And the reason I call that out is Cohesity. Cohesity has an extremely high churn rate here about 17% and unlike Rubrik, they were not on the utilization side. So Rubrik is seeing both, Cohesity is not. It's not being utilized, but it's seeing a high churn. So that's the way you can look at this data and say, "Hm." Same thing with Puppet. You noticed that it was on the other slide. It's also on this one. So basically what it means is a lot of people are giving Puppet a shot, but it's starting to churn, which means it's not as sticky as we would like. One that was surprising on here for me was Tanium. It's kind of jumbled in there. It's hard to see in the middle, but Tanium, I was very surprised to see as high of a churn because what I do hear from our end user community is that people that use it, like it. It really kind of spreads into not only vulnerability management, but also that endpoint detection and response side. So I was surprised by that one, mostly to see Tanium in here. Mural, again, was another one of those application design softwares that's seeing a very high churn as well. >> So you're saying if you're in both... Alex, bring that back up if you would. So if you're in both like MariaDB is for example, I think, yeah, they're in both. They're both green in the previous one and red here, that's not as bad. You mentioned Rubrik is going to be in both. Cohesity is a bit of a concern. Cohesity just brought on Sanjay Poonen. So this could be a go to market issue, right? I mean, 'cause Cohesity has got a great product and they got really happy customers. So they're just maybe having to figure out, okay, what's the right ideal customer profile and Sanjay Poonen, I guarantee, is going to have that company cranking. I mean they had been doing very well on the surveys and had fallen off of a bit. The other interesting things wondering the previous survey I saw Cvent, which is an event platform. My only reason I pay attention to that is 'cause we actually have an event platform. We don't sell it separately. We bundle it as part of our offerings. And you see Hopin on here. Hopin raised a billion dollars during the pandemic. And we were like, "Wow, that's going to blow up." And so you see Hopin on the churn and you didn't see 'em in the previous chart, but that's sort of interesting. Like you said, let's not kind of dwell on the negative, but you really don't. You know, churn is a real big concern. Okay, now we're going to drill down into two sectors, security and data. Where data comprises three areas, database and data warehousing, machine learning and AI and big data analytics. So first let's take a look at the security sector. Now this is interesting because not only is it a sector drill down, but also gives an indicator of how much money the firm has raised, which is the size of that bubble. And to tell us if a company is punching above its weight and efficiently using its venture capital. Erik, take us through this slide. Explain the dots, the size of the dots. Set this up please. >> Yeah. So again, the axis is still the same, net sentiment and mindshare, but what we've done this time is we've taken publicly available information on how much capital company is raised and that'll be the size of the circle you see around the name. And then whether it's green or red is basically saying relative to the amount of money they've raised, how are they doing in our data? So when you see a Netskope, which has been around forever, raised a lot of money, that's why you're going to see them more leading towards red, 'cause it's just been around forever and kind of would expect it. Versus a name like SecurityScorecard, which is only raised a little bit of money and it's actually performing just as well, if not better than a name, like a Netskope. OneTrust doing absolutely incredible right now. BeyondTrust. We've seen the issues with Okta, right. So those are two names that play in that space that obviously are probably getting some looks about what's going on right now. Wiz, we've all heard about right? So raised a ton of money. It's doing well on net sentiment, but the mindshare isn't as well as you'd want, which is why you're going to see a little bit of that red versus a name like Aqua, which is doing container and application security. And hasn't raised as much money, but is really neck and neck with a name like Wiz. So that is why on a relative basis, you'll see that more green. As we all know, information security is never going away. But as we'll get to later in the program, Dave, I'm not sure in this current market environment, if people are as willing to do POCs and switch away from their security provider, right. There's a little bit of tepidness out there, a little trepidation. So right now we're seeing overall a slight pause, a slight cooling in overall evaluations on the security side versus historical levels a year ago. >> Now let's stay on here for a second. So a couple things I want to point out. So it's interesting. Now Snyk has raised over, I think $800 million but you can see them, they're high on the vertical and the horizontal, but now compare that to Lacework. It's hard to see, but they're kind of buried in the middle there. That's the biggest dot in this whole thing. I think I'm interpreting this correctly. They've raised over a billion dollars. It's a Mike Speiser company. He was the founding investor in Snowflake. So people watch that very closely, but that's an example of where they're not punching above their weight. They recently had a layoff and they got to fine tune things, but I'm still confident they they're going to do well. 'Cause they're approaching security as a data problem, which is probably people having trouble getting their arms around that. And then again, I see Arctic Wolf. They're not red, they're not green, but they've raised fair amount of money, but it's showing up to the right and decent level there. And a couple of the other ones that you mentioned, Netskope. Yeah, they've raised a lot of money, but they're actually performing where you want. What you don't want is where Lacework is, right. They've got some work to do to really take advantage of the money that they raised last November and prior to that. >> Yeah, if you're seeing that more neutral color, like you're calling out with an Arctic Wolf, like that means relative to their peers, this is where they should be. It's when you're seeing that red on a Lacework where we all know, wow, you raised a ton of money and your mindshare isn't where it should be. Your net sentiment is not where it should be comparatively. And then you see these great standouts, like Salt Security and SecurityScorecard and Abnormal. You know they haven't raised that much money yet, but their net sentiment's higher and their mindshare's doing well. So those basically in a nutshell, if you're a PE or a VC and you see a small green circle, then you're doing well, then it means you made a good investment. >> Some of these guys, I don't know, but you see these small green circles. Those are the ones you want to start digging into and maybe help them catch a wave. Okay, let's get into the data discussion. And again, three areas, database slash data warehousing, big data analytics and ML AI. First, we're going to look at the database sector. So Alex, thank you for bringing that up. Alright, take us through this, Erik. Actually, let me just say Postgres SQL. I got to ask you about this. It shows some funding, but that actually could be a mix of EDB, the company that commercializes Postgres and Postgres the open source database, which is a transaction system and kind of an open source Oracle. You see MariaDB is a database, but open source database. But the companies they've raised over $200 million and they filed an S-4. So Erik looks like this might be a little bit of mashup of companies and open source products. Help us understand this. >> Yeah, it's tough when you start dealing with the open source side and I'll be honest with you, there is a little bit of a mashup here. There are certain names here that are a hundred percent for profit companies. And then there are others that are obviously open source based like Redis is open source, but Redis Labs is the one trying to monetize the support around it. So you're a hundred percent accurate on this slide. I think one of the things here that's important to note though, is just how important open source is to data. If you're going to be going to any of these areas, it's going to be open source based to begin with. And Neo4j is one I want to call out here. It's not one everyone's familiar with, but it's basically geographical charting database, which is a name that we're seeing on a net sentiment side actually really, really high. When you think about it's the third overall net sentiment for a niche database play. It's not as big on the mindshare 'cause it's use cases aren't as often, but third biggest play on net sentiment. I found really interesting on this slide. >> And again, so MariaDB, as I said, they filed an S-4 I think $50 million in revenue, that might even be ARR. So they're not huge, but they're getting there. And by the way, MariaDB, if you don't know, was the company that was formed the day that Oracle bought Sun in which they got MySQL and MariaDB has done a really good job of replacing a lot of MySQL instances. Oracle has responded with MySQL HeatWave, which was kind of the Oracle version of MySQL. So there's some interesting battles going on there. If you think about the LAMP stack, the M in the LAMP stack was MySQL. And so now it's all MariaDB replacing that MySQL for a large part. And then you see again, the red, you know, you got to have some concerns about there. Aerospike's been around for a long time. SingleStore changed their name a couple years ago, last year. Yellowbrick Data, Fire Bolt was kind of going after Snowflake for a while, but yeah, you want to get out of that red zone. So they got some work to do. >> And Dave, real quick for the people that aren't aware, I just want to let them know that we can cut this data with the public company data as well. So we can cross over this with that because some of these names are competing with the larger public company names as well. So we can go ahead and cross reference like a MariaDB with a Mongo, for instance, or of something of that nature. So it's not in this slide, but at another point we can certainly explain on a relative basis how these private names are doing compared to the other ones as well. >> All right, let's take a quick look at analytics. Alex, bring that up if you would. Go ahead, Erik. >> Yeah, I mean, essentially here, I can't see it on my screen, my apologies. I just kind of went to blank on that. So gimme one second to catch up. >> So I could set it up while you're doing that. You got Grafana up and to the right. I mean, this is huge right. >> Got it thank you. I lost my screen there for a second. Yep. Again, open source name Grafana, absolutely up and to the right. But as we know, Grafana Labs is actually picking up a lot of speed based on Grafana, of course. And I think we might actually hear some noise from them coming this year. The names that are actually a little bit more disappointing than I want to call out are names like ThoughtSpot. It's been around forever. Their mindshare of course is second best here but based on the amount of time they've been around and the amount of money they've raised, it's not actually outperforming the way it should be. We're seeing Moogsoft obviously make some waves. That's very high net sentiment for that company. It's, you know, what, third, fourth position overall in this entire area, Another name like Fivetran, Matillion is doing well. Fivetran, even though it's got a high net sentiment, again, it's raised so much money that we would've expected a little bit more at this point. I know you know this space extremely well, but basically what we're looking at here and to the bottom left, you're going to see some names with a lot of red, large circles that really just aren't performing that well. InfluxData, however, second highest net sentiment. And it's really pretty early on in this stage and the feedback we're getting on this name is the use cases are great, the efficacy's great. And I think it's one to watch out for. >> InfluxData, time series database. The other interesting things I just noticed here, you got Tamer on here, which is that little small green. Those are the ones we were saying before, look for those guys. They might be some of the interesting companies out there and then observe Jeremy Burton's company. They do observability on top of Snowflake, not green, but kind of in that gray. So that's kind of cool. Monte Carlo is another one, they're sort of slightly green. They are doing some really interesting things in data and data mesh. So yeah, okay. So I can spend all day on this stuff, Erik, phenomenal data. I got to get back and really dig in. Let's end with machine learning and AI. Now this chart it's similar in its dimensions, of course, except for the money raised. We're not showing that size of the bubble, but AI is so hot. We wanted to cover that here, Erik, explain this please. Why TensorFlow is highlighted and walk us through this chart. >> Yeah, it's funny yet again, right? Another open source name, TensorFlow being up there. And I just want to explain, we do break out machine learning, AI is its own sector. A lot of this of course really is intertwined with the data side, but it is on its own area. And one of the things I think that's most important here to break out is Databricks. We started to cover Databricks in machine learning, AI. That company has grown into much, much more than that. So I do want to state to you Dave, and also the audience out there that moving forward, we're going to be moving Databricks out of only the MA/AI into other sectors. So we can kind of value them against their peers a little bit better. But in this instance, you could just see how dominant they are in this area. And one thing that's not here, but I do want to point out is that we have the ability to break this down by industry vertical, organization size. And when I break this down into Fortune 500 and Fortune 1000, both Databricks and Tensorflow are even better than you see here. So it's quite interesting to see that the names that are succeeding are also succeeding with the largest organizations in the world. And as we know, large organizations means large budgets. So this is one area that I just thought was really interesting to point out that as we break it down, the data by vertical, these two names still are the outstanding players. >> I just also want to call it H2O.ai. They're getting a lot of buzz in the marketplace and I'm seeing them a lot more. Anaconda, another one. Dataiku consistently popping up. DataRobot is also interesting because all the kerfuffle that's going on there. The Cube guy, Cube alum, Chris Lynch stepped down as executive chairman. All this stuff came out about how the executives were taking money off the table and didn't allow the employees to participate in that money raising deal. So that's pissed a lot of people off. And so they're now going through some kind of uncomfortable things, which is unfortunate because DataRobot, I noticed, we haven't covered them that much in "Breaking Analysis", but I've noticed them oftentimes, Erik, in the surveys doing really well. So you would think that company has a lot of potential. But yeah, it's an important space that we're going to continue to watch. Let me ask you Erik, can you contextualize this from a time series standpoint? I mean, how is this changed over time? >> Yeah, again, not show here, but in the data. I'm sorry, go ahead. >> No, I'm sorry. What I meant, I should have interjected. In other words, you would think in a downturn that these emerging companies would be less interesting to buyers 'cause they're more risky. What have you seen? >> Yeah, and it was interesting before we went live, you and I were having this conversation about "Is the downturn stopping people from evaluating these private companies or not," right. In a larger sense, that's really what we're doing here. How are these private companies doing when it comes down to the actual practitioners? The people with the budget, the people with the decision making. And so what I did is, we have historical data as you know, I went back to the Emerging Technology Survey we did in November of 21, right at the crest right before the market started to really fall and everything kind of started to fall apart there. And what I noticed is on the security side, very much so, we're seeing less evaluations than we were in November 21. So I broke it down. On cloud security, net sentiment went from 21% to 16% from November '21. That's a pretty big drop. And again, that sentiment is our one aggregate metric for overall positivity, meaning utilization and actual evaluation of the name. Again in database, we saw it drop a little bit from 19% to 13%. However, in analytics we actually saw it stay steady. So it's pretty interesting that yes, cloud security and security in general is always going to be important. But right now we're seeing less overall net sentiment in that space. But within analytics, we're seeing steady with growing mindshare. And also to your point earlier in machine learning, AI, we're seeing steady net sentiment and mindshare has grown a whopping 25% to 30%. So despite the downturn, we're seeing more awareness of these companies in analytics and machine learning and a steady, actual utilization of them. I can't say the same in security and database. They're actually shrinking a little bit since the end of last year. >> You know it's interesting, we were on a round table, Erik does these round tables with CISOs and CIOs, and I remember one time you had asked the question, "How do you think about some of these emerging tech companies?" And one of the executives said, "I always include somebody in the bottom left of the Gartner Magic Quadrant in my RFPs. I think he said, "That's how I found," I don't know, it was Zscaler or something like that years before anybody ever knew of them "Because they're going to help me get to the next level." So it's interesting to see Erik in these sectors, how they're holding up in many cases. >> Yeah. It's a very important part for the actual IT practitioners themselves. There's always contracts coming up and you always have to worry about your next round of negotiations. And that's one of the roles these guys play. You have to do a POC when contracts come up, but it's also their job to stay on top of the new technology. You can't fall behind. Like everyone's a software company. Now everyone's a tech company, no matter what you're doing. So these guys have to stay in on top of it. And that's what this ETS can do. You can go in here and look and say, "All right, I'm going to evaluate their technology," and it could be twofold. It might be that you're ready to upgrade your technology and they're actually pushing the envelope or it simply might be I'm using them as a negotiation ploy. So when I go back to the big guy who I have full intentions of writing that contract to, at least I have some negotiation leverage. >> Erik, we got to leave it there. I could spend all day. I'm going to definitely dig into this on my own time. Thank you for introducing this, really appreciate your time today. >> I always enjoy it, Dave and I hope everyone out there has a great holiday weekend. Enjoy the rest of the summer. And, you know, I love to talk data. So anytime you want, just point the camera on me and I'll start talking data. >> You got it. I also want to thank the team at ETR, not only Erik, but Darren Bramen who's a data scientist, really helped prepare this data, the entire team over at ETR. I cannot tell you how much additional data there is. We are just scratching the surface in this "Breaking Analysis". So great job guys. I want to thank Alex Myerson. Who's on production and he manages the podcast. Ken Shifman as well, who's just coming back from VMware Explore. Kristen Martin and Cheryl Knight help get the word out on social media and in our newsletters. And Rob Hof is our editor in chief over at SiliconANGLE. Does some great editing for us. Thank you. All of you guys. Remember these episodes, they're all available as podcast, wherever you listen. All you got to do is just search "Breaking Analysis" podcast. I publish each week on wikibon.com and siliconangle.com. Or you can email me to get in touch david.vellante@siliconangle.com. You can DM me at dvellante or comment on my LinkedIn posts and please do check out etr.ai for the best survey data in the enterprise tech business. This is Dave Vellante for Erik Bradley and The Cube Insights powered by ETR. Thanks for watching. Be well. And we'll see you next time on "Breaking Analysis". (upbeat music)

Published Date : Sep 7 2022

SUMMARY :

bringing you data driven it's called the Emerging Great to see you too, Dave, so much in the mainstream, not only for the ITDMs themselves It is the heart of innovation So the net sentiment is a very So a lot of names that we And then of course you have AnyScale, That's the bad zone, I guess, So the gray dots that you're rates, adoption and the all And on the lower side, Vena, Acton, in the green. are in the enterprise already. So now let's look at the churn So that's the way you can look of dwell on the negative, So again, the axis is still the same, And a couple of the other And then you see these great standouts, Those are the ones you want to but Redis Labs is the one And by the way, MariaDB, So it's not in this slide, Alex, bring that up if you would. So gimme one second to catch up. So I could set it up but based on the amount of time Those are the ones we were saying before, And one of the things I think didn't allow the employees to here, but in the data. What have you seen? the market started to really And one of the executives said, And that's one of the Thank you for introducing this, just point the camera on me We are just scratching the surface

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
ErikPERSON

0.99+

Alex MyersonPERSON

0.99+

Ken ShifmanPERSON

0.99+

Sanjay PoonenPERSON

0.99+

Dave VellantePERSON

0.99+

DavePERSON

0.99+

Erik BradleyPERSON

0.99+

November 21DATE

0.99+

Darren BramenPERSON

0.99+

AlexPERSON

0.99+

Cheryl KnightPERSON

0.99+

PostgresORGANIZATION

0.99+

DatabricksORGANIZATION

0.99+

NetskopeORGANIZATION

0.99+

AdobeORGANIZATION

0.99+

Rob HofPERSON

0.99+

FivetranORGANIZATION

0.99+

$50 millionQUANTITY

0.99+

21%QUANTITY

0.99+

Chris LynchPERSON

0.99+

19%QUANTITY

0.99+

Jeremy BurtonPERSON

0.99+

$800 millionQUANTITY

0.99+

6,000QUANTITY

0.99+

OracleORGANIZATION

0.99+

Redis LabsORGANIZATION

0.99+

November '21DATE

0.99+

ETRORGANIZATION

0.99+

FirstQUANTITY

0.99+

25%QUANTITY

0.99+

last yearDATE

0.99+

OneTrustORGANIZATION

0.99+

two dimensionsQUANTITY

0.99+

two groupsQUANTITY

0.99+

November of 21DATE

0.99+

bothQUANTITY

0.99+

BostonLOCATION

0.99+

more than 400 companiesQUANTITY

0.99+

Kristen MartinPERSON

0.99+

MySQLTITLE

0.99+

MoogsoftORGANIZATION

0.99+

The CubeORGANIZATION

0.99+

thirdQUANTITY

0.99+

GrafanaORGANIZATION

0.99+

H2OORGANIZATION

0.99+

Mike SpeiserPERSON

0.99+

david.vellante@siliconangle.comOTHER

0.99+

secondQUANTITY

0.99+

twoQUANTITY

0.99+

firstQUANTITY

0.99+

28%QUANTITY

0.99+

16%QUANTITY

0.99+

SecondQUANTITY

0.99+

Day 1 Keynote Analysis | Snowflake Summit 2022


 

>>Good morning live from Las Vegas, Lisa Martin and Dave Lanta here covering snowflake summit 22. Dave, it's great to be here in person. The keynote we just came from was standing room only. In fact, there was overflow. People are excited to be back and to hear from the company in person the first time, since the IPO, >>Lots of stuff, lots of deep technical dives, uh, you know, they took the high end of the pyramid and then dove down deep in the keynotes. It >>Was good. They did. And we've got Doug Hench with us to break this down in the next eight to 10 minutes, VP and principle analyst at constellation research. Doug, welcome to the cube. >>Great to be here. >>All right, so guys, I was telling Dave, as we were walking back from the keynote, this was probably the most technical keynote I've seen in a very long time. Obviously in person let's break down some of the key announcements. What were some of the things Dave that stood out to you and what they announced just in the last hour and a half alone? >>Well, I, you know, we had a leave before they did it, but the unit store piece was really interesting to me cuz you know, the big criticism is, oh, say snowflake, that doesn't do transaction data. It's just a data warehouse. And now they're sort of reaching out. We're seeing the evolution of the ecosystem. Uh, sluman said it was by design. It was one of the questions I had for them. Is this just kind of happen or is it by design? So that's one of many things that, that we can unpack. I mean the security workload, uh, the, the Apache tables, we were just talking about thatt, which not a lot of hands went up when they said, who uses Apache tables, but, but a lot of the things they're doing seem to me anyway, to be trying to counteract the narrative, that snow, I mean that data bricks is put out there about you guys. Aren't open, you're a walled garden and now they're saying, Hey, we're we're as open as anybody, but what are your thoughts, Doug? >>Well, that's the, the iceberg announcement, uh, also, uh, the announcement of, of uni store being able to reach out to, to any source. Uh, you know, I think the big theme here was this, this contrast you constantly see with snowflake between their effort to democratize and simplify and disrupt the market by bringing in a great big tent. And you saw that great big tent here today, 7,000 people, 2,007,000 plus I'm told 2000 just three years ago. So this company is growing hugely quickly, >>Unprecedented everybody. >>Yeah. Uh, fastest company to a billion in revenue is Frank Salman said in his keynote today. Um, you know, and I think that there's, there's that great big tent. And then there's the innovations they're delivering. And a lot of their announcements are way ahead of the J general availability. A lot of the things they talked about today, Python support and some, some other aspects they're just getting into public preview. And many of the things that they're announcing today are in private preview. So it could be six, 12 months be before they're generally available. So they're here educating a lot of these customers. What is iceberg? You know, they're letting them know about, Hey, we're not just the data warehouse. We're not just letting you migrate your old workloads into the cloud. We're helping you innovate with things like the data marketplace. I see the data marketplace is really crucial to a lot of the announcements they're making today. Particularly the native apps, >>You know, what was interesting sluman in his keynote said we don't use the term data mesh, cuz that means has meaning to the people, lady from Geico stood up and said, we're building a data mesh. And when you think about, you know, the, those Gemma Dani's definition of data mesh, Snowflake's actually ticking a lot of boxes. I mean, it's it's is it a decentralized architecture? You could argue that it's sort of their own wall garden, but things like data as product we heard about building data products, uh, uh, self-serve infrastructure, uh, computational governance, automated governance. So those are all principles of Gemma's data mesh. So I there's close as anybody that, that I've seen with the exception of it's all in the data cloud. >>Why do you think he was very particular in saying we're not gonna call it a data mesh? I, >>I think he's respecting the principles that have been put forth by the data mesh community generally and specifically Jamma Dani. Uh, and they don't want to, you know, they don't want to data mesh wash. I mean, I, I, I think that's a good call. >>Yeah, that's it's a little bit out there and, and it, they didn't talk about data mesh so much as Geico, uh, the keynote or mentioned their building one. So again, they have this mix of the great big tent of customers and then very forward looking very sophisticated customers. And that's who they're speaking to with some of these announcements, like the native apps and the uni store to bring transactional data, bring more data in and innovate, create new apps. And the key to the apps is that they're made available through the marketplace. Things like data sharing. That's pretty simple. A lot of, uh, of their competitors are talking about, Hey, we can data share, but they don't have the things that make it easy, like the way to distribute the data, the way to monetize the data. So now they're looking forward monetizing apps, they changed the name from the data marketplace to the, to the snowflake marketplace. So it'll be apps. It will be data. It'll be all sorts of innovative products. >>We talk about Geico, uh, JPMC is speaking at this conference, uh, and the lead technical person of their data mesh initiative. So it's like, they're some of their customers that they're putting forth. So it's kind of interesting. And then Doug, something else that you and I have talked about on the, some of the panels that we've done is you've got an application development stack, you got the database over there and then you have the data analytics stack and we've, I've said, well, those things come together. Then people have said, yeah, they have to. And this is what snowflake seems to be driving towards. >>Well with uni store, they're reaching out and trying to bring transactional data in, right? Hey, don't limit this to analytical information. And there's other ways to do that, like CDC and streaming, but they're very closely tying that again to that marketplace, with the idea of bring your data over here and you can monetize it. Don't just leave it in that transactional database. So a, another reach to a broader play across a big community that they're >>Building different than what we saw last week at Mongo, different than what you know, Oracle does with, with heat wave. A lot of ways to skin a cat. >>That was gonna be my next question to both of you is talk to me about all the announcements that we saw. And, and like we said, we didn't actually get to see the entire keynote had come back here. Where are they from a differentiation perspective in terms of the competitive market? You mentioned Doug, a lot of the announcements in either private preview or soon to be public preview early. Talk to me about your thoughts where they are from a competitive standpoint. >>Again, it's that dichotomy between their very forward looking announcements. They're just coming on with things like Python support. That's just becoming generally available. They're just introducing, uh, uh, machine learning algorithms, like time series built into the database. So in some ways they're catching up while painting this vision of future capabilities and talking about things that are in development or in private preview that won't be here for a year or two, but they're so they're out there, uh, talking about a BLE bleeding edge story yet the reality is the product sometimes are lagging behind. Yeah, >>It's interesting. I mean, they' a lot of companies choose not to announce anything until it's ready to ship. Yeah. Typically that's a technique used by the big whales to try to freeze the market, but I think it's different here. And the strategy is to educate customers on what's possible because snowflake really does have, you know, they're trying to differentiate from, Hey, we're not just a data warehouse. We have a highly differentiatable strategy from whether it's Oracle or certainly, you know, Mongo is more transactional, but, but you know, whether it's couch base or Redis or all the other databases out there, they're saying we're not a database, we're a data cloud. <laugh> right. Right. Okay. What is that? Well, look at all the things that you can do with the data cloud, but to me, the most interesting is you can actually build data products and you can monetize that. And their, the emphasis on ecosystem, you, they look at Salman's previous company would ServiceNow took a long time for them to build an ecosystem. It was a lot of SI in smaller SI and they finally kind of took off, but this is exceeding my expectations and ecosystem is critical because they can't do it all. You know, they're gonna O otherwise they're gonna spread themselves to >>That. That's what I think some competitors just don't get about snowflake. They don't get that. It's all about the community, about their network that they're building and the relationships between these customers. And that they're facilitating that with distribution, with monetization, things that are hard. So you can't just add sharing, or you can share data from one of their, uh, legacy competitors, uh, in, in somebody else's marketplace that doesn't facilitate the transaction that doesn't, you know, build on the community. Well, >>And you know, one of the criticisms too, of the criticism on snowflake goes, they don't, you know, they can't do complex joins. They don't do workload management. And I think their answer to that is, well, we're gonna look to the ecosystem to do that. Or you, you saw some kind of, um, cost governance today in the, in the keynote, we're gonna help you optimize your spend, um, a little different than workload management, but related >>Part of their governance was having a, a, a node, uh, for every workload. So workload isolation in that way, but that led to the cost problems, you know, like too many nodes with not enough optimization. So here too, you saw a lot of, uh, announcements around cost controls, budgets, new features, uh, user groups that you could bring, uh, caps and guardrails around those costs. >>In the last couple minutes, guys talk about their momentum. Franks Lutman showed a slide today that showed over 5,900 customers. I was looking at some stats, uh, in the last couple of days that showed that there is an over 1200% increase in the number of customers with a million plus ARR. Talk about their momentum, what you expect to see here. A lot of people here, people are ready to hear what they're doing in person. >>Well, I think this, the stats say it all, uh, fastest company to a, to a billion in revenue. Uh, you see the land and expand experience that many companies have and in the cost control, uh, announcements they were making, they showed the typical curve like, and he talked about it being a roller coaster, and we wanna help you level that out. Uh, so that's, uh, a matter of maturation. Uh, that's one of the downsides of this rapid growth. You know, you have customers adding new users, adding new clusters, multi clusters, and the costs get outta control. They want to help customers even that out, uh, with reporting with these budget and cost control measures. So, uh, one of the growing pains that comes with, uh, adding so many customers so quickly, and those customers adding so many users and new, uh, workloads quickly, >>I know we gotta break, but last point I'll make about the key. Uh, keynote is SL alluded to the fact that they're not taking the foot off the gas. They don't see any reason to, despite the narrative in the press, they have inherent profitability. If they want to be more profitable, they could be, but they're going for growth >>Going for growth. There is so much to unpack in the next three days. You won't wanna miss it. The Cube's wall to oil coverage, Lisa Martin for Dave Valenti, Doug hen joined us in our keynote analysis. Thanks so much for walking, watching stick around. Our first guest is up in just a few minutes.

Published Date : Jun 14 2022

SUMMARY :

22. Dave, it's great to be here in person. Lots of stuff, lots of deep technical dives, uh, you know, they took the high end of the pyramid and then dove down deep And we've got Doug Hench with us to break this down in the next eight to 10 minutes, stood out to you and what they announced just in the last hour and a half alone? but, but a lot of the things they're doing seem to me anyway, to be trying to counteract the narrative, Uh, you know, I think the big theme here was this, And many of the things that they're announcing today are in private preview. And when you think about, you know, the, those Gemma Dani's definition of data mesh, Uh, and they don't want to, you know, And the key to the apps is that they're made available through the marketplace. And then Doug, something else that you and I have talked about on the, some of the panels that we've done is you've So a, another reach to a broader play across a big community that Building different than what we saw last week at Mongo, different than what you know, Oracle does with, That was gonna be my next question to both of you is talk to me about all the announcements that we saw. into the database. Well, look at all the things that you can do with the data cloud, but to me, the most interesting is you So you can't just add sharing, or you can share data from one of their, And you know, one of the criticisms too, of the criticism on snowflake goes, they don't, you know, they can't do complex joins. new features, uh, user groups that you could bring, uh, A lot of people here, people are ready to hear what they're doing they showed the typical curve like, and he talked about it being a roller coaster, and we wanna help you level that Uh, keynote is SL alluded to the fact that they're There is so much to unpack in the next three days.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

Dave LantaPERSON

0.99+

Frank SalmanPERSON

0.99+

Doug HenchPERSON

0.99+

DavePERSON

0.99+

DougPERSON

0.99+

Dave ValentiPERSON

0.99+

Las VegasLOCATION

0.99+

2,007,000QUANTITY

0.99+

7,000 peopleQUANTITY

0.99+

PythonTITLE

0.99+

sixQUANTITY

0.99+

todayDATE

0.99+

OracleORGANIZATION

0.99+

last weekDATE

0.99+

a yearQUANTITY

0.99+

SalmanPERSON

0.99+

oneQUANTITY

0.99+

Doug henPERSON

0.99+

bothQUANTITY

0.99+

first guestQUANTITY

0.99+

2000DATE

0.99+

over 5,900 customersQUANTITY

0.99+

Jamma DaniPERSON

0.98+

three years agoDATE

0.98+

over 1200%QUANTITY

0.98+

first timeQUANTITY

0.98+

GeicoORGANIZATION

0.98+

MongoORGANIZATION

0.98+

JPMCORGANIZATION

0.97+

10 minutesQUANTITY

0.97+

slumanPERSON

0.96+

Gemma DaniPERSON

0.96+

Day 1QUANTITY

0.96+

Snowflake Summit 2022EVENT

0.95+

twoQUANTITY

0.95+

eightQUANTITY

0.89+

ServiceNowORGANIZATION

0.87+

12 monthsQUANTITY

0.86+

a million plusQUANTITY

0.84+

ApacheORGANIZATION

0.81+

GemmaPERSON

0.78+

a billionQUANTITY

0.76+

next three daysDATE

0.72+

CDCORGANIZATION

0.71+

RedisORGANIZATION

0.7+

Franks LutmanPERSON

0.67+

last couple of daysDATE

0.66+

summit 22LOCATION

0.64+

uni storeORGANIZATION

0.63+

and a halfQUANTITY

0.53+

CubeORGANIZATION

0.52+

last hourDATE

0.52+

uniTITLE

0.51+

storeORGANIZATION

0.47+

billionQUANTITY

0.46+

SnowflakeTITLE

0.39+

coupleDATE

0.36+

Breaking Analysis: How Lake Houses aim to be the Modern Data Analytics Platform


 

from the cube studios in palo alto in boston bringing you data driven insights from the cube and etr this is breaking analysis with dave vellante earnings season has shown a conflicting mix of signals for software companies well virtually all firms are expressing caution over so-called macro headwinds we're talking about ukraine inflation interest rates europe fx headwinds supply chain just overall i.t spend mongodb along with a few other names appeared more sanguine thanks to a beat in the recent quarter and a cautious but upbeat outlook for the near term hello and welcome to this week's wikibon cube insights powered by etr in this breaking analysis ahead of mongodb world 2022 we drill into mongo's business and what etr survey data tells us in the context of overall demand and the patterns that we're seeing from other software companies and we're seeing some distinctly different results from major firms these days we'll talk more about [ __ ] in this session which beat eps by 30 cents in revenue by more than 18 million dollars salesforce had a great quarter and its diversified portfolio is paying off as seen by the stocks noticeable uptick post earnings uipath which had been really beaten down prior to this quarter it's brought in a new co-ceo and it's business is showing a nice rebound with a small three cent eps beat and a nearly 20 million dollar top line beat crowdstrike is showing strength as well meanwhile managements at microsoft workday and snowflake expressed greater caution about the macroeconomic climate and especially on investors minds his concern about consumption pricing models snowflake in particular which had a small top-line beat cited softness and effects from reduced consumption especially from certain consumer-facing customers which has analysts digging more deeply into the predictability of their models in fact barclays analyst ramo lenchow published an especially thoughtful piece on this topic concluding that [ __ ] was less susceptible to consumption headwinds than for example snowflake essentially for a few reasons one because atlas mongo's cloud managed service which is the consumption model comprises only about 60 percent of mongo's revenue second is the premise that [ __ ] is supporting core operational applications that can't be easily dialed down or turned off and three that snowflake customers it sounds like has a more concentrated customer base and due to that fact there's a preponderance of its revenue is consumption driven and would be more sensitive to swings in these consumption patterns now i'll say this first consumption pricing models are here to stay and the much preferred model for customers is consumption the appeal of consumption is i can actually dial down turn off if i need to and stop spending for a while which happened or at least happened to a certain extent this quarter for certain companies but to the point about [ __ ] supporting core applications i do believe that over time you're going to see the increased emergence of data products that will become core monetization drivers in snowflake along with other data platforms is going to feed those data products and services and become over time maybe less susceptible and less sensitive to these consumption patterns it'll always be there but i think increasingly it's going to be tied to operational revenue last two points here in this slide software evaluations have reverted to their historical mean which is a good thing in our view we've taken some air out of the bubble and returned to more normalized valuations was really predicted and looked forward to look we're still in a lousy market for stocks it's really a bear market for tech the market tends to be at least six months ahead of the economy and often not always but often is a good predictor we've had some tough compares relative to the pandemic days in tech and we'll be watching next quarter very closely because the macro headwinds have now been firmly inserted into the guidance of software companies okay let's have a look at how certain names have performed relative to a software index benchmark so far this year here's a year-to-date chart comparing microsoft salesforce [ __ ] and snowflake to the igv software heavy etf which is shown in the darker blue line which by the way it does not own the ctf does not own snowflake or [ __ ] you can see that these big super caps have fared pretty well whereas [ __ ] and especially snowflake those higher growth companies have been much more negatively impacted year to date from a stock price standpoint now let's move on let's take a financial snapshot of [ __ ] and put it next to snowflake so we can compare these two higher growth names what we've done here in this chart has taken the most recent quarters revenue and multiplied it by 4x to get a revenue run rate and we've parenthetically added a projection for the full year revenue [ __ ] as you see will do north of a billion dollars in revenue while snowflake will begin to approach three billion dollars 2.7 and run right through that that four quarter run rate that they just had last quarter and you can see snowflake is growing faster than [ __ ] at 85 percent this past quarter and we took now these most of these profit of these next profitability ratios off the current quarter with one exception both companies have high gross margins of course you'd expect that but as we've discussed not as high as some traditional software companies in part because of their cloud costs but also you know their maturity or lack thereof both [ __ ] and snowflake because they are in high growth mode have thin operating margins they spend nearly half or more than half of their revenue on growth that's the sg a line mostly the s the sales and marketing is really where they're spending money uh and and they're specialists so they spend a fair amount of their revenue on r d but maybe not as high as you might think but a pretty hefty percentage the free cash flow as a percentage of revenue line we calculated off the full year projections because there was a kind of an anomaly this quarter in the in the snowflake numbers and you can see snowflakes free cash flow uh which again was abnormally high this quarter is going to settle in around 16 this year versus mongo's six percent so strong focus by snowflake on free cash flow and its management snowflake is about four billion dollars in cash and marketable securities on its balance sheet with little or no debt whereas [ __ ] has about two billion dollars on its balance sheet with a little bit of longer term debt and you can see snowflakes market cap is about double that of mongos so you're paying for higher growth with snowflake you're paying for the slootman scarpelli execution engine the expectation there a stronger balance sheet etc but snowflake is well off its roughly 100 billion evaluation which it touched during the peak days of tech during the pandemic and just that as an aside [ __ ] has around 33 000 customers about five times the number of customers snowflake has so a bit of a different customer mix and concentration but both companies in our view have no lack of market in terms of tam okay now let's dig a little deeper into mongo's business and bring in some etr data this colorful chart shows the breakdown of mongo's net score net score is etr's proprietary methodology that measures the percent of customers in the etr survey that are adding the platform new that's the lime green at nine percent existing customers that are spending six percent or more on the platform that's the forest green at 37 spending flat that's the gray at 46 percent decreasing spend that's the pinkish at around 5 and churning that's only 3 that's the bright red for [ __ ] subtract the red from the greens and you net out to a 38 which is a very solid net score figure note this is a survey of 1500 or so organizations and it includes 150 mongodb customers which includes by the way 68 global 2000 customers and they show a spending velocity or a net score of 44 so notably higher among the larger clients and while it's a smaller sample only 27 emea's net score for [ __ ] is 33 now that's down from 60 last quarter note that [ __ ] cited softness in its european business on its earning calls so that aligns to the gtr data okay now let's plot [ __ ] relative to some other data platforms these don't all necessarily compete head to head with [ __ ] but they are in data and database platforms in the etr data set and that's what this chart shows it's an xy graph with net score or as we say spending momentum on the vertical axis and overlap or presence or pervasiveness in the data set on the horizontal axis see that red dotted line there at 40 that indicates an elevated level of spending anything above that is highly elevated we've highlighted [ __ ] in that red box which is very close to that 40 percent line it has a pretty strong presence on the x-axis right there with gcp snowflake as we've reported has come down to earth but still well elevated again that aligns with the earnings releases uh aws and microsoft they have many data platforms especially aws so their plot position reflects their broad portfolio massive size on the x-axis um that's the presence and and very impressive on the vertical axis so despite that size they have strong spending momentum and you can see the pack of others including cockroach small on the verdict on the horizontal but elevated on the vertical couch base is creeping up since its ipo redis maria db which was launched the day that oracle bought sun and and got my sequel and some legacy platforms including the leader in database oracle as well as ibm and teradata's both cloud and on-prem platforms now one interesting side note here is on mongo's earning call it clearly cited the advantages of its increasingly all-in-one approach relative to others that offer a portfolio of bespoke or what we some sometimes call horses for courses databases [ __ ] cited the advantages of its simplicity and lower costs as it adds more and more functionality this is an argument often made by oracle and they often target aws as the company with too many databases and of course [ __ ] makes that argument uh as well but they also make the argument that oracle they don't necessarily call them out but they talk about traditional relational databases of course they're talking about oracle and others they say that's more complex less flexible and less appealing to developers than is [ __ ] now oracle of course would retur we retort saying hey we now support a mongodb api so why go anywhere else we're the most robust and the best for mission critical but this gives credence to the fact that if oracle is trying to capture business by offering a [ __ ] api for example that [ __ ] must be doing something right okay let's look at why they buy [ __ ] here's an etr chart that addresses that question it's it's mongo's feature breadth is the number one reason lower cost or better roi is number two integrations and stack alignment is third and mongo's technology lead is fourth those four kind of stand out with notice on the right hand side security and vision much lower there in the right that doesn't necessarily mean that [ __ ] doesn't have good security and and good vision although it has been cited uh security concerns um and and so we keep an eye on that but look [ __ ] has a document database it's become a viable alternative to traditional relational databases meaning you have much more flexibility over your schema um and in fact you know it's kind of schema-less you can pretty much put anything into a document database uh developers seem to love it generally it's fair to say mongo's architecture would favor consistency over availability because it uses a single master architecture as a primary and you can create secondary nodes in the event of a primary failure but you got to think about that and how to architect availability into the platform and got to consider recovery more carefully now now no schema means it's not a tables and rows structure and you can again shove anything you want into the database but you got to think about how to optimize performance um on queries now [ __ ] has been hard at work evolving the platform from the early days when you go back and look at its roadmap it's been you know started as a document database purely it added graph processing time series it's made search you know much much easier and more fundamental it's added atlas that fully managed cloud database uh service which we said now comprises 60 of its revenue it's you know kubernetes integrations and kind of the modern microservices stack and dozens and dozens and dozens of other features mongo's done a really fine job we think of creating a leading database platform today that is loved by customers loved by developers and is highly functional and next week the cube will be at mongodb world and we'll be looking for some of these items that we're showing here and this this chart this always going to be main focus on developers [ __ ] prides itself on being a developer friendly platform we're going to look for new features especially around security and governance and simplification of configurations and cluster management [ __ ] is likely going to continue to advance its all-in-one appeal and add more capabilities that reduce the need to to spin up bespoke platforms and we would expect enhance enhancements to atlas further enhancements there is atlas really is the future you know maybe adding you know more cloud native features and integrations and perhaps simplified ways to migrate to the cloud to atlas and improve access to data sources generally making the lives of developers and data analysts easier that's going to be we think a big theme at the event so these are the main things that we'll be scoping out at the event so please stop by if you're in new york city new york city at mongodb world or tune in to thecube.net okay that's it for today thanks to my colleagues stephanie chan who helps research breaking analysis from time to time alex meyerson is on production as today is as is andrew frick sarah kenney steve conte conte anderson hill and the entire team in palo alto thank you kristen martin and cheryl knight helped get the word out and rob hof is our editor-in-chief over there at siliconangle remember all these episodes are available as podcasts wherever you listen just search breaking analysis podcast we do publish each week on wikibon.com and siliconangle.com want to reach me email me david.velante siliconangle.com or dm me at divalante or a comment on my linkedin post and please do check out etr.ai for the best survey data in the enterprise tech business this is dave vellante for the cube insights powered by etr thanks for watching see you next time [Music] you

Published Date : Jun 3 2022

SUMMARY :

into the platform and got to consider

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
nine percentQUANTITY

0.99+

30 centsQUANTITY

0.99+

six percentQUANTITY

0.99+

46 percentQUANTITY

0.99+

ramo lenchowPERSON

0.99+

new yorkLOCATION

0.99+

next weekDATE

0.99+

thecube.netOTHER

0.99+

85 percentQUANTITY

0.99+

microsoftORGANIZATION

0.99+

40 percentQUANTITY

0.99+

six percentQUANTITY

0.99+

cherylPERSON

0.99+

andrew frickPERSON

0.99+

three billion dollarsQUANTITY

0.99+

more than 18 million dollarsQUANTITY

0.99+

dave vellantePERSON

0.99+

oracleORGANIZATION

0.99+

this yearDATE

0.99+

stephanie chanPERSON

0.99+

alex meyersonPERSON

0.99+

next quarterDATE

0.99+

37QUANTITY

0.99+

44QUANTITY

0.99+

last quarterDATE

0.99+

bostonLOCATION

0.99+

60QUANTITY

0.99+

both companiesQUANTITY

0.99+

38QUANTITY

0.99+

david.velanteOTHER

0.99+

todayDATE

0.99+

about two billion dollarsQUANTITY

0.99+

dozensQUANTITY

0.98+

about four billion dollarsQUANTITY

0.98+

rob hofPERSON

0.98+

33QUANTITY

0.98+

firstQUANTITY

0.98+

each weekQUANTITY

0.98+

around 33 000 customersQUANTITY

0.98+

27QUANTITY

0.98+

secondQUANTITY

0.98+

4xQUANTITY

0.97+

150 mongodb customersQUANTITY

0.97+

threeQUANTITY

0.97+

more than halfQUANTITY

0.97+

fourthQUANTITY

0.96+

awsORGANIZATION

0.96+

this weekDATE

0.96+

bothQUANTITY

0.96+

nearly 20 million dollarQUANTITY

0.96+

anderson hillPERSON

0.96+

2022DATE

0.95+

palo altoORGANIZATION

0.94+

mongoORGANIZATION

0.94+

sarah kenneyPERSON

0.94+

kristen martinPERSON

0.93+

about 60 percentQUANTITY

0.93+

oneQUANTITY

0.93+

40QUANTITY

0.93+

one exceptionQUANTITY

0.93+

2.7QUANTITY

0.93+

thirdQUANTITY

0.93+

fourQUANTITY

0.93+

atlasTITLE

0.92+

two higher growthQUANTITY

0.92+

about five timesQUANTITY

0.92+

3QUANTITY

0.91+

etrORGANIZATION

0.91+

pandemicEVENT

0.91+

atlas mongoORGANIZATION

0.91+

this quarterDATE

0.9+

ukraineLOCATION

0.9+

siliconangle.comOTHER

0.89+

2000 customersQUANTITY

0.88+

paloLOCATION

0.88+

around 5QUANTITY

0.87+

Breaking Analysis: What you May not Know About the Dell Snowflake Deal


 

>> From theCUBE Studios in Palo Alto, in Boston bringing you Data Driven Insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante. >> In the pre-cloud era hardware companies would run benchmarks, showing how database and or application performance ran better on their systems relative to competitors or previous generation boxes. And they would make a big deal out of it. And the independent software vendors, you know they'd do a little golf clap if you will, in the form of a joint press release it became a game of leaprog amongst hardware competitors. That was pretty commonplace over the years. The Dell Snowflake Deal underscores that the value proposition between hardware companies and ISVs is changing and has much more to do with distribution channels, volumes and the amount of data that lives On-Prem in various storage platforms. For cloud native ISVs like Snowflake they're realizing that despite their Cloud only dogma they have to grit their teeth and deal with On-premises data or risk getting shut out of evolving architectures. Hello and welcome to this week's Wikibon Cube Insights powered by ETR. In this breaking analysis, we unpack what little is known about the Snowflake announcement from Dell Technologies World and discuss the implications of a changing Cloud landscape. We'll also share some new data for Cloud and Database platforms from ETR that shows Snowflake has actually entered the Earth's orbit when it comes to spending momentum on its platform. Now, before we get into the news I want you to listen to Frank's Slootman's answer to my question as to whether or not Snowflake would ever architect the platform to run On-Prem because it's doable technically, here's what he said, play the clip >> Forget it, this will only work in the Public Cloud. Because it's, this is how the utility model works, right. I think everybody is coming through this realization, right? I mean, excuses are running out at this point. You know, we think that it'll, people will come to the Public Cloud a lot sooner than we will ever come to the Private Cloud. It's not that we can't run a private Cloud. It's just diminishes the potential and the value that we bring. >> So you may be asking yourselves how do you square that circle? Because basically the Dell Snowflake announcement is about bringing Snowflake to the private cloud, right? Or is it let's get into the news and we'll find out. Here's what we know at Dell Technologies World. One of the more buzzy announcements was the, by the way this was a very well attended vet event. I should say about I would say 8,000 people by my estimates. But anyway, one of the more buzzy announcements was Snowflake can now run analytics on Non-native Snowflake data that lives On-prem in a Dell object store Dell's ECS to start with. And eventually it's software defined object store. Here's Snowflake's clark, Snowflake's Clark Patterson describing how it works this past week on theCUBE. Play the clip. The way it works is I can now access Non-native Snowflake data using what materialized views, external tables How does that work? >> Some combination of the, all the above. So we've had in Snowflake, a capability called External Tables, which you refer to, it goes hand in hand with this notion of external stages. Basically there's a through the combination of those two capabilities, it's a metadata layer on data, wherever it resides. So customers have actually used this in Snowflake for data lake data outside of Snowflake in the Cloud, up until this point. So it's effectively an extension of that functionality into the Dell On-Premises world, so that we can tap into those things. So we use the external stages to expose all the metadata about what's in the Dell environment. And then we build external tables in Snowflake. So that data looks like it is in Snowflake. And then the experience for the analyst or whomever it is, is exactly as though that data lives in the Snowflake world. >> So as Clark explained, this capability of External tables has been around in the Cloud for a while, mainly to suck data out of Cloud data lakes. Snowflake External Tables use file level metadata, for instance, the name of the file and the versioning so that it can be queried in a stage. A stage is just an external location outside of Snowflake. It could be an S3 bucket or an Azure Blob and it's soon will be a Dell object store. And in using this feature, the Dell looks like it lives inside of Snowflake and Clark essentially, he's correct to say to an analyst that looks exactly like the data is in Snowflake, but uh, not exactly the data's read only which means you can't do what are called DML operations. DML stands for Data Manipulation Language and allows for things like inserting data into tables or deleting and modifying existing data. But the data can be queried. However, the performance of those queries to External Tables will almost certainly be slower. Now users can build things like materialized views which are going to speed things up a bit, but at the end of the day, it's going to run faster than the Cloud. And you can be almost certain that's where Snowflake wants it to run, but some organizations can't or won't move data into the Cloud for a variety of reasons, data sovereignty, compliance security policies, culture, you know, whatever. So data can remain in place On-prem, or it can be moved into the Public Cloud with this new announcement. Now, the compute today presumably is going to be done in the Public Cloud. I don't know where else it's going to be done. They really didn't talk about the compute side of things. Remember, one of Snowflake's early innovations was to separate compute from storage. And what that gave them is you could more efficiently scale with unlimited resources when you needed them. And you could shut off the compute when you don't need us. You didn't have to buy, and if you need more storage you didn't have to buy more compute and vice versa. So everybody in the industry has copied that including AWS with Redshift, although as we've reported not as elegantly as Snowflake did. RedShift's more of a storage tiering solution which minimizes the compute required but you can't really shut it off. And there are companies like Vertica with Eon Mode that have enabled this capability to be done On-prem, you know, but of course in that instance you don't have unlimited elastic compute scale on-Prem but with solutions like Dell Apex and HPE GreenLake, you can certainly, you can start to simulate that Cloud elasticity On-prem. I mean, it's not unlimited but it's sort of gets you there. According to a Dell Snowflake joint statement, the companies the quote, the companies will pursue product integrations and joint go to market efforts in the second half of 2022. So that's a little vague and kind of benign. It's not really clear when this is going to be available based on that statement from the two first, but, you know, we're left wondering will Dell develop an On-Prem compute capability and enable queries to run locally maybe as part of an extended apex offering? I mean, we don't know really not sure there's even a market for that but it's probably a good bet that again, Snowflake wants that data to land in the Snowflake data Cloud kind of makes you wonder how this deal came about. You heard Sloop on earlier Snowflake has always been pretty dogmatic about getting data into its native snowflake format to enable the best performance as we talked about but also data sharing and governance. But you could imagine that data architects they're building out their data mesh we've reported on this quite extensively and their data fabric and those visions around that. And they're probably telling Snowflake, Hey if you want to be a strategic partner of ours you're going to have to be more inclusive of our data. That for whatever reason we're not putting in your Cloud. So Snowflake had to kind of hold its nose and capitulate. Now the good news is it further opens up Snowflakes Tam the total available market. It's obviously good marketing posture. And ultimately it provides an on ramp to the Cloud. And we're going to come back to that shortly but let's look a little deeper into what's happening with data platforms and to do that we'll bring in some ETR data. Now, let me just say as companies like Dell, IBM, Cisco, HPE, Lenovo, Pure and others build out their hybrid Clouds. The cold hard fact is not only do they have to replicate the Cloud Operating Model. You will hear them talk about that a lot, but they got to do that. So it, and that's critical from a user experience but in order to gain that flywheel momentum they need to build a robust ecosystem that goes beyond their proprietary portfolios. And, you know, honestly they're really not even in the first inning most companies and for the likes of Snowflake to sort of flip this, they've had to recognize that not everything is moving into the Cloud. Now, let's bring up the next slide. One of the big areas of discussion at Dell Tech World was Apex. That's essentially Dell's nascent as a service offering. Apex is infrastructure as a Service Cloud On-prem and obviously has the vision of connecting to the Cloud and across Clouds and out to the Edge. And it's no secret that database is one of the most important ingredients of infrastructure as a service generally in Cloud Infrastructure specifically. So this chart here shows the ETR data for data platforms inside of Dell accounts. So the beauty of ETR platform is you can cut data a million different ways. So we cut it. We said, okay, give us the Cloud platforms inside Dell accounts, how are they performing? Now, this is a two dimensional graphic. You got net score or spending momentum on the vertical axis and what ETR now calls Overlap formally called Market Share which is a measure of pervasiveness in the survey. That's on the horizontal axis that red dotted line at 40% represents highly elevated spending on the Y. The table insert shows the raw data for how the dots are positioned. Now, the first call out here is Snowflake. According to ETR quote, after 13 straight surveys of astounding net scores, Snowflake has finally broken the trend with its net score dropping below the 70% mark among all respondents. Now, as you know, net score is measured by asking customers are you adding the platform new? That's the lime green in the bar that's pointing from Snowflake in the graph and or are you increasing spend by 6% or more? That's the forest green is spending flat that's the gray is you're spend decreasing by 6% or worse. That's the pinkish or are you decommissioning the platform bright red which is essentially zero for Snowflake subtract the reds from the greens and you get a net score. Now, what's somewhat interesting is that snowflakes net score overall in the survey is 68 which is still huge, just under 70%, but it's net score inside the Dell account base drops to the low sixties. Nonetheless, this chart tells you why Snowflake it's highly elevated spending momentum combined with an increasing presence in the market over the past two years makes it a perfect initial data platform partner for Dell. Now and in the Ford versus Ferrari dynamic. That's going on between the likes of Dell's apex and HPE GreenLake database deals are going to become increasingly important beyond what we're seeing with this recent Snowflake deal. Now noticed by the way HPE is positioned on this graph with its acquisition of map R which is now part of HPE Ezmeral. But if these companies want to be taken seriously as Cloud players, they need to further expand their database affinity to compete ideally spinning up databases as part of their super Clouds. We'll come back to that that span multiple Clouds and include Edge data platforms. We're a long ways off from that. But look, there's Mongo, there's Couchbase, MariaDB, Cloudera or Redis. All of those should be on the short list in my view and why not Microsoft? And what about Oracle? Look, that's to be continued on maybe as a future topic in a, in a Breaking Analysis but I'll leave you with this. There are a lot of people like John Furrier who believe that Dell is playing with fire in the Snowflake deal because he sees it as a one way ticket to the Cloud. He calls it a one way door sometimes listen to what he said this past week. >> I would say that that's a dangerous game because we've seen that movie before, VMware and AWS. >> Yeah, but that we've talked about this don't you think that was the right move for VMware? >> At the time, but if you don't nurture the relationship AWS will take all those customers ultimately from VMware. >> Okay, so what does the data say about what John just said? How is VMware actually doing in Cloud after its early missteps and then its subsequent embracing of AWS and other Clouds. Here's that same XY graphic spending momentum on the Y and pervasiveness on the X and the same table insert that plots the dots and the, in the breakdown of Dell's net score granularity. You see that at the bottom of the chart in those colors. So as usual, you see Azure and AWS up and to the right with Google well behind in a distant third, but still in the mix. So very impressive for Microsoft and AWS to have both that market presence in such elevated spending momentum. But the story here in context is that the VMware Cloud on AWS and VMware's On-Prem Cloud like VMware Cloud Foundation VCF they're doing pretty well in the market. Look, at HPE, gaining some traction in Cloud. And remember, you may not think HPE and Dell and VCF are true Cloud but these are customers answering the survey. So their perspective matters more than the purest view. And the bad news is the Dell Cloud is not setting the world on fire from a momentum standpoint on the vertical axis but it's above the line of zero and compared to Dell's overall net score of 20 you could see it's got some work to do. Okay, so overall Dell's got a pretty solid net score to you know, positive 20, as I say their Cloud perception needs to improve. Look, Apex has to be the Dell Cloud brand not Dell reselling VMware. And that requires more maturity of Apex it's feature sets, its selling partners, its compensation models and it's ecosystem. And I think Dell clearly understands that. I think they're pretty open about that. Now this includes partners that go beyond being just sellers has to include more tech offerings in the marketplace. And actually they got to build out a marketplace like Cloud Platform. So they got a lot of work to do there. And look, you've got Oracle coming up. I mean they're actually kind of just below the magic 40% in the line which is pro it's pretty impressive. And we've been telling you for years, you can hate Oracle all you want. You can hate its price, it's closed system all of that it's red stack shore. You can say it's legacy. You can say it's old and outdated, blah, blah, blah. You can say Oracle is irrelevant in trouble. You are dead wrong. When it comes to mission critical workloads. Oracle is the king of the hill. They're a founder led company that knows exactly what it's doing and they're showing Cloud momentum. Okay, the last point is that while Microsoft AWS and Google have major presence as shown on the X axis. VMware and Oracle now have more than a hundred citations in the survey. You can see that on the insert in the right hand, right most column. And IBM had better keep the momentum from last quarter going, or it won't be long before they get passed by Dell and HP in Cloud. So look, John might be right. And I would think Snowflake quietly agrees that this Dell deal is all about access to Dell's customers and their data. So they can Hoover it into the Snowflake Data Cloud but the data right now, anyway doesn't suggest that's happening with VMware. Oh, by the way, we're keeping an eye close eye on NetApp who last September ink, a similar deal to VMware Cloud on AWS to see how that fares. Okay, let's wrap with some closing thoughts on what this deal means. We learned a lot from the Cloud generally in AWS, specifically in two pizza teams, working backwards, customer obsession. We talk about flywheel all the time and we've been talking today about marketplaces. These have all become common parlance and often fundamental narratives within strategic plans investor decks and customer presentations. Cloud ecosystems are different. They take both competition and partnerships to new heights. You know, when I look at Azure service offerings like Apex, GreenLake and similar services and I see the vendor noise or hear the vendor noise that's being made around them. I kind of shake my head and ask, you know which movie were these companies watching last decade? I really wish we would've seen these initiatives start to roll out in 2015, three years before AWS announced Outposts not three years after but Hey, the good news is that not only was Outposts a wake up call for the On-Prem crowd but it's showing how difficult it is to build a platform like Outposts and bring it to On-Premises. I mean, Outpost isn't currently even a rounding era in the marketplace. It really doesn't do much in terms of database support and support of other services. And, you know, it's unclear where that that is going. And I don't think it has much momentum. And so the Hybrid Cloud Vendors they've had time to figure it out. But now it's game on, companies like Dell they're promising a consistent experience between On-Prem into the Cloud, across Clouds and out to the Edge. They call it MultCloud which by the way my view has really been multi-vendor Chuck, Chuck Whitten. Who's the new co-COO of Dell called it Multi-Cloud by default. (laughing) That's really, I think an accurate description of that. I call this new world Super Cloud. To me, it's different than MultiCloud. It's a layer that runs on top of hyperscale infrastructure kind of hides the underlying complexity of the Cloud. It's APIs, it's primitives. And it stretches not only across Clouds but out to the Edge. That's a big vision and that's going to require some seriously intense engineering to build out. It's also going to require partnerships that go beyond the portfolios of companies like Dell like their own proprietary stacks if you will. It's going to have to replicate the Cloud Operating Model and to do that, you're going to need more and more deals like Snowflake and even deeper than Snowflake, not just in database. Sure, you'll need to have a catalog of databases that run in your On-Prem and Hybrid and Super Cloud but also other services that customers can tap. I mean, can you imagine a day when Dell offers and embraces a directly competitive service inside of apex. I have trouble envisioning that, you know not with their historical posture, you think about companies like, you know, Nutanix, you know, or Cisco where they really, you know those relationships cooled quite quickly but you know, look, think about it. That's what AWS does. It offers for instance, Redshift and Snowflake side by side happily and the Redshift guys they probably hate Snowflake. I wouldn't blame them, but the EC Two Folks, they love them. And Adam SloopesKy understands that ISVs like Snowflake are a key part of the Cloud ecosystem. Again, I have a hard time envisioning that occurring with Dell or even HPE, you know maybe less so with HPE, but what does this imply that the Edge will allow companies like Dell to a reach around on the Cloud and somehow create a new type of model that begrudgingly accommodates the Public Cloud but drafts of the new momentum of the Edge, which right now to these companies is kind of mostly telco and retail. It's hard to see that happening. I think it's got to evolve in a more comprehensive and inclusive fashion. What's much more likely is companies like Dell are going to substantially replicate that Cloud Operating Model for the pieces that they own pieces that they control which admittedly are big pieces of the market. But unless they're able to really tap that ecosystem magic they're not going to be able to grow much beyond their existing install bases. You take that lime green we showed you earlier that new adoption metric from ETR as an example, by my estimates, AWS and Azure are capturing new accounts at a rate between three to five times faster than Dell and HPE. And in the more mature US and mere markets it's probably more like 10 X and a major reason is because of the Cloud's robust ecosystem and the optionality and simplicity of transaction that that is bringing to customers. Now, Dell for its part is a hundred billion dollar revenue company. And it has the capability to drive that kind of dynamic. If it can pivot its partner ecosystem mindset from kind of resellers to Cloud services and technology optionality. Okay, that's it for now? Thanks to my colleagues, Stephanie Chan who helped research topics for Breaking Analysis. Alex Myerson is on the production team. Kristen Martin and Cheryl Knight and Rob Hof, on editorial they helped get the word out and thanks to Jordan Anderson for the new Breaking Analysis branding and graphics package. Remember these episodes are all available as podcasts wherever you listen. All you do is search Breaking Analysis podcasts. You could check out ETR website @etr.ai. We publish a full report every week on wikibon.com and siliconangle.com. You want to get in touch. @dave.vellente @siliconangle.com. You can DM me @dvellante. You can make a comment on our LinkedIn posts. This is Dave Vellante for the Cube Insights powered by ETR. Have a great week, stay safe, be well. And we'll see you next time. (upbeat music)

Published Date : May 7 2022

SUMMARY :

bringing you Data Driven and the amount of data that lives On-Prem and the value that we bring. One of the more buzzy into the Dell On-Premises world, Now and in the Ford I would say that At the time, but if you And it has the capability to

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jordan AndersonPERSON

0.99+

Stephanie ChanPERSON

0.99+

IBMORGANIZATION

0.99+

DellORGANIZATION

0.99+

Clark PattersonPERSON

0.99+

Alex MyersonPERSON

0.99+

Dave VellantePERSON

0.99+

AWSORGANIZATION

0.99+

Rob HofPERSON

0.99+

LenovoORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

JohnPERSON

0.99+

MicrosoftORGANIZATION

0.99+

John FurrierPERSON

0.99+

OracleORGANIZATION

0.99+

2015DATE

0.99+

GoogleORGANIZATION

0.99+

Cheryl KnightPERSON

0.99+

ClarkPERSON

0.99+

HPORGANIZATION

0.99+

Palo AltoLOCATION

0.99+

BostonLOCATION

0.99+

HPEORGANIZATION

0.99+

6%QUANTITY

0.99+

FordORGANIZATION

0.99+

threeQUANTITY

0.99+

40%QUANTITY

0.99+

Chuck WhittenPERSON

0.99+

VMwareORGANIZATION

0.99+

NutanixORGANIZATION

0.99+

Kristen MartinPERSON

0.99+

FerrariORGANIZATION

0.99+

Adam SloopesKyPERSON

0.99+

EarthLOCATION

0.99+

13 straight surveysQUANTITY

0.99+

70%QUANTITY

0.99+

firstQUANTITY

0.99+

68QUANTITY

0.99+

last quarterDATE

0.99+

RedshiftTITLE

0.99+

siliconangle.comOTHER

0.99+

theCUBE StudiosORGANIZATION

0.99+

SnowflakeEVENT

0.99+

SnowflakeTITLE

0.99+

8,000 peopleQUANTITY

0.99+

bothQUANTITY

0.99+

20QUANTITY

0.99+

VCFORGANIZATION

0.99+

SnowflakeORGANIZATION

0.99+

Analyst Power Panel: Future of Database Platforms


 

(upbeat music) >> Once a staid and boring business dominated by IBM, Oracle, and at the time newcomer Microsoft, along with a handful of wannabes, the database business has exploded in the past decade and has become a staple of financial excellence, customer experience, analytic advantage, competitive strategy, growth initiatives, visualizations, not to mention compliance, security, privacy and dozens of other important use cases and initiatives. And on the vendor's side of the house, we've seen the rapid ascendancy of cloud databases. Most notably from Snowflake, whose massive raises leading up to its IPO in late 2020 sparked a spate of interest and VC investment in the separation of compute and storage and all that elastic resource stuff in the cloud. The company joined AWS, Azure and Google to popularize cloud databases, which have become a linchpin of competitive strategies for technology suppliers. And if I get you to put your data in my database and in my cloud, and I keep innovating, I'm going to build a moat and achieve a hugely attractive lifetime customer value in a really amazing marginal economics dynamic that is going to fund my future. And I'll be able to sell other adjacent services, not just compute and storage, but machine learning and inference and training and all kinds of stuff, dozens of lucrative cloud offerings. Meanwhile, the database leader, Oracle has invested massive amounts of money to maintain its lead. It's building on its position as the king of mission critical workloads and making typical Oracle like claims against the competition. Most were recently just yesterday with another announcement around MySQL HeatWave. An extension of MySQL that is compatible with on-premises MySQLs and is setting new standards in price performance. We're seeing a dramatic divergence in strategies across the database spectrum. On the far left, we see Amazon with more than a dozen database offerings each with its own API and primitives. AWS is taking a right tool for the right job approach, often building on open source platforms and creating services that it offers to customers to solve very specific problems for developers. And on the other side of the line, we see Oracle, which is taking the Swiss Army Knife approach, converging database functionality, enabling analytic and transactional workloads to run in the same data store, eliminating the need to ETL, at the same time adding capabilities into its platform like automation and machine learning. Welcome to this database Power Panel. My name is Dave Vellante, and I'm so excited to bring together some of the most respected industry analyst in the community. Today we're going to assess what's happening in the market. We're going to dig into the competitive landscape and explore the future of database and database platforms and decode what it means to customers. Let me take a moment to welcome our guest analyst today. Matt Kimball is a vice president and principal analysts at Moor Insights and Strategy, Matt. He knows products, he knows industry, he's got real world IT expertise, and he's got all the angles 25 plus years of experience in all kinds of great background. Matt, welcome. Thanks very much for coming on theCUBE. Holgar Mueller, friend of theCUBE, vice president and principal analyst at Constellation Research in depth knowledge on applications, application development, knows developers. He's worked at SAP and Oracle. And then Bob Evans is Chief Content Officer and co-founder of the Acceleration Economy, founder and principle of Cloud Wars. Covers all kinds of industry topics and great insights. He's got awesome videos, these three minute hits. If you haven't seen 'em, checking them out, knows cloud companies, his Cloud Wars minutes are fantastic. And then of course, Marc Staimer is the founder of Dragon Slayer Research. A frequent contributor and guest analyst at Wikibon. He's got a wide ranging knowledge across IT products, knows technology really well, can go deep. And then of course, Ron Westfall, Senior Analyst and Director Research Director at Futurum Research, great all around product trends knowledge. Can take, you know, technical dives and really understands competitive angles, knows Redshift, Snowflake, and many others. Gents, thanks so much for taking the time to join us in theCube today. It's great to have you on, good to see you. >> Good to be here, thanks for having us. >> Thanks, Dave. >> All right, let's start with an around the horn and briefly, if each of you would describe, you know, anything I missed in your areas of expertise and then you answer the following question, how would you describe the state of the database, state of platform market today? Matt Kimball, please start. >> Oh, I hate going first, but that it's okay. How would I describe the world today? I would just in one sentence, I would say, I'm glad I'm not in IT anymore, right? So, you know, it is a complex and dangerous world out there. And I don't envy IT folks I'd have to support, you know, these modernization and transformation efforts that are going on within the enterprise. It used to be, you mentioned it, Dave, you would argue about IBM versus Oracle versus this newcomer in the database space called Microsoft. And don't forget Sybase back in the day, but you know, now it's not just, which SQL vendor am I going to go with? It's all of these different, divergent data types that have to be taken, they have to be merged together, synthesized. And somehow I have to do that cleanly and use this to drive strategic decisions for my business. That is not easy. So, you know, you have to look at it from the perspective of the business user. It's great for them because as a DevOps person, or as an analyst, I have so much flexibility and I have this thing called the cloud now where I can go get services immediately. As an IT person or a DBA, I am calling up prevention hotlines 24 hours a day, because I don't know how I'm going to be able to support the business. And as an Oracle or as an Oracle or a Microsoft or some of the cloud providers and cloud databases out there, I'm licking my chops because, you know, my market is expanding and expanding every day. >> Great, thank you for that, Matt. Holgar, how do you see the world these days? You always have a good perspective on things, share with us. >> Well, I think it's the best time to be in IT, I'm not sure what Matt is talking about. (laughing) It's easier than ever, right? The direction is going to cloud. Kubernetes has won, Google has the best AI for now, right? So things are easier than ever before. You made commitments for five plus years on hardware, networking and so on premise, and I got gray hair about worrying it was the wrong decision. No, just kidding. But you kind of both sides, just to be controversial, make it interesting, right. So yeah, no, I think the interesting thing specifically with databases, right? We have this big suite versus best of breed, right? Obviously innovation, like you mentioned with Snowflake and others happening in the cloud, the cloud vendors server, where to save of their databases. And then we have one of the few survivors of the old guard as Evans likes to call them is Oracle who's doing well, both their traditional database. And now, which is really interesting, remarkable from that because Oracle it was always the power of one, have one database, add more to it, make it what I call the universal database. And now this new HeatWave offering is coming and MySQL open source side. So they're getting the second (indistinct) right? So it's interesting that older players, traditional players who still are in the market are diversifying their offerings. Something we don't see so much from the traditional tools from Oracle on the Microsoft side or the IBM side these days. >> Great, thank you Holgar. Bob Evans, you've covered this business for a while. You've worked at, you know, a number of different outlets and companies and you cover the competition, how do you see things? >> Dave, you know, the other angle to look at this from is from the customer side, right? You got now CEOs who are any sort of business across all sorts of industries, and they understand that their future success is going to be dependent on their ability to become a digital company, to understand data, to use it the right way. So as you outline Dave, I think in your intro there, it is a fantastic time to be in the database business. And I think we've got a lot of new buyers and influencers coming in. They don't know all this history about IBM and Microsoft and Oracle and you know, whoever else. So I think they're going to take a long, hard look, Dave, at some of these results and who is able to help these companies not serve up the best technology, but who's going to be able to help their business move into the digital future. So it's a fascinating time now from every perspective. >> Great points, Bob. I mean, digital transformation has gone from buzzword to imperative. Mr. Staimer, how do you see things? >> I see things a little bit differently than my peers here in that I see the database market being segmented. There's all the different kinds of databases that people are looking at for different kinds of data, and then there is databases in the cloud. And so database as cloud service, I view very differently than databases because the traditional way of implementing a database is changing and it's changing rapidly. So one of the premises that you stated earlier on was that you viewed Oracle as a database company. I don't view Oracle as a database company anymore. I view Oracle as a cloud company that happens to have a significant expertise and specialty in databases, and they still sell database software in the traditional way, but ultimately they're a cloud company. So database cloud services from my point of view is a very distinct market from databases. >> Okay, well, you gave us some good meat on the bone to talk about that. Last but not least-- >> Dave did Marc, just say Oracle's a cloud company? >> Yeah. (laughing) Take away the database, it would be interesting to have that discussion, but let's let Ron jump in here. Ron, give us your take. >> That's a great segue. I think it's truly the era of the cloud database, that's something that's rising. And the key trends that come with it include for example, elastic scaling. That is the ability to scale on demand, to right size workloads according to customer requirements. And also I think it's going to increase the prioritization for high availability. That is the player who can provide the highest availability is going to have, I think, a great deal of success in this emerging market. And also I anticipate that there will be more consolidation across platforms in order to enable cost savings for customers, and that's something that's always going to be important. And I think we'll see more of that over the horizon. And then finally security, security will be more important than ever. We've seen a spike (indistinct), we certainly have seen geopolitical originated cybersecurity concerns. And as a result, I see database security becoming all the more important. >> Great, thank you. Okay, let me share some data with you guys. I'm going to throw this at you and see what you think. We have this awesome data partner called Enterprise Technology Research, ETR. They do these quarterly surveys and each period with dozens of industry segments, they track clients spending, customer spending. And this is the database, data warehouse sector okay so it's taxonomy, so it's not perfect, but it's a big kind of chunk. They essentially ask customers within a category and buy a specific vendor, you're spending more or less on the platform? And then they subtract the lesses from the mores and they derive a metric called net score. It's like NPS, it's a measure of spending velocity. It's more complicated and granular than that, but that's the basis and that's the vertical axis. The horizontal axis is what they call market share, it's not like IDC market share, it's just pervasiveness in the data set. And so there are a couple of things that stand out here and that we can use as reference point. The first is the momentum of Snowflake. They've been off the charts for many, many, for over two years now, anything above that dotted red line, that 40%, is considered by ETR to be highly elevated and Snowflake's even way above that. And I think it's probably not sustainable. We're going to see in the next April survey, next month from those guys, when it comes out. And then you see AWS and Microsoft, they're really pervasive on the horizontal axis and highly elevated, Google falls behind them. And then you got a number of well funded players. You got Cockroach Labs, Mongo, Redis, MariaDB, which of course is a fork on MySQL started almost as protest at Oracle when they acquired Sun and they got MySQL and you can see the number of others. Now Oracle who's the leading database player, despite what Marc Staimer says, we know, (laughs) and they're a cloud player (laughing) who happens to be a leading database player. They dominate in the mission critical space, we know that they're the king of that sector, but you can see here that they're kind of legacy, right? They've been around a long time, they get a big install base. So they don't have the spending momentum on the vertical axis. Now remember this is, just really this doesn't capture spending levels, so that understates Oracle but nonetheless. So it's not a complete picture like SAP for instance is not in here, no Hana. I think people are actually buying it, but it doesn't show up here, (laughs) but it does give an indication of momentum and presence. So Bob Evans, I'm going to start with you. You've commented on many of these companies, you know, what does this data tell you? >> Yeah, you know, Dave, I think all these compilations of things like that are interesting, and that folks at ETR do some good work, but I think as you said, it's a snapshot sort of a two-dimensional thing of a rapidly changing, three dimensional world. You know, the incidents at which some of these companies are mentioned versus the volume that happens. I think it's, you know, with Oracle and I'm not going to declare my religious affiliation, either as cloud company or database company, you know, they're all of those things and more, and I think some of our old language of how we classify companies is just not relevant anymore. But I want to ask too something in here, the autonomous database from Oracle, nobody else has done that. So either Oracle is crazy, they've tried out a technology that nobody other than them is interested in, or they're onto something that nobody else can match. So to me, Dave, within Oracle, trying to identify how they're doing there, I would watch autonomous database growth too, because right, it's either going to be a big plan and it breaks through, or it's going to be caught behind. And the Snowflake phenomenon as you mentioned, that is a rare, rare bird who comes up and can grow 100% at a billion dollar revenue level like that. So now they've had a chance to come in, scare the crap out of everybody, rock the market with something totally new, the data cloud. Will the bigger companies be able to catch up and offer a compelling alternative, or is Snowflake going to continue to be this outlier. It's a fascinating time. >> Really, interesting points there. Holgar, I want to ask you, I mean, I've talked to certainly I'm sure you guys have too, the founders of Snowflake that came out of Oracle and they actually, they don't apologize. They say, "Hey, we not going to do all that complicated stuff that Oracle does, we were trying to keep it real simple." But at the same time, you know, they don't do sophisticated workload management. They don't do complex joints. They're kind of relying on the ecosystems. So when you look at the data like this and the various momentums, and we talked about the diverging strategies, what does this say to you? >> Well, it is a great point. And I think Snowflake is an example how the cloud can turbo charge a well understood concept in this case, the data warehouse, right? You move that and you find steroids and you see like for some players who've been big in data warehouse, like Sentara Data, as an example, here in San Diego, what could have been for them right in that part. The interesting thing, the problem though is the cloud hides a lot of complexity too, which you can scale really well as you attract lots of customers to go there. And you don't have to build things like what Bob said, right? One of the fascinating things, right, nobody's answering Oracle on the autonomous database. I don't think is that they cannot, they just have different priorities or the database is not such a priority. I would dare to say that it's for IBM and Microsoft right now at the moment. And the cloud vendors, you just hide that right through scripts and through scale because you support thousands of customers and you can deal with a little more complexity, right? It's not against them. Whereas if you have to run it yourself, very different story, right? You want to have the autonomous parts, you want to have the powerful tools to do things. >> Thank you. And so Matt, I want to go to you, you've set up front, you know, it's just complicated if you're in IT, it's a complicated situation and you've been on the customer side. And if you're a buyer, it's obviously, it's like Holgar said, "Cloud's supposed to make this stuff easier, but the simpler it gets the more complicated gets." So where do you place your bets? Or I guess more importantly, how do you decide where to place your bets? >> Yeah, it's a good question. And to what Bob and Holgar said, you know, the around autonomous database, I think, you know, part of, as I, you know, play kind of armchair psychologist, if you will, corporate psychologists, I look at what Oracle is doing and, you know, databases where they've made their mark and it's kind of, that's their strong position, right? So it makes sense if you're making an entry into this cloud and you really want to kind of build momentum, you go with what you're good at, right? So that's kind of the strength of Oracle. Let's put a lot of focus on that. They do a lot more than database, don't get me wrong, but you know, I'm going to short my strength and then kind of pivot from there. With regards to, you know, what IT looks at and what I would look at you know as an IT director or somebody who is, you know, trying to consume services from these different cloud providers. First and foremost, I go with what I know, right? Let's not forget IT is a conservative group. And when we look at, you know, all the different permutations of database types out there, SQL, NoSQL, all the different types of NoSQL, those are largely being deployed by business users that are looking for agility or businesses that are looking for agility. You know, the reason why MongoDB is so popular is because of DevOps, right? It's a great platform to develop on and that's where it kind of gained its traction. But as an IT person, I want to go with what I know, where my muscle memory is, and that's my first position. And so as I evaluate different cloud service providers and cloud databases, I look for, you know, what I know and what I've invested in and where my muscle memory is. Is there enough there and do I have enough belief that that company or that service is going to be able to take me to, you know, where I see my organization in five years from a data management perspective, from a business perspective, are they going to be there? And if they are, then I'm a little bit more willing to make that investment, but it is, you know, if I'm kind of going in this blind or if I'm cloud native, you know, that's where the Snowflakes of the world become very attractive to me. >> Thank you. So Marc, I asked Andy Jackson in theCube one time, you have all these, you know, data stores and different APIs and primitives and you know, very granular, what's the strategy there? And he said, "Hey, that allows us as the market changes, it allows us to be more flexible. If we start building abstractions layers, it's harder for us." I think also it was not a good time to market advantage, but let me ask you, I described earlier on that spectrum from AWS to Oracle. We just saw yesterday, Oracle announced, I think the third major enhancement in like 15 months to MySQL HeatWave, what do you make of that announcement? How do you think it impacts the competitive landscape, particularly as it relates to, you know, converging transaction and analytics, eliminating ELT, I know you have some thoughts on this. >> So let me back up for a second and defend my cloud statement about Oracle for a moment. (laughing) AWS did a great job in developing the cloud market in general and everything in the cloud market. I mean, I give them lots of kudos on that. And a lot of what they did is they took open source software and they rent it to people who use their cloud. So I give 'em lots of credit, they dominate the market. Oracle was late to the cloud market. In fact, they actually poo-pooed it initially, if you look at some of Larry Ellison's statements, they said, "Oh, it's never going to take off." And then they did 180 turn, and they said, "Oh, we're going to embrace the cloud." And they really have, but when you're late to a market, you've got to be compelling. And this ties into the announcement yesterday, but let's deal with this compelling. To be compelling from a user point of view, you got to be twice as fast, offer twice as much functionality, at half the cost. That's generally what compelling is that you're going to capture market share from the leaders who established the market. It's very difficult to capture market share in a new market for yourself. And you're right. I mean, Bob was correct on this and Holgar and Matt in which you look at Oracle, and they did a great job of leveraging their database to move into this market, give 'em lots of kudos for that too. But yesterday they announced, as you said, the third innovation release and the pace is just amazing of what they're doing on these releases on HeatWave that ties together initially MySQL with an integrated builtin analytics engine, so a data warehouse built in. And then they added automation with autopilot, and now they've added machine learning to it, and it's all in the same service. It's not something you can buy and put on your premise unless you buy their cloud customers stuff. But generally it's a cloud offering, so it's compellingly better as far as the integration. You don't buy multiple services, you buy one and it's lower cost than any of the other services, but more importantly, it's faster, which again, give 'em credit for, they have more integration of a product. They can tie things together in a way that nobody else does. There's no additional services, ETL services like Glue and AWS. So from that perspective, they're getting better performance, fewer services, lower cost. Hmm, they're aiming at the compelling side again. So from a customer point of view it's compelling. Matt, you wanted to say something there. >> Yeah, I want to kind of, on what you just said there Marc, and this is something I've found really interesting, you know. The traditional way that you look at software and, you know, purchasing software and IT is, you look at either best of breed solutions and you have to work on the backend to integrate them all and make them all work well. And generally, you know, the big hit against the, you know, we have one integrated offering is that, you lose capability or you lose depth of features, right. And to what you were saying, you know, that's the thing I found interesting about what Oracle is doing is they're building in depth as they kind of, you know, build that service. It's not like you're losing a lot of capabilities, because you're going to one integrated service versus having to use A versus B versus C, and I love that idea. >> You're right. Yeah, not only you're not losing, but you're gaining functionality that you can't get by integrating a lot of these. I mean, I can take Snowflake and integrate it in with machine learning, but I also have to integrate in with a transactional database. So I've got to have connectors between all of this, which means I'm adding time. And what it comes down to at the end of the day is expertise, effort, time, and cost. And so what I see the difference from the Oracle announcements is they're aiming at reducing all of that by increasing performance as well. Correct me if I'm wrong on that but that's what I saw at the announcement yesterday. >> You know, Marc, one thing though Marc, it's funny you say that because I started out saying, you know, I'm glad I'm not 19 anymore. And the reason is because of exactly what you said, it's almost like there's a pseudo level of witchcraft that's required to support the modern data environment right in the enterprise. And I need simpler faster, better. That's what I need, you know, I am no longer wearing pocket protectors. I have turned from, you know, break, fix kind of person, to you know, business consultant. And I need that point and click simplicity, but I can't sacrifice, you know, a depth of features of functionality on the backend as I play that consultancy role. >> So, Ron, I want to bring in Ron, you know, it's funny. So Matt, you mentioned Mongo, I often and say, if Oracle mentions you, you're on the map. We saw them yesterday Ron, (laughing) they hammered RedShifts auto ML, they took swipes at Snowflake, a little bit of BigQuery. What were your thoughts on that? Do you agree with what these guys are saying in terms of HeatWaves capabilities? >> Yes, Dave, I think that's an excellent question. And fundamentally I do agree. And the question is why, and I think it's important to know that all of the Oracle data is backed by the fact that they're using benchmarks. For example, all of the ML and all of the TPC benchmarks, including all the scripts, all the configs and all the detail are posted on GitHub. So anybody can look at these results and they're fully transparent and replicate themselves. If you don't agree with this data, then by all means challenge it. And we have not really seen that in all of the new updates in HeatWave over the last 15 months. And as a result, when it comes to these, you know, fundamentals in looking at the competitive landscape, which I think gives validity to outcomes such as Oracle being able to deliver 4.8 times better price performance than Redshift. As well as for example, 14.4 better price performance than Snowflake, and also 12.9 better price performance than BigQuery. And so that is, you know, looking at the quantitative side of things. But again, I think, you know, to Marc's point and to Matt's point, there are also qualitative aspects that clearly differentiate the Oracle proposition, from my perspective. For example now the MySQL HeatWave ML capabilities are native, they're built in, and they also support things such as completion criteria. And as a result, that enables them to show that hey, when you're using Redshift ML for example, you're having to also use their SageMaker tool and it's running on a meter. And so, you know, nobody really wants to be running on a meter when, you know, executing these incredibly complex tasks. And likewise, when it comes to Snowflake, they have to use a third party capability. They don't have the built in, it's not native. So the user, to the point that he's having to spend more time and it increases complexity to use auto ML capabilities across the Snowflake platform. And also, I think it also applies to other important features such as data sampling, for example, with the HeatWave ML, it's intelligent sampling that's being implemented. Whereas in contrast, we're seeing Redshift using random sampling. And again, Snowflake, you're having to use a third party library in order to achieve the same capabilities. So I think the differentiation is crystal clear. I think it definitely is refreshing. It's showing that this is where true value can be assigned. And if you don't agree with it, by all means challenge the data. >> Yeah, I want to come to the benchmarks in a minute. By the way, you know, the gentleman who's the Oracle's architect, he did a great job on the call yesterday explaining what you have to do. I thought that was quite impressive. But Bob, I know you follow the financials pretty closely and on the earnings call earlier this month, Ellison said that, "We're going to see HeatWave on AWS." And the skeptic in me said, oh, they must not be getting people to come to OCI. And then they, you remember this chart they showed yesterday that showed the growth of HeatWave on OCI. But of course there was no data on there, it was just sort of, you know, lines up and to the right. So what do you guys think of that? (Marc laughs) Does it signal Bob, desperation by Oracle that they can't get traction on OCI, or is it just really a smart tame expansion move? What do you think? >> Yeah, Dave, that's a great question. You know, along the way there, and you know, just inside of that was something that said Ellison said on earnings call that spoke to a different sort of philosophy or mindset, almost Marc, where he said, "We're going to make this multicloud," right? With a lot of their other cloud stuff, if you wanted to use any of Oracle's cloud software, you had to use Oracle's infrastructure, OCI, there was no other way out of it. But this one, but I thought it was a classic Ellison line. He said, "Well, we're making this available on AWS. We're making this available, you know, on Snowflake because we're going after those users. And once they see what can be done here." So he's looking at it, I guess you could say, it's a concession to customers because they want multi-cloud. The other way to look at it, it's a hunting expedition and it's one of those uniquely I think Oracle ways. He said up front, right, he doesn't say, "Well, there's a big market, there's a lot for everybody, we just want on our slice." Said, "No, we are going after Amazon, we're going after Redshift, we're going after Aurora. We're going after these users of Snowflake and so on." And I think it's really fairly refreshing these days to hear somebody say that, because now if I'm a buyer, I can look at that and say, you know, to Marc's point, "Do they measure up, do they crack that threshold ceiling? Or is this just going to be more pain than a few dollars savings is worth?" But you look at those numbers that Ron pointed out and that we all saw in that chart. I've never seen Dave, anything like that. In a substantive market, a new player coming in here, and being able to establish differences that are four, seven, eight, 10, 12 times better than competition. And as new buyers look at that, they're going to say, "What the hell are we doing paying, you know, five times more to get a poor result? What's going on here?" So I think this is going to rattle people and force a harder, closer look at what these alternatives are. >> I wonder if the guy, thank you. Let's just skip ahead of the benchmarks guys, bring up the next slide, let's skip ahead a little bit here, which talks to the benchmarks and the benchmarking if we can. You know, David Floyer, the sort of semiretired, you know, Wikibon analyst said, "Dave, this is going to force Amazon and others, Snowflake," he said, "To rethink actually how they architect databases." And this is kind of a compilation of some of the data that they shared. They went after Redshift mostly, (laughs) but also, you know, as I say, Snowflake, BigQuery. And, like I said, you can always tell which companies are doing well, 'cause Oracle will come after you, but they're on the radar here. (laughing) Holgar should we take this stuff seriously? I mean, or is it, you know, a grain salt? What are your thoughts here? >> I think you have to take it seriously. I mean, that's a great question, great point on that. Because like Ron said, "If there's a flaw in a benchmark, we know this database traditionally, right?" If anybody came up that, everybody will be, "Oh, you put the wrong benchmark, it wasn't audited right, let us do it again," and so on. We don't see this happening, right? So kudos to Oracle to be aggressive, differentiated, and seem to having impeccable benchmarks. But what we really see, I think in my view is that the classic and we can talk about this in 100 years, right? Is the suite versus best of breed, right? And the key question of the suite, because the suite's always slower, right? No matter at which level of the stack, you have the suite, then the best of breed that will come up with something new, use a cloud, put the data warehouse on steroids and so on. The important thing is that you have to assess as a buyer what is the speed of my suite vendor. And that's what you guys mentioned before as well, right? Marc said that and so on, "Like, this is a third release in one year of the HeatWave team, right?" So everybody in the database open source Marc, and there's so many MySQL spinoffs to certain point is put on shine on the speed of (indistinct) team, putting out fundamental changes. And the beauty of that is right, is so inherent to the Oracle value proposition. Larry's vision of building the IBM of the 21st century, right from the Silicon, from the chip all the way across the seven stacks to the click of the user. And that what makes the database what Rob was saying, "Tied to the OCI infrastructure," because designed for that, it runs uniquely better for that, that's why we see the cross connect to Microsoft. HeatWave so it's different, right? Because HeatWave runs on cheap hardware, right? Which is the breadth and butter 886 scale of any cloud provider, right? So Oracle probably needs it to scale OCI in a different category, not the expensive side, but also allow us to do what we said before, the multicloud capability, which ultimately CIOs really want, because data gravity is real, you want to operate where that is. If you have a fast, innovative offering, which gives you more functionality and the R and D speed is really impressive for the space, puts away bad results, then it's a good bet to look at. >> Yeah, so you're saying, that we versus best of breed. I just want to sort of play back then Marc a comment. That suite versus best of breed, there's always been that trade off. If I understand you Holgar you're saying that somehow Oracle has magically cut through that trade off and they're giving you the best of both. >> It's the developing velocity, right? The provision of important features, which matter to buyers of the suite vendor, eclipses the best of breed vendor, then the best of breed vendor is in the hell of a potential job. >> Yeah, go ahead Marc. >> Yeah and I want to add on what Holgar just said there. I mean the worst job in the data center is data movement, moving the data sucks. I don't care who you are, nobody likes it. You never get any kudos for doing it well, and you always get the ah craps, when things go wrong. So it's in- >> In the data center Marc all the time across data centers, across cloud. That's where the bleeding comes. >> It's right, you get beat up all the time. So nobody likes to move data, ever. So what you're looking at with what they announce with HeatWave and what I love about HeatWave is it doesn't matter when you started with it, you get all the additional features they announce it's part of the service, all the time. But they don't have to move any of the data. You want to analyze the data that's in your transactional, MySQL database, it's there. You want to do machine learning models, it's there, there's no data movement. The data movement is the key thing, and they just eliminate that, in so many ways. And the other thing I wanted to talk about is on the benchmarks. As great as those benchmarks are, they're really conservative 'cause they're underestimating the cost of that data movement. The ETLs, the other services, everything's left out. It's just comparing HeatWave, MySQL cloud service with HeatWave versus Redshift, not Redshift and Aurora and Glue, Redshift and Redshift ML and SageMaker, it's just Redshift. >> Yeah, so what you're saying is what Oracle's doing is saying, "Okay, we're going to run MySQL HeatWave benchmarks on analytics against Redshift, and then we're going to run 'em in transaction against Aurora." >> Right. >> But if you really had to look at what you would have to do with the ETL, you'd have to buy two different data stores and all the infrastructure around that, and that goes away so. >> Due to the nature of the competition, they're running narrow best of breed benchmarks. There is no suite level benchmark (Dave laughs) because they created something new. >> Well that's you're the earlier point they're beating best of breed with a suite. So that's, I guess to Floyer's earlier point, "That's going to shake things up." But I want to come back to Bob Evans, 'cause I want to tap your Cloud Wars mojo before we wrap. And line up the horses, you got AWS, you got Microsoft, Google and Oracle. Now they all own their own cloud. Snowflake, Mongo, Couchbase, Redis, Cockroach by the way they're all doing very well. They run in the cloud as do many others. I think you guys all saw the Andreessen, you know, commentary from Sarah Wang and company, to talk about the cost of goods sold impact of cloud. So owning your own cloud has to be an advantage because other guys like Snowflake have to pay cloud vendors and negotiate down versus having the whole enchilada, Safra Catz's dream. Bob, how do you think this is going to impact the market long term? >> Well, Dave, that's a great question about, you know, how this is all going to play out. If I could mention three things, one, Frank Slootman has done a fantastic job with Snowflake. Really good company before he got there, but since he's been there, the growth mindset, the discipline, the rigor and the phenomenon of what Snowflake has done has forced all these bigger companies to really accelerate what they're doing. And again, it's an example of how this intense competition makes all the different cloud vendors better and it provides enormous value to customers. Second thing I wanted to mention here was look at the Adam Selipsky effect at AWS, took over in the middle of May, and in Q2, Q3, Q4, AWS's growth rate accelerated. And in each of those three quotas, they grew faster than Microsoft's cloud, which has not happened in two or three years, so they're closing the gap on Microsoft. The third thing, Dave, in this, you know, incredibly intense competitive nature here, look at Larry Ellison, right? He's got his, you know, the product that for the last two or three years, he said, "It's going to help determine the future of the company, autonomous database." You would think he's the last person in the world who's going to bring in, you know, in some ways another database to think about there, but he has put, you know, his whole effort and energy behind this. The investments Oracle's made, he's riding this horse really hard. So it's not just a technology achievement, but it's also an investment priority for Oracle going forward. And I think it's going to form a lot of how they position themselves to this new breed of buyer with a new type of need and expectations from IT. So I just think the next two or three years are going to be fantastic for people who are lucky enough to get to do the sorts of things that we do. >> You know, it's a great point you made about AWS. Back in 2018 Q3, they were doing about 7.4 billion a quarter and they were growing in the mid forties. They dropped down to like 29% Q4, 2020, I'm looking at the data now. They popped back up last quarter, last reported quarter to 40%, that is 17.8 billion, so they more doubled and they accelerated their growth rate. (laughs) So maybe that pretends, people are concerned about Snowflake right now decelerating growth. You know, maybe that's going to be different. By the way, I think Snowflake has a different strategy, the whole data cloud thing, data sharing. They're not trying to necessarily take Oracle head on, which is going to make this next 10 years, really interesting. All right, we got to go, last question. 30 seconds or less, what can we expect from the future of data platforms? Matt, please start. >> I have to go first again? You're killing me, Dave. (laughing) In the next few years, I think you're going to see the major players continue to meet customers where they are, right. Every organization, every environment is, you know, kind of, we use these words bespoke in Snowflake, pardon the pun, but Snowflakes, right. But you know, they're all opinionated and unique and what's great as an IT person is, you know, there is a service for me regardless of where I am on my journey, in my data management journey. I think you're going to continue to see with regards specifically to Oracle, I think you're going to see the company continue along this path of being all things to all people, if you will, or all organizations without sacrificing, you know, kind of richness of features and sacrificing who they are, right. Look, they are the data kings, right? I mean, they've been a database leader for an awful long time. I don't see that going away any time soon and I love the innovative spirit they've brought in with HeatWave. >> All right, great thank you. Okay, 30 seconds, Holgar go. >> Yeah, I mean, the interesting thing that we see is really that trend to autonomous as Oracle calls or self-driving software, right? So the database will have to do more things than just store the data and support the DVA. It will have to show it can wide insights, the whole upside, it will be able to show to one machine learning. We haven't really talked about that. How in just exciting what kind of use case we can get of machine learning running real time on data as it changes, right? So, which is part of the E5 announcement, right? So we'll see more of that self-driving nature in the database space. And because you said we can promote it, right. Check out my report about HeatWave latest release where I post in oracle.com. >> Great, thank you for that. And Bob Evans, please. You're great at quick hits, hit us. >> Dave, thanks. I really enjoyed getting to hear everybody's opinion here today and I think what's going to happen too. I think there's a new generation of buyers, a new set of CXO influencers in here. And I think what Oracle's done with this, MySQL HeatWave, those benchmarks that Ron talked about so eloquently here that is going to become something that forces other companies, not just try to get incrementally better. I think we're going to see a massive new wave of innovation to try to play catch up. So I really take my hat off to Oracle's achievement from going to, push everybody to be better. >> Excellent. Marc Staimer, what do you say? >> Sure, I'm going to leverage off of something Matt said earlier, "Those companies that are going to develop faster, cheaper, simpler products that are going to solve customer problems, IT problems are the ones that are going to succeed, or the ones who are going to grow. The one who are just focused on the technology are going to fall by the wayside." So those who can solve more problems, do it more elegantly and do it for less money are going to do great. So Oracle's going down that path today, Snowflake's going down that path. They're trying to do more integration with third party, but as a result, aiming at that simpler, faster, cheaper mentality is where you're going to continue to see this market go. >> Amen brother Marc. >> Thank you, Ron Westfall, we'll give you the last word, bring us home. >> Well, thank you. And I'm loving it. I see a wave of innovation across the entire cloud database ecosystem and Oracle is fueling it. We are seeing it, with the native integration of auto ML capabilities, elastic scaling, lower entry price points, et cetera. And this is just going to be great news for buyers, but also developers and increased use of open APIs. And so I think that is really the key takeaways. Just we're going to see a lot of great innovation on the horizon here. >> Guys, fantastic insights, one of the best power panel as I've ever done. Love to have you back. Thanks so much for coming on today. >> Great job, Dave, thank you. >> All right, and thank you for watching. This is Dave Vellante for theCube and we'll see you next time. (soft music)

Published Date : Mar 31 2022

SUMMARY :

and co-founder of the and then you answer And don't forget Sybase back in the day, the world these days? and others happening in the cloud, and you cover the competition, and Oracle and you know, whoever else. Mr. Staimer, how do you see things? in that I see the database some good meat on the bone Take away the database, That is the ability to scale on demand, and they got MySQL and you I think it's, you know, and the various momentums, and Microsoft right now at the moment. So where do you place your bets? And to what Bob and Holgar said, you know, and you know, very granular, and everything in the cloud market. And to what you were saying, you know, functionality that you can't get to you know, business consultant. you know, it's funny. and all of the TPC benchmarks, By the way, you know, and you know, just inside of that was of some of the data that they shared. the stack, you have the suite, and they're giving you the best of both. of the suite vendor, and you always get the ah In the data center Marc all the time And the other thing I wanted to talk about and then we're going to run 'em and all the infrastructure around that, Due to the nature of the competition, I think you guys all saw the Andreessen, And I think it's going to form I'm looking at the data now. and I love the innovative All right, great thank you. and support the DVA. Great, thank you for that. And I think what Oracle's done Marc Staimer, what do you say? or the ones who are going to grow. we'll give you the last And this is just going to Love to have you back. and we'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
David FloyerPERSON

0.99+

Dave VellantePERSON

0.99+

Ron WestfallPERSON

0.99+

DavePERSON

0.99+

Marc StaimerPERSON

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

MarcPERSON

0.99+

EllisonPERSON

0.99+

Bob EvansPERSON

0.99+

OracleORGANIZATION

0.99+

MattPERSON

0.99+

Holgar MuellerPERSON

0.99+

AWSORGANIZATION

0.99+

Frank SlootmanPERSON

0.99+

RonPERSON

0.99+

StaimerPERSON

0.99+

Andy JacksonPERSON

0.99+

BobPERSON

0.99+

Matt KimballPERSON

0.99+

GoogleORGANIZATION

0.99+

100%QUANTITY

0.99+

Sarah WangPERSON

0.99+

San DiegoLOCATION

0.99+

AmazonORGANIZATION

0.99+

RobPERSON

0.99+

Keynote Enabling Business and Developer Success | Open Cloud Innovations


 

(upbeat music) >> Hello, and welcome to this startup showcase. It's great to be here and talk about some of the innovations we are doing at AWS, how we work with our partner community, especially our open source partners. My name is Deepak Singh. I run our compute services organization, which is a very vague way of saying that I run a number of things that are connected together through compute. Very specifically, I run a container services organization. So for those of you who are into containers, ECS, EKS, fargate, ECR, App Runner Those are all teams that are within my org. I also run the Amazon Linux and BottleRocketing. So anything AWS does with Linux, both externally and internally, as well as our high-performance computing team. And perhaps very relevant to this discussion, I run the Amazon open source program office. Serving at AWS for over 13 years, almost 14, involved with compute in various ways, including EC2. What that has done has given me a vantage point of seeing how our customers use the services that we build for them, how they leverage various partner solutions, and along the way, how AWS itself has gotten involved with opensource. And I'll try and talk to you about some of those factors and how they impact, how you consume our services. So why don't we get started? So for many of you, you know, one of the things, there's two ways to look at AWS and open-source and Amazon in general. One is the number of contributors you may have. And the number of repositories that contribute to. Those are just a couple of measures. There are people that I work with on a regular basis, who will remind you that, those are not perfect measures. Sometimes you could just contribute to one thing and have outsized impact because of the nature of that thing. But it address being what it is, increasingly we'll look at different ways in which we can help contribute and enhance open source 'cause we consume a lot of it as well. I'll talk about it very specifically from the space that I work in the container space in particular, where we've worked a lot with people in the Kubernetes community. We've worked a lot with people in the broader CNCF community, as well as, you know, small projects that our customers might have got started off with. For example, I want to like talking about is Argo CD from Intuit. We were very actively involved with helping them figure out what to do with it. And it was great to see how into it. And we worked, etc, came together to think about get-ups at the Kubernetes level. And while those are their projects, we've always been involved with them. So we try and figure out what's important to our customers, how we can help and then take because of that. Well, let's talk about a little bit more, here's some examples of the kinds of open source projects that Amazon and AWS contribute to. They arranged from the open JDK. I think we even now have our own implementation of Java, the Corretto open source project. We contribute to projects like rust, where we are very active in the rest foundation from a leadership role as well, the robot operating system, just to pick some, we collaborate with Facebook and actively involved with the pirates project. And there's many others. You can see all the logos in here where we participate either because they're important to us as AWS in the services that we run or they're important to our customers and the services that they consume or the open source projects they care about and how we get to those. How we get and make those decisions is often depends on the importance of that particular project. At that point in time, how much impact they're having to AWS customers, or sometimes very feel that us contributing to that project is super critical because it helps us build more robust services. I'll talk about it in a completely, you know, somewhat different basis. You may have heard of us talk about our new next generation of Amazon Linux 2022, which is based on fedora as its sub stream. One of the reasons we made this decision was it allows us to go and participate in the preneurial project and make sure that the upstream project is robust, stays robust. And that, that what that ends up being is that Amazon Linux 2022 will be a robust operating system with the kinds of capabilities that our customers are asking for. That's just one example of how we think about it. So for example, you know, the Python software foundation is something that we work with very closely because so many of our customers use Python. So we help run something like PyPy which is many, you know, if you're a Python developer, I happened to be a Ruby one, but lots of our customers use Python and helping the Python project be robust by making sure PyPy is available to everybody is something that we help provide credits for help support in other ways. So it's not just code. It can mean many different ways of contributing as well, but in the end code and operations is where we hang our happens. Good examples of this is projects that we will create an open source because it makes sense to make sure that we open source some of the core primitives or foundations that are part of our own services. A great example of that, whether this be things that we open source or things that we contribute to. And I'll talk about both and I'll talk about things near and dear to my heart. There's many examples I've picked the two that I like talking about. The first of these is firecracker. Many of you have heard about it, a firecracker for those of you who don't know is a very lightweight virtual machine manager, which allows you to run these micro VMs. And why was this important many years ago when we started Lambda and quite honestly, Fugate and foggy, it still runs quite a bit in that mode, we used to have to run on VMs like everything else and finding the right VM for the size of tasks that somebody asks for the size of function that somebody asks for is requires us to provision capacity ahead of time. And it also wastes a lot of capacity because Lambda function is small. You won't even if you find the smallest VM possible, those can be a little that can be challenging. And you know, there's a lot of resources that are being wasted. VM start at a particular speed because they have to do a whole bunch of things before the operating system spins up and the virtual machine spins up and we asked ourselves, can we do better? come up with something that allows us to create right size, very lightweight, very fast booting. What's your machines, micro virtual machine that we ended up calling them. That's what led to firecracker. And we open source the project. And today firecrackers use, not just by AWS Lambda or foggy, but by a number of other folks, there's companies like fly IO that are using it. We know people using firecracker to run Kubernetes on prem on bare metal as an example. So we've seen a lot of other folks embrace it and use it as the foundation for building their own serverless services, their own container services. And we think there's a lot of value and learnings that we can bring to the table because we get the experience of operating at scale, but other people can bring to the table cause they may have specific requirements that we may not find it as important from an AWS perspective. So that's firecracker an example of a project where we contribute because we feel it's fundamentally important to us as continually. We were found, you know, we've been involved with continuity from the beginning. Today, we are a whole team that does nothing else, but contribute to container D because container D underlies foggy. It underlies our Kubernetes offerings. And it's increasingly being used by customers directly by their placement. You know, where they're running container D instead of running a full on Docker or similar container engine, what it has allowed us to do is focus on what's important so that we can operate continuously at scale, keep it robust and secure, add capabilities to it that AWS customers need manifested often through foggy Kubernetes, but in the end, it's a win-win for everybody. It makes continuously better. If you want to use containers for yourself on AWS, that's a great way to you. You know, you still, you still benefit from all the work that we're doing. The decision we took was since it's so important to us and our customers, we wanted a team that lived in breathed container D and made sure a super robust and there's many, many examples like that. No, that we ended up participating in, either by taking a project that exists or open sourcing our own. Here's an example of some of the open source projects that we have done from an AWS on Amazon perspective. And there's quite a few when I was looking at this list, I was quite surprised, not quite surprised I've seen the reports before, but every time I do, I have to recount and say, that's a lot more than one would have thought, even though I'd been looking at it for such a long time, examples of this in my world alone are things like, you know, what work had to do with Amazon Linux BottleRocket, which is a container host operating system. That's been open-sourced from day one. Firecracker is something we talked about. We have a project called AWS peril cluster, which allows you to spin up high performance computing clusters on AWS using the kind of schedulers you may use to use like slum. And that's an open source project. We have plenty of source projects in the web development space, in the security space. And more recently things like the open 3d engine, which is something that we are very excited about and that'd be open sourced a few months ago. And so there's a number of these projects that cover everything from tooling to developer, application frameworks, all the way to database and analytics and machine learning. And you'll notice that in a few areas, containers, as an example, machine learning as an example, our default is to go with open source option is where we can open source. And it makes sense for us to do so where we feel the product community might benefit from it. That's our default stance. The CNCF, the cloud native computing foundation is something that we've been involved with quite a bit. You know, we contribute to Kubernetes, be contribute to Envoy. I talked about continuity a bit. We've also contributed projects like CDK 8, which marries the AWS cloud development kit with Kubernetes. It's now a sandbox project in Kubernetes, and those are some of the areas. CNCF is such a wide surface area. We don't contribute to everything, but we definitely participate actively in CNCF with projects like HCB that are critical to eat for us. We are very, very active in just how the project evolves, but also try and see which of the projects that are important to our customers who are running Kubernetes maybe by themselves or some other project on AWS. Envoy is a good example. Kubernetes itself is a good example because in the end, we want to make sure that people running Kubernetes on AWS, even if they are not using our services are successful and we can help them, or we can work on the projects that are important to them. That's kind of how we think about the world. And it's worked pretty well for us. We've done a bunch of work on the Kubernetes side to make sure that we can integrate and solve a customer problem. We've, you know, from everything from models to work that we have done with gravity on our arm processor to a virtual GPU plugin that allows you to share and media GPU resources to the elastic fabric adapter, which are the network device for high performance computing that it can use at Kubernetes on AWS, along with things that directly impact Kubernetes customers like the CDKs project. I talked about work that we do with the container networking interface to the Amazon control of a Kubernetes, which is an open source project that allows you to use other AWS services directly from Kubernetes clusters. Again, you notice success, Kubernetes, not EKS, which is a managed Kubernetes service, because if we want you to be successful with Kubernetes and AWS, whether using our managed service or running your own, or some third party service. Similarly, we worked with premetheus. We now have a managed premetheus service. And at reinvent last year, we announced the general availability of this thing called carpenter, which is a provisioning and auto-scaling engine for Kubernetes, which is also an open source project. But here's the beauty of carpenter. You don't have to be using EKS to use it. Anyone running Kubernetes on AWS can leverage it. We focus on the AWS provider, but we've built it in such a way that if you wanted to take carpenter and implemented on prem or another cloud provider, that'd be completely okay. That's how it's designed and what we anticipated people may want to do. I talked a little bit about BottleRocket it's our Linux-based open-source operating system. And the thing that we have done with BottleRocket is make sure that we focus on security and the needs of customers who want to run orchestrated container, very focused on that problem. So for example, BottleRocket only has essential software needed to run containers, se Linux. I just notice it says that's the lineups, but I'm sure that, you know, Lena Torvalds will be pretty happy. And seeing that SE linux is enabled by default, we use things like DM Verity, and it has a read only root file system, no shell, you can assess it. You can install it if you wanted to. We allowed it to create different bill types, variants as we call them, you can create a variant for a non AWS resource as well. If you have your own homegrown container orchestrator, you can create a variant for that. It's designed to be used in many different contexts and all of that is open sourced. And then we use the update framework to publish and secure repository and kind of how this transactional system way of updating the software. And it's something that we didn't invent, but we have embraced wholeheartedly. It's a bottle rockets, completely open source, you know, have partners like Aqua, where who develop security tools for containers. And for them, you know, something I bought in rocket is a natural partnership because people are running a container host operating system. You can use Aqua tooling to make sure that they have a secure Indiana environment. And we see many more examples like that. You may think so over us, it's all about AWS proprietary technology because Lambda is a proprietary service. But you know, if you look peek under the covers, that's not necessarily true. Lambda runs on top of firecracker, as we've talked about fact crackers and open-source projects. So the foundation of Lambda in many ways is open source. What it also allows people to do is because Lambda runs at such extreme scale. One of the things that firecracker is really good for is running at scale. So if you want to build your own firecracker base at scale service, you can have most of the confidence that as long as your workload fits the design parameters, a firecracker, the battle hardening the robustness is being proved out day-to-day by services at scale like Lambda and foggy. For those of you who don't know service support services, you know, in the end, our goal with serverless is to make sure that you don't think about all the infrastructure that your applications run on. We focus on business logic as much as you can. That's how we think about it. And serverless has become its own quote-unquote "Sort of environment." The number of partners and open-source frameworks and tools that are spun up around serverless. In which case mostly, I mean, Lambda, API gateway. So it says like that is pretty high. So, you know, number of open source projects like Zappa server serverless framework, there's so many that have come up that make it easier for our customers to consume AWS services like Lambda and API gateway. We've also done some of our own tooling and frameworks, a serverless application model, AWS jealous. If you're a Python developer, we have these open service runtimes for Lambda, rust dot other options. We have amount of number of tools that we opened source. So in general, you'll find that tooling that we do runtime will tend to be always be open-sourced. We will often take some of the guts of the things that we use to build our systems like firecracker and open-source them while the control plane, etc, AWS services may end up staying proprietary, which is the case in Lambda. Increasingly our customers build their applications and leverage the broader AWS partner network. The AWS partner network is a network of partnerships that we've built of trusted partners. when you go to the APN website and find a partner, they know that that partner meets a certain set of criteria that AWS has developed, and you can rely on those partners for your own business. So whether you're a little tiny business that wants some function fulfill that you don't have the resources for or large enterprise that wants all these applications that you've been using on prem for a long time, and want to keep leveraging them in the cloud, you can go to APN and find that partner and then bring their solution on as part of your cloud infrastructure and could even be a systems integrator, for example, to help you solve this specific development problem that you may have a need for. Increasingly, you know, one of the things we like to do is work with an apartment community that is full of open-source providers. So a great one, there's so many, and you have, we have a panel discussion with many other partners as well, who make it easier for you to build applications on AWS, all open source and built on open source. But I like to call it a couple of them. The first one of them is TIDELIFT. TIDELIFT, For those of you who don't know is a company that provides SAS based tools to curate track, manage open source catalogs. You know, they have a whole network of maintainers and providers. They help, if you're an independent open developer, or a smart team should probably get to know TIDELIFT. They provide you benefits and, you know, capabilities as a developer and maintainer that are pretty unique and really help. And I've seen a number of our open source community embraced TIDELIFT quite honestly, even before they were part of the APN. But as part of the partner network, they get to participate in things like ISP accelerate and they get to they're officially an advanced tier partner because they are, they migrated the SAS offering onto AWS. But in the end, if you're part of the open source supply chain, you're a maintainer, you are a developer. I would recommend working with TIDELIFT because their goal is making all of you who are developing open source solutions, especially on AWS, more successful. And that's why I enjoy this partnership with them. And I'm looking to do a lot more because I think as a company, we want to make sure that open source developers don't feel like they are not supported because all you have to do is read various forums. It's challenging often to be a maintainer, especially of a small project. So I think with helping with licensing license management, security identification remediation, helping these maintainers is a big part of what TIDELIFT to us and it was great to see them as part of a partner network. Another partner that I like to call sysdig. I actually got introduced to them many years ago when they first launched. And one of the things that happened where they were super interested in some of our serverless stuff. And we've been trying to figure out how we can work together because all of our customers are interested in the capabilities that cystic provides. And over the last few years, he found a number of areas where we can collaborate. So sysdig, I know them primarily in a security company. So people use cystic to secure the bills, detect, you know, do threat response, threat detection, completely continuously validate their posture, get this continuous analytics signal on how they're doing and monitor performance. At the end of it, it's a SAS platform. They have a very nice open source security stack. The one I'm most familiar with. And I think most of you are probably familiar with is Falco. You know, sysdig, a CNCF project has been super popular. It's just to go SSS what 3, 37, 40 million downloads by now. So that's pretty, pretty cool. And they have been a great partner because we've had to do make sure that their solution works at target, which is not a natural place for their software to run, but there was enough demand and interest from our customers that, you know, or both companies leaned in to make sure they can be successful. So last year sister got a security competency. We have a number of specific competencies that we for our partners, they have integration and security hub is great. partners are lean in the way cystic has onto making our customer successful. And working with us are the best partners that we have. And there's a number of open source companies out there built on open source where their entire portfolio is built on open source software or the active participants like we are that we love working with on a day to day basis. So, you know, I think the thing I would like to, as we wind this out in this presentation is, you know, AWS is constantly looking for partnerships because our partners enable our customers. They could be with companies like Redis with Mongo, confluent with Databricks customers. Your default reaction might be, "Hey, these are companies that maybe compete with AWS." but no, I mean, I think we are partners as well, like from somebody at the lower end of the spectrum where people run on top of the services that I own on Linux and containers are SE 2, For us, these partners are just as important customers as any AWS service or any third party, 20 external customer. And so it's not a zero sum game. We look forward to working with all these companies and open source projects from an AWS perspective, a big part of how, where my open source program spends its time is making it easy for our developers to contribute, to open source, making it easy for AWS teams to decide when to open source software or participate in open source projects. Over the last few years, we've made significant changes in how we reduce the friction. And I think you can see it in the results that I showed you earlier in this stock. And the last one is one of the most important things that I say and I'll keep saying that, that we do as AWS is carry the pager. There's a lot of open source projects out there, operationalizing them, running them at scale is not easy. It's not all for whatever reason. It may not have anything to do with the software itself. But our core competency is taking that and being really good at operating it and becoming experts at operating it. And then ideally taking that expertise and experience and operating that project, that software and contributing back upstream. Cause that makes it better for everybody. And I think you'll see us do a lot more of that going forward. We've been doing that for the last few years, you know, in the container space, we do it every day. And I'm excited about the possibilities. With that. Thank you very much. And I hope you enjoy the rest of the showcase. >> Okay. Welcome back. We have Deepak sing here. We just had the keynote closing keynote vice-president of compute services. Deepak. Great to a great keynote, great wisdom and insight from that session. A very notable highlights and cutting edge trends and product information. Thanks for sharing. >> No, anytime it's always good to be here. It's too bad that we still doing this virtually, but always good to talk to you, John. >> We'll get hopefully through this way pretty quickly, I want to jump right in. Cause we don't have a lot of time. I want to get some quick question. You've brought up a good things. Open source innovation. Okay. Going next level. You've seen the rise of super clouds and super apps developing at open source. You're seeing big companies contributing, you know, you mentioned Argo into it. You're seeing that dynamic where companies are forming around this. This is a rising tide. This is, this is actually real. It's not the old school of, okay, here's a project. And then someone manages support and commercialization of it. It's actually platform in cloud scale. This is next gen. >> Yeah. And actually I think it started a few years ago. We can talk about a company that, you know, you're very familiar with as part of this event, which is armory many years ago, Netflix spun off this project called Spinnaker. A Spinnaker is CISED you know, CSED system that was developed at Netflix for their own purposes, but they chose to open solicit. And since then, it's become very popular with customers who want to use it even on prem. And you have a company that spun up on it. I think what's making this world very unique is you have very large companies like Facebook that will build things for themselves like VITAS or Netflix with Spinnaker and open source them. And you can have a lot of discussion about why they chose to do so, etc. But increasingly that's becoming the default when Amazon or Netflix or Facebook or Mehta, I guess you call them these days, build something for themselves for their own needs. The first question we ask ourselves is, should it be opensource? And increasingly we are all saying yes. And here's what happens because of that. It gives an opportunity depending on how you open source it for innovation through commercial deployments, so that you get SaaS companies, you know, that are going to take that product and make it relevant and useful to a very broad number of customers. You build partnerships with cloud providers like AWS, because our customers love this open source project and they need help. And they may choose an AWS managed service, or they may end up working with this partner on a day-to-day basis. And we want to work with that partner because they're making our customers successful, which is one reason all of us are here. So you're having this set of innovation from large companies from, you know, whether they are just consumer companies like Metta infrastructure companies like us, or just random innovation that's happening in an open source project that which ends up in companies being spun up and that foster that innovative innovation and that flywheel that's happening right now. And I think you said that like, this is unique. I mean, you never saw this happen before from so many different directions. >> It really is a nice progression on the business model side as well. You mentioned Argo, which is a great organic thing that was Intuit developed. We just interviewed code fresh. They just presented here in the showcase as well. You seeing the formation around these projects develop now in the community at a different scale. I mean, look at code fresh. I mean, Intuit did it Argo and they're not just supporting it. They're building a platform. So you seeing the dynamics of tools and now emerging the platforms, you mentioned Lambda, okay. Which is proprietary for AWS and your talk powered by open source. So again, open source combined with cloud scale allows for new potential super applications or super clouds that are developing. This is a new phenomenon. This isn't just lift and shift and host on the cloud. This is actually a construction production developer workflow. >> Yeah. And you are seeing consumers, large companies, enterprises, startups, you know, it used to be that startups would be comfortable adopting some of these solutions, but now you see companies of all sizes doing so. And I said, it's not just software it's software, the services increasingly becoming the way these are given, delivered to customers. I actually think the innovation is just getting going, which is why we have this. We have so many partners here who are all in inventing and innovating on top of open source, whether it's developed by them or a broader community. >> Yeah. I liked, I liked the represent container. Do you guys have, did that drove that you've seen a lot of changes and again, with cloud scale and open source, you seeing the dynamics change, whether you're enabling that, and then you see kind of like real big change. So let's take snowflake, a big customer of AWS. They started out as a startup too, but they weren't a data warehouse. They were bringing data warehouse like functionality and then changing everything differently and making it consumable for the cloud. And hence they're huge. So that's a disruption into an incumbent leader or sector. Then you've got new capabilities emerging. What's your thoughts, Deepak? Can you share your vision on how you have the disruption to existing leaders, old guard, if you will, as you guys call them and then new capabilities as these new platforms emerge at a net new functionality, how do you see that emerging? >> Yeah. So I speak from my side of the world. I've lived in over the last few years, which has containers and serverless, right? There's a lot of, if you go to any enterprise and ask them, do you want to modernize the infrastructure? Do you want to take advantage of automated software delivery, continuous delivery infrastructure as code modern observability, all of them will say yes, but they also are still a large enterprise, which has these enterprise level requirements. I'm using the word enterprise a lot. And I usually it's a trigger word for me because so many customers have similar requirements, but I'm using it here as large company with a lot of existing software and existing practices. I think the innovation that's coming and I see a lot of companies doing that is saying, "Hey, we understand the problems you want to solve. We understand the world where you live in, which could be regulated." You want to use all these new modalities. How do we allow you to use all of them? Keep the advantages of switching to a Lambda or switching to, and a service running on far gate, but give you the same capabilities. And I think I'll bring up cystic here because we work so closely with them on Falco. As an example, I just talked about them in my keynote. They could have just said, "Oh no, we'll just support the SE2 and be done with it." They said, "No, we're going to make sure that serverless containers in particular are something that you're going to be really good at because our customers want to use them, but requires us to think differently. And then they ended up developing new things like Falco that are born in this new world, but understand the requirements of the old world. If you get what I'm saying. And I think that a real example. >> Yeah. Oh, well, I mean, first of all, they're smart. So that was pretty obvious for most people that know, sees that you can connect the dots on serverless, which is a great point, but not everyone can see that again, this is what's new and and systig was just found in his backyard. As I found out on my interview, a great, great founder, they would do a new thing. So it was a very easy to connect the dots there again, that's the trend. Well, I got to ask if they're doing that for serverless, you mentioned graviton in your speech and what came out of you mentioned graviton in your speech and what came out of re-invent this past year was all the innovation going on at the compute level with gravitron at many levels in the Silicon. How should companies and open source developers think about how to innovate with graviton? >> Yeah, I mean, you've seen examples from people blogging and tweeting about how fast their applications run and grab it on the price performance benefits that they get, whether it's on, you know, whether it's an observability or other places. something that AWS is going to embrace across a compute something that AWS is going to embrace across a compute portfolio. Obviously you can go find EC2 instances, the gravitron two instances and run on them and that'll be great. But we know that most of our customers, many of our customers are building new applications on serverless containers and serveless than even as containers increasingly with things like foggy, where they don't want to operate the underlying infrastructure. A big part of what we're doing is to make sure that graviton is available to you on every compute modality. You can run it on a C2 forever. You've been running, being able to use ECS and EKS and run and grab it on almost since launch. What do you want me to take it a step further? You elastic Beanstalk customers, elastic Beanstalk has been around for a decade, but you can now use it with graviton. people running ECS on for gate can now use graviton. Lambda customers can pick graviton as well. So we're taking this price performance benefits that you get So we're taking this price performance benefits that you get from graviton and basically putting it across the entire compute portfolio. What it means is every high level service that gets built on compute infrastructure. And you get the price performance benefits, you get the price performance benefits of the lower power consumption of arm processes. So I'm personally excited like crazy. And you know, this has graviton 2 graviton 3 is coming. >> That's incredible. It's an opportunity like serverless was it's pretty obvious. And I think hopefully everyone will jump on that final question as the time's ticking here. I want to get your thoughts quickly. If you look at what's happened with containers over the past say eight years since the original founding of the first Docker instance, if you will, to how that's evolved and then the introduction of Kubernetes and the cloud native wave we're seeing now, what is, how would you describe the relationship between the success Docker, seeing now with Kubernetes in the cloud native construct what's different and why is this combination so successful? >> Yeah. I often say that containers would have, let me rephrase that. what I say is that people would have adopted sort of the modern way of running applications, whether containers came around or not. But the fact that containers came around made that migration and that journey is so much more efficient for people. So right from, I still remember the first doc that Solomon gave Billy announced DACA and starting to use it on customers, starting to get interested all the way to the more sort of advanced orchestration that we have now for containers across the board. And there's so many examples of the way you can do that. Kubernetes being the most, most well-known one. Here's the thing that I think has changed. I think what Kubernetes or Docker, or the whole sort of modern way of building applications has done is it's taken people who would have taken years adopting these practices and by bringing it right to the fingertips and rebuilding it into the APIs. And in the case of Kubernetes building an entire sort of software world around it, the number of, I would say number of decisions people have to take has gone smaller in many ways. There's so many options, the number of decisions that become higher, but the com the speed at which they can get to a result and a production version of an application that works for them is way low. I have not seen anything like what I've seen in the last 6, 7, 8 years of how quickly the most you know, the most I would say is, you know, a company that you would think would never adopt modern technology has been able to go from, this is interesting to getting a production really quickly. And I think it's because the tooling makes it So, and the fact that you see the adoption that you see right and the fact that you see the adoption that you see right from the fact that you could do Docker run Docker, build Docker, you know, so easily back in the day, all the way to all the advanced orchestration you can do with container orchestrator is today. sort of taking all of that away as well. there's never been a better time to be a developer independent of whatever you're trying to build. And I think containers are a big central part of why that's happened. >> Like the recipe, the combination of cloud-scale, the timing of Kubernetes and the containerization concepts just explode as a beautiful thing. And it creates more opportunities and will challenges, which are opportunities that are net new, but it solves the automation piece that we're seeing this again, it's only makes things go faster. >> Yes. >> And that's the key trend. Deepak, thank you so much for coming on. We're seeing tons of open cloud innovations, thanks to the success of your team at AWS and being great participants in the community. We're seeing innovations from startups. You guys are helping enabling that. Of course, they want to live on their own and be successful and build their super clouds and super app. So thank you for spending the time with us. Appreciate. >> Yeah. Anytime. And thank you. And you know, this is a great event. So I look forward to people running software and building applications, using AWS services and all these wonderful partners that we have. >> Awesome, great stuff. Great startups, great next generation leaders emerging. When you see startups, when they get successful, they become the modern software applications platforms out there powering business and changing the world. This is the cube you're watching the AWS startup showcase. Season two episode one open cloud innovations on John Furrier your host, see you next time.

Published Date : Jan 26 2022

SUMMARY :

And the thing that we have We just had the keynote closing but always good to talk to you, John. It's not the old school And I think you said that So you seeing the dynamics but now you see companies and then you see kind How do we allow you to use all of them? sees that you can connect is available to you on Kubernetes and the cloud of the way you can do that. but it solves the automation And that's the key trend. And you know, and changing the world.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmazonORGANIZATION

0.99+

AWSORGANIZATION

0.99+

DeepakPERSON

0.99+

Lena TorvaldsPERSON

0.99+

FalcoORGANIZATION

0.99+

NetflixORGANIZATION

0.99+

JohnPERSON

0.99+

Deepak SinghPERSON

0.99+

MehtaORGANIZATION

0.99+

twoQUANTITY

0.99+

FacebookORGANIZATION

0.99+

LambdaTITLE

0.99+

firstQUANTITY

0.99+

John FurrierPERSON

0.99+

JavaTITLE

0.99+

PythonTITLE

0.99+

SolomonPERSON

0.99+

two waysQUANTITY

0.99+

OneQUANTITY

0.99+

PyPyTITLE

0.99+

last yearDATE

0.99+

over 13 yearsQUANTITY

0.99+

LinuxTITLE

0.99+

TodayDATE

0.99+

IndianaLOCATION

0.99+

DatabricksORGANIZATION

0.99+

bothQUANTITY

0.99+

How Open Source is Changing the Corporate and Startup Enterprises | Open Cloud Innovations


 

(gentle upbeat music) >> Hello, and welcome to theCUBE presentation of the AWS Startup Showcase Open Cloud Innovations. This is season two episode one of an ongoing series covering setting status from the AWS ecosystem. Talking about innovation, here it's open source for this theme. We do this every episode, we pick a theme and have a lot of fun talking to the leaders in the industry and the hottest startups. I'm your host John Furrier here with Lisa Martin in our Palo Alto studios. Lisa great series, great to see you again. >> Good to see you too. Great series, always such spirited conversations with very empowered and enlightened individuals. >> I love the episodic nature of these events, we get more stories out there than ever before. They're the hottest startups in the AWS ecosystem, which is dominating the cloud sector. And there's a lot of them really changing the game on cloud native and the enablement, the stories that are coming out here are pretty compelling, not just from startups they're actually penetrating the enterprise and the buyers are changing their architectures, and it's just really fun to catch the wave here. >> They are, and one of the things too about the open source community is these companies embracing that and how that's opening up their entry to your point into the enterprise. I was talking with several customers, companies who were talking about the 70% of their pipeline comes from the open source community. That's using the premium version of the technology. So, it's really been a very smart, strategic way into the enterprise. >> Yeah, and I love the format too. We get the keynote we're doing now, opening keynote, some great guests. We have Sir John on from AWS started program, he is the global startups lead. We got Swami coming on and then closing keynote with Deepak Singh. Who's really grown in the Amazon organization from containers now, compute services, which now span how modern applications are being built. And I think the big trend that we're seeing that these startups are riding on that big wave is cloud natives driving the modern architecture for software development, not just startups, but existing, large ISV and software companies are rearchitecting and the customers who buy their products and services in the cloud are rearchitecting too. So, it's a whole new growth wave coming in, the modern era of cloud some say, and it's exciting a small startup could be the next big name tomorrow. >> One of the things that kind of was a theme throughout the conversations that I had with these different guests was from a modern application security perspective is, security is key, but it's not just about shifting lab. It's about doing so empowering the developers. They don't have to be security experts. They need to have a developer brain and a security heart, and how those two organizations within companies can work better together, more collaboratively, but ultimately empowering those developers, which goes a long way. >> Well, for the folks who are watching this, the format is very simple. We have a keynote, editorial keynote speakers come in, and then we're going to have a bunch of companies who are going to present their story and their showcase. We've interviewed them, myself, you Dave Vallante and Dave Nicholson from theCUBE team. They're going to tell their stories and between the companies and the AWS heroes, 14 companies are represented and some of them new business models and Deepak Singh who leads the AWS team, he's going to have the closing keynote. He talks about the new changing business model in open source, not just the tech, which has a lot of tech, but how companies are being started around the new business models around open source. It's really, really amazing. >> I bet, and does he see any specific verticals that are taking off? >> Well, he's seeing the contribution from big companies like AWS and the Facebook's of the world and large companies, Netflix, Intuit, all contributing content to the open source and then startups forming around them. So Netflix does some great work. They donated to open source and next thing you know a small group of people get together entrepreneurs, they form a company and they create a platform around it with unification and scale. So, the cloud is enabling this new super application environment, superclouds as we call them, that's emerging and this new supercloud and super applications are scaling data-driven machine learning and AI that's the new formula for success. >> The new formula for success also has to have that velocity that developers expect, but also that the consumerization of tech has kind of driven all of us to expect things very quickly. >> Well, we're going to bring in Serge Shevchenko, AWS Global Startup program into the program. Serge is our partner. He is the leader at AWS who has been working on this program Serge, great to see you. Thanks for coming on. >> Yeah, likewise, John, thank you for having me very excited to be here. >> We've been working together on collaborating on this for over a year. Again, season two of this new innovative program, which is a combination of CUBE Media partnership, and AWS getting the stories out. And this has been a real success because there's a real hunger to discover content. And then in the marketplace, as these new solutions coming from startups are the next big thing coming. So, you're starting to see this going on. So I have to ask you, first and foremost, what's the AWS startup showcase about. Can you explain in your terms, your team's vision behind it, and why those startup focus? >> Yeah, absolutely. You know John, we curated the AWS Startup Showcase really to bring meaningful and oftentimes educational content to our customers and partners highlighting innovative solutions within these themes and ultimately to help customers find the best solutions for their use cases, which is a combination of AWS and our partners. And really from pre-seed to IPO, John, the world's most innovative startups build on AWS. From leadership downward, very intentional about cultivating vigorous AWS community and since 2019 at re:Invent at the launch of the AWS Global Startup program, we've helped hundreds of startups accelerate their growth through product development support, go to market and co-sell programs. >> So Serge question for you on the theme of today, John mentioned our showcases having themes. Today's theme is going to cover open source software. Talk to us about how Amazon thinks about opensource. >> Sure, absolutely. And I'll just touch on it briefly, but I'm very excited for the keynote at the end of today, that will be delivered by Deepak the VP of compute services at AWS. We here at Amazon believe in open source. In fact, Amazon contributes to open source in multiple ways, whether that's through directly contributing to third-party project, repos or significant code contributions to Kubernetes, Rust and other projects. And all the way down to leadership participation in organizations such as the CNCF. And supporting of dozens of ISV myself over the years, I've seen explosive growth when it comes to open source adoption. I mean, look at projects like Checkov, within 12 months of launching their open source project, they had about a million users. And another great example is Falco within, under a decade actually they've had about 37 million downloads and that's about 300% increase since it's become an incubating project in the CNCF. So, very exciting things that we're seeing here at AWS. >> So explosive growth, lot of content. What do you hope that our viewers and our guests are going to be able to get out of today? >> Yeah, great question, Lisa. I really hope that today's event will help customers understand why AWS is the best place for them to run open source, commercial and which partner solutions will help them along their journey. I think that today the lineup through the partner solutions and Deepak at the end with the ending keynote is going to present a very valuable narrative for customers and startups in selecting where and which projects to run on AWS. >> That's great stuff Serge would love to have you on and again, I want to just say really congratulate your team and we enjoy working with them. We think this showcase does a great service for the community. It's kind of open source in its own way if I can co contributing working on out there, but you're really getting the voices out at scale. We've got companies like Armory, Kubecost, Sysdig, Tidelift, Codefresh. I mean, these are some of the companies that are changing the game. We even had Patreon a customer and one of the partners sneak with security, all the big names in the startup scene. Plus AWS Deepak saying Swami is going to be on the AWS Heroes. I mean really at scale and this is really a great. So, thank you so much for participating and enabling all of this. >> No, thank you to theCUBE. You've been a great partner in this whole process, very excited for today. >> Thanks Serge really appreciate it. Lisa, what a great segment that was kicking off the event. We've got a great lineup coming up. We've got the keynote, final keynote fireside chat with Deepak Singh a big name at AWS, but Serge in the startup showcase really innovative. >> Very innovative and in a short time period, he talked about the launch of this at re:Invent 2019. They've helped hundreds of startups. We've had over 50 I think on the showcase in the last year or so John. So we really gotten to cover a lot of great customers, a lot of great stories, a lot of great content coming out of theCUBE. >> I love the openness of it. I love the scale, the storytelling. I love the collaboration, a great model, Lisa, great to work with you. We also Dave Vallante and Dave Nicholson interview. They're not here, but let's kick off the show. Let's get started with our next guest Swami. The leader at AWS Swami just got promoted to VP of the database, but also he ran machine learning and AI at AWS. He is a leader. He's the author of the original DynamoDB paper, which is celebrating its 10th year anniversary really impacted distributed computing and open source. Swami's introduced many opensource aspects of products within AWS and has been a leader in the engineering side for many, many years at AWS, from an intern to now an executive. Swami, great to see you. Thanks for coming on our AWS startup showcase. Thanks for spending the time with us. >> My pleasure, thanks again, John. Thanks for having me. >> I wanted to just, if you don't mind asking about the database market over the past 10 to 20 years cloud and application development as you see, has changed a lot. You've been involved in so many product launches over the years. Cloud and machine learning are the biggest waves happening to your point to what you're doing now. Software is under the covers it's powering it all infrastructure is code. Open source has been a big part of it and it continues to grow and change. Deepak Singh from AWS talks about the business model transformation of how like Netflix donates to the open source. Then a company starts around it and creates more growth. Machine learnings and all the open source conversations around automation as developers and builders, like software as cloud and machine learning become the key pistons in the engine. This is a big wave, what's your view on this? How how has cloud scale and data impacting the software market? >> I mean, that's a broad question. So I'm going to break it down to kind of give some of the back data. So now how we are thinking about it first, I'd say when it comes to the open source, I'll start off by saying first the longevity and by ability of open sources are very important to our customers and that is why we have been a significant contributor and supporter of these communities. I mean, there are several efforts in open source, even internally by actually open sourcing some of our key Amazon technologies like Firecracker or BottleRocket or our CDK to help advance the industry. For example, CDK itself provides some really powerful way to build and configure cloud services as well. And we also contribute to a lot of different open source projects that are existing ones, open telemetries and Linux, Java, Redis and Kubernetes, Grafana and Kafka and Robotics Operating System and Hadoop, Leucine and so forth. So, I think, I can go on and on, but even now I'd say the database and observability space say machine learning we have always started with embracing open source in a big material way. If you see, even in deep learning framework, we championed MX Linux and some of the core components and we open sourced our auto ML technology auto Glue on, and also be open sourced and collaborated with partners like Facebook Meta on Fighter showing some major components and there, and then we are open search Edge Compiler. So, I would say the number one thing is, I mean, we are actually are very, very excited to partner with broader community on problems that really mattered to the customers and actually ensure that they are able to get amazing benefit of this. >> And I see machine learning is a huge thing. If you look at how cloud group and when you had DynamoDB paper, when you wrote it, that that was the beginning of, I call the cloud surge. It was the beginning of not just being a resource versus building a data center, certainly a great alternative. Every startup did it. That's history phase one inning and a half, first half inning. Then it became a large scale. Machine learning feels like the same way now. You feel like you're seeing a lot of people using it. A lot of people are playing around with it. It's evolving. It's been around as a science, but combined with cloud scale, this is a big thing. What should people who are in the enterprise think about how should they think about machine learning? How has some of your top customers thought about machine learning as they refactor their applications? What are some of the things that you can share from your experience and journey here? >> I mean, one of the key things I'd say just to set some context on scale and numbers. More than one and a half million customers use our database analytics or ML services end-to-end. Part of which machine learning services and capabilities are easily used by more than a hundred thousand customers at a really good scale. However, I still think in Amazon, we tend to use the phrase, "It's day one in the age of internet," even though it's an, or the phrase, "Now, but it's a golden one," but I would say in the world of machine learning, yes it's day one but I also think we just woke up and we haven't even had a cup of coffee yet. That's really that early, so. And, but when you it's interesting, you've compared it to where cloud was like 10, 12 years ago. That's early days when I used to talk to engineering leaders who are running their own data center and then we talked about cloud and various disruptive technologies. I still used to get a sense about like why cloud and basic and whatnot at that time, Whereas now with machine learning though almost every CIO, CEO, all of them never asked me why machine learning. Instead, the number one question, I get is, how do I get started with it? What are the best use cases? which is great, and this is where I always tell them one of the learnings that we actually learned in Amazon. So again, a few years ago, probably seven or eight years ago, and Amazon itself realized as a company, the impact of what machine learning could do in terms of changing how we actually run our business and what it means to provide better customer experience optimize our supply chain and so far we realized that the we need to help our builders learn machine learning and the help even our business leaders understand the power of machine learning. So we did two things. One, we actually, from a bottom-up level, we built what I call as machine learning university, which is run in my team. It's literally stocked with professors and teachers who offer curriculum to builders so that they get educated on machine learning. And now from a top-down level we also, in our yearly planning process, we call it the operational planning process where we write Amazon style narratives six pages and then answer FAQ's. We asked everyone to answer one question around, like how do you plan to leverage machine learning in your business? And typically when someone says, I really don't play into our, it does not apply. It's usually it doesn't go well. So we kind of politely encourage them to do better and come back with a better answer. This kind of dynamic on top-down and bottom-up, changed the conversation and we started seeing more and more measurable growth. And these are some of the things you're starting to see more and more among our customers too. They see the business benefit, but this is where to address the talent gap. We also made machine learning university curriculum actually now open source and freely available. And we launched SageMaker Studio Lab, which is a no cost, no set up SageMaker notebook service for educating learner profiles and all the students as well. And we are excited to also announce AIMLE scholarship for underrepresented students as well. So, so much more we can do well. >> Well, congratulations on the DynamoDB paper. That's the 10 year anniversary, which is a revolutionary product, changed the game that did change the world and that a huge impact. And now as machine learning goes to the next level, the next intern out there is at school with machine learning. They're going to be writing that next paper, your advice to them real quick. >> My biggest advice is, always, I encourage all the builders to always dream big, and don't be hesitant to speak your mind as long as you have the right conviction saying you're addressing a real customer problem. So when you feel like you have an amazing solution to address a customer problem, take the time to articulate your thoughts better, and then feel free to speak up and communicate to the folks you're working with. And I'm sure any company that nurtures good talent and knows how to hire and develop the best they will be willing to listen and then you will be able to have an amazing impact in the industry. >> Swami, great to know you're CUBE alumni love our conversations from intern on the paper of DynamoDB to the technical leader at AWS and database analyst machine learning, congratulations on all your success and continue innovating on behalf of the customers and the industry. Thanks for spending the time here on theCUBE and our program, appreciate it. >> Thanks again, John. Really appreciate it. >> Okay, now let's kick off our program. That ends the keynote track here on the AWS startup showcase. Season two, episode one, enjoy the program and don't miss the closing keynote with Deepak Singh. He goes into great detail on the changing business models, all the exciting open source innovation. (gentle bright music)

Published Date : Jan 26 2022

SUMMARY :

of the AWS Startup Showcase Good to see you too. and the buyers are changing and one of the things too Yeah, and I love the format too. One of the things and the AWS heroes, like AWS and the Facebook's of the world but also that the consumerization of tech He is the leader at AWS who has thank you for having me and AWS getting the stories out. at the launch of the AWS Talk to us about how Amazon And all the way down to are going to be able to get out of today? and Deepak at the end and one of the partners in this whole process, but Serge in the startup in the last year or so John. Thanks for spending the time with us. Thanks for having me. and data impacting the software market? but even now I'd say the database are in the enterprise and all the students as well. on the DynamoDB paper. take the time to articulate and the industry. Thanks again, John. and don't miss the closing

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SergePERSON

0.99+

Dave NicholsonPERSON

0.99+

Dave VallantePERSON

0.99+

AWSORGANIZATION

0.99+

Dave NicholsonPERSON

0.99+

Lisa MartinPERSON

0.99+

Deepak SinghPERSON

0.99+

JohnPERSON

0.99+

AmazonORGANIZATION

0.99+

SwamiPERSON

0.99+

NetflixORGANIZATION

0.99+

John FurrierPERSON

0.99+

CodefreshORGANIZATION

0.99+

DeepakPERSON

0.99+

ArmoryORGANIZATION

0.99+

LisaPERSON

0.99+

SysdigORGANIZATION

0.99+

Serge ShevchenkoPERSON

0.99+

KubecostORGANIZATION

0.99+

TideliftORGANIZATION

0.99+

FacebookORGANIZATION

0.99+

14 companiesQUANTITY

0.99+

six pagesQUANTITY

0.99+

one questionQUANTITY

0.99+

12 monthsQUANTITY

0.99+

more than a hundred thousand customersQUANTITY

0.99+

OneQUANTITY

0.99+

two thingsQUANTITY

0.99+

todayDATE

0.98+

last yearDATE

0.98+

CNCFORGANIZATION

0.98+

More than one and a half million customersQUANTITY

0.98+

two organizationsQUANTITY

0.98+

TodayDATE

0.98+

CDKORGANIZATION

0.98+

IntuitORGANIZATION

0.98+

DynamoDBTITLE

0.98+

first half inningQUANTITY

0.98+

Breaking Analysis: AWS & Azure Accelerate Cloud Momentum


 

>> From theCUBE studios in Palo Alto in Boston, bringing you data-driven insights from theCUBE in ETR. This is "Breaking Analysis" with Dave Vellante. >> Despite all the talk about repatriation, hybrid and multi-Cloud opportunities, and Cloud is an increasingly expensive option for customers, the data continues to show the importance of public Cloud to the digital economy. Moreover, the two leaders, AWS and Azure, are showing signs of accelerated momentum that point to those two giants pulling away from the pack in the years ahead, with each firm's showing broad based momentum across their respective product lines. It's unclear if anything, other than government intervention or self-inflicted wounds will slow these two companies down this decade. Despite their commanding lead, a winning strategy for companies that don't run their own Cloud continues to be innovating on top of their massive CapEx investments. The most notable example here being Snowflake. Hello, everyone. Welcome to this week's Wikibon CUBE insights powered by ETR. In this breaking analysis, we provide our quarterly market share update for the big four hyperscale Cloud providers. And we'll share some new ETR data from their most recent survey. And we'll drill into some of the reasons for the momentum of these two companies and drill further into the database and data warehouse sector to see what, if anything, has changed in that space. First, let's look at some of the noteworthy comments from AWS and Microsoft in their recent earnings updates. We heard from Amazon, the following, "AWS has seen a reacceleration of revenue growth as customers have expanded their commitment to the Cloud and selected AWS as their Cloud partner." Notably, AWS revenues increased 39% in Q3 2021. That's a thousand basis point increase in growth relative to Q3 2020. That's an astounding milestone for a company that we expect to surpass $60 billion in revenue this year. Further, AWS touted the adoption of its custom silicon, and specifically its Graviton2 processors. AWS is fond of emphasizing Graviton's 40% price performance improvements relative to x86 processors, something we've reported on quite extensively. AWS is investing in custom silicon, encouraging ISVs to port their code to the platform so that customers will experience little or no code changes when they migrate. Again, we believe this is a secret weapon for AWS as its cost structure will continue to improve at a rate faster than competitors that don't have the resources or the skills or the stomach to develop such capabilities. Microsoft, for its part, also saw astoundingly good growth of 48% this past quarter for Azure. This is a company that we forecast will approach $40 billion in IaaS and PaaS public Cloud revenue this year. Microsoft's CEO, Satya Nadella, on its earnings call, emphasized the changing nature of Cloud expanding in a distributed fashion to the edge. He referenced Azure as the world's computer. Building on his statements last year that Microsoft is building out a powerful, ubiquitous, intelligent, sensing and predictive Cloud. Yes, folks, it does feel like we're entering the so-called Metaverse, doesn't it? Okay, to underscore the momentum of these two companies, let's take a look at the ETR breakdown of Net score, which measures spending momentum. This chart will be familiar to our listeners. It shows the breakdown of net score for AWS, with the lime green showing new adoptions. That's 11%. The forest green is spending more than 6% relative to the first half of this year. That's a very robust 53%. The gray is flat spending. That's 30% on a very, very large base. And the pink is spending declines of minus 6% or worse. That's 4%. And the bright red is defections i.e those leaving AWS. That's 1%. That's virtually non-existent. You subtract the reds from the greens and you get a net score of 59. Remember, anything over 40, we can still consider to be elevated. Let's look at that same data for Microsoft again. You have some new ads that lime green, that's 7%. The forest green is at 46% of customers spending more, which is an incredible figure for a company with revenues that will in the near term surpass $200 billion. And the red is in the low single digits. Buffered by its enormous PC software profits over the years, Microsoft is powered through its Window's Dogma and transitioned into a Cloud powerhouse. Let's now share some of our latest numbers for the big four hyperscale players, AWS, Azure, Alibaba and Google. Here, we show data for these companies from 2018 and our estimates for 2021. This data includes our final figures for AWS, Azure and GCP for Q3 with Alibaba yet to report. Remember, only AWS and Alibaba report IaaS revenue cleanly with Microsoft and Google, they give us a little breadcrumb nuggets that allow us to triangulate with our survey data and other intelligence. But it's our attempt to do an apples to apples comparison for those four companies using AWS and it's reporting as a baseline. In Q3, AWS reported more than $16 billion in revenue. We estimate Azure at 10 billion, Alibaba, we expect to come in at just under 3 billion, and GCP at 2.5 billion for the quarter. With three quarters of data in, with the exception of Alibaba, we're forecasting AWS to capture 51% of the big four revenue, the hyperscale revenue. And really we believe these are the only four hyperscalers. AWS will surpass 60 billion with Azure just under 40 billion, Alibaba approaching 11 billion, and Google coming in just under 10 billion for the year is our expectation. We forecast these four will account for $120 billion this year. That's a 41% increase over 2020 and the same collective growth rate as 2020 relative to 2019. We expect Azure to be 63% of the size of AWS revenue. So it is gaining share. Both of those companies, however, saw accelerated growth this past quarter with Alibaba and GCP's growth rates decelerating relative to last year. Now, let's take a closer look at those growth rates. This chart shows the quarterly growth rates for each of the four going back to the beginning of 2019. Both GCP and Alibaba are showing dramatic declines in growth rates, whereas, this past quarter Azure saw accelerated growth and AWS has now seen an increased rate of growth for the past two quarters. In fact, AWS' growth is about where it was in 2019 when it was around half of its current revenue size. And in 2019 growth was decelerating through the quarters as you can see where today that trend has reversed. It's quite amazing. All right, let's take a look at the broader Cloud landscape and bring back some ETR data. This chart that we're showing here, it shows net score or spending momentum on the vertical axis and market share or presence in the dataset on the horizontal axis. Note that red dotted line, anything above that we can still consider elevated and impressive. As when we've previously shared this data, AWS and Microsoft Azure are up and to the right. Now remember, this chart is not just counting IaaS and PaaS as we showed you earlier, it's however the customers views whatever they think Cloud is. And so they're likely including Microsoft SaaS in this picture. Which is why Microsoft shows larger than AWS despite what we showed you earlier. Nonetheless, these two are well ahead of the pack and the growth rates indicate that they're pulling away. But we've added some of the other players, most notably VMware Cloud on AWS. It's showing momentum as is VMware Cloud, which is VMware Cloud foundation and other on-prem Cloud offerings, even though it's below the red line for the on-prem piece, it's very respectable. The VMware Cloud on AWS has been consistently up above that red line. Has popped beneath it in some quarters, but it's very, very strong. As is, you know, Red Hat OpenShift, it's a little bit below the line, but it is respectable. We've superimposed this by the way. Red Hat OpenShift in the ETR platform is under the container orchestration taxonomy, but we'd like to put it in next to the Cloud players for context. That's how Red Hat sort of thinks about this as well. They think about OpenShift as Cloud. And then you can see the other players. Alibaba has got a small sample in the ETR dataset. Just does not enough presence in China. But Dell and HPE have started to show up in the Cloud taxonomy. So buyers are associating their private Clouds with Cloud. So Dell's Apex, HPE's GreenLake. So that's a positive. And you can see Oracle, which of course is OCI, Oracle Cloud infrastructure. And then IBM with its public Cloud. So, it's a positive that these on-prem players are showing up in this data, but the reality is the hyperscalers are growing collectively at 40% annually and the on-prem players are growing in the low single digits. So, and if you carve out the IaaS business of AWS and Azure, they're larger than most of the on-premises infrastructure players. And all the on-prem players are moving toward an as a service model, as I just alluded to. So, undoubtedly, hybrid multicloud edge are going to present opportunities for the likes of Dell, HPE, Cisco, VMware, IBM, Red Hat, et cetera. But they also present opportunities for the public Cloud players who have vibrant ecosystems and marketplaces much more diverse and deep than the traditional vendors. You know, we have a clearer picture of Microsoft's sort of hybrid and edge strategy because the company has such an enormous legacy business, it really had to think about that much more deeply. It wasn't a blank sheet of paper like AWS. It's going to be interesting at reinvent this year if new CEO, Adam Selipsky, will talk about this. And it will be good to hear how he's thinking about the next decade, how AWS thinks about hybrid and edge, I guarantee that with their developer affinity and custom Silicon capabilities, they're thinking about it differently than traditional enterprise players. And as we've stressed in this segment, they have across the board momentum. Now to quantify that, let's take a look at AWS as portfolio in the spending momentum within its product segments. This chart shows AWS's net scores or spending momentum in the areas where AWS participates in the ETR taxonomy. Again, note that red line. Anything above 40% is considered an elevated watermark. We're showing data from last October, this past July and the latest October 21 survey. That yellow line or a bar. What's notable is the yellow versus the gray bars up across the board for the most part, other than chime... And by the way, other than chime, everything is above the 40% mark as well. Now, we've highlighted database because we feel it's one of the most strategic sectors in a real battleground. So we want to drill into that a bit. Here's our familiar X Y graph showing Net score on the Y axis, remember, that's, again, spending momentum and market share or pervasiveness in the survey on the horizontal axis. This data, by the way, includes on-prem and Cloud database data warehouse. So keep that in mind. Let's start with one of our favorite topics; Snowflake. We've reported again and again and again, that we've never seen anything like this. The company's net score has moderated ever so slightly this quarter, but it's still just below 80%. Very highly elevated. Well, above that 40% mark. It's Snowflake's presence continues to grow as a gain share in the market. Snowflake is growing revenue in the triple digits. It's an insane pace, hence its current $115 billion market cap as of this episode. Now that said, all three US-based Cloud players there are above the 40% line with AWS and Microsoft having significant presence on the horizontal axis. You see Cockroach Labs, Redis, Couchbase, they're all elevated or highly elevated. Couchbase just went public this summer. So that may help with its presence. MongoDB, they're killing it. They have a $37 billion market cap as of this episode. The stock has been on a tear. You see MariaDB was also in the mix. And then of course you have Oracle, the database leader. Look, they continue to invest in making the Oracle database and other software like MySQL, the best solution for mission critical workloads, and they're investing in their Cloud. But you can see overall, they just don't have the momentum from a spending standpoint that the others do because the declines in their legacy business. And they've been around a long time. Those declines are not fully offset by the growth in Cloud database and Cloud migration. But look, Oracle is a financial powerhouse with a $250 billion plus market cap. And the stock has done very well this past year. Up over 60%. Cloudera is going private. So it can hide the pain of the transitions that it's undergoing between the legacy install bases of Cloudera and Hortonworks. It's just a tough situation. When the companies came together, Cloudera essentially had a dead end. Each of those respective platforms and migrate their customers to a more modern stack as part of its Cloud strategy. Ironic that it's name is Cloudera. You know, that's always a difficult thing to do. So as a private company, Cloudera can maybe get off that 90 day shot clock and buy some time to invest without getting hammered by the street. And you know, Teradata consistently has not shown up well in the ETR dataset. It's transitioned to Cloud and cross-Cloud still hasn't shown momentum in the surveys. So, look right now, it's looking like the rich get richer. So just to quantify that a little bit, let's line up some of the database players and look a little bit more closely at net score. This chart shows the spending momentum or lack thereof with the net score or spending velocity granularity that we described before. Remember, green is spending more, red is spending less, bright red is leaving the platform, bright green is adding the platform. You take red, subtract red from the green, and that gives you a net score. Snowflake, as we said, tops the list. You can see the granularity there. You can compare the performance. In a little different view to understand how these scores are derived, look, the ideal profile is a solid lime green, a big forest green, a not too large gray and ideally little or no bright red AKA defections. And you can see the green funnel in the gray increasing prominence as the vendor momentum declines. Interestingly, with the exception of Cloudera and Teradata, defections are all in the single digits or nonexistent. In the case of Snowflake, Redis, red is no red at all, but small sample, Couchbase has no defections and very little defection for the giant Microsoft. Incredibly impressive. This speaks to how hard it is to migrate off of a database no matter how disgruntled you are. The more common scenario is to isolate the database and build new functionality on modern platforms. Okay, so what to watch out for. Well, reinvent this coming up next month. Oh this month. It's the first time someone other than Andy Jassy will be keynoting as CEO. 15 years of Cloud, this is the 10th re-invent, which is always a market for the direction of the industry. I've said many times that the last decade was largely about IT transformation powered by the Cloud. I believe we're entering a new era of business transformation where the Cloud is going to play a significant role. But the Cloud is evolving from a set of remote services out there in the Cloud to an omnipresent platform on top of which many customers and technology companies can innovate. And virtually every industry will be impacted by Cloud. However it evolves in the coming decade. The question will be, how fast can you go? And how will players like AWS and Microsoft and many others that are building on top of these platforms make it easier for you to go fast? That's what I'll be watching for at re-invent and beyond. Okay, that's a wrap for today. Remember, these episodes, they're all available as podcasts, wherever you listen. All you got to do is search Breaking Analysis podcasts. Check out ETR's website at etr.plus. We also publish a full report every week on wikibon.com and siliconangle.com. You can get in touch with me, david.vellante@siliconangle.com. You can DM me @dvellante or comment on our LinkedIn posts. This is Dave Vellante for theCUBE insights powered by ETR. Have a great week, everybody. Stay safe, be well. And we'll see you next time. We'll see you at re-invent. (soft upbeat music)

Published Date : Nov 13 2021

SUMMARY :

This is "Breaking Analysis" and GCP at 2.5 billion for the quarter.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Adam SelipskyPERSON

0.99+

Dave VellantePERSON

0.99+

MicrosoftORGANIZATION

0.99+

AWS'ORGANIZATION

0.99+

AWSORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

IBMORGANIZATION

0.99+

AlibabaORGANIZATION

0.99+

Satya NadellaPERSON

0.99+

2019DATE

0.99+

$115 billionQUANTITY

0.99+

Andy JassyPERSON

0.99+

ChinaLOCATION

0.99+

CiscoORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

DellORGANIZATION

0.99+

2018DATE

0.99+

$200 billionQUANTITY

0.99+

$37 billionQUANTITY

0.99+

OracleORGANIZATION

0.99+

2020DATE

0.99+

Palo AltoLOCATION

0.99+

41%QUANTITY

0.99+

2.5 billionQUANTITY

0.99+

10 billionQUANTITY

0.99+

GCPORGANIZATION

0.99+

53%QUANTITY

0.99+

2021DATE

0.99+

40%QUANTITY

0.99+

51%QUANTITY

0.99+

63%QUANTITY

0.99+

$250 billionQUANTITY

0.99+

30%QUANTITY

0.99+

4%QUANTITY

0.99+

48%QUANTITY

0.99+

60 billionQUANTITY

0.99+

7%QUANTITY

0.99+

$60 billionQUANTITY

0.99+

two companiesQUANTITY

0.99+

two leadersQUANTITY

0.99+

$120 billionQUANTITY

0.99+

39%QUANTITY

0.99+

more than $16 billionQUANTITY

0.99+

HPEORGANIZATION

0.99+

Survey Data Shows no Slowdown in AWS & Cloud Momentum


 

from the cube studios in palo alto in boston bringing you data-driven insights from the cube and etr this is breaking analysis with dave vellante despite all the chatter about cloud repatriation and the exorbitant cost of cloud computing customer spending momentum continues to accelerate in the post-isolation economy if the pandemic was good for the cloud it seems that the benefits of cloud migration remain lasting in the late stages of covid and beyond and we believe this stickiness is going to continue for quite some time we expect i asked revenue for the big four hyperscalers to surpass 115 billion dollars in 2021 moreover the strength of aws specifically as well as microsoft azure remain notable such large organizations showing elevated spending momentum as shown in the etr survey results is perhaps unprecedented in the technology sector hello everyone and welcome to this week's wikibon cube insights powered by etr in this breaking analysis we'll share some some fresh july survey data that indicates accelerating momentum for the largest cloud computing firms importantly not only is the momentum broad-based but it's also notable in key strategic sectors namely ai and database there seems to be no stopping the cloud momentum there's certainly plenty of buzz about the cloud tax so-called cloud tax but other than wildly assumptive valuation models and some pockets of anecdotal evidence you don't really see the supposed backlash impacting cloud momentum our forecast calls for the big four hyperscalers aws azure alibaba and gcp to surpass 115 billion as we said in is revenue this year the latest etr survey results show that aws lambda has retaken the lead among all major cloud services tracked in the data set as measured in spending momentum this is the service with the most elevated scores azure overall azure functions vmware cloud on aws and aws overall also demonstrate very highly elevated performance all above that of gcp now impressively aws momentum in the all-important fortune 500 where it has always showed strength is also accelerating one concern in the most recent survey data is that the on-prem clouds and so-called hybrid platforms which we had previously reported as showing an upward spending trajectory seem to have cooled off a bit but the data is mixed and it's a little bit too early to draw firm conclusions nonetheless while hyperscalers are holding steady the spending data appears to be somewhat tepid for the on-prem players you know particularly for their cloud we'll study that further after etr drops its full results on july 23rd now turning our attention back to aws the aws cloud is showing strength across its entire portfolio and we're going to show you that shortly in particular we see notable strength relative to others in analytics ai and the all-important database category aurora and redshift are particularly strong but several other aws database services are showing elevated spending velocity which we'll quantify in a moment all that said snowflake continues to lead all database suppliers in spending momentum by a wide margin which again will quantify in this episode but before we dig into the survey let's take a look at our latest projections for the big four hyperscalers in is as you know we track quarterly revenues for the hyperscalers remember aws and alibaba ias data is pretty clean and reported in their respective earnings reports azure and gcp we have to extrapolate and strip out all a lot of the the apps and other certain revenue to make an apples-to-apples comparison with aws and alibaba and as you can see we have the 2021 market exceeding 115 billion dollars worldwide that's a torrid 35 growth rate on top of 41 in 2020 relative to 2019. aggressive yes but the data continues to point us in this direction until we see some clearer headwinds for the cloud players this is the call we're making aws is perhaps losing a sharepoint or so but it's also is so large that its annual incremental revenue is comparable to alibaba's and google's respective cloud business in total is business in total the big three u.s cloud companies all report at the end of july while alibaba is mid mid-august so we'll update these figures at that time okay let's move on and dig into the survey data we don't have the data yet on alibaba and we're limited as to what we can share until etr drops its research update on on the 23rd but here's a look at the net score timeline in the fortune 500 specifically so we filter the fortune 500 for cloud computing you got azure and the yellow aws and the black and gcp in blue so two points here stand out first is that aws and microsoft are converging and remember the customers who respond to the survey they probably include a fair amount of application software spending in their cloud answers so it favors microsoft in that respect and gcp second point is showing notable deceleration relative to the two leaders and the green call out is because this cut is from an aws point of view so in other words gcp declines are a positive for aws so that's how it should be interpreted now let's take a moment to better understand the idea of net score this is one of the fundamental metrics of the etr methodology here's the data for aws so we use that as a as a reference point net score is calculated by asking customers if they're adding a platform new that's the lime green bar that you see here in the current survey they're asking are you spending six percent or more in the second half relative to the first half of the year that's the forest green they're also asking is spending flat that's the gray or are you spending less that's the pink or are you replacing the platform i.e repatriating so not much spending going on in replacements now in fairness one percent of aws is half a billion dollars so i can see where some folks would get excited about that but in the grand scheme of things it's a sliver so again we don't see repatriation in the numbers okay back to net score subtract the reds from the greens and you get net score which in the case of aws is 61 now just for reference my personal subjective elevated net score level is 40 so anything above that is really impressive based on my experience and to have a company of this size be so elevated is meaningful same for microsoft by the way which is consistently well above the 50 mark in net score in the etr surveys so that's you can think about it that's even more impressive perhaps than aws because it's triple the revenue okay let's stay with aws and take a look at the portfolio and the strength across the board this chart shows net score for the past three surveys serverless is on fire by the way not just aws but azure and gcp functions as well but look at the aws portfolio every category is well above the 40 percent elevated red line the only exception is chime and even chime is showing an uptick and chime is meh if you've ever used chime every other category is well above 50 percent next net score very very strong for aws now as we've frequently reported ai is one of the four biggest focus areas from a spending standpoint along with cloud containers and rpa so it stands to reason that the company with the best ai and ml and the greatest momentum in that space has an advantage because ai is being embedded into apps data processes machines everywhere this chart compares the ai players on two dimensions net score on the vertical axis and market share or presence in the data set on the horizontal axis for companies with more than 15 citations in the survey aws has the highest net score and what's notable is the presence on the horizontal axis databricks is a company where high on also shows elevated scores above both google and microsoft who are showing strength in their own right and then you can see data iq data robot anaconda and salesforce with einstein all above that 40 percent mark and then below you can see the position of sap with leonardo ibm watson and oracle which is well below the 40 line all right let's look at at the all-important database category for a moment and we'll first take a look at the aws database portfolio this chart shows the database services in aws's arsenal and breaks down the net score components with the total net score superimposed on top of the bars point one is aurora is highly elevated with a net score above 70 percent that's due to heavy new adoptions redshift is also very strong as are virtually all aws database offerings with the exception of neptune which is the graph database rds dynamodb elastic document db time stream and quantum ledger database all show momentum above that all important 40 line so while a lot of people criticize the fragmentation of the aws data portfolio and their right tool for the right job approach the spending spending metrics tell a story and that that the strategy is working now let's take a look at the microsoft database portfolio there's a story here similar similar to that of aws azure sql and cosmos db microsoft's nosql distributed database are both very highly elevated as are azure database for mysql and mariadb azure cash for redis and azure for cassandra also microsoft is giving look at microsoft's giving customers a lot of options which is kind of interesting you know we've often said that oracle's strategy because we think about oracle they're building the oracle database cloud we've said oracle strategy should be to not just be the cloud for oracle databases but be the cloud for all databases i mean oracle's got a lot of specialty capability there but it looks like microsoft is beating oracle to that punch not that oracle is necessarily going there but we think it should to expand the appeal of its cloud okay last data chart that we'll show and then and then this one looks at database disruption the chart shows how the cloud database companies are doing in ibm oracle teradata in cloudera accounts the bars show the net score granularity as we described earlier and the etr callouts are interesting so first remember this is an aws this is in an aws context so with 47 responses etr rightly indicates that aws is very well positioned in these accounts with the 68 net score but look at snowflake it has an 81 percent net score which is just incredible and you can see google database is also very strong and the high 50 percent range while microsoft even though it's above the 40 percent mark is noticeably lower than the others as is mongodb with presumably atlas which is surprisingly low frankly but back to snowflake so the etr callout stresses that snowflake doesn't have a strong as strong a presence in the legacy database vendor accounts yet now i'm not sure i would put cloudair in the legacy database category but okay whatever cloudera they're positioning cdp is a hybrid platform as are all the on-prem players with their respective products and platforms but it's going to be interesting to see because snowflake has flat out said it's not straddling the cloud and on-prem rather it's all in on cloud but there is a big opportunity to connect on-prem to the cloud and across clouds which snowflake is pursuing that that ladder the cross-cloud the multi-cloud and snowflake is betting on incremental use cases that involve data sharing and federated governance while traditional players they're protecting their turf at the same time trying to compete in cloud native and of course across cloud i think there's room for both but clearly as we've shown cloud has the spending velocity and a tailwind at its back and aws along with microsoft seem to be getting stronger especially in the all-important categories related to machine intelligence ai and database now to be an essential infrastructure technology player in the data era it would seem obvious that you have to have database and or data management intellectual property in your portfolio or you're going to be less valuable to customers and investors okay we're going to leave it there for today remember these episodes they're all available as podcasts wherever you listen all you do is search breaking analysis podcast and please subscribe to the series check out etr's website at etr dot plus plus etr plus we also publish a full report every week on wikibon.com and siliconangle.com you can get in touch with me david.velante at siliconangle.com you can dm me at d vallante or you can hit hit me up on our linkedin post this is dave vellante for the cube insights powered by etr have a great week stay safe be well and we'll see you next time you

Published Date : Jul 16 2021

SUMMARY :

that the company with the best ai and ml

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
alibabaORGANIZATION

0.99+

six percentQUANTITY

0.99+

81 percentQUANTITY

0.99+

2020DATE

0.99+

2021DATE

0.99+

2019DATE

0.99+

40 percentQUANTITY

0.99+

july 23rdDATE

0.99+

microsoftORGANIZATION

0.99+

115 billionQUANTITY

0.99+

dave vellantePERSON

0.99+

50 percentQUANTITY

0.99+

41QUANTITY

0.99+

61QUANTITY

0.99+

47 responsesQUANTITY

0.99+

bostonLOCATION

0.99+

one percentQUANTITY

0.99+

second halfQUANTITY

0.99+

awsORGANIZATION

0.99+

40QUANTITY

0.99+

two leadersQUANTITY

0.99+

second pointQUANTITY

0.99+

115 billion dollarsQUANTITY

0.99+

firstQUANTITY

0.99+

half a billion dollarsQUANTITY

0.99+

more than 15 citationsQUANTITY

0.98+

mid mid-augustDATE

0.98+

two pointsQUANTITY

0.98+

googleORGANIZATION

0.98+

siliconangle.comOTHER

0.98+

end of julyDATE

0.98+

david.velantePERSON

0.97+

julyDATE

0.97+

50QUANTITY

0.97+

40 percentQUANTITY

0.97+

this yearDATE

0.97+

bothQUANTITY

0.96+

oracleORGANIZATION

0.95+

sqlTITLE

0.95+

mysqlTITLE

0.95+

first halfQUANTITY

0.95+

palo altoORGANIZATION

0.95+

pandemicEVENT

0.95+

35QUANTITY

0.94+

this weekDATE

0.93+

etrORGANIZATION

0.93+

four biggest focus areasQUANTITY

0.91+

aws azureORGANIZATION

0.91+

azureORGANIZATION

0.91+

oneQUANTITY

0.91+

23rdDATE

0.9+

40 lineQUANTITY

0.89+

Breaking Analysis: How JPMC is Implementing a Data Mesh Architecture on the AWS Cloud


 

>> From theCUBE studios in Palo Alto and Boston, bringing you data-driven insights from theCUBE and ETR. This is braking analysis with Dave Vellante. >> A new era of data is upon us, and we're in a state of transition. You know, even our language reflects that. We rarely use the phrase big data anymore, rather we talk about digital transformation or digital business, or data-driven companies. Many have come to the realization that data is a not the new oil, because unlike oil, the same data can be used over and over for different purposes. We still use terms like data as an asset. However, that same narrative, when it's put forth by the vendor and practitioner communities, includes further discussions about democratizing and sharing data. Let me ask you this, when was the last time you wanted to share your financial assets with your coworkers or your partners or your customers? Hello everyone, and welcome to this week's Wikibon Cube Insights powered by ETR. In this breaking analysis, we want to share our assessment of the state of the data business. We'll do so by looking at the data mesh concept and how a leading financial institution, JP Morgan Chase is practically applying these relatively new ideas to transform its data architecture. Let's start by looking at what is the data mesh. As we've previously reported many times, data mesh is a concept and set of principles that was introduced in 2018 by Zhamak Deghani who's director of technology at ThoughtWorks, it's a global consultancy and software development company. And she created this movement because her clients, who were some of the leading firms in the world had invested heavily in predominantly monolithic data architectures that had failed to deliver desired outcomes in ROI. So her work went deep into trying to understand that problem. And her main conclusion that came out of this effort was the world of data is distributed and shoving all the data into a single monolithic architecture is an approach that fundamentally limits agility and scale. Now a profound concept of data mesh is the idea that data architectures should be organized around business lines with domain context. That the highly technical and hyper specialized roles of a centralized cross functional team are a key blocker to achieving our data aspirations. This is the first of four high level principles of data mesh. So first again, that the business domain should own the data end-to-end, rather than have it go through a centralized big data technical team. Second, a self-service platform is fundamental to a successful architectural approach where data is discoverable and shareable across an organization and an ecosystem. Third, product thinking is central to the idea of data mesh. In other words, data products will power the next era of data success. And fourth data products must be built with governance and compliance that is automated and federated. Now there's lot more to this concept and there are tons of resources on the web to learn more, including an entire community that is formed around data mesh. But this should give you a basic idea. Now, the other point is that, in observing Zhamak Deghani's work, she is deliberately avoided discussions around specific tooling, which I think has frustrated some folks because we all like to have references that tie to products and tools and companies. So this has been a two-edged sword in that, on the one hand it's good, because data mesh is designed to be tool agnostic and technology agnostic. On the other hand, it's led some folks to take liberties with the term data mesh and claim mission accomplished when their solution, you know, maybe more marketing than reality. So let's look at JP Morgan Chase in their data mesh journey. Is why I got really excited when I saw this past week, a team from JPMC held a meet up to discuss what they called, data lake strategy via data mesh architecture. I saw that title, I thought, well, that's a weird title. And I wondered, are they just taking their legacy data lakes and claiming they're now transformed into a data mesh? But in listening to the presentation, which was over an hour long, the answer is a definitive no, not at all in my opinion. A gentleman named Scott Hollerman organized the session that comprised these three speakers here, James Reid, who's a divisional CIO at JPMC, Arup Nanda who is a technologist and architect and Serita Bakst who is an information architect, again, all from JPMC. This was the most detailed and practical discussion that I've seen to date about implementing a data mesh. And this is JP Morgan's their approach, and we know they're extremely savvy and technically sound. And they've invested, it has to be billions in the past decade on data architecture across their massive company. And rather than dwell on the downsides of their big data past, I was really pleased to see how they're evolving their approach and embracing new thinking around data mesh. So today, we're going to share some of the slides that they use and comment on how it dovetails into the concept of data mesh that Zhamak Deghani has been promoting, and at least as we understand it. And dig a bit into some of the tooling that is being used by JP Morgan, particularly around it's AWS cloud. So the first point is it's all about business value, JPMC, they're in the money business, and in that world, business value is everything. So Jr Reid, the CIO showed this slide and talked about their overall goals, which centered on a cloud first strategy to modernize the JPMC platform. I think it's simple and sensible, but there's three factors on which he focused, cut costs always short, you got to do that. Number two was about unlocking new opportunities, or accelerating time to value. But I was really happy to see number three, data reuse. That's a fundamental value ingredient in the slide that he's presenting here. And his commentary was all about aligning with the domains and maximizing data reuse, i.e. data is not like oil and making sure there's appropriate governance around that. Now don't get caught up in the term data lake, I think it's just how JP Morgan communicates internally. It's invested in the data lake concept, so they use water analogies. They use things like data puddles, for example, which are single project data marts or data ponds, which comprise multiple data puddles. And these can feed in to data lakes. And as we'll see, JPMC doesn't strive to have a single version of the truth from a data standpoint that resides in a monolithic data lake, rather it enables the business lines to create and own their own data lakes that comprise fit for purpose data products. And they do have a single truth of metadata. Okay, we'll get to that. But generally speaking, each of the domains will own end-to-end their own data and be responsible for those data products, we'll talk about that more. Now the genesis of this was sort of a cloud first platform, JPMC is leaning into public cloud, which is ironic since the early days, in the early days of cloud, all the financial institutions were like never. Anyway, JPMC is going hard after it, they're adopting agile methods and microservices architectures, and it sees cloud as a fundamental enabler, but it recognizes that on-prem data must be part of the data mesh equation. Here's a slide that starts to get into some of that generic tooling, and then we'll go deeper. And I want to make a couple of points here that tie back to Zhamak Deghani's original concept. The first is that unlike many data architectures, this puts data as products right in the fat middle of the chart. The data products live in the business domains and are at the heart of the architecture. The databases, the Hadoop clusters, the files and APIs on the left-hand side, they serve the data product builders. The specialized roles on the right hand side, the DBA's, the data engineers, the data scientists, the data analysts, we could have put in quality engineers, et cetera, they serve the data products. Because the data products are owned by the business, they inherently have the context that is the middle of this diagram. And you can see at the bottom of the slide, the key principles include domain thinking, an end-to-end ownership of the data products. They build it, they own it, they run it, they manage it. At the same time, the goal is to democratize data with a self-service as a platform. One of the biggest points of contention of data mesh is governance. And as Serita Bakst said on the Meetup, metadata is your friend, and she kind of made a joke, she said, "This sounds kind of geeky, but it's important to have a metadata catalog to understand where data resides and the data lineage in overall change management. So to me, this really past the data mesh stink test pretty well. Let's look at data as products. CIO Reid said the most difficult thing for JPMC was getting their heads around data product, and they spent a lot of time getting this concept to work. Here's the slide they use to describe their data products as it related to their specific industry. They set a common language and taxonomy is very important, and you can imagine how difficult that was. He said, for example, it took a lot of discussion and debate to define what a transaction was. But you can see at a high level, these three product groups around wholesale, credit risk, party, and trade and position data as products, and each of these can have sub products, like, party, we'll have to know your customer, KYC for example. So a key for JPMC was to start at a high level and iterate to get more granular over time. So lots of decisions had to be made around who owns the products and the sub-products. The product owners interestingly had to defend why that product should even exist, what boundaries should be in place and what data sets do and don't belong in the various products. And this was a collaborative discussion, I'm sure there was contention around that between the lines of business. And which sub products should be part of these circles? They didn't say this, but tying it back to data mesh, each of these products, whether in a data lake or a data hub or a data pond or data warehouse, data puddle, each of these is a node in the global data mesh that is discoverable and governed. And supporting this notion, Serita said that, "This should not be infrastructure-bound, logically, any of these data products, whether on-prem or in the cloud can connect via the data mesh." So again, I felt like this really stayed true to the data mesh concept. Well, let's look at some of the key technical considerations that JPM discussed in quite some detail. This chart here shows a diagram of how JP Morgan thinks about the problem, and some of the challenges they had to consider were how to write to various data stores, can you and how can you move data from one data store to another? How can data be transformed? Where's the data located? Can the data be trusted? How can it be easily accessed? Who has the right to access that data? These are all problems that technology can help solve. And to address these issues, Arup Nanda explained that the heart of this slide is the data in ingestor instead of ETL. All data producers and contributors, they send their data to the ingestor and the ingestor then registers the data so it's in the data catalog. It does a data quality check and it tracks the lineage. Then, data is sent to the router, which persists the data in the data store based on the best destination as informed by the registration. This is designed to be a flexible system. In other words, the data store for a data product is not fixed, it's determined at the point of inventory, and that allows changes to be easily made in one place. The router simply reads that optimal location and sends it to the appropriate data store. Nowadays you see the schema infer there is used when there is no clear schema on right. In this case, the data product is not allowed to be consumed until the schema is inferred, and then the data goes into a raw area, and the inferer determines the schema and then updates the inventory system so that the data can be routed to the proper location and properly tracked. So that's some of the detail of how the sausage factory works in this particular use case, it was very interesting and informative. Now let's take a look at the specific implementation on AWS and dig into some of the tooling. As described in some detail by Arup Nanda, this diagram shows the reference architecture used by this group within JP Morgan, and it shows all the various AWS services and components that support their data mesh approach. So start with the authorization block right there underneath Kinesis. The lake formation is the single point of entitlement and has a number of buckets including, you can see there the raw area that we just talked about, a trusted bucket, a refined bucket, et cetera. Depending on the data characteristics at the data catalog registration block where you see the glue catalog, that determines in which bucket the router puts the data. And you can see the many AWS services in use here, identity, the EMR, the elastic MapReduce cluster from the legacy Hadoop work done over the years, the Redshift Spectrum and Athena, JPMC uses Athena for single threaded workloads and Redshift Spectrum for nested types so they can be queried independent of each other. Now remember very importantly, in this use case, there is not a single lake formation, rather than multiple lines of business will be authorized to create their own lakes, and that creates a challenge. So how can that be done in a flexible and automated manner? And that's where the data mesh comes into play. So JPMC came up with this federated lake formation accounts idea, and each line of business can create as many data producer or consumer accounts as they desire and roll them up into their master line of business lake formation account. And they cross-connect these data products in a federated model. And these all roll up into a master glue catalog so that any authorized user can find out where a specific data element is located. So this is like a super set catalog that comprises multiple sources and syncs up across the data mesh. So again to me, this was a very well thought out and practical application of database. Yes, it includes some notion of centralized management, but much of that responsibility has been passed down to the lines of business. It does roll up to a master catalog, but that's a metadata management effort that seems compulsory to ensure federated and automated governance. As well at JPMC, the office of the chief data officer is responsible for ensuring governance and compliance throughout the federation. All right, so let's take a look at some of the suspects in this world of data mesh and bring in the ETR data. Now, of course, ETR doesn't have a data mesh category, there's no such thing as that data mesh vendor, you build a data mesh, you don't buy it. So, what we did is we use the ETR dataset to select and filter on some of the culprits that we thought might contribute to the data mesh to see how they're performing. This chart depicts a popular view that we often like to share. It's a two dimensional graphic with net score or spending momentum on the vertical axis and market share or pervasiveness in the data set on the horizontal axis. And we filtered the data on sectors such as analytics, data warehouse, and the adjacencies to things that might fit into data mesh. And we think that these pretty well reflect participation that data mesh is certainly not all compassing. And it's a subset obviously, of all the vendors who could play in the space. Let's make a few observations. Now as is often the case, Azure and AWS, they're almost literally off the charts with very high spending velocity and large presence in the market. Oracle you can see also stands out because much of the world's data lives inside of Oracle databases. It doesn't have the spending momentum or growth, but the company remains prominent. And you can see Google Cloud doesn't have nearly the presence in the dataset, but it's momentum is highly elevated. Remember that red dotted line there, that 40% line, anything over that indicates elevated spending momentum. Let's go to Snowflake. Snowflake is consistently shown to be the gold standard in net score in the ETR dataset. It continues to maintain highly elevated spending velocity in the data. And in many ways, Snowflake with its data marketplace and its data cloud vision and data sharing approach, fit nicely into the data mesh concept. Now, a caution, Snowflake has used the term data mesh in it's marketing, but in our view, it lacks clarity, and we feel like they're still trying to figure out how to communicate what that really is. But is really, we think a lot of potential there to that vision. Databricks is also interesting because the firm has momentum and we expect further elevated levels in the vertical axis in upcoming surveys, especially as it readies for its IPO. The firm has a strong product and managed service, and is really one to watch. Now we included a number of other database companies for obvious reasons like Redis and Mongo, MariaDB, Couchbase and Terradata. SAP as well is in there, but that's not all database, but SAP is prominent so we included them. As is IBM more of a database, traditional database player also with the big presence. Cloudera includes Hortonworks and HPE Ezmeral comprises the MapR business that HPE acquired. So these guys got the big data movement started, between Cloudera, Hortonworks which is born out of Yahoo, which was the early big data, sorry early Hadoop innovator, kind of MapR when it's kind of owned course, and now that's all kind of come together in various forms. And of course, we've got Talend and Informatica are there, they are two data integration companies that are worth noting. We also included some of the AI and ML specialists and data science players in the mix like DataRobot who just did a monster $250 million round. Dataiku, H2O.ai and ThoughtSpot, which is all about democratizing data and injecting AI, and I think fits well into the data mesh concept. And you know we put VMware Cloud in there for reference because it really is the predominant on-prem infrastructure platform. All right, let's wrap with some final thoughts here, first, thanks a lot to the JP Morgan team for sharing this data. I really want to encourage practitioners and technologists, go to watch the YouTube of that meetup, we'll include it in the link of this session. And thank you to Zhamak Deghani and the entire data mesh community for the outstanding work that you're doing, challenging the established conventions of monolithic data architectures. The JPM presentation, it gives you real credibility, it takes Data Mesh well beyond concept, it demonstrates how it can be and is being done. And you know, this is not a perfect world, you're going to start somewhere and there's going to be some failures, the key is to recognize that shoving everything into a monolithic data architecture won't support massive scale and agility that you're after. It's maybe fine for smaller use cases in smaller firms, but if you're building a global platform in a data business, it's time to rethink data architecture. Now much of this is enabled by the cloud, but cloud first doesn't mean cloud only, doesn't mean you'll leave your on-prem data behind, on the contrary, you have to include non-public cloud data in your Data Mesh vision just as JPMC has done. You've got to get some quick wins, that's crucial so you can gain credibility within the organization and grow. And one of the key takeaways from the JP Morgan team is, there is a place for dogma, like organizing around data products and domains and getting that right. On the other hand, you have to remain flexible because technologies is going to come, technology is going to go, so you got to be flexible in that regard. And look, if you're going to embrace the metaphor of water like puddles and ponds and lakes, we suggest maybe a little tongue in cheek, but still we believe in this, that you expand your scope to include data ocean, something John Furry and I have talked about and laughed about extensively in theCUBE. Data oceans, it's huge. It's the new data lake, go transcend data lake, think oceans. And think about this, just as we're evolving our language, we should be evolving our metrics. Much the last the decade of big data was around just getting the stuff to work, getting it up and running, standing up infrastructure and managing massive, how much data you got? Massive amounts of data. And there were many KPIs built around, again, standing up that infrastructure, ingesting data, a lot of technical KPIs. This decade is not just about enabling better insights, it's a more than that. Data mesh points us to a new era of data value, and that requires the new metrics around monetizing data products, like how long does it take to go from data product conception to monetization? And how does that compare to what it is today? And what is the time to quality if the business owns the data, and the business has the context? the quality that comes out of them, out of the shoot should be at a basic level, pretty good, and at a higher mark than out of a big data team with no business context. Automation, AI, and very importantly, organizational restructuring of our data teams will heavily contribute to success in the coming years. So we encourage you, learn, lean in and create your data future. Okay, that's it for now, remember these episodes, they're all available as podcasts wherever you listen, all you got to do is search, breaking analysis podcast, and please subscribe. Check out ETR's website at etr.plus for all the data and all the survey information. We publish a full report every week on wikibon.com and siliconangle.com. And you can get in touch with us, email me david.vellante@siliconangle.com, you can DM me @dvellante, or you can comment on my LinkedIn posts. This is Dave Vellante for theCUBE insights powered by ETR. Have a great week everybody, stay safe, be well, and we'll see you next time. (upbeat music)

Published Date : Jul 12 2021

SUMMARY :

This is braking analysis and the adjacencies to things

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JPMCORGANIZATION

0.99+

Dave VellantePERSON

0.99+

2018DATE

0.99+

Zhamak DeghaniPERSON

0.99+

James ReidPERSON

0.99+

JP MorganORGANIZATION

0.99+

JP MorganORGANIZATION

0.99+

ClouderaORGANIZATION

0.99+

Serita BakstPERSON

0.99+

IBMORGANIZATION

0.99+

HPEORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Scott HollermanPERSON

0.99+

HortonworksORGANIZATION

0.99+

BostonLOCATION

0.99+

40%QUANTITY

0.99+

JP Morgan ChaseORGANIZATION

0.99+

SeritaPERSON

0.99+

YahooORGANIZATION

0.99+

Arup NandaPERSON

0.99+

eachQUANTITY

0.99+

ThoughtWorksORGANIZATION

0.99+

firstQUANTITY

0.99+

OracleORGANIZATION

0.99+

Palo AltoLOCATION

0.99+

david.vellante@siliconangle.comOTHER

0.99+

each lineQUANTITY

0.99+

TerradataORGANIZATION

0.99+

RedisORGANIZATION

0.99+

$250 millionQUANTITY

0.99+

first pointQUANTITY

0.99+

three factorsQUANTITY

0.99+

SecondQUANTITY

0.99+

MapRORGANIZATION

0.99+

todayDATE

0.99+

InformaticaORGANIZATION

0.99+

TalendORGANIZATION

0.99+

John FurryPERSON

0.99+

Zhamak DeghaniPERSON

0.99+

first platformQUANTITY

0.98+

YouTubeORGANIZATION

0.98+

fourthQUANTITY

0.98+

singleQUANTITY

0.98+

OneQUANTITY

0.98+

ThirdQUANTITY

0.97+

CouchbaseORGANIZATION

0.97+

three speakersQUANTITY

0.97+

two dataQUANTITY

0.97+

first strategyQUANTITY

0.96+

oneQUANTITY

0.96+

one placeQUANTITY

0.96+

Jr ReidPERSON

0.96+

single lakeQUANTITY

0.95+

SAPORGANIZATION

0.95+

wikibon.comOTHER

0.95+

siliconangle.comOTHER

0.94+

AzureORGANIZATION

0.93+

Breaking Analysis: Chasing Snowflake in Database Boomtown


 

(upbeat music) >> From theCUBE studios in Palo Alto, in Boston bringing you data-driven insights from theCUBE and ETR. This is braking analysis with Dave Vellante. >> Database is the heart of enterprise computing. The market is both exploding and it's evolving. The major force is transforming the space include Cloud and data, of course, but also new workloads, advanced memory and IO capabilities, new processor types, a massive push towards simplicity, new data sharing and governance models, and a spate of venture investment. Snowflake stands out as the gold standard for operational excellence and go to market execution. The company has attracted the attention of customers, investors, and competitors and everyone from entrenched players to upstarts once in the act. Hello everyone and welcome to this week's Wikibon CUBE Insights powered by ETR. In this breaking analysis, we'll share our most current thinking on the database marketplace and dig into Snowflake's execution. Some of its challenges and we'll take a look at how others are making moves to solve customer problems and try to get a piece of the growing database pie. Let's look at some of the factors that are driving market momentum. First, customers want lower license costs. They want simplicity. They want to avoid database sprawl. They want to run anywhere and manage new data types. These needs often are divergent and they pull vendors and technologies in different direction. It's really hard for any one platform to accommodate every customer need. The market is large and it's growing. Gardner has it at around 60 to 65 billion with a CAGR of somewhere around 20% over the next five years. But the market, as we know it is being redefined. Traditionally, databases have served two broad use cases, OLTP or transactions and reporting like data warehouses. But a diversity of workloads and new architectures and innovations have given rise to a number of new types of databases to accommodate all these diverse customer needs. Many billions have been spent over the last several years in venture money and it continues to pour in. Let me just give you some examples. Snowflake prior to its IPO, raised around 1.4 billion. Redis Labs has raised more than 1/2 billion dollars so far, Cockroach Labs, more than 350 million, Couchbase, 250 million, SingleStore formerly MemSQL, 238 million, Yellowbrick Data, 173 million. And if you stretch the definition of database a little bit to including low-code or no-code, Airtable has raised more than 600 million. And that's by no means a complete list. Now, why is all this investment happening? Well, in a large part, it's due to the TAM. The TAM is huge and it's growing and it's being redefined. Just how big is this market? Let's take a look at a chart that we've shown previously. We use this chart to Snowflakes TAM, and it focuses mainly on the analytics piece, but we'll use it here to really underscore the market potential. So the actual database TAM is larger than this, we think. Cloud and Cloud-native technologies have changed the way we think about databases. Virtually 100% of the database players that they're are in the market have pivoted to a Cloud first strategy. And many like Snowflake, they're pretty dogmatic and have a Cloud only strategy. Databases has historically been very difficult to manage, they're really sensitive to latency. So that means they require a lot of tuning. Cloud allows you to throw virtually infinite resources on demand and attack performance problems and scale very quickly, minimizing the complexity and tuning nuances. This idea, this layer of data as a service we think of it as a staple of digital transformation. Is this layer that's forming to support things like data sharing across ecosystems and the ability to build data products or data services. It's a fundamental value proposition of Snowflake and one of the most important aspects of its offering. Snowflake tracks a metric called edges, which are external connections in its data Cloud. And it claims that 15% of its total shared connections are edges and that's growing at 33% quarter on quarter. This notion of data sharing is changing the way people think about data. We use terms like data as an asset. This is the language of the 2010s. We don't share our assets with others, do we? No, we protect them, we secure or them, we even hide them. But we absolutely don't want to share those assets but we do want to share our data. I had a conversation recently with Forrester analyst, Michelle Goetz. And we both agreed we're going to scrub data as an asset from our phrasiology. Increasingly, people are looking at sharing as a way to create, as I said, data products or data services, which can be monetized. This is an underpinning of Zhamak Dehghani's concept of a data mesh, make data discoverable, shareable and securely governed so that we can build data products and data services that can be monetized. This is where the TAM just explodes and the market is redefining. And we think is in the hundreds of billions of dollars. Let's talk a little bit about the diversity of offerings in the marketplace. Again, databases used to be either transactional or analytic. The bottom lines and top lines. And this chart here describe those two but the types of databases, you can see the middle of mushrooms, just looking at this list, blockchain is of course a specialized type of database and it's also finding its way into other database platforms. Oracle is notable here. Document databases that support JSON and graph data stores that assist in visualizing data, inference from multiple different sources. That's is one of the ways in which adtech has taken off and been so effective. Key Value stores, log databases that are purpose-built, machine learning to enhance insights, spatial databases to help build the next generation of products, the next automobile, streaming databases to manage real time data flows and time series databases. We might've missed a few, let us know if you think we have, but this is a kind of pretty comprehensive list that is somewhat mind boggling when you think about it. And these unique requirements, they've spawned tons of innovation and companies. Here's a small subset on this logo slide. And this is by no means an exhaustive list, but you have these companies here which have been around forever like Oracle and IBM and Teradata and Microsoft, these are the kind of the tier one relational databases that have matured over the years. And they've got properties like atomicity, consistency, isolation, durability, what's known as ACID properties, ACID compliance. Some others that you may or may not be familiar with, Yellowbrick Data, we talked about them earlier. It's going after the best price, performance and analytics and optimizing to take advantage of both hybrid installations and the latest hardware innovations. SingleStore, as I said, formerly known as MemSQL is a very high end analytics and transaction database, supports mixed workloads, extremely high speeds. We're talking about trillions of rows per second that could be ingested in query. Couchbase with hybrid transactions and analytics, Redis Labs, open source, no SQL doing very well, as is Cockroach with distributed SQL, MariaDB with its managed MySQL, Mongo and document database has a lot of momentum, EDB, which supports open source Postgres. And if you stretch the definition a bit, Splunk, for log database, why not? ChaosSearch, really interesting startup that leaves data in S-3 and is going after simplifying the ELK stack, New Relic, they have a purpose-built database for application performance management and we probably could have even put Workday in the mix as it developed a specialized database for its apps. Of course, we can't forget about SAP with how not trying to pry customers off of Oracle. And then the big three Cloud players, AWS, Microsoft and Google with extremely large portfolios of database offerings. The spectrum of products in this space is very wide, with you've got AWS, which I think we're up to like 16 database offerings, all the way to Oracle, which has like one database to do everything not withstanding MySQL because it owns MySQL got that through the Sun Acquisition. And it recently, it made some innovations there around the heat wave announcement. But essentially Oracle is investing to make its database, Oracle database run any workload. While AWS takes the approach of the right tool for the right job and really focuses on the primitives for each database. A lot of ways to skin a cat in this enormous and strategic market. So let's take a look at the spending data for the names that make it into the ETR survey. Not everybody we just mentioned will be represented because they may not have quite the market presence of the ends in the survey, but ETR that capture a pretty nice mix of players. So this chart here, it's one of the favorite views that we like to share quite often. It shows the database players across the 1500 respondents in the ETR survey this past quarter and it measures their net score. That's spending momentum and is shown on the vertical axis and market share, which is the pervasiveness in the data set is on the horizontal axis. The Snowflake is notable because it's been hovering around 80% net score since the survey started picking them up. Anything above 40%, that red line there, is considered by us to be elevated. Microsoft and AWS, they also stand out because they have both market presence and they have spending velocity with their platforms. Oracle is very large but it doesn't have the spending momentum in the survey because nearly 30% of Oracle installations are spending less, whereas only 22% are spending more. Now as a caution, this survey doesn't measure dollar spent and Oracle will be skewed toward the big customers with big budgets. So you got to consider that caveat when evaluating this data. IBM is in a similar position although its market share is not keeping up with Oracle's. Google, they've got great tech especially with BigQuery and it has elevated momentum. So not a bad spot to be in although I'm sure it would like to be closer to AWS and Microsoft on the horizontal axis, so it's got some work to do there. And some of the others we mentioned earlier, like MemSQL, Couchbase. As shown MemSQL here, they're now SingleStore. Couchbase, Reddis, Mongo, MariaDB, all very solid scores on the vertical axis. Cloudera just announced that it was selling to private equity and that will hopefully give it some time to invest in this platform and get off the quarterly shot clock. MapR was acquired by HPE and it's part of HPE's Ezmeral platform, their data platform which doesn't yet have the market presence in the survey. Now, something that is interesting in looking at in Snowflakes earnings last quarter, is this laser focused on large customers. This is a hallmark of Frank Slootman and Mike Scarpelli who I know they don't have a playbook but they certainly know how to go whale hunting. So this chart isolates the data that we just showed you to the global 1000. Note that both AWS and Snowflake go up higher on the X-axis meaning large customers are spending at a faster rate for these two companies. The previous chart had an end of 161 for Snowflake, and a 77% net score. This chart shows the global 1000, in the end there for Snowflake is 48 accounts and the net score jumps to 85%. We're not going to show it here but when you isolate the ETR data, nice you can just cut it, when you isolate it on the fortune 1000, the end for Snowflake goes to 59 accounts in the data set and Snowflake jumps another 100 basis points in net score. When you cut the data by the fortune 500, the Snowflake N goes to 40 accounts and the net score jumps another 200 basis points to 88%. And when you isolate on the fortune 100 accounts is only 18 there but it's still 18, their net score jumps to 89%, almost 90%. So it's very strong confirmation that there's a proportional relationship between larger accounts and spending momentum in the ETR data set. So Snowflakes large account strategy appears to be working. And because we think Snowflake is sticky, this probably is a good sign for the future. Now we've been talking about net score, it's a key measure in the ETR data set, so we'd like to just quickly remind you what that is and use Snowflake as an example. This wheel chart shows the components of net score, that lime green is new adoptions. 29% of the customers in the ETR dataset that are new to Snowflake. That's pretty impressive. 50% of the customers are spending more, that's the forest green, 20% are flat, that's the gray, and only 1%, the pink, are spending less. And 0% zero or replacing Snowflake, no defections. What you do here to get net scores, you subtract the red from the green and you get a net score of 78%. Which is pretty sick and has been sick as in good sick and has been steady for many, many quarters. So that's how the net score methodology works. And remember, it typically takes Snowflake customers many months like six to nine months to start consuming it's services at the contracted rate. So those 29% new adoptions, they're not going to kick into high gear until next year, so that bodes well for future revenue. Now, it's worth taking a quick snapshot at Snowflakes most recent quarter, there's plenty of stuff out there that you can you can google and get a summary but let's just do a quick rundown. The company's product revenue run rate is now at 856 million they'll surpass $1 billion on a run rate basis this year. The growth is off the charts very high net revenue retention. We've explained that before with Snowflakes consumption pricing model, they have to account for retention differently than what a SaaS company. Snowflake added 27 net new $1 million accounts in the quarter and claims to have more than a hundred now. It also is just getting its act together overseas. Slootman says he's personally going to spend more time in Europe, given his belief, that the market is huge and they can disrupt it and of course he's from the continent. He was born there and lived there and gross margins expanded, do in a large part to renegotiation of its Cloud costs. Welcome back to that in a moment. Snowflake it's also moving from a product led growth company to one that's more focused on core industries. Interestingly media and entertainment is one of the largest along with financial services and it's several others. To me, this is really interesting because Disney's example that Snowflake often puts in front of its customers as a reference. And it seems to me to be a perfect example of using data and analytics to both target customers and also build so-called data products through data sharing. Snowflake has to grow its ecosystem to live up to its lofty expectations and indications are that large SIS are leaning in big time. Deloitte cross the $100 million in deal flow in the quarter. And the balance sheet's looking good. Thank you very much with $5 billion in cash. The snarks are going to focus on the losses, but this is all about growth. This is a growth story. It's about customer acquisition, it's about adoption, it's about loyalty and it's about lifetime value. Now, as I said at the IPO, and I always say this to young people, don't buy a stock at the IPO. There's probably almost always going to be better buying opportunities ahead. I'm not always right about that, but I often am. Here's a chart of Snowflake's performance since IPO. And I have to say, it's held up pretty well. It's trading above its first day close and as predicted there were better opportunities than day one but if you have to make a call from here. I mean, don't take my stock advice, do your research. Snowflake they're priced to perfection. So any disappointment is going to be met with selling. You saw that the day after they beat their earnings last quarter because their guidance in revenue growth,. Wasn't in the triple digits, it sort of moderated down to the 80% range. And they pointed, they pointed to a new storage compression feature that will lower customer costs and consequently, it's going to lower their revenue. I swear, I think that that before earnings calls, Scarpelli sits back he's okay, what kind of creative way can I introduce the dampen enthusiasm for the guidance. Now I'm not saying lower storage costs will translate into lower revenue for a period of time. But look at dropping storage prices, customers are always going to buy more, that's the way the storage market works. And stuff like did allude to that in all fairness. Let me introduce something that people in Silicon Valley are talking about, and that is the Cloud paradox for SaaS companies. And what is that? I was a clubhouse room with Martin Casado of Andreessen when I first heard about this. He wrote an article with Sarah Wang, calling it to question the merits of SaaS companies sticking with Cloud at scale. Now the basic premise is that for startups in early stages of growth, the Cloud is a no brainer for SaaS companies, but at scale, the cost of Cloud, the Cloud bill approaches 50% of the cost of revenue, it becomes an albatross that stifles operating leverage. Their conclusion ended up saying that as much as perhaps as much as the back of the napkin, they admitted that, but perhaps as much as 1/2 a trillion dollars in market cap is being vacuumed away by the hyperscalers that could go to the SaaS providers as cost savings from repatriation. And that Cloud repatriation is an inevitable path for large SaaS companies at scale. I was particularly interested in this as I had recently put on a post on the Cloud repatriation myth. I think in this instance, there's some merit to their conclusions. But I don't think it necessarily bleeds into traditional enterprise settings. But for SaaS companies, maybe service now has it right running their own data centers or maybe a hybrid approach to hedge bets and save money down the road is prudent. What caught my attention in reading through some of the Snowflake docs, like the S-1 in its most recent 10-K were comments regarding long-term purchase commitments and non-cancelable contracts with Cloud companies. And the companies S-1, for example, there was disclosure of $247 million in purchase commitments over a five plus year period. And the company's latest 10-K report, that same line item jumped to 1.8 billion. Now Snowflake is clearly managing these costs as it alluded to when its earnings call. But one has to wonder, at some point, will Snowflake follow the example of say Dropbox which Andreessen used in his blog and start managing its own IT? Or will it stick with the Cloud and negotiate hard? Snowflake certainly has the leverage. It has to be one of Amazon's best partners and customers even though it competes aggressively with Redshift but on the earnings call, CFO Scarpelli said, that Snowflake was working on a new chip technology to dramatically increase performance. What the heck does that mean? Is this Snowflake is not becoming a hardware company? So I going to have to dig into that a little bit and find out what that it means. I'm guessing, it means that it's taking advantage of ARM-based processes like graviton, which many ISVs ar allowing their software to run on that lower cost platform. Or maybe there's some deep dark in the weeds secret going on inside Snowflake, but I doubt it. We're going to leave all that for there for now and keep following this trend. So it's clear just in summary that Snowflake they're the pace setter in this new exciting world of data but there's plenty of room for others. And they still have a lot to prove. For instance, one customer in ETR, CTO round table express skepticism that Snowflake will live up to its hype because its success is going to lead to more competition from well-established established players. This is a common theme you hear it all the time. It's pretty easy to reach that conclusion. But my guess is this the exact type of narrative that fuels Slootman and sucked him back into this game of Thrones. That's it for now, everybody. Remember, these episodes they're all available as podcasts, wherever you listen. All you got to do is search braking analysis podcast and please subscribe to series. Check out ETR his website at etr.plus. We also publish a full report every week on wikinbon.com and siliconangle.com. You can get in touch with me, Email is David.vellante@siliconangle.com. You can DM me at DVelante on Twitter or comment on our LinkedIn posts. This is Dave Vellante for theCUBE Insights powered by ETR. Have a great week everybody, be well and we'll see you next time. (upbeat music)

Published Date : Jun 5 2021

SUMMARY :

This is braking analysis and the net score jumps to 85%.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Michelle GoetzPERSON

0.99+

AWSORGANIZATION

0.99+

Mike ScarpelliPERSON

0.99+

Dave VellantePERSON

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

Sarah WangPERSON

0.99+

AmazonORGANIZATION

0.99+

50%QUANTITY

0.99+

Palo AltoLOCATION

0.99+

AndreessenPERSON

0.99+

EuropeLOCATION

0.99+

40 accountsQUANTITY

0.99+

$1 billionQUANTITY

0.99+

Frank SlootmanPERSON

0.99+

SlootmanPERSON

0.99+

OracleORGANIZATION

0.99+

Redis LabsORGANIZATION

0.99+

ScarpelliPERSON

0.99+

TAMORGANIZATION

0.99+

sixQUANTITY

0.99+

33%QUANTITY

0.99+

$5 billionQUANTITY

0.99+

80%QUANTITY

0.99+

GoogleORGANIZATION

0.99+

1.8 billionQUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

59 accountsQUANTITY

0.99+

Cockroach LabsORGANIZATION

0.99+

DisneyORGANIZATION

0.99+

TeradataORGANIZATION

0.99+

18QUANTITY

0.99+

77%QUANTITY

0.99+

85%QUANTITY

0.99+

29%QUANTITY

0.99+

20%QUANTITY

0.99+

BostonLOCATION

0.99+

78%QUANTITY

0.99+

Martin CasadoPERSON

0.99+

48 accountsQUANTITY

0.99+

856 millionQUANTITY

0.99+

1500 respondentsQUANTITY

0.99+

nine monthsQUANTITY

0.99+

Zhamak DehghaniPERSON

0.99+

0%QUANTITY

0.99+

wikinbon.comOTHER

0.99+

88%QUANTITY

0.99+

twoQUANTITY

0.99+

$100 millionQUANTITY

0.99+

89%QUANTITY

0.99+

AirtableORGANIZATION

0.99+

next yearDATE

0.99+

SnowflakeORGANIZATION

0.99+

two companiesQUANTITY

0.99+

DeloitteORGANIZATION

0.99+

200 basis pointsQUANTITY

0.99+

FirstQUANTITY

0.99+

HPEORGANIZATION

0.99+

15%QUANTITY

0.99+

more than 600 millionQUANTITY

0.99+

last quarterDATE

0.99+

161QUANTITY

0.99+

David.vellante@siliconangle.comOTHER

0.99+

$247 millionQUANTITY

0.99+

27 netQUANTITY

0.99+

2010sDATE

0.99+

siliconangle.comOTHER

0.99+

ForresterORGANIZATION

0.99+

MemSQLTITLE

0.99+

Yellowbrick DataORGANIZATION

0.99+

more than 1/2 billion dollarsQUANTITY

0.99+

DropboxORGANIZATION

0.99+

MySQLTITLE

0.99+

BigQueryTITLE

0.99+

Bob Wise, AWS & Peder Ulander, AWS | Red Hat Summit 2021 Virtual Experience


 

(smart gentle music) >> Hey, welcome back everyone to theCUBE's coverage of Red Hat Summit 2021 virtual. I'm John Furrier, host of theCUBE, got two great guests here from AWS, Bob Wise, General Manager of Kubernetes for Amazon Web Services and Peder Ulander, Head of product marketing for the enterprise developer and open-source at AWS. Gentlemen, you guys are the core leaders in the AWS open-source initiatives. Thanks for joining us on theCUBE here for Red Hat Summit. >> Thanks for having us, John. >> Good to be here. >> So the innovation that's come from people building on top of the cloud has just been amazing. You guys, props to Amazon Web Services for constantly adding more and raising the bar on more services every year. You guys do that, and now public cloud has become so popular, and so important that now Hybrid has pushed the Edge. You got outpost with Amazon you see everyone following suit. It's pretty much clear vote of confidence from the customers that, Hybrid is the operating model of the future. And that really is about the Edge. So I want to chat with you about the open-source intersection there, so let's get into it. So we're here at Red Hat Summit. So Red Hat's an open-source company and timing is great for them. Now, part of IBM you guys have had a relationship with Red Hat for some time. Can you tell us about the partnership and how it's working together? >> Yeah, absolutely. Why don't I take that one? AWS and Red Hat have been strategic partners since, shoot, I think it's 2008 or so in the early days of AWS, when engaging with customers, we wanted to ensure that AWS was the best place for enterprises to run their Red Hat workloads. And this is super important when you think about, what Red Hat has accomplished with RHEL in the enterprise, it's running SAP, it's running Oracle's, it's running all different types of core business applications, as well as a lot of the new things that customers are innovating. And so having that relationship to ensure that not only did it work on AWS, but it actually scaled we had integration of services, we had the performance, the price all of the things that were so critical to customers was critical from day one. And we continue to evolve this relationship over time. As you see us coming into Red Hat Summit this year. >> Well, again, to the hard news here also the new service Red Hat OpenShift servers on AWS known as ROSA, the A for Amazon Red Hat OpenShift, A for Amazon Web Services, a clever acronym but really it's on AWS. What exactly is this service? What does it do? And who is it designed for? >> Well, I'll let me jump in on this one. Maybe let's start with the why? Why ROSA? Customers love using OpenShift, but they also want to use AWS. They want the best of both. So they want their peanut butter and their chocolate together in a single confection. A lot of those customers have deployed AWS, have deployed OpenShift on AWS. They want managed service simplified supply chain. We want to be able to streamline moving on premises, OpenShift workloads to AWS, naturally want good integration with AWS services. So as to the, what? Our new service jointly operated is supported by Red Hat and AWS to provide a fully managed to OpenShifts on AWS. So again, like lot of customers have been running OpenShift on AWS before this time, but of course they were managing it themselves typically. And so now they get a fully managed option with also simplified supply chain. Single support channels, single billing. >> You know, were talking before we came on camera about the acronym on AWS and people build on the clouds kind of like it's no big deal to say that, but I know it means something. I want to explain, you guys to explain this on because I know I've been scolded saying things on theCUBE that were kind of misspoken because it's easy to say, Oh yeah, I built that app. We built all this stuff on theCUBE was on AWS, but it's not on AWS. It means something from a designation standpoint what does on AWS mean? 'Cause this is OpenShift servers on AWS, we see this other companies have their products on AWS. This is specific designation. Can you share, please. >> John, when you see the branding of something like Red Hat on AWS, what that basically signals to our customers is that this is joint engineering work. This is the top of the strategic partners where we actually do a lot of joint engineering and work to make sure that we're driving the right integrations and the right experience, make sure that these things are accessible and discoverable in our console. They're treated effectively as a first-class service inside of the AWS ecosystem. So it's, there's not many of the on's, if you will. You think about SAP on VMware cloud, on AWS, and now Red Hat OpenShift on AWS, it really is that signal that helps give customers the confidence of tested, tried, trued, supported and validated service on top of AWS. And we think that's significantly better than anything else. It's easy to run an image on a VM and stuffed it into a cloud service to make it available, but customers want better, customer want tighter experiences. They want to be able to take advantage of all the great things that we have from a scale availability and performance perspective. And that's really what we're pushing towards. >> Yeah. I've seen examples specifically where when partners work with Amazon at that level of joint engineering, deeper partnerships. The results were pretty significant on the business side. So congratulations to you guys working with OpenShift and Red Hat, that's real testament to their product. But I got to ask you guys, pull the Amazon playbook out and challenge you guys, or just, create a new some commentary around the process of working backwards. Every time I talked to Andy Jassy, he always says, we work backwards from the customer and we get the requirements, and we're listening to customers. Okay, great. He loves that, he loves to say that it's true. I know that I've seen that. What is the customer work backwards document look like here? What is the, what was the need and what made this become such an important part of AWS? What was the, and then what are they saying now, now that the products out there? >> Well, OpenShift has a very wide footprint as does AWS. Some working backwards documents kind of write themselves, because now the customer demand is so strong that there's just no avoiding it. Now, it really just becomes about making sure you have a good plan so it becomes much more operational at that point. ROSA's definitely one of those services. We had so much demand and as a result, no surprise that we're getting a lot of enthusiasm for customers because so many of them asked us for it. (crosstalk) >> What's been the reaction in asking demand. That's kind of got the sense of that, but okay. So there's demand now, what's the what's the use cases? What are customers saying? What's the reaction been? >> Lot of the use cases are these Hybrid kind of use cases where a customer has a big OpenShift footprint. What we see from a lot of these customers is a strong demand for consistency in order to reduce IT sprawl. What they really want to do is have the smallest number of simplest environments they can. And so when customers that standardized on OpenShift really wants to be able to standardize OpenShifts, both in their on premises environment and on AWS and get managed service options just to remove the undifferentiated heavy lifting. >> Hey, what's your take on the product marketing side of this, where you got open-source becoming very enterprise specific, Red Hat's been there for a very long time. I've been user of Red Hat since the beginning and following them, and Linux, obviously is Linux where that's come from. But what features specifically jump out in this offering that customers are resonating around? What's the vibe here? >> John, you kind of alluded to it early on, which is I don't know that I'd necessarily call it Hybrid but the reality is our customers have environments that are on premises in the cloud and all the way out to the Edge. Today, when you think of a lot of solutions and services, it's a fractured experience that they have between those three locations. And one of our biggest commitments to our customers, just to make things super simple, remove the complexity do all of the hard work, which means, customers are looking for a consistent experience environment and tooling that spans data center to cloud, to Edge. And that's probably the biggest kind of core asset here for customers who might have standardized on OpenShift in the data centers. They come to the cloud, they want to continue to leverage those skills. I think probably one of the, an interesting one is we headed down in this path, we all know Delta Airlines. Delta is a great example of a customer who, joint customer, who have been doing stuff inside of AWS for a long time. They've been standardizing on Red Hat for a long time and bringing this together just gave them that simple extension to take their investment in Red Hat OpenShift and leverage their experience. And again, the scale and performance of what AWS brings them. >> Next question, what's next for a Red Hat OpenShift on AWS in your work with Red Hat. Where does this go next? What's the big to-do item, what do you guys see as the vision? >> I'm glad you mentioned open-source collaboration at the start there. We're taking to point out is that AWS works on the Kubernetes project upstream as does the Red Hat teams. So one of the ways that we collaborate with the Red Hat team is in open-source. One of those projects is on a new project called ACK. It was on controllers for Kubernetes and this is a kind of Kubernetes friendly way for my customers to use an API to manage AWS services. So that's one of the things that we're looking forward to as that goes GA wobbling out into both ROSA and onto our other services. >> Awesome. I got to ask you guys this while you're here, because it's very rare to get two luminaries within AWS on the open-source side. This has been a huge build-out over the many, many years for AWS, and some people really kind of don't understand kind of the position. So take a minute to clarify the position of AWS on open-source. You guys are very active in a lot of projects. You mentioned upstream with Kubernetes in other areas. I've had many countries with Adrian Cockcroft on this, as well as others within AWS. Huge proponents web services, I mean, you go back to the original Amazon. I mean, Jeff Barr was saying 15 years ago some of those API's are still in play here. API's back in 15 years ago, that was kind of not main stream at that time. So you had open standards, really made Amazon web services successful and you guys are continuing it but as the modern era is very enterprise, like and you see a lot of legacy, you seeing a lot more operations that they're going to be driven by open technologies that you guys are investing in. I'll take a minute to explain what AWS is doing and what you guys care about and your mission? >> Yeah. Well, why don't I start? And then we'll kick it over to Bob 'cause I think Bob can also talk about some of the key contribution sides, but the best way to think about it is kind of in three different pillars. So let's start with the first one, which is, around the fact of ensuring that our customer's favorite open-source projects run best on AWS. Since 2006, we've been helping our customers operationalize their open-source investments and really kind of achieve that scale and focus more on how they use and innovate on the products versus how they set up and run. And for myself being an open-source since the late 90s, the biggest opportunity, yet challenge was the access to the technology, but it still required you as a customer to learn how to set up, configure, operationalized support and sustain. AWS removes that heavy lifting and, again, back to that earlier point from the beginning of AWS, we helped customers scale and implement their Apache services, their database services, all of these different types of open-source projects to make them really work exceptionally well on AWS. And back to that point, make sure that AWS was the best place for their open-source projects. I think the second thing that we do, and you're seeing that today with what we're doing with ROSA and Red Hat is we partner with open-source leaders from Red Hat to Redis and Confluent to a number of different players out there, Grafana, and Prometheus, to even foundations like the LF and the CNCF. We partner with these leaders to ensure that we're working together to grow grow the overall experience and the overall the overall pie, if you will. And this kind of gets into that point you were making John in that, the old world legacy proprietary stuff, there's a huge chance for refresh and new opportunity and rethinking or modernization if you will, as you come into the cloud having the expertise and the partnerships with these key players is as enterprises move in, is so crucial. And then the third piece I'd like to talk about that's important to our open-source strategies is really around contribution. We have a number of projects that we've delivered ourselves. I think the two most recent ones that really come top of mind for me is, what we did with Babel Fish, as well as with OpenSearch. So contributing and driving a true open-source project that helps our customers, take advantage of things like an SQL, a proprietary to open-source SQL conversion tool, or what we're doing to make Elasticsearch, the opportune or the primary open platform for our customers. But it's not just about those services, it's also collaborating with key industry initiatives. Bob's at the forefront of that with what we're doing with the CNCF around things, like Kubernetes and Prometheus et cetera, Bob you want to jump in on some of that? >> Sure, I think the one thing I would add here is that customers love using those open-source projects. The one of the challenges with them frequently is security. And this is job zero to AWS. So a lot of the collaboration work we do, a lot of the work that we do on upstream projects is go specifically around kind of security oriented things because that is what customers expect when they come to get a managed service at AWS. Some of those efforts are somewhat unsung because you generally do more work and less talk, in security oriented things. But projects across AWS, that's always a key contribution focus for us. >> Good way to call out security too. I think that's being built-in to the everything now, that's an operating model. People call it shift-left day two operations. Whatever you want to look at it. You got this nice formation going between under the hood kind of programmability of the infrastructure at scale. And then you have the modern application development which is just beginning, programmable DevSecOps. It's funny, Bob, I'd love to get your take on this because I remember in the 80s and during the Unix generation I used to peddle software under the table. Like, here's a copy of, you just don't tell anyone, people in the younger generation don't get the fact that it wasn't always open. And so now you have open and you have this idea of an enterprise that's going to be a system management system view. So you got engineering and you got computer science kind of coming together, this SRE middle layer. You're hearing that as a, kind of a new discipline. So DevOps kind of has won. I mean, we kind of knew this for many, many years. I said this in 2013 on theCUBE actually at re-inventing. I just recently shared that clip. But okay, now you've got SecOps, DevSecOps. So now you have an era where it's a system thinking and open-source is driving all of that. So can you share your perspective because this is kind of where the puck is going. It's an open to open world. That's going to have to be open and scalable. How does open-source and you guys take it to the next level to give that same scale and reliability? What's your vision? >> The key here is really around automation and what we're seeing you could look at Kubernetes. Kubernetes, is essentially a robot. It was like the early design of it was built around robotics principles. So it's a giant software robot and the world has changed. If you just look at the influx of all kinds of automation to not just the DevOps world but to all industries, you see a similar kind of trend. And so the world of IT operations person is changing from doing the work that the robot did and replacing it with the robot to managing large numbers of robots. And in this case, the robots are like a little early and a little hard to talk to. And so, you end up using languages like YAML and other things, but it turns out robots still just do what you tell them to do. And so one of the things you have to do is be really, really careful because robots will go and do whatever it is you ask them to do. On the other hand, they're really, really good at doing that. So in the security area, they take the research points to the largest single source of security issues, being people making manual mistakes. And a lot of people are still a little bit terrified if human beings aren't touching things on the way to production. In AWS, we're terrified if humans aren't touching it. And that is a super hard chasm to cross and open-source projects have really, are really playing a big role in what's really a IT wide migration to a whole new set of, not just tools, but organizational approaches. >> What's your reaction to that? Because we're talking that essentially software concepts, because if you write bad code, the code will execute what you did. So assuming it compiles left in the old days. Now, if you're going to scale a large scale operations that has dynamic capabilities, services being initiated in terminating tear down up started, you need the automation, but if you really don't design it right, you could be screwed. This is a huge deal. >> This is one reason why we've put so much effort into getops that you can think of it as a more narrowly defined subset of the DevOps world with a specific set of principles around using kind of simplified declarative approaches, along with robots that converge the desired state, converge the system to the desired state. And when you get into large distributed systems, you end up needing to take those kinds of approaches to get it to work at scale. Otherwise you have problems. >> Yeah, just adding to that. And it's funny, you said DevOps has won. I actually think DevOps has won, but DevOps hasn't changed (indistinct) Bob, you were right, the reality is it was founded back what quite a while ago, it was more around CICD in the enterprise and the closed data center. And it was one of those where automation and runbooks took addressed the fact that, every pair of hands between service requests and service delivery recreated or created an issue. So that growth and that mental model of moving from a waterfall, agile to DevOps, you built it, you run it, type of a model, I think is really, really important. But as it comes out into the cloud, you no longer have those controls of the data center and you actually have infinite scale. So back to your point of you got to get this right. You have to architect correctly you have to make sure that your code is good, you have to make sure that you have full visibility. This is where it gets really interesting at AWS. And some of the things that we're tying in. So whether we're talking about getops like what Bob just went through, or what you brought up with DevSecOps, you also have things like, AIOps. And so looking at how we take our machine learning tools to really implement the appropriate types of code reviews to assessing your infrastructure or your choices against well-architected principles and providing automated remediation is key, adding to that is observability, developers, especially in a highly distributed environment need to have better understanding, fidelity and touchpoints of what's going on with our application as it runs in production. And so what we do with regards to the work we have in observability around Grafana and Prometheus projects only accelerate that co-whole concept of continuous monitoring and continuous observability, and then kind of really, adding to that, I think it was last month, we introduce our fault injection simulator, a chaos engineering tool that, again takes advantage of all of this automation and machine learning to really help our developers, our customers operate at scale. And make sure that when they are releasing code, they're releasing code that is not just great in a small sense, it works on my laptop, but it works great in a highly distributed massively scaled environment around the globe. >> You know, this is one of the things that impresses me about Red Hat this year. And I've said this before all the covers events I've covered with them is that they get the cloud scale piece and I think their relationship with you guys shows that I think, DevOps has won, but it's the gift that keeps giving in open-source because what you have here is no longer a conversation about the cloud moving to the cloud. It's the cloud has become the operating model. So the conversation shifts to much more complicated enterprise or, and or intelligent Edge, and whether it's industrial or human or whatever, you got a data problem. So that's about a programmability issue at scale. So what's interesting is that Red Hat is on those bandwagon. It's an operating system. I mean, basically it's a distributed computing paradigm, essentially ala AWS concept as a cloud. Now it goes to the Edge, it's just distributed services via an open-source. So what's your reaction to that? >> Yeah, it's back to the original point, John where I said, any CIO is thinking about their IT environment from data center to cloud, to Edge and the more consistency automation and, kind of tools that they're at their disposal to enable them to create that kind of, I think you started to talk about an infrastructure the whole as code infrastructure's code, it's now, almost everything is code. And that starts with the operating system, obviously. And that's why this is so critical that we're partnering with companies like Red Hat on our vision and their vision, because they aligned to where our customers were ultimately going. Bob, you want to, you want to add to that? >> Bob: No, I think you said it. >> John: You guys are crushing it. Bob, one quick question for you, while I got you here. You mentioned getops, I've heard this before, I kind of understand it. Can you just quickly define from your perspective. What is getops? >> Sure, well, getops is really taking the, I said before it's a kind of narrowed version of DevOps. Sure, it's infrastructure is code. Sure, you're doing things incrementally but the getops principle, it's back to like, what are the good, what are the best practices we are managing large numbers, large numbers of robots. And in this case, it's around this idea of declarative intent. So instead of having systems that reach into production and change things, what you do is you set up the defined declared state of the system that you want and then leave the robots to constantly work to converge the state there. That seems kind of nebulous. Let me give you like a really concrete example from Kubernetes, by the way the entire Kubernetes system design is based on this. You say, I want five pods running in production and that's running my application. So what Kubernetes does is it sits there and it constantly checks, Oh, I'm supposed to have five pods. Do I have five? Well, what happens if the machine running one of those pods goes away. Now, suddenly it goes and checks and says, Oh, I'm supposed to have five pods, but there's four pods. What action do I take to now try to get the system back to the state. So you don't have a system running, reaching out and checking externally to Kubernetes, you let Kubernetes do the heavy lifting there. And so it goes through, goes through a loop of, Oh, I need to start a new pod and then it converges the system state back to running five pods. So it's really taking that kind of declarative intent combined with constant convergence loops to fully production at scale. >> That's awesome. Well, we do a whole segment on state and stateless future, but we don't have time. I do want to summarize real quick. We're here at the Red Hat Summit 2021. You got Red Hat OpenShift on AWS. The big news, Bob and Peder tell us quickly in summary, why AWS? Why Red Hat? Why better together? Give the quick overview, Bob, we'll start with you. >> Bob, you want to kick us off? >> I'm going to repeat peanut butter and chocolate. Customers love OpenShift, they love managed services. They want a simplified operations, simplified supply chain. So you get the best of both worlds. You get the OpenShift that you want fully managed on AWS, where you get all of the security and scale. Yeah, I can't add much to that. Other than saying, Red Hat is powerhouse obviously on data centers it is the operating system of the data center. Bringing together the best in the cloud, with the best in the data center is such a huge benefit to our customers. Because back to your point, John, our customers are thinking about what are they doing from data center to cloud, to Edge and bringing the best of those pieces together in a seamless solution is so, so critical. And that that's why AW. (indistinct) >> Thanks for coming on, I really appreciate it. I just want to give you guys a plug for you and being humble, but you've worked in the CNCF and standards bodies has been well, well known and I'm getting the word out. Congratulations for the commitment to open-source. Really appreciate the community. Thanks you, thank you for your time. >> Thanks, John. >> Okay, Cube coverage here, covering Red Hat Summit 2021. I'm John Ferry, host of theCUBE. Thanks for watching. (smart gentle music)

Published Date : Apr 27 2021

SUMMARY :

in the AWS open-source initiatives. And that really is about the Edge. And so having that relationship to ensure also the new service Red Red Hat and AWS to kind of like it's no big deal to say that, of the on's, if you will. But I got to ask you guys, pull the Amazon because now the customer That's kind of got the Lot of the use cases are of this, where you got do all of the hard work, which what do you guys see as the vision? So one of the ways that we collaborate I got to ask you guys this the overall pie, if you will. So a lot of the collaboration work we do, And so now you have open And so one of the things you have to do the code will execute what you did. into getops that you can of the data center and you So the conversation shifts to and the more consistency automation and, I kind of understand it. of the system that you want We're here at the Red Hat Summit 2021. in the cloud, with the best I just want to give you guys a I'm John Ferry, host of theCUBE.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmazonORGANIZATION

0.99+

JohnPERSON

0.99+

Jeff BarrPERSON

0.99+

AWSORGANIZATION

0.99+

John FerryPERSON

0.99+

ROSAORGANIZATION

0.99+

Adrian CockcroftPERSON

0.99+

Bob WisePERSON

0.99+

BobPERSON

0.99+

RedisORGANIZATION

0.99+

twoQUANTITY

0.99+

IBMORGANIZATION

0.99+

2013DATE

0.99+

Andy JassyPERSON

0.99+

John FurrierPERSON

0.99+

Red HatORGANIZATION

0.99+

DeltaORGANIZATION

0.99+

Amazon Web ServicesORGANIZATION

0.99+

2008DATE

0.99+

LFORGANIZATION

0.99+

fiveQUANTITY

0.99+

Amazon Web ServicesORGANIZATION

0.99+

Delta AirlinesORGANIZATION

0.99+

CNCFORGANIZATION

0.99+

five podsQUANTITY

0.99+

Red Hat OpenShiftTITLE

0.99+

GrafanaORGANIZATION

0.99+

Red HatTITLE

0.99+

five podsQUANTITY

0.99+

Amazon Web ServicesORGANIZATION

0.99+

KubernetesORGANIZATION

0.99+

Another test of transitions


 

>> Hi, my name is Andy Clemenko. I'm a Senior Solutions Engineer at StackRox. Thanks for joining us today for my talk on labels, labels, labels. Obviously, you can reach me at all the socials. Before we get started, I like to point you to my GitHub repo, you can go to andyc.info/dc20, and it'll take you to my GitHub page where I've got all of this documentation, socials. Before we get started, I like to point you to my GitHub repo, you can go to andyc.info/dc20, (upbeat music) >> Hi, my name is Andy Clemenko. I'm a Senior Solutions Engineer at StackRox. Thanks for joining us today for my talk on labels, labels, labels. Obviously, you can reach me at all the socials. Before we get started, I like to point you to my GitHub repo, you can go to andyc.info/dc20, and it'll take you to my GitHub page where I've got all of this documentation, I've got the Keynote file there. YAMLs, I've got Dockerfiles, Compose files, all that good stuff. If you want to follow along, great, if not go back and review later, kind of fun. So let me tell you a little bit about myself. I am a former DOD contractor. This is my seventh DockerCon. I've spoken, I had the pleasure to speak at a few of them, one even in Europe. I was even a Docker employee for quite a number of years, providing solutions to the federal government and customers around containers and all things Docker. So I've been doing this a little while. One of the things that I always found interesting was the lack of understanding around labels. So why labels, right? Well, as a former DOD contractor, I had built out a large registry. And the question I constantly got was, where did this image come from? How did you get it? What's in it? Where did it come from? How did it get here? And one of the things we did to kind of alleviate some of those questions was we established a baseline set of labels. Labels really are designed to provide as much metadata around the image as possible. I ask everyone in attendance, when was the last time you pulled an image and had 100% confidence, you knew what was inside it, where it was built, how it was built, when it was built, you probably didn't, right? The last thing we obviously want is a container fire, like our image on the screen. And one kind of interesting way we can kind of prevent that is through the use of labels. We can use labels to address security, address some of the simplicity on how to run these images. So think of it, kind of like self documenting, Think of it also as an audit trail, image provenance, things like that. These are some interesting concepts that we can definitely mandate as we move forward. What is a label, right? Specifically what is the Schema? It's just a key-value. All right? It's any key and pretty much any value. What if we could dump in all kinds of information? What if we could encode things and store it in there? And I've got a fun little demo to show you about that. Let's start off with some of the simple keys, right? Author, date, description, version. Some of the basic information around the image. That would be pretty useful, right? What about specific labels for CI? What about a, where's the version control? Where's the source, right? Whether it's Git, whether it's GitLab, whether it's GitHub, whether it's Gitosis, right? Even SPN, who cares? Where are the source files that built, where's the Docker file that built this image? What's the commit number? That might be interesting in terms of tracking the resulting image to a person or to a commit, hopefully then to a person. How is it built? What if you wanted to play with it and do a git clone of the repo and then build the Docker file on your own? Having a label specifically dedicated on how to build this image might be interesting for development work. Where it was built, and obviously what build number, right? These kind of all, not only talk about continuous integration, CI but also start to talk about security. Specifically what server built it. The version control number, the version number, the commit number, again, how it was built. What's the specific build number? What was that job number in, say, Jenkins or GitLab? What if we could take it a step further? What if we could actually apply policy enforcement in the build pipeline, looking specifically for some of these specific labels? I've got a good example of, in my demo of a policy enforcement. So let's look at some sample labels. Now originally, this idea came out of label-schema.org. And then it was a modified to opencontainers, org.opencontainers.image. There is a link in my GitHub page that links to the full reference. But these are some of the labels that I like to use, just as kind of like a standardization. So obviously, Author's, an email address, so now the image is attributable to a person, that's always kind of good for security and reliability. Where's the source? Where's the version control that has the source, the Docker file and all the assets? How it was built, build number, build server the commit, we talked about, when it was created, a simple description. A fun one I like adding in is the healthZendpoint. Now obviously, the health check directive should be in the Docker file. But if you've got other systems that want to ping your applications, why not declare it and make it queryable? Image version, obviously, that's simple declarative And then a title. And then I've got the two fun ones. Remember, I talked about what if we could encode some fun things? Hypothetically, what if we could encode the Compose file of how to build the stack in the first image itself? And conversely the Kubernetes? Well, actually, you can and I have a demo to show you how to kind of take advantage of that. So how do we create labels? And really creating labels as a function of build time okay? You can't really add labels to an image after the fact. The way you do add labels is either through the Docker file, which I'm a big fan of, because it's declarative. It's in version control. It's kind of irrefutable, especially if you're tracking that commit number in a label. You can extend it from being a static kind of declaration to more a dynamic with build arguments. And I can show you, I'll show you in a little while how you can use a build argument at build time to pass in that variable. And then obviously, if you did it by hand, you could do a docker build--label key equals value. I'm not a big fan of the third one, I love the first one and obviously the second one. Being dynamic we can take advantage of some of the variables coming out of version control. Or I should say, some of the variables coming out of our CI system. And that way, it self documents effectively at build time, which is kind of cool. How do we view labels? Well, there's two major ways to view labels. The first one is obviously a docker pull and docker inspect. You can pull the image locally, you can inspect it, you can obviously, it's going to output as JSON. So you going to use something like JQ to crack it open and look at the individual labels. Another one which I found recently was Skopeo from Red Hat. This allows you to actually query the registry server. So you don't even have to pull the image initially. This can be really useful if you're on a really small development workstation, and you're trying to talk to a Kubernetes cluster and wanting to deploy apps kind of in a very simple manner. Okay? And this was that use case, right? Using Kubernetes, the Kubernetes demo. One of the interesting things about this is that you can base64 encode almost anything, push it in as text into a label and then base64 decode it, and then use it. So in this case, in my demo, I'll show you how we can actually use a kubectl apply piped from the base64 decode from the label itself from skopeo talking to the registry. And what's interesting about this kind of technique is you don't need to store Helm charts. You don't need to learn another language for your declarative automation, right? You don't need all this extra levels of abstraction inherently, if you use it as a label with a kubectl apply, It's just built in. It's kind of like the kiss approach to a certain extent. It does require some encoding when you actually build the image, but to me, it doesn't seem that hard. Okay, let's take a look at a demo. And what I'm going to do for my demo, before we actually get started is here's my repo. Here's a, let me actually go to the actual full repo. So here's the repo, right? And I've got my Jenkins pipeline 'cause I'm using Jenkins for this demo. And in my demo flask, I've got the Docker file. I've got my compose and my Kubernetes YAML. So let's take a look at the Docker file, right? So it's a simple Alpine image. The org statements are the build time arguments that are passed in. Label, so again, I'm using the org.opencontainers.image.blank, for most of them. There's a typo there. Let's see if you can find it, I'll show you it later. My source, build date, build number, commit. Build number and get commit are derived from the Jenkins itself, which is nice. I can just take advantage of existing URLs. I don't have to create anything crazy. And again, I've got my actual Docker build command. Now this is just a label on how to build it. And then here's my simple Python, APK upgrade, remove the package manager, kind of some security stuff, health check getting Python through, okay? Let's take a look at the Jenkins pipeline real quick. So here is my Jenkins pipeline and I have four major stages, four stages, I have built. And here in build, what I do is I actually do the Git clone. And then I do my docker build. From there, I actually tell the Jenkins StackRox plugin. So that's what I'm using for my security scanning. So go ahead and scan, basically, I'm staging it to scan the image. I'm pushing it to Hub, okay? Where I can see the, basically I'm pushing the image up to Hub so such that my StackRox security scanner can go ahead and scan the image. I'm kicking off the scan itself. And then if everything's successful, I'm pushing it to prod. Now what I'm doing is I'm just using the same image with two tags, pre-prod and prod. This is not exactly ideal, in your environment, you probably want to use separate registries and non-prod and a production registry, but for demonstration purposes, I think this is okay. So let's go over to my Jenkins and I've got a deliberate failure. And I'll show you why there's a reason for that. And let's go down. Let's look at my, so I have a StackRox report. Let's look at my report. And it says image required, required image label alert, right? Request that the maintainer, add the required label to the image, so we're missing a label, okay? One of the things we can do is let's flip over, and let's look at Skopeo. Right? I'm going to do this just the easy way. So instead of looking at org.zdocker, opencontainers.image.authors. Okay, see here it says build signature? That was the typo, we didn't actually pass in. So if we go back to our repo, we didn't pass in the the build time argument, we just passed in the word. So let's fix that real quick. That's the Docker file. Let's go ahead and put our dollar sign in their. First day with the fingers you going to love it. And let's go ahead and commit that. Okay? So now that that's committed, we can go back to Jenkins, and we can actually do another build. And there's number 12. And as you can see, I've been playing with this for a little bit today. And while that's running, come on, we can go ahead and look at the Console output. Okay, so there's our image. And again, look at all the build arguments that we're passing into the build statement. So we're passing in the date and the date gets derived on the command line. With the build arguments, there's the base64 encoded of the Compose file. Here's the base64 encoding of the Kubernetes YAML. We do the build. And then let's go down to the bottom layer exists and successful. So here's where we can see no system policy violations profound marking stack regimes security plugin, build step as successful, okay? So we're actually able to do policy enforcement that that image exists, that that label sorry, exists in the image. And again, we can look at the security report and there's no policy violations and no vulnerabilities. So that's pretty good for security, right? We can now enforce and mandate use of certain labels within our images. And let's flip back over to Skopeo, and let's go ahead and look at it. So we're looking at the prod version again. And there's it is in my email address. And that validated that that was valid for that policy. So that's kind of cool. Now, let's take it a step further. What if, let's go ahead and take a look at all of the image, all the labels for a second, let me remove the dash org, make it pretty. Okay? So we have all of our image labels. Again, author's build, commit number, look at the commit number. It was built today build number 12. We saw that right? Delete, build 12. So that's kind of cool dynamic labels. Name, healthz, right? But what we're looking for is we're going to look at the org.zdockerketers label. So let's go look at the label real quick. Okay, well that doesn't really help us because it's encoded but let's base64 dash D, let's decode it. And I need to put the dash r in there 'cause it doesn't like, there we go. So there's my Kubernetes YAML. So why can't we simply kubectl apply dash f? Let's just apply it from standard end. So now we've actually used that label. From the image that we've queried with skopeo, from a remote registry to deploy locally to our Kubernetes cluster. So let's go ahead and look everything's up and running, perfect. So what does that look like, right? So luckily, I'm using traefik for Ingress 'cause I love it. And I've got an object in my Kubernetes YAML called flask.doctor.life. That's my Ingress object for traefik. I can go to flask.docker.life. And I can hit refresh. Obviously, I'm not a very good web designer 'cause the background image in the text. We can go ahead and refresh it a couple times we've got Redis storing a hit counter. We can see that our server name is roundrobing. Okay? That's kind of cool. So let's kind of recap a little bit about my demo environment. So my demo environment, I'm using DigitalOcean, Ubuntu 19.10 Vms. I'm using K3s instead of full Kubernetes either full Rancher, full Open Shift or Docker Enterprise. I think K3s has some really interesting advantages on the development side and it's kind of intended for IoT but it works really well and it deploys super easy. I'm using traefik for Ingress. I love traefik. I may or may not be a traefik ambassador. I'm using Jenkins for CI. And I'm using StackRox for image scanning and policy enforcement. One of the things to think about though, especially in terms of labels is none of this demo stack is required. You can be in any cloud, you can be in CentOs, you can be in any Kubernetes. You can even be in swarm, if you wanted to, or Docker compose. Any Ingress, any CI system, Jenkins, circle, GitLab, it doesn't matter. And pretty much any scanning. One of the things that I think is kind of nice about at least StackRox is that we do a lot more than just image scanning, right? With the policy enforcement things like that. I guess that's kind of a shameless plug. But again, any of this stack is completely replaceable, with any comparative product in that category. So I'd like to, again, point you guys to the andyc.infodc20, that's take you right to the GitHub repo. You can reach out to me at any of the socials @clemenko or andy@stackrox.com. And thank you for attending. I hope you learned something fun about labels. And hopefully you guys can standardize labels in your organization and really kind of take your images and the image provenance to a new level. Thanks for watching. (upbeat music) >> Narrator: Live from Las Vegas It's theCUBE. Covering AWS re:Invent 2019. Brought to you by Amazon Web Services and Intel along with it's ecosystem partners. >> Okay, welcome back everyone theCUBE's live coverage of AWS re:Invent 2019. This is theCUBE's 7th year covering Amazon re:Invent. It's their 8th year of the conference. I want to just shout out to Intel for their sponsorship for these two amazing sets. Without their support we wouldn't be able to bring our mission of great content to you. I'm John Furrier. Stu Miniman. We're here with the chief of AWS, the chief executive officer Andy Jassy. Tech athlete in and of himself three hour Keynotes. Welcome to theCUBE again, great to see you. >> Great to be here, thanks for having me guys. >> Congratulations on a great show a lot of great buzz. >> Andy: Thank you. >> A lot of good stuff. Your Keynote was phenomenal. You get right into it, you giddy up right into it as you say, three hours, thirty announcements. You guys do a lot, but what I liked, the new addition, the last year and this year is the band; house band. They're pretty good. >> Andy: They're good right? >> They hit the queen notes, so that keeps it balanced. So we're going to work on getting a band for theCUBE. >> Awesome. >> So if I have to ask you, what's your walk up song, what would it be? >> There's so many choices, it depends on what kind of mood I'm in. But, uh, maybe Times Like These by the Foo Fighters. >> John: Alright. >> These are unusual times right now. >> Foo Fighters playing at the Amazon Intersect Show. >> Yes they are. >> Good plug Andy. >> Headlining. >> Very clever >> Always getting a good plug in there. >> My very favorite band. Well congratulations on the Intersect you got a lot going on. Intersect is a music festival, I'll get to that in a second But, I think the big news for me is two things, obviously we had a one-on-one exclusive interview and you laid out, essentially what looks like was going to be your Keynote, and it was. Transformation- >> Andy: Thank you for the practice. (Laughter) >> John: I'm glad to practice, use me anytime. >> Yeah. >> And I like to appreciate the comments on Jedi on the record, that was great. But I think the transformation story's a very real one, but the NFL news you guys just announced, to me, was so much fun and relevant. You had the Commissioner of NFL on stage with you talking about a strategic partnership. That is as top down, aggressive goal as you could get to have Rodger Goodell fly to a tech conference to sit with you and then bring his team talk about the deal. >> Well, ya know, we've been partners with the NFL for a while with the Next Gen Stats that they use on all their telecasts and one of the things I really like about Roger is that he's very curious and very interested in technology and the first couple times I spoke with him he asked me so many questions about ways the NFL might be able to use the Cloud and digital transformation to transform their various experiences and he's always said if you have a creative idea or something you think that could change the world for us, just call me he said or text me or email me and I'll call you back within 24 hours. And so, we've spent the better part of the last year talking about a lot of really interesting, strategic ways that they can evolve their experience both for fans, as well as their players and the Player Health and Safety Initiative, it's so important in sports and particularly important with the NFL given the nature of the sport and they've always had a focus on it, but what you can do with computer vision and machine learning algorithms and then building a digital athlete which is really like a digital twin of each athlete so you understand, what does it look like when they're healthy and compare that when it looks like they may not be healthy and be able to simulate all kinds of different combinations of player hits and angles and different plays so that you could try to predict injuries and predict the right equipment you need before there's a problem can be really transformational so we're super excited about it. >> Did you guys come up with the idea or was it a collaboration between them? >> It was really a collaboration. I mean they, look, they are very focused on players safety and health and it's a big deal for their- you know, they have two main constituents the players and fans and they care deeply about the players and it's a-it's a hard problem in a sport like Football, I mean, you watch it. >> Yeah, and I got to say it does point out the use cases of what you guys are promoting heavily at the show here of the SageMaker Studio, which was a big part of your Keynote, where they have all this data. >> Andy: Right. >> And they're data hoarders, they hoard data but the manual process of going through the data was a killer problem. This is consistent with a lot of the enterprises that are out there, they have more data than they even know. So this seems to be a big part of the strategy. How do you get the customers to actually wake up to the fact that they got all this data and how do you tie that together? >> I think in almost every company they know they have a lot of data. And there are always pockets of people who want to do something with it. But, when you're going to make these really big leaps forward; these transformations, the things like Volkswagen is doing where they're reinventing their factories and their manufacturing process or the NFL where they're going to radically transform how they do players uh, health and safety. It starts top down and if the senior leader isn't convicted about wanting to take that leap forward and trying something different and organizing the data differently and organizing the team differently and using machine learning and getting help from us and building algorithms and building some muscle inside the company it just doesn't happen because it's not in the normal machinery of what most companies do. And so it always, almost always, starts top down. Sometimes it can be the Commissioner or CEO sometimes it can be the CIO but it has to be senior level conviction or it doesn't get off the ground. >> And the business model impact has to be real. For NFL, they know concussions, hurting their youth pipe-lining, this is a huge issue for them. This is their business model. >> They lose even more players to lower extremity injuries. And so just the notion of trying to be able to predict injuries and, you know, the impact it can have on rules and the impact it can have on the equipment they use, it's a huge game changer when they look at the next 10 to 20 years. >> Alright, love geeking out on the NFL but Andy, you know- >> No more NFL talk? >> Off camera how about we talk? >> Nobody talks about the Giants being 2 and 10. >> Stu: We're both Patriots fans here. >> People bring up the undefeated season. >> So Andy- >> Everybody's a Patriot's fan now. (Laughter) >> It's fascinating to watch uh, you and your three hour uh, Keynote, uh Werner in his you know, architectural discussion, really showed how AWS is really extending its reach, you know, it's not just a place. For a few years people have been talking about you know, Cloud is an operational model its not a destination or a location but, I felt it really was laid out is you talked about Breadth and Depth and Werner really talked about you know, Architectural differentiation. People talk about Cloud, but there are very-there are a lot of differences between the vision for where things are going. Help us understand why, I mean, Amazon's vision is still a bit different from what other people talk about where this whole Cloud expansion, journey, put ever what tag or label you want on it but you know, the control plane and the technology that you're building and where you see that going. >> Well I think that, we've talked about this a couple times we have two macro types of customers. We have those that really want to get at the low level building blocks and stitch them together creatively however they see fit to create whatever's in their-in their heads. And then we have the second segment of customers that say look, I'm willing to give up some of that flexibility in exchange for getting 80% of the way there much faster. In an abstraction that's different from those low level building blocks. And both segments of builders we want to serve and serve well and so we've built very significant offerings in both areas. I think when you look at microservices um, you know, some of it has to do with the fact that we have this very strongly held belief born out of several years of Amazon where you know, the first 7 or 8 years of Amazon's consumer business we basically jumbled together all of the parts of our technology in moving really quickly and when we wanted to move quickly where you had to impact multiple internal development teams it was so long because it was this big ball, this big monolithic piece. And we got religion about that in trying to move faster in the consumer business and having to tease those pieces apart. And it really was a lot of impetus behind conceiving AWS where it was these low level, very flexible building blocks that6 don't try and make all the decisions for customers they get to make them themselves. And some of the microservices that you saw Werner talking about just, you know, for instance, what we-what we did with Nitro or even what we did with Firecracker those are very much about us relentlessly working to continue to uh, tease apart the different components. And even things that look like low level building blocks over time, you build more and more features and all of the sudden you realize they have a lot of things that are combined together that you wished weren't that slow you down and so, Nitro was a completely re imagining of our Hypervisor and Virtualization layer to allow us, both to let customers have better performance but also to let us move faster and have a better security story for our customers. >> I got to ask you the question around transformation because I think that all points, all the data points, you got all the references, Goldman Sachs on stage at the Keynote, Cerner, I mean healthcare just is an amazing example because I mean, that's demonstrating real value there there's no excuse. I talked to someone who wouldn't be named last night, in and around the area said, the CIA has a cost bar like this a cost-a budget like this but the demand for mission based apps is going up exponentially, so there's need for the Cloud. And so, you see more and more of that. What is your top down, aggressive goals to fill that solution base because you're also a very transformational thinker; what is your-what is your aggressive top down goals for your organization because you're serving a market with trillions of dollars of spend that's shifting, that's on the table. >> Yeah. >> A lot of competition now sees it too, they're going to go after it. But at the end of the day you have customers that have a demand for things, apps. >> Andy: Yeah. >> And not a lot of budget increase at the same time. This is a huge dynamic. >> Yeah. >> John: What's your goals? >> You know I think that at a high level our top down aggressive goals are that we want every single customer who uses our platform to have an outstanding customer experience. And we want that outstanding customer experience in part is that their operational performance and their security are outstanding, but also that it allows them to build, uh, build projects and initiatives that change their customer experience and allow them to be a sustainable successful business over a long period of time. And then, we also really want to be the technology infrastructure platform under all the applications that people build. And we're realistic, we know that you know, the market segments we address with infrastructure, software, hardware, and data center services globally are trillions of dollars in the long term and it won't only be us, but we have that goal of wanting to serve every application and that requires not just the security operational premise but also a lot of functionality and a lot of capability. We have by far the most amount of capability out there and yet I would tell you, we have 3 to 5 years of items on our roadmap that customers want us to add. And that's just what we know today. >> And Andy, underneath the covers you've been going through some transformation. When we talked a couple of years ago, about how serverless is impacting things I've heard that that's actually, in many ways, glue behind the two pizza teams to work between organizations. Talk about how the internal transformations are happening. How that impacts your discussions with customers that are going through that transformation. >> Well, I mean, there's a lot of- a lot of the technology we build comes from things that we're doing ourselves you know? And that we're learning ourselves. It's kind of how we started thinking about microservices, serverless too, we saw the need, you know, we would have we would build all these functions that when some kind of object came into an object store we would spin up, compute, all those tasks would take like, 3 or 4 hundred milliseconds then we'd spin it back down and yet, we'd have to keep a cluster up in multiple availability zones because we needed that fault tolerance and it was- we just said this is wasteful and, that's part of how we came up with Lambda and you know, when we were thinking about Lambda people understandably said, well if we build Lambda and we build this serverless adventure in computing a lot of people were keeping clusters of instances aren't going to use them anymore it's going to lead to less absolute revenue for us. But we, we have learned this lesson over the last 20 years at Amazon which is, if it's something that's good for customers you're much better off cannibalizing yourself and doing the right thing for customers and being part of shaping something. And I think if you look at the history of technology you always build things and people say well, that's going to cannibalize this and people are going to spend less money, what really ends up happening is they spend less money per unit of compute but it allows them to do so much more that they ultimately, long term, end up being more significant customers. >> I mean, you are like beating the drum all the time. Customers, what they say, we encompass the roadmap, I got that you guys have that playbook down, that's been really successful for you. >> Andy: Yeah. >> Two years ago you told me machine learning was really important to you because your customers told you. What's the next traunch of importance for customers? What's on top of mind now, as you, look at- >> Andy: Yeah. >> This re:Invent kind of coming to a close, Replay's tonight, you had conversations, you're a tech athlete, you're running around, doing speeches, talking to customers. What's that next hill from if it's machine learning today- >> There's so much I mean, (weird background noise) >> It's not a soup question (Laughter) And I think we're still in the very early days of machine learning it's not like most companies have mastered it yet even though they're using it much more then they did in the past. But, you know, I think machine learning for sure I think the Edge for sure, I think that um, we're optimistic about Quantum Computing even though I think it'll be a few years before it's really broadly useful. We're very um, enthusiastic about robotics. I think the amount of functions that are going to be done by these- >> Yeah. >> robotic applications are much more expansive than people realize. It doesn't mean humans won't have jobs, they're just going to work on things that are more value added. We're believers in augmented virtual reality, we're big believers in what's going to happen with Voice. And I'm also uh, I think sometimes people get bored you know, I think you're even bored with machine learning already >> Not yet. >> People get bored with the things you've heard about but, I think just what we've done with the Chips you know, in terms of giving people 40% better price performance in the latest generation of X86 processors. It's pretty unbelievable in the difference in what people are going to be able to do. Or just look at big data I mean, big data, we haven't gotten through big data where people have totally solved it. The amount of data that companies want to store, process, analyze, is exponentially larger than it was a few years ago and it will, I think, exponentially increase again in the next few years. You need different tools and services. >> Well I think we're not bored with machine learning we're excited to get started because we have all this data from the video and you guys got SageMaker. >> Andy: Yeah. >> We call it the stairway to machine learning heaven. >> Andy: Yeah. >> You start with the data, move up, knock- >> You guys are very sophisticated with what you do with technology and machine learning and there's so much I mean, we're just kind of, again, in such early innings. And I think that, it was so- before SageMaker, it was so hard for everyday developers and data scientists to build models but the combination of SageMaker and what's happened with thousands of companies standardizing on it the last two years, plus now SageMaker studio, giant leap forward. >> Well, we hope to use the data to transform our experience with our audience. And we're on Amazon Cloud so we really appreciate that. >> Andy: Yeah. >> And appreciate your support- >> Andy: Yeah, of course. >> John: With Amazon and get that machine learning going a little faster for us, that would be better. >> If you have requests I'm interested, yeah. >> So Andy, you talked about that you've got the customers that are builders and the customers that need simplification. Traditionally when you get into the, you know, the heart of the majority of adoption of something you really need to simplify that environment. But when I think about the successful enterprise of the future, they need to be builders. how'l I normally would've said enterprise want to pay for solutions because they don't have the skill set but, if they're going to succeed in this new economy they need to go through that transformation >> Andy: Yeah. >> That you talk to, so, I mean, are we in just a total new era when we look back will this be different than some of these previous waves? >> It's a really good question Stu, and I don't think there's a simple answer to it. I think that a lot of enterprises in some ways, I think wish that they could just skip the low level building blocks and only operate at that higher level abstraction. That's why people were so excited by things like, SageMaker, or CodeGuru, or Kendra, or Contact Lens, these are all services that allow them to just send us data and then run it on our models and get back the answers. But I think one of the big trends that we see with enterprises is that they are taking more and more of their development in house and they are wanting to operate more and more like startups. I think that they admire what companies like AirBnB and Pintrest and Slack and Robinhood and a whole bunch of those companies, Stripe, have done and so when, you know, I think you go through these phases and eras where there are waves of success at different companies and then others want to follow that success and replicate it. And so, we see more and more enterprises saying we need to take back a lot of that development in house. And as they do that, and as they add more developers those developers in most cases like to deal with the building blocks. And they have a lot of ideas on how they can creatively stich them together. >> Yeah, on that point, I want to just quickly ask you on Amazon versus other Clouds because you made a comment to me in our interview about how hard it is to provide a service to other people. And it's hard to have a service that you're using yourself and turn that around and the most quoted line of my story was, the compression algorithm- there's no compression algorithm for experience. Which to me, is the diseconomies of scale for taking shortcuts. >> Andy: Yeah. And so I think this is a really interesting point, just add some color commentary because I think this is a fundamental difference between AWS and others because you guys have a trajectory over the years of serving, at scale, customers wherever they are, whatever they want to do, now you got microservices. >> Yeah. >> John: It's even more complex. That's hard. >> Yeah. >> John: Talk about that. >> I think there are a few elements to that notion of there's no compression algorithm for experience and I think the first thing to know about AWS which is different is, we just come from a different heritage and a different background. We ran a business for a long time that was our sole business that was a consumer retail business that was very low margin. And so, we had to operate at very large scale given how many people were using us but also, we had to run infrastructure services deep in the stack, compute storage and database, and reliable scalable data centers at very low cost and margins. And so, when you look at our business it actually, today, I mean its, its a higher margin business in our retail business, its a lower margin business in software companies but at real scale, it's a high volume, relatively low margin business. And the way that you have to operate to be successful with those businesses and the things you have to think about and that DNA come from the type of operators we have to be in our consumer retail business. And there's nobody else in our space that does that. So, you know, the way that we think about costs, the way we think about innovation in the data center, um, and I also think the way that we operate services and how long we've been operating services as a company its a very different mindset than operating package software. Then you look at when uh, you think about some of the uh, issues in very large scale Cloud, you can't learn some of those lessons until you get to different elbows of the curve and scale. And so what I was telling you is, its really different to run your own platform for your own users where you get to tell them exactly how its going to be done. But that's not the way the real world works. I mean, we have millions of external customers who use us from every imaginable country and location whenever they want, without any warning, for lots of different use cases, and they have lots of design patterns and we don't get to tell them what to do. And so operating a Cloud like that, at a scale that's several times larger than the next few providers combined is a very different endeavor and a very different operating rigor. >> Well you got to keep raising the bar you guys do a great job, really impressed again. Another tsunami of announcements. In fact, you had to spill the beans earlier with Quantum the day before the event. Tight schedule. I got to ask you about the musical festival because, I think this is a very cool innovation. It's the inaugural Intersect conference. >> Yes. >> John: Which is not part of Replay, >> Yes. >> John: Which is the concert tonight. Its a whole new thing, big music act, you're a big music buff, your daughter's an artist. Why did you do this? What's the purpose? What's your goal? >> Yeah, it's an experiment. I think that what's happened is that re:Invent has gotten so big, we have 65 thousand people here, that to do the party, which we do every year, its like a 35-40 thousand person concert now. Which means you have to have a location that has multiple stages and, you know, we thought about it last year and when we were watching it and we said, we're kind of throwing, like, a 4 hour music festival right now. There's multiple stages, and its quite expensive to set up that set for a party and we said well, maybe we don't have to spend all that money for 4 hours and then rip it apart because actually the rent to keep those locations for another two days is much smaller than the cost of actually building multiple stages and so we thought we would try it this year. We're very passionate about music as a business and I think we-I think our customers feel like we've thrown a pretty good music party the last few years and we thought we would try it at a larger scale as an experiment. And if you look at the economics- >> At the headliners real quick. >> The Foo Fighters are headlining on Saturday night, Anderson Paak and the Free Nationals, Brandi Carlile, Shawn Mullins, um, Willy Porter, its a good set. Friday night its Beck and Kacey Musgraves so it's a really great set of um, about thirty artists and we're hopeful that if we can build a great experience that people will want to attend that we can do it at scale and it might be something that both pays for itself and maybe, helps pay for re:Invent too overtime and you know, I think that we're also thinking about it as not just a music concert and festival the reason we named it Intersect is that we want an intersection of music genres and people and ethnicities and age groups and art and technology all there together and this will be the first year we try it, its an experiment and we're really excited about it. >> Well I'm gone, congratulations on all your success and I want to thank you we've been 7 years here at re:Invent we've been documenting the history. You got two sets now, one set upstairs. So appreciate you. >> theCUBE is part of re:Invent, you know, you guys really are apart of the event and we really appreciate your coming here and I know people appreciate the content you create as well. >> And we just launched CUBE365 on Amazon Marketplace built on AWS so thanks for letting us- >> Very cool >> John: Build on the platform. appreciate it. >> Thanks for having me guys, I appreciate it. >> Andy Jassy the CEO of AWS here inside theCUBE, it's our 7th year covering and documenting the thunderous innovation that Amazon's doing they're really doing amazing work building out the new technologies here in the Cloud computing world. I'm John Furrier, Stu Miniman, be right back with more after this short break. (Outro music)

Published Date : Sep 29 2020

SUMMARY :

at org the org to the andyc and it was. of time. That's hard. I think that

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Andy ClemenkoPERSON

0.99+

AndyPERSON

0.99+

Stu MinimanPERSON

0.99+

Amazon Web ServicesORGANIZATION

0.99+

Andy JassyPERSON

0.99+

CIAORGANIZATION

0.99+

John FurrierPERSON

0.99+

AWSORGANIZATION

0.99+

EuropeLOCATION

0.99+

JohnPERSON

0.99+

3QUANTITY

0.99+

StackRoxORGANIZATION

0.99+

80%QUANTITY

0.99+

4 hoursQUANTITY

0.99+

100%QUANTITY

0.99+

AmazonORGANIZATION

0.99+

VolkswagenORGANIZATION

0.99+

Rodger GoodellPERSON

0.99+

AirBnBORGANIZATION

0.99+

RogerPERSON

0.99+

40%QUANTITY

0.99+

Brandi CarlilePERSON

0.99+

PintrestORGANIZATION

0.99+

PythonTITLE

0.99+

two daysQUANTITY

0.99+

4 hourQUANTITY

0.99+

7th yearQUANTITY

0.99+

Willy PorterPERSON

0.99+

Friday nightDATE

0.99+

andy@stackrox.comOTHER

0.99+

7 yearsQUANTITY

0.99+

Goldman SachsORGANIZATION

0.99+

two tagsQUANTITY

0.99+

IntelORGANIZATION

0.99+

millionsQUANTITY

0.99+

Foo FightersORGANIZATION

0.99+

last yearDATE

0.99+

GiantsORGANIZATION

0.99+

todayDATE

0.99+

andyc.info/dc20OTHER

0.99+

65 thousand peopleQUANTITY

0.99+

Saturday nightDATE

0.99+

SlackORGANIZATION

0.99+

two setsQUANTITY

0.99+

flask.docker.lifeOTHER

0.99+

WernerPERSON

0.99+

two thingsQUANTITY

0.99+

Shawn MullinsPERSON

0.99+

RobinhoodORGANIZATION

0.99+

IntersectORGANIZATION

0.99+

thousandsQUANTITY

0.99+

Kacey MusgravesPERSON

0.99+

4 hundred millisecondsQUANTITY

0.99+

first imageQUANTITY

0.99+

Andy


 

>> Hi, my name is Andy Clemenko. I'm a Senior Solutions Engineer at StackRox. Thanks for joining us today for my talk on labels, labels, labels. Obviously, you can reach me at all the socials. Before we get started, I like to point you to my GitHub repo, you can go to andyc.info/dc20, and it'll take you to my GitHub page where I've got all of this documentation, I've got the Keynote file there. YAMLs, I've got Dockerfiles, Compose files, all that good stuff. If you want to follow along, great, if not go back and review later, kind of fun. So let me tell you a little bit about myself. I am a former DOD contractor. This is my seventh DockerCon. I've spoken, I had the pleasure to speak at a few of them, one even in Europe. I was even a Docker employee for quite a number of years, providing solutions to the federal government and customers around containers and all things Docker. So I've been doing this a little while. One of the things that I always found interesting was the lack of understanding around labels. So why labels, right? Well, as a former DOD contractor, I had built out a large registry. And the question I constantly got was, where did this image come from? How did you get it? What's in it? Where did it come from? How did it get here? And one of the things we did to kind of alleviate some of those questions was we established a baseline set of labels. Labels really are designed to provide as much metadata around the image as possible. I ask everyone in attendance, when was the last time you pulled an image and had 100% confidence, you knew what was inside it, where it was built, how it was built, when it was built, you probably didn't, right? The last thing we obviously want is a container fire, like our image on the screen. And one kind of interesting way we can kind of prevent that is through the use of labels. We can use labels to address security, address some of the simplicity on how to run these images. So think of it, kind of like self documenting, Think of it also as an audit trail, image provenance, things like that. These are some interesting concepts that we can definitely mandate as we move forward. What is a label, right? Specifically what is the Schema? It's just a key-value. All right? It's any key and pretty much any value. What if we could dump in all kinds of information? What if we could encode things and store it in there? And I've got a fun little demo to show you about that. Let's start off with some of the simple keys, right? Author, date, description, version. Some of the basic information around the image. That would be pretty useful, right? What about specific labels for CI? What about a, where's the version control? Where's the source, right? Whether it's Git, whether it's GitLab, whether it's GitHub, whether it's Gitosis, right? Even SPN, who cares? Where are the source files that built, where's the Docker file that built this image? What's the commit number? That might be interesting in terms of tracking the resulting image to a person or to a commit, hopefully then to a person. How is it built? What if you wanted to play with it and do a git clone of the repo and then build the Docker file on your own? Having a label specifically dedicated on how to build this image might be interesting for development work. Where it was built, and obviously what build number, right? These kind of all, not only talk about continuous integration, CI but also start to talk about security. Specifically what server built it. The version control number, the version number, the commit number, again, how it was built. What's the specific build number? What was that job number in, say, Jenkins or GitLab? What if we could take it a step further? What if we could actually apply policy enforcement in the build pipeline, looking specifically for some of these specific labels? I've got a good example of, in my demo of a policy enforcement. So let's look at some sample labels. Now originally, this idea came out of label-schema.org. And then it was a modified to opencontainers, org.opencontainers.image. There is a link in my GitHub page that links to the full reference. But these are some of the labels that I like to use, just as kind of like a standardization. So obviously, Author's, an email address, so now the image is attributable to a person, that's always kind of good for security and reliability. Where's the source? Where's the version control that has the source, the Docker file and all the assets? How it was built, build number, build server the commit, we talked about, when it was created, a simple description. A fun one I like adding in is the healthZendpoint. Now obviously, the health check directive should be in the Docker file. But if you've got other systems that want to ping your applications, why not declare it and make it queryable? Image version, obviously, that's simple declarative And then a title. And then I've got the two fun ones. Remember, I talked about what if we could encode some fun things? Hypothetically, what if we could encode the Compose file of how to build the stack in the first image itself? And conversely the Kubernetes? Well, actually, you can and I have a demo to show you how to kind of take advantage of that. So how do we create labels? And really creating labels as a function of build time okay? You can't really add labels to an image after the fact. The way you do add labels is either through the Docker file, which I'm a big fan of, because it's declarative. It's in version control. It's kind of irrefutable, especially if you're tracking that commit number in a label. You can extend it from being a static kind of declaration to more a dynamic with build arguments. And I can show you, I'll show you in a little while how you can use a build argument at build time to pass in that variable. And then obviously, if you did it by hand, you could do a docker build--label key equals value. I'm not a big fan of the third one, I love the first one and obviously the second one. Being dynamic we can take advantage of some of the variables coming out of version control. Or I should say, some of the variables coming out of our CI system. And that way, it self documents effectively at build time, which is kind of cool. How do we view labels? Well, there's two major ways to view labels. The first one is obviously a docker pull and docker inspect. You can pull the image locally, you can inspect it, you can obviously, it's going to output as JSON. So you going to use something like JQ to crack it open and look at the individual labels. Another one which I found recently was Skopeo from Red Hat. This allows you to actually query the registry server. So you don't even have to pull the image initially. This can be really useful if you're on a really small development workstation, and you're trying to talk to a Kubernetes cluster and wanting to deploy apps kind of in a very simple manner. Okay? And this was that use case, right? Using Kubernetes, the Kubernetes demo. One of the interesting things about this is that you can base64 encode almost anything, push it in as text into a label and then base64 decode it, and then use it. So in this case, in my demo, I'll show you how we can actually use a kubectl apply piped from the base64 decode from the label itself from skopeo talking to the registry. And what's interesting about this kind of technique is you don't need to store Helm charts. You don't need to learn another language for your declarative automation, right? You don't need all this extra levels of abstraction inherently, if you use it as a label with a kubectl apply, It's just built in. It's kind of like the kiss approach to a certain extent. It does require some encoding when you actually build the image, but to me, it doesn't seem that hard. Okay, let's take a look at a demo. And what I'm going to do for my demo, before we actually get started is here's my repo. Here's a, let me actually go to the actual full repo. So here's the repo, right? And I've got my Jenkins pipeline 'cause I'm using Jenkins for this demo. And in my demo flask, I've got the Docker file. I've got my compose and my Kubernetes YAML. So let's take a look at the Docker file, right? So it's a simple Alpine image. The org statements are the build time arguments that are passed in. Label, so again, I'm using the org.opencontainers.image.blank, for most of them. There's a typo there. Let's see if you can find it, I'll show you it later. My source, build date, build number, commit. Build number and get commit are derived from the Jenkins itself, which is nice. I can just take advantage of existing URLs. I don't have to create anything crazy. And again, I've got my actual Docker build command. Now this is just a label on how to build it. And then here's my simple Python, APK upgrade, remove the package manager, kind of some security stuff, health check getting Python through, okay? Let's take a look at the Jenkins pipeline real quick. So here is my Jenkins pipeline and I have four major stages, four stages, I have built. And here in build, what I do is I actually do the Git clone. And then I do my docker build. From there, I actually tell the Jenkins StackRox plugin. So that's what I'm using for my security scanning. So go ahead and scan, basically, I'm staging it to scan the image. I'm pushing it to Hub, okay? Where I can see the, basically I'm pushing the image up to Hub so such that my StackRox security scanner can go ahead and scan the image. I'm kicking off the scan itself. And then if everything's successful, I'm pushing it to prod. Now what I'm doing is I'm just using the same image with two tags, pre-prod and prod. This is not exactly ideal, in your environment, you probably want to use separate registries and non-prod and a production registry, but for demonstration purposes, I think this is okay. So let's go over to my Jenkins and I've got a deliberate failure. And I'll show you why there's a reason for that. And let's go down. Let's look at my, so I have a StackRox report. Let's look at my report. And it says image required, required image label alert, right? Request that the maintainer, add the required label to the image, so we're missing a label, okay? One of the things we can do is let's flip over, and let's look at Skopeo. Right? I'm going to do this just the easy way. So instead of looking at org.zdocker, opencontainers.image.authors. Okay, see here it says build signature? That was the typo, we didn't actually pass in. So if we go back to our repo, we didn't pass in the the build time argument, we just passed in the word. So let's fix that real quick. That's the Docker file. Let's go ahead and put our dollar sign in their. First day with the fingers you going to love it. And let's go ahead and commit that. Okay? So now that that's committed, we can go back to Jenkins, and we can actually do another build. And there's number 12. And as you can see, I've been playing with this for a little bit today. And while that's running, come on, we can go ahead and look at the Console output. Okay, so there's our image. And again, look at all the build arguments that we're passing into the build statement. So we're passing in the date and the date gets derived on the command line. With the build arguments, there's the base64 encoded of the Compose file. Here's the base64 encoding of the Kubernetes YAML. We do the build. And then let's go down to the bottom layer exists and successful. So here's where we can see no system policy violations profound marking stack regimes security plugin, build step as successful, okay? So we're actually able to do policy enforcement that that image exists, that that label sorry, exists in the image. And again, we can look at the security report and there's no policy violations and no vulnerabilities. So that's pretty good for security, right? We can now enforce and mandate use of certain labels within our images. And let's flip back over to Skopeo, and let's go ahead and look at it. So we're looking at the prod version again. And there's it is in my email address. And that validated that that was valid for that policy. So that's kind of cool. Now, let's take it a step further. What if, let's go ahead and take a look at all of the image, all the labels for a second, let me remove the dash org, make it pretty. Okay? So we have all of our image labels. Again, author's build, commit number, look at the commit number. It was built today build number 12. We saw that right? Delete, build 12. So that's kind of cool dynamic labels. Name, healthz, right? But what we're looking for is we're going to look at the org.zdockerketers label. So let's go look at the label real quick. Okay, well that doesn't really help us because it's encoded but let's base64 dash D, let's decode it. And I need to put the dash r in there 'cause it doesn't like, there we go. So there's my Kubernetes YAML. So why can't we simply kubectl apply dash f? Let's just apply it from standard end. So now we've actually used that label. From the image that we've queried with skopeo, from a remote registry to deploy locally to our Kubernetes cluster. So let's go ahead and look everything's up and running, perfect. So what does that look like, right? So luckily, I'm using traefik for Ingress 'cause I love it. And I've got an object in my Kubernetes YAML called flask.doctor.life. That's my Ingress object for traefik. I can go to flask.docker.life. And I can hit refresh. Obviously, I'm not a very good web designer 'cause the background image in the text. We can go ahead and refresh it a couple times we've got Redis storing a hit counter. We can see that our server name is roundrobing. Okay? That's kind of cool. So let's kind of recap a little bit about my demo environment. So my demo environment, I'm using DigitalOcean, Ubuntu 19.10 Vms. I'm using K3s instead of full Kubernetes either full Rancher, full Open Shift or Docker Enterprise. I think K3s has some really interesting advantages on the development side and it's kind of intended for IoT but it works really well and it deploys super easy. I'm using traefik for Ingress. I love traefik. I may or may not be a traefik ambassador. I'm using Jenkins for CI. And I'm using StackRox for image scanning and policy enforcement. One of the things to think about though, especially in terms of labels is none of this demo stack is required. You can be in any cloud, you can be in CentOs, you can be in any Kubernetes. You can even be in swarm, if you wanted to, or Docker compose. Any Ingress, any CI system, Jenkins, circle, GitLab, it doesn't matter. And pretty much any scanning. One of the things that I think is kind of nice about at least StackRox is that we do a lot more than just image scanning, right? With the policy enforcement things like that. I guess that's kind of a shameless plug. But again, any of this stack is completely replaceable, with any comparative product in that category. So I'd like to, again, point you guys to the andyc.infodc20, that's take you right to the GitHub repo. You can reach out to me at any of the socials @clemenko or andy@stackrox.com. And thank you for attending. I hope you learned something fun about labels. And hopefully you guys can standardize labels in your organization and really kind of take your images and the image provenance to a new level. Thanks for watching. (upbeat music)

Published Date : Sep 28 2020

SUMMARY :

at org the org to the andyc

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Andy ClemenkoPERSON

0.99+

EuropeLOCATION

0.99+

100%QUANTITY

0.99+

StackRoxORGANIZATION

0.99+

two tagsQUANTITY

0.99+

PythonTITLE

0.99+

flask.docker.lifeOTHER

0.99+

andy@stackrox.comOTHER

0.99+

AndyPERSON

0.99+

andyc.info/dc20OTHER

0.99+

DockerORGANIZATION

0.99+

todayDATE

0.99+

flask.doctor.lifeOTHER

0.99+

third oneQUANTITY

0.99+

DockerfilesTITLE

0.99+

seventhQUANTITY

0.99+

KubernetesTITLE

0.98+

first oneQUANTITY

0.98+

second oneQUANTITY

0.98+

label-schema.orgOTHER

0.98+

OneQUANTITY

0.98+

KeynoteTITLE

0.98+

andyc.infodc20OTHER

0.98+

first imageQUANTITY

0.98+

First dayQUANTITY

0.97+

CentOsTITLE

0.97+

StackRoxTITLE

0.97+

SkopeoORGANIZATION

0.96+

Red HatORGANIZATION

0.96+

GitTITLE

0.96+

Ubuntu 19.10 VmsTITLE

0.95+

oneQUANTITY

0.95+

build 12OTHER

0.95+

JQTITLE

0.95+

base64TITLE

0.93+

JenkinsTITLE

0.93+

build number 12OTHER

0.91+

org.opencontainers.image.OTHER

0.91+

IngressORGANIZATION

0.89+

DODORGANIZATION

0.89+

opencontainers.image.authors.OTHER

0.89+

a secondQUANTITY

0.89+

two major waysQUANTITY

0.89+

Jenkins StackRoxTITLE

0.88+

GitosisTITLE

0.86+

GitLabORGANIZATION

0.86+

GitHubORGANIZATION

0.86+

two fun onesQUANTITY

0.84+

GitLabTITLE

0.82+

skopeoORGANIZATION

0.82+

DockerTITLE

0.81+

JSONTITLE

0.81+

traefikTITLE

0.77+

skopeoTITLE

0.76+

@clemenkoPERSON

0.74+

RancherTITLE

0.74+

IngressTITLE

0.73+

org.zdockerOTHER

0.72+

RedisTITLE

0.72+

DigitalOceanTITLE

0.71+

org.opencontainers.image.blankOTHER

0.71+

KuberORGANIZATION

0.69+

Breaking Analysis: Google Rides the Cloud Wave but Remains a Distant Third


 

>> From The Cube Studios in Palo Alto and Boston, bringing you data driven insights from The Cube and ETR, this is Breaking Analysis with Dave Vellante. >> Despite it's faster growth and infrastructure as a service, relative to AWS and Azure, Google Cloud platform remains a third wheel in the race for cloud dominance. Google begins its Cloud Next online event starting July fourteenth in a series of nine rolling sessions that go through early September. Ahead of that, we want to update you on our most current data on Google's cloud business. Hello everyone, this is Dave Vellante, and welcome to this week's Wikibon Cube insights, powered by ETR. In this session, we'll review the current state of cloud, and Google's position in the market. We'll drill into the ETR data and share fresh insights from our partner and the Cube community. So let's get right into it. You know, Google, if you think about it, was actually very early into the cloud game. Google's 2004 IPO was a milestone event for the tech industry, and in you know many ways, it really marked the end of the post-dotcom malaise. It signaled the beginning of a new era of innovation. During this time, Google was busy building out its massive, global cloud infrastructure, probably the largest in the world, with undersea cables, global data centers, and tools like the Google file system, and of course Bigtable. But it took many years for Google to pull its head out of its ad serving butt and realize the opportunity to sell its cloud services to global enterprises. Bigtable, Google's no-sequel database, for example, was released in 2005, but it wasn't until 2015 that Google made this service available to its customers. That was the same year Google brought in VMware founder, Diane Greene to begin its enterprise journey in earnest. Now Google, they have a dizzying array of services in compute, storage, database, networking, IT ops, dev tools, machine learning, AI, analytics, big data, security, on and on and on. Name a category and it's likely that Google has something in it as a cloud service. But Google, to this day, still hasn't figured out how to sell to the enterprise. It really struggles to find the right formula. So, as you know, Google brought in Thomas Kurian from Oracle, to figure this out. Of course Kurian is, he's going to go with Google's strengths like analytics and database, but it has to have differentiation, so it comes up with unique pricing models like sustained discounts, which automatically apply discount for heavy usage, as opposed to forcing users to buy reserved instances such as what AWS does. You know Google is more aggressive partnering around multi-cloud, for instance, with Anthos, and it's smartly open-sourced Kubernetes really to minimize the importance of, physically, where workloads run. The bottom-line, however, is that these moves are necessary for Google to compete because it lags behind the leaders. And it has a long way to go before it's going to be satisfied with its cloud business. Let's look at the IaaS market in context. Now, I don't want to say it's all gloom and doom for Google. Far from it. Earnings for Q2, they're going to start rolling out later this month, but this chart shows our latest estimates of IaaS and PaaS for the big three cloud players. Now, I got to caution you, as I did before, other than AWS, which reports very clean numbers each quarter on IaaS and PaaS, we have to estimate Azure and GCP revenue because they bundle in other things. I'll give an example. Google reports its overall cloud numbers which include G Suite. Microsoft reports a category they call intelligent cloud. Now that includes public, private clouds, hybrid, sequel server, Windows server, system center, GitHub, enterprise support and consulting services. And Azure, the IaaS and PaaS numbers are also in there too. So what we have to do is to squint through the earnings reports and the 10 Ks and try to get a clean IaaS and PaaS figure for these players, and that's what we show here. Now there's really two points that we want to stress with this data. First, on a trailing 12 month basis, the big three cloud players now account for nearly 60 billion dollars in IaaS and PaaS revenue. And this 60 billion dollars, on a weighted average basis, is growing in the mid 40% range. So well on its way to being a 100 billion dollar business. Just for these three firms. And as we've reported, that's eating directly into the on-premises infrastructure install base, which is a flat to declining market. And that trend is going to play out in a big way this decade. We've predicted that public cloud is going to out pace on-prem infrastructure by more that 1800 basis points over the next 10 years, from a spending standpoint. Now the second point that I want to make relates to Google IaaS and PaaS growth. We peg it at greater than 70%, based on public statements, reading the 10 Ks and ETR data, which we'll discuss in a moment. So, very healthy growth, but from a much smaller install base than, or base than AWS and Azure. But in our view it's not enough, because AWS and Azure are so large and strong still, growth wise, that we feel Google is going to remain a distant third, really indefinitely. Nonetheless, a lot of companies would be thrilled to have a four billion dollar cloud business and there's certainly good news in the data for Google. So let's look at some of that survey data. Now, as we've reported in the past, Google pushes G Suite very hard, as part of its cloud story, and it leads often times with G Suite in its messaging. You know, but to us that's never really been that compelling. So let me start with some anecdotal data from ETR. ETR runs a regular program, they call it VENN, and in the VENN they invite clients into a private session to listen to named CIOs talk about their experience with vendors and overall spending intentions. It's a facilitated session. And we've had ETR's Eric Bradley on as a guest who directs the VENN program, and does much of the facilitation, and here's a statement from a recent VENN session quoting a CIO at a midsize Telco, that I think sums it up nicely. He says Google's G Suite is fine and dandy, but I don't see that truly as an enterprise solution. And frankly, it's still not of the quality of an Office application, talking about Microsoft. All in all I really like the infrastructure-as-a-service and the platform-as-a-service components that GCP had. And I thought they were coming along very very well in that space. Now, the reason that I share this is because the IT buyers that we speak with, you know they're very serious about exploring Google. They want options other than Azure and AWS and they see Google as having great tech and as a viable alternative. So let's talk about GCP and the enterprise. We looking, when we look into the ETR data for the most recent survey, which ran in June and early July, GCP is showing strength in one really important bellwether category, the giant public and private companies. These are the largest firms in the ETR dataset and often point to secular trends. Now, before we get into that, let's look at the picture for GCP using ETR's net score up methodology. This is fundamental to the ETR approach, and remember, each quarter ETR goes out and asks its respondents, are you planning to spend more or less? In its July survey, ETR focuses on second half spending. The next chart captures results across Google's entire portfolio. So here's the breakdown for, for Google across all sectors. 14% of the respondents are adopting new, that's the lime green. 39% plan to increase spending in the second half versus the first half, that's the forest green. Then there's a big fat middle, that's flat, and you see that in the gray area. And the 7% are spending less, with 2% replacing, that's the pinkish and dark red, respectively. So, I would say this result is mixed, in my opinion. Yeah, it's not bad, don't get me wrong, and we've, we'll see once ETR comes out of its quite period, how this compares to Azure and AWR, so remember, I can only share limited data until ETR clients get the data and have time to act on it. But this calculates out to a net score of 44%, which is respectable, but frankly not overly inspiring. So let's look across the GCP portfolio using the ETR taxonomy and see what it looks like. This chart shows the net score comparisons across three different surveys, October 19, April 20, and July 20. So reading the bars left to right, you can see Google's strong suit really is machine learning and AI. Container platforms are also very strong, as are functions, or server-less, and databases, very solid, we'll talk more about that in a minute. You know, video conferencing was just added by ETR and sure it pops up with the work from home. Cloud is actually holding firm when compared to October of last year. But surprisingly, analytics is looking a bit softer. And ETR for the first time added G Suite with, it shows a 26% net score, first time out, which is pretty tepid. I mean not very impressive at all. But overall, the picture looks pretty good for Google. So let's dig further into the giant public and private sector, that bellwether I talked about. And let's peal the onion a bit and look closer at the results from the largest companies in the dataset. So this chart shows the giant public, plus private organizations. So it would include like monster public companies but also large companies like a Cargill or a Coke Industries, if in fact they responded in this survey. And you can see, in that all important sector, it's a story of a lot of green with hardly any red, so quite a positive sign for Google within those bellwethers. Here's what I think is happening here. Is these large, and often far flung organizations, have realized that they have multiple cloud vendors, and they're asking their senior IT leadership to bring some consistency and sanity to their cloud strategies. So they look at the big three and say, okay, what's the best strategic fit for each workload? So they might say for instance let's use AWS for core IaaS, let's use Azure for productivity workloads, and we'll sprinkle some Google in for machine learning and related projects. So we do see some real strength in some of the larger strongholds for Google, although interestingly ETR sort of tells me that there's softness in the midsize and smaller companies that have powered AWS for so many years. And of course this, with Google's base, but compare that to AWS and AWS is much stronger in those smaller companies, start-ups and the like, and of course COVID's the wild car in all this. You know, we have to take that into account, and we will with Sagar Kadakia, who's ETR's director of research in the coming weeks. But I want to look at Google in the all important database category. So before we wrap, let's look at database. You remember, Google's playing catch up in the cloud and its marketing takes a more open posture around partners and things like multi-cloud and you know you can contrast that with AWS for example, but look, make no mistake, Google wants you data in their cloud, and that's why database is so strategic and so important. Look, it's the mother of all lock specs. All you got to do is look at Oracle and their success. Now, as we've reported many times, there's a new workload emerging in the cloud around this idea of the modern data warehouse. I mean I don't even like that term anymore, data warehouse, because it sounds just so static. But anyway, any rate, I'm talking about workloads that bring database, machine learning, AI, data science, compute and storage along with visualization tools to deliver real-time insights and operational analytics. Database is at the heart of everything here. Win the database and everything else falls into place. Now, Google has six or seven database products and one of the most impressive, in my opinion, is BigQuery. I mean, for those who have followed me over the years you know I love the technology behind Google's banner, but BigQuery is where much of the action is around this new workload that I'm talking about. So, let's look at, deeper at Google's position in database. This chart shows one of my favorite views. On the Y axis is the net score, or spending momentum, and on the X axis is market share or pervasiveness in the ETR dataset. The chart plots various database companies and their position within the all important giant public plus private sector. So these are the companies in the ETR survey that are the largest, and oftentimes, again, are a bellwether. And you can see Microsoft and Oracle and AWS have very strong presence on the horizontal axis. Mongo, MongoDB looms large, MemSQL, they just raised 50 million dollars this past May, MariaDB just raised another 25 million this month. You can see Couchbase and Redis, they show up, and they're on my radar. I'm learning more about those companies. Folks, database is hot. VC's are pouring money in and it's something that's very important to the Cube community to look at. And of course you see Google in the chart, with a strong net score, you know, but not the type of market presence that you see from the other big cloud players. In fact, they've pulled back a little somewhat in this last ETR survey. So despite some bright spots in the enterprise in terms of spending momentum, just not quite enough presence yet. Oh, by the way, look who's right there with Google. I know I sound like a broken record, but Snowflake is everywhere. You'll find them in AWS, you'll find them in Azure and on GCP. Now remember, Snowflake is only about one tenth the size of Google's IaaS and PaaS business. But it has stronger spending momentum than all the big guys, and it continues to creep its way to the right in terms of market share or presence. You know, but Google has great database tech and BigQuery is at the heart of its strategy to support analytics at scale, and automate the data pipeline. BigQuery's very well designed, it started as a cloud native database, it's based on server-less, it's highly scalable, and it's very cost-effective. In fact, ESG, enterprise strategy group, wrote a report comparing the TCO of the cloud databases. Let me pull that up and show you. Now the report was commissioned by Google, so I got to caution you there. But it was very well done in my opinion by a guy named Aviv Kaufmann, and you can see here it compares BigQuery with the other cloud databases, and of course, you know, BigQuery wins, got the lowest TCO, but again I thought the report was really detailed and well researched. I have no doubt that Snowflake has an answer for the big brown bar, which is on-demand cloud cost. I think ESG was making certain assumptions, maybe worst case assumptions, about the need to over-provision resources for Snowflake, which I'm sure ESG can defend, but I'll bet dollars to donuts that Snowflake, you know, has an answer to that or a comeback. I'm going to ask them. But the point I want to make here is that BigQuery was designed from day one, again, as a cloud-native database. We've been talking about that a lot. It's very efficient and is going to be competitive. So you can see, there are some bright spots in the enterprise, for Google. Okay, let's wrap up. Now, having called out some of the positives, and there are many, Google is still not getting it done in the enterprise, in my opinion. I certainly would not say too little too late, but I would say they spotted the competition a huge lead, and the only reason is Google just didn't act on the opportunity staring them in the face, within the enterprise, fast enough, and they finally woke up. But enterprise sales are, they're really hard. Thomas Kurian, for all his experience, is coming from way, way behind with regard to the enterprise go to market, systems and processes, pricing, partnerships, special deals for the enterprise. Google's still learning how to sell the business outcomes and is relying far too much on its technology chops, which, while impressive, are not going to win the day without better enterprise sales, marketing, and ecosystem integration. Now I feel like for years, Google has said to the enterprise market, give me heat and I'll add the wood. Meaning we have the best tech, go ahead and use it. That strategy just doesn't work in the enterprise. Kurian knows it and I suspect that's why Google's showing some strength within these large, giant public and private companies. They're probably applying focused sales resources to nail customer success with some of its top accounts where they have a presence, and then once they nail that they'll broaden to the market. But they got to move fast. We'll learn more about Google's intentions and its progress over the next few, next few months as they try their online event experiment, and of course we'll be there providing our wall to wall coverage. Remember, these Breaking Analysis episodes, they're all available as podcasts. ETR is shortly exiting its quiet period, this week, and will be rolling out the data, so check out etr.plus. I publish weekly on wikibon.com and siloconeangle.com and as always please comment on my LinkedIn posts, I really appreciate the feedback. This is Dave Vellante for the Cube Insights, powered by ETR. Thanks for watching everyone. We'll see you next time.

Published Date : Jul 13 2020

SUMMARY :

in Palo Alto and Boston, and realize the opportunity to sell

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

KurianPERSON

0.99+

MicrosoftORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

Thomas KurianPERSON

0.99+

JuneDATE

0.99+

2005DATE

0.99+

AWSORGANIZATION

0.99+

Eric BradleyPERSON

0.99+

sixQUANTITY

0.99+

OctoberDATE

0.99+

Diane GreenePERSON

0.99+

OracleORGANIZATION

0.99+

12 monthQUANTITY

0.99+

October 19DATE

0.99+

2015DATE

0.99+

July 20DATE

0.99+

39%QUANTITY

0.99+

JulyDATE

0.99+

April 20DATE

0.99+

2%QUANTITY

0.99+

second pointQUANTITY

0.99+

7%QUANTITY

0.99+

BostonLOCATION

0.99+

TelcoORGANIZATION

0.99+

second halfQUANTITY

0.99+

60 billion dollarsQUANTITY

0.99+

14%QUANTITY

0.99+

CargillORGANIZATION

0.99+

FirstQUANTITY

0.99+

siloconeangle.comOTHER

0.99+

first halfQUANTITY

0.99+

44%QUANTITY

0.99+

Palo AltoLOCATION

0.99+

G SuiteTITLE

0.99+

ESGORGANIZATION

0.99+

Coke IndustriesORGANIZATION

0.99+

26%QUANTITY

0.99+

two pointsQUANTITY

0.99+

100 billion dollarQUANTITY

0.99+

Breaking Analysis: Emerging Tech sees Notable Decline post Covid-19


 

>> Announcer: From theCUBE studios in Palo Alto in Boston, connecting with thought leaders all around the world, this is a CUBE conversation. >> As you may recall, coming into the second part of 2019 we reported, based on ETR Survey data, that there was a narrowing of spending on emerging tech and an unplugging of a lot of legacy systems. This was really because people were going from experimentation into operationalizing their digital initiatives. When COVID hit, conventional wisdom suggested that there would be a flight to safety. Now, interestingly, we reported with Eric Bradley, based on one of the Venns, that a lot of CIOs were still experimenting with emerging vendors. But this was very anecdotal. Today, we have more data, fresh data, from the ETR Emerging Technology Study on private companies, which really does suggest that there's a notable decline in experimentation, and that's affecting emerging technology vendors. Hi, everybody, this is Dave Vellante, and welcome to this week's Wikibon Cube Insights, powered by ETR. Once again, Sagar Kadakia is joining us. Sagar is the Director of Research at ETR. Sagar, good to see you. Thanks for coming on. >> Good to see you again. Thanks for having me, Dave. >> So, it's really important to point out, this Emerging Tech Study that you guys do, it's different from your quarterly Technology Spending Intention Survey. Take us through the methodology. Guys, maybe you could bring up the first chart. And, Sagar, walk us through how you guys approach this. >> No problem. So, a lot of the viewers are used to seeing a lot of the results from the Technology Spending Intention Survey, or the TSIS, as we call it. That study, as the title says, it really tracks spending intentions on more pervasive vendors, right, Microsoft, AWS, as an example. What we're going to look at today is our Emerging Technology Study, which we conduct biannually, in May and November. This study is a little bit different. We ask CIOs around evaluations, awareness, planned evaluations, so think of this as pre-spend, right. So that's a major differentiator from the TSIS. That, and this study, really focuses on private emerging providers. We're really only focused on those really emerging private companies, say, like your Series B to Series G or H, whatever it may be, so, two big differences within those studies. And then today what we're really going to look at is the results from the Emerging Technology Study. Just a couple of quick things here. We had 811 CIOs participate, which represents about 380 billion in annual IT spend, so the results from this study matter. We had almost 75 Fortune 100s take it. So, again, we're really measuring how private emerging providers are doing in the largest organizations. And so today we're going to be reviewing notable sectors, but largely this survey tracks roughly 356 private technologies and frameworks. >> All right, guys, bring up the pie chart, the next slide. Now, Sagar, this is sort of a snapshot here, and it basically says that 44% of CIOs agree that COVID has decreased the organization's evaluation and utilization of emerging tech, despite what I mentioned, Eric Bradley's Venn, which suggested one CIO in particular said, "Hey, I always pick somebody in the lower left "of the magic quadrant." But, again, this is a static view. I know we have some other data, but take us through this, and how this compares to other surveys that you've done. >> No problem. So let's start with the high level takeaways. And I'll actually kind of get into to the point that Eric was debating, 'cause that point is true. It's just really how you kind of slice and dice the data to get to that. So, what you're looking at here, and what the overall takeaway from the Emerging Technology Study was, is, you know, you are going to see notable declines in POCs, of proof-of-concepts, any valuations because of COVID-19. Even though we had been communicating for quite some time, you know, the last few months, that there's increasing pressure for companies to further digitize with COVID-19, there are IT budget constraints. There is a huge pivot in IT resources towards supporting remote employees, a decrease in risk tolerance, and so that's why what you're seeing here is a rather notable number of CIOs, 44%, that said that they are decreasing their organization's evaluation and utilization of private emerging providers. So that is notable. >> Now, as you pointed out, you guys run this survey a couple of times a year. So now let's look at the time series. Guys, if you bring up the next chart. We can see how the sentiment has changed since last year. And, of course, we're isolating here on some of larger companies. So, take us through what this data means. >> No problem. So, how do we quantify what we just saw in the prior slide? We saw 44% of CIOs indicating that they are going to be decreasing their evaluations. But what exactly does that mean? We can pretty much determine that by looking at a lot of the data that we captured through our Emerging Technology Study. There's a lot going on in this slide, but I'll walk you through it. What you're looking at here is Fortune 1000 organizations, so we've really isolated the data to those organizations that matter. So, let's start with the teal, kind of green line first, because I think it's a little bit easier to understand. What you're looking at, Fortune 1000 evaluations, both planned and current, okay? And you're looking at a time series, one year ago and six months ago. So, two of the answer options that we provide CIOs in this survey, right, think about the survey as a grid, where you have seven answer options going horizontally, and then 300-plus vendors and technologies going vertically. For any given vendor, they can essentially indicate one of these options, two of them being on currently evaluating them or I plan to evaluate them in six months. So what you're looking at here is effectively the aggregate number, or the average number of Fortune 1000 evaluations. So if you look into May 2019, all the way on the left of that chart, that 24% roughly means that a quarter of selections made by Fortune 1000 of the survey, they selected plan to evaluate or currently evaluating. If you fast-forward six months, to the middle of the chart, November '19, it's roughly the same, one in four technologies that are Fortune 1000 selected, they indicated that I plan or am currently evaluating them. But now look at that big drop off going into May 2020, the 17%, right? So now one out of every six technologies, or one out of every selections that they made was an evaluation. So a very notable drop. And then if you look at the blue line, this is another answer option that we provided CIOs: I'm aware of the technology but I have no plans to evaluate. So this answer option essentially tracks awareness levels. If you look at the last six months, look at that big uptick from 44% to over 50%, right? So now, essentially one out of every two technologies, or private technologies that a CIO is aware of, they have no plans to evaluate. So this is going to have an impact on the general landscape, when we think about those private emerging providers. But there is one caveat, and, Dave, this is what you mentioned earlier, this is what Eric was talking about. The providers that are doing well are the ones that are work-from-home aligned. And so, just like a few years ago, we were really analyzing results based on are you cloud-native or are you Cloud-aligned, because those technologies are going to do the best, what we're seeing in the emerging space is now the same thing. Those emerging providers that enable organizations to maintain productivity for their employees, essentially allowing their employees to work remotely, those emerging providers are still doing well. And that is probably the second biggest takeaway from this study. >> So now what we're seeing here is this flight to perceive safety, which, to your point, Sagar, doesn't necessarily mean good news for all enterprise tech vendors, but certainly for those that are positioned for the work-from-home pivot. So now let's take a look at a couple of sectors. We'll start with information security. We've reported for years about how the perimeter's been broken down, and that more spend was going to shift from inside the moat to a distributed network, and that's clearly what's happened as a result of COVID. Guys, if you bring up the next chart. Sagar, you take us through this. >> No problem. And as you imagine, I think that the big theme here is zero trust. So, a couple of things here. And let me just explain this chart a little bit, because we're going to be going through a couple of these. What you're seeing on the X-axis here, is this is effectively what we're classifying as near term growth opportunity from all customers. The way we measure that effectively is we look at all the evaluations, current evaluations, planned evaluations, we look at people who are evaluated and plan to utilize these vendors. The more indications you get on that the more to the top right you're going to be. The more indications you get around I'm aware of but I don't plan to evaluate, or I'm replacing this early-stage vendor, the further down and on the left you're going to be. So, on the X-axis you have near term growth opportunity from all customers, and on the Y-axis you have near term growth opportunity from, really, the biggest shops in the world, your Global 2000, your Forbes Private 225, like Cargill, as an example, and then, of course, your federal agencies. So you really want to be positioned up and to the right here. So, the big takeaway here is zero trust. So, just a couple of things on this slide when we think about zero trust. As organizations accelerate their Cloud and Saas spend because of COVID-19, and, you know, what we were talking about earlier, Dave, remote work becomes the new normal, that perimeter security approach is losing appeal, because the perimeter's less defined, right? Apps and data are increasingly being stored in the Cloud. That, and employees are working remotely from everywhere, and they're accessing all of these items. And so what we're seeing now is a big move into zero trust. So, if we look at that chart again, what you're going to see in that upper right quadrant are a lot of identity and access management players. And look at the bifurcation in general. This is what we were talking about earlier in terms of the landscape not doing well. Most security vendors are in that red area, you know, in the middle to the bottom. But if you look at the top right, what are you seeing here? Unify ID, Auth0, WSO2, right, all identity and access management players. These are critical in your zero trust approach, and this is one of the few area where we are seeing upticks. You also see here BitSight, Lucideus. So that's going to be security assessment. You're seeing VECTRA and Netskope and Darktrace, and a few others here. And Cloud Security and IDPS, Intrusion Detection and Prevention System. So, very few sectors are seeing an uptick, very few security sectors actually look pretty good, based on opportunities that are coming. But, essentially, all of them are in that work-from-home aligned security stack, so to speak. >> Right, and of course, as we know, as we've been reporting, buyers have options, from both established companies and these emerging companies that are public, Okta, CrowdStrike, Zscaler. We've seen the work-from-home pivot benefit those guys, but even Palo Alto Networks, even CISCO, I asked (other speaker drowns out speech) last week, I said, "Hey, what about this pivot to work from home? "What about this zero trust?" And he said, "Look, the reality is, yes, "a big part of our portfolio is exposed "to that traditional infrastructure, "but we have options for zero trust as well." So, from a buyer's standpoint, that perceived flight to safety, you have a lot of established vendors, and that clearly is showing up in your data. Now, the other sector that we want to talk about is database. We've been reporting a lot on database, data warehouse. So, why don't you take us through the next graphic here, if you would. >> Sagar: No problem. So, our theme here is that Snowflake is really separating itself from the pack, and, again, you can see that here. Private database and data warehousing vendors really continue to impact a lot of their public peers, and Snowflake is leading the way. We expect Snowflake to gain momentum in the next few years. And, look, there's some rumors that IPOing soon. And so when we think about that set-up, we like it, because as organizations transition away from hybrid Cloud architectures to 100% or near-100% public Cloud, Snowflake is really going to benefit. So they look good, their data stacks look pretty good, right, that's resiliency, redundancy across data centers. So we kind of like them as well. Redis Labs bring a DB and they look pretty good here on the opportunity side, but we are seeing a little bit of churn, so I think probably Snowflake and DataStax are probably our two favorites here. And again, when you think about Snowflake, we continue to think more pervasive vendors, like Paradata and Cloudera, and some of the other larger database firms, they're going to continue seeing wallet and market share losses due to some of these emerging providers. >> Yeah. If you could just keep that slide up for a second, I would point out, in many ways Snowflake is kind of a safer bet, you know, we talk about flight to safety, because they're well-funded, they're established. You can go from zero to Snowflake very quickly, that's sort of their mantra, if you will. But I want to point out and recognize that it is somewhat oranges and tangerines here, Snowflake being an analytical database. You take MariaDB, for instance, I look at that, anyway, as relational and operational. And then you mentioned DataStax. I would say Couchbase, Redis Labs, Aerospike. Cockroach is really a... EValue Store. You've got some non-relational databases in there. But we're looking at the entire sector of databases, which has become a really interesting market. But again, some of those established players are going to do very well, and I would put Snowflake on that cusp. As you pointed out, Bloomberg broke the story, I think last week, that they were contemplating an IPO, which we've known for a while. >> Yeah. And just one last thing on that. We do like some of the more pervasive players, right. Obviously, AWS, all their products, Redshift and DynamoDB. Microsoft looks really good. It's just really some of the other legacy ones, like the Teradatas, the Oracles, the Hadoops, right, that we are going to be impacted. And so the claw providers look really good. >> So, the last decade has really brought forth this whole notion of DevOps, infrastructure as code, the whole API economy. And that's the piece we want to jump into now. And there are some real stand-outs here, you know, despite the early data that we showed you, where CIOs are less prone to look at emerging vendors. There are some, for instance, if you bring up the next chart, guys, like Hashi, that really are standing out, aren't they? >> That's right, Dave. So, again, what you're seeing here is you're seeing that bifurcation that we were talking about earlier. There are a lot of infrastructure software vendors that are not positioned well, but if you look at the ones at the top right that are positioned well... We have two kind of things on here, starting with infrastructure automation. We think a winner here is emerging with Terraform. Look all the way up to the right, how well-positioned they are, how many opportunities they're getting. And for the second straight survey now, Terraform is leading along their peers, Chef, Puppet, SaltStack. And they're leading their peers in so many different categories, notably on allocating more spend, which is obviously very important. For Chef, Puppet and SaltStack, which you can see a little bit below, probably a little bit higher than the middle, we are seeing some elevator churn levels. And so, really, Terraform looks like they're kind of separating themselves. And we've got this great quote from the CIO just a few months ago, on why Terraform is likely pulling away, and I'll read it out here quickly. "The Terraform tool creates "an entire infrastructure in a box. "Unlike vendors that use procedural languages, "like Ants, Bull and Chef, "it will show you the infrastructure "in the way you want it to be. "You don't have to worry about "the things that happen underneath." I know some companies where you can put your entire Amazon infrastructure through Terraform. If Amazon disappears, if your availability drops, load balancers, RDS, everything, you just run Terraform and everything will be created in 10 to 15 minutes. So that shows you the power of Terraform and why we think it's ranked better than some of the other vendors. >> Yeah, I think that really does sum it up. And, actually, guys, if you don't mind bringing that chart back up again. So, a point out, so, Mitchell Hashimoto, Hashi, really, I believe I'm correct, talking to Stu about this a little bit, he sort of led the Terraform project, which is an Open Source project, and, to your point, very easy to deploy. Chef, Puppet, Salt, they were largely disrupted by Cloud, because they're designed to automate deployment largely on-prem and DevOps, and now Terraform sort of packages everything up into a platform. So, Hashi actually makes money, and you'll see it on this slide, and things, Vault, which is kind of their security play. You see GitLab on here. That's really application tooling to deploy code. You see Docker containers, you know, Docker, really all about open source, and they've had great adoption, Docker's challenge has always been monetization. You see Turbonomic on here, which is application resource management. You can't go too deep on these things, but it's pretty deep within this sector. But we are comparing different types of companies, but just to give you a sense as to where the momentum is. All right, let's wrap here. So maybe some final thoughts, Sagar, on the Emerging Technology Study, and then what we can expect in the coming month here, on the update in the Technology Spending Intention Study, please. >> Yeah, no problem. One last thing on the zero trust side that has been a big issue that we didn't get to cover, is VPN spend. Our data is pointing that, yes, even though VPN spend did increase the last few months because of remote work, we actually think that people are going to move away from that as they move onto zero trust. So just one last point on that, just in terms of overall thoughts, you know, again, as we cover it, you can see how bifurcated all these spaces are. Really, if we were to go sector by sector by sector, right, storage and block chain and MLAI and all that stuff, you would see there's a few or maybe one or two vendors doing well, and the majority of vendors are not seeing as many opportunities. And so, again, are you work-from-home aligned? Are you the best vendor of all the other emerging providers? And if you fit those two criteria then you will continue seeing POCs and evaluations. And if you don't fit that criteria, unfortunately, you're going to see less opportunities. So think that's really the big takeaway on that. And then, just in terms of next steps, we're already transitioning now to our next Technology Spending Intention Survey. That launched last week. And so, again, we're going to start getting a feel for how CIOs are spending in 2H-20, right, so, for the back half of the year. And our question changes a little bit. We ask them, "How do you plan on spending in the back half year "versus how you actually spent "in the first half of the year, or 1H-20?" So, we're kind of, tighten the screw, so to speak, and really getting an idea of what's spend going to look like in the back half, and we're also going to get some updates as it relates to budget impacts from COVID-19, as well as how vendor-relationships have changed, as well as business impacts, like layoffs and furloughs, and all that stuff. So we have a tremendous amount of data that's going to be coming in the next few weeks, and it should really prepare us for what to see over the summer and into the fall. >> Yeah, very excited, Sagar, to see that. I just wanted to double down on what you said about changes in networking. We've reported with you guys on NPLS networks, shifting to SD-WAN. But even VPN and SD-WAN are being called into question as the internet becomes the new private network. And so lots of changes there. And again, very excited to see updated data, return of post-COVID, as we exit this isolation economy. Really want to point out to folks that this is not a snapshot survey, right? This is an ongoing exercise that ETR runs, and grateful for our partnership with you guys. Check out ETR.plus, that's the ETR website. I publish weekly on Wikibon.com and SiliconANGLE.com. Sagar, thanks so much for coming on. Once again, great to have you. >> Thank you so much, for having me, Dave. I really appreciate it, as always. >> And thank you for watching this episode of theCube Insights, powered by ETR. This Dave Vellante. We'll see you next time. (gentle music)

Published Date : Jun 22 2020

SUMMARY :

leaders all around the world, Sagar is the Director of Research at ETR. Good to see you again. So, it's really important to point out, So, a lot of the viewers that COVID has decreased the of slice and dice the data So now let's look at the time series. by looking at a lot of the data is this flight to perceive safety, and on the Y-axis you have Now, the other sector that we and Snowflake is leading the way. And then you mentioned DataStax. And so the claw providers And that's the piece we "in the way you want it to be. but just to give you a sense and the majority of vendors are not seeing on what you said about Thank you so much, for having me, Dave. And thank you for watching this episode

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SagarPERSON

0.99+

Dave VellantePERSON

0.99+

EricPERSON

0.99+

May 2019DATE

0.99+

CISCOORGANIZATION

0.99+

DavePERSON

0.99+

twoQUANTITY

0.99+

May 2020DATE

0.99+

Eric BradleyPERSON

0.99+

Palo AltoLOCATION

0.99+

TerraformORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

Mitchell HashimotoPERSON

0.99+

100%QUANTITY

0.99+

ZscalerORGANIZATION

0.99+

oneQUANTITY

0.99+

44%QUANTITY

0.99+

ETRORGANIZATION

0.99+

AWSORGANIZATION

0.99+

last yearDATE

0.99+

November '19DATE

0.99+

Palo Alto NetworksORGANIZATION

0.99+

24%QUANTITY

0.99+

10QUANTITY

0.99+

17%QUANTITY

0.99+

MayDATE

0.99+

AmazonORGANIZATION

0.99+

last weekDATE

0.99+

Redis LabsORGANIZATION

0.99+

CouchbaseORGANIZATION

0.99+

OktaORGANIZATION

0.99+

AerospikeORGANIZATION

0.99+

COVID-19OTHER

0.99+

ParadataORGANIZATION

0.99+

811 CIOsQUANTITY

0.99+

HashiPERSON

0.99+

CrowdStrikeORGANIZATION

0.99+

one caveatQUANTITY

0.99+

NovemberDATE

0.99+

two criteriaQUANTITY

0.99+

Series GOTHER

0.99+

BostonLOCATION

0.99+

X-axisORGANIZATION

0.99+

todayDATE

0.99+

bothQUANTITY

0.99+

BloombergORGANIZATION

0.99+

ClouderaORGANIZATION

0.99+

DataStaxORGANIZATION

0.99+

two kindQUANTITY

0.99+

six months agoDATE

0.99+

15 minutesQUANTITY

0.99+

TodayDATE

0.99+

six monthsQUANTITY

0.98+

Sagar KadakiaPERSON

0.98+

about 380 billionQUANTITY

0.98+

OraclesORGANIZATION

0.98+

one year agoDATE

0.98+

MariaDBTITLE

0.98+

over 50%QUANTITY

0.98+

zero trustQUANTITY

0.98+

two vendorsQUANTITY

0.98+

Series BOTHER

0.98+

first chartQUANTITY

0.98+

UNLIST TILL 4/2 - Vertica @ Uber Scale


 

>> Sue: Hi, everybody. Thank you for joining us today, for the Virtual Vertica BDC 2020. This breakout session is entitled "Vertica @ Uber Scale" My name is Sue LeClaire, Director of Marketing at Vertica. And I'll be your host for this webinar. Joining me is Girish Baliga, Director I'm sorry, user, Uber Engineering Manager of Big Data at Uber. Before we begin, I encourage you to submit questions or comments during the virtual session. You don't have to wait, just type your question or comment in the question box below the slides and click Submit. There will be a Q and A session, at the end of the presentation. We'll answer as many questions as we're able to during that time. Any questions that we don't address, we'll do our best to answer offline. Alternately, you can also Vertica forums to post your questions there after the session. Our engineering team is planning to join the forums to keep the conversation going. And as a reminder, you can maximize your screen by clicking the double arrow button, in the lower right corner of the slides. And yet, this virtual session is being recorded, and you'll be able to view on demand this week. We'll send you a notification as soon as it's ready. So let's get started. Girish over to you. >> Girish: Thanks a lot Sue. Good afternoon, everyone. Thanks a lot for joining this session. My name is Girish Baliga. And as Sue mentioned, I manage interactive and real time analytics teams at Uber. Vertica is one of the main platforms that we support, and Vertica powers a lot of core business use cases. In today's talk, I wanted to cover two main things. First, how Vertica is powering critical business use cases, across a variety of orgs in the company. And second, how we are able to do this at scale and with reliability, using some of the additional functionalities and systems that we have built into the Vertica ecosystem at Uber. And towards the end, I also have a little extra bonus for all of you. I will be sharing an easy way for you to take advantage of, many of the ideas and solutions that I'm going to present today, that you can apply to your own Vertica deployments in your companies. So stick around and put on your seat belts, and let's go start on the ride. At Uber, our mission is to ignite opportunity by setting the world in motion. So we are focused on solving mobility problems, and enabling people all over the world to solve their local problems, their local needs, their local issues, in a manner that's efficient, fast and reliable. As our CEO Dara has said, we want to become the mobile operating system of local cities and communities throughout the world. As of today, Uber is operational in over 10,000 cities around the world. So, across our various business lines, we have over 110 million monthly users, who use our rides, services, or eat services, and a whole bunch of other services that we provide to Uber. And just to give you a scale of our daily operations, we in the ride business, have over 20 million trips per day. And that each business is also catching up, particularly during the recent times that we've been having. And so, I hope these numbers give you a scale of the amount of data, that we process each and every day. And support our users in their analytical and business reporting needs. So who are these users at Uber? Let's take a quick look. So, Uber to describe it very briefly, is a lot like Amazon. We are largely an operation and logistics company. And employee work based reflects that. So over 70% of our employees work in teams, which come under the umbrella of Community Operations and Centers of Excellence. So these are all folks working in various cities and towns that we operate around the world, and run the Uber businesses, as somewhat local businesses responding to local needs, local market conditions, local regulation and so forth. And Vertica is one of the most important tools, that these folks use in their day to day business activities. So they use Vertica to get insights into how their businesses are going, to deeply into any issues that they want to triage , to generate reports, to plan for the future, a whole lot of use cases. The second big class of users, are in our marketplace team. So marketplace is the engineering team, that backs our ride shared business. And as part of this, running this business, a key problem that they have to solve, is how to determine what prices to set, for particular rides, so that we have a good match between supply and demand. So obviously the real time pricing decisions they're made by serving systems, with very detailed and well crafted machine learning models. However, the training data that goes into this models, the historical trends, the insights that go into building these models, a lot of these things are powered by the data that we store, and serve out of Vertica. Similarly, in each business, we have use cases spanning all the way from engineering and back-end systems, to support operations, incentives, growth, and a whole bunch of other domains. So the big class of applications that we support across a lot of these business lines, is dashboards and reporting. So we have a lot of dashboards, which are built by core data analysts teams and shared with a whole bunch of our operations and other teams. So these are dashboards and reports that run, periodically say once a week or once a day even, depending on the frequency of data that they need. And many of these are powered by the data, and the analytics support that we provide on our Vertica platform. Another big category of use cases is for growth marketing. So this is to understand historical trends, figure out what are various business lines, various customer segments, various geographical areas, doing in terms of growth, where it is necessary for us to reinvest or provide some additional incentives, or marketing support, and so forth. So the analysis that backs a lot of these decisions, is powered by queries running on Vertica. And finally, the heart and soul of Uber is data science. So data science is, how we provide best in class algorithms, pricing, and matching. And a lot of the analysis that goes into, figuring out how to build these systems, how to build the models, how to build the various coefficients and parameters that go into making real time decisions, are based on analysis that data scientists run on Vertica systems. So as you can see, Vertica usage spans a whole bunch of organizations and users, all across the different Uber teams and ecosystems. Just to give you some quick numbers, we have over 5000 weekly active, people who run queries at least once a week, to do some critical business role or problem to solve, that they have in their day to day operations. So next, let's see how Vertica fits into the Uber data ecosystem. So when users open up their apps, and request for a ride or order food delivery on each platform, the apps are talking to our serving systems. And the serving systems use online storage systems, to store the data as the trips and eat orders are getting processed in real time. So for this, we primarily use an in house built, key value storage system called Schemaless, and an open source system called Cassandra. We also have other systems like MySQL and Redis, which we use for storing various bits of data to support serving systems. So all of this operations generates a lot of data, that we then want to process and analyze, and use for our operational improvements. So, we have ingestion systems that periodically pull in data from our serving systems and land them in our data lake. So at Uber a data lake is powered by Hadoop, with files stored on HDFS clusters. So once the raw data lines on the data lake, we then have ETL jobs that process these raw datasets, and generate, modeled and customize datasets which we then use for further analysis. So once these model datasets are available, we load them into our data warehouse, which is entirely powered by Vertica. So then we have a business intelligence layer. So with internal tools, like QueryBuilder, which is a UI interface to write queries, and look at results. And it read over the front-end sites, and Dashbuilder, which is a dash, board building tool, and report management tool. So these are all various tools that we have built within Uber. And these can talk to Vertica and run SQL queries to power, whatever, dashboards and reports that they are supporting. So this is what the data ecosystem looks like at Uber. So why Vertica and what does it really do for us? So it powers insights, that we show on dashboards as folks use, and it also powers reports that we run periodically. But more importantly, we have some core, properties and core feature sets that Vertica provides, which allows us to support many of these use cases, very well and at scale. So let me take a brief tour of what these are. So as I mentioned, Vertica powers Uber's data warehouse. So what this means is that we load our core fact and dimension tables onto Vertica. The core fact tables are all the trips, all the each orders and all these other line items for various businesses from Uber, stored as partitioned tables. So think of having one partition per day, as well as dimension tables like cities, users, riders, career partners and so forth. So we have both these two kinds of datasets, which will load into Vertica. And we have full historical data, all the way since we launched these businesses to today. So that folks can do deeper longitudinal analysis, so they can look at patterns, like how the business has grown from month to month, year to year, the same month, over a year, over multiple years, and so forth. And, the really powerful thing about Vertica, is that most of these queries, you run the deep longitudinal queries, run very, very fast. And that's really why we love Vertica. Because we see query latency P90s. That is 90 percentile of all queries that we run on our platform, typically finish in under a minute. So that's very important for us because Vertica is used, primarily for interactive analytics use cases. And providing SQL query execution times under a minute, is critical for our users and business owners to get the most out of analytics and Big Data platforms. Vertica also provides a few advanced features that we use very heavily. So as you might imagine, at Uber, one of the most important set of use cases we have is around geospatial analytics. In particular, we have some critical internal dashboards, that rely very heavily on being able to restrict datasets by geographic areas, cities, source destination pairs, heat maps, and so forth. And Vertica has a rich array of functions that we use very heavily. We also have, support for custom projections in Vertica. And this really helps us, have very good performance for critical datasets. So for instance, in some of our core fact tables, we have done a lot of query and analysis to figure out, how users run their queries, what kind of columns they use, what combination of columns they use, and what joints they do for typical queries. And then we have laid out our custom projections to maximize performance on these particular dimensions. And the ability to do that through Vertica, is very valuable for us. So we've also had some very successful collaborations, with the Vertica engineering team. About a year and a half back, we had open-sourced a Python Client, that we had built in house to talk to Vertica. We were using this Python Client in our business intelligence layer that I'd shown on the previous slide. And we had open-sourced it after working closely with Eng team. And now Vertica formally supports the Python Client as an open-source project, which you can download to and integrate into your systems. Another more recent example of collaboration is the Vertica Eon mode on GCP. So as most of or at least some of you know, Vertica Eon mode is formally supported on AWS. And at Uber, we were also looking to see if we could run our data infrastructure on GCP. So Vertica team hustled on this, and provided us early preview version, which we've been testing out to see how performance, is impacted by running on the Cloud, and on GCP. And so far, I think things are going pretty well, but we should have some numbers about this very soon. So here I have a visualization of an internal dashboard, that is powered solely by data and queries running on Vertica. So this GIF has sequence have different visualizations supported by this tool. So for instance, here you see a heat map, downgrading heat map of source of traffic demand for ride shares. And then you will see a bunch of arrows here about source destination pairs and the trip lines. And then you can see how demand moves around. So, as the cycles through the various animations, you can basically see all the different kinds of insights, and query shapes that we send to Vertica, which powers this critical business dashboard for our operations teams. All right, so now how do we do all of this at scale? So, we started off with a single Vertica cluster, a few years back. So we had our data lake, the data would land into Vertica. So these are the core fact and dimension tables that I just spoke about. And then Vertica powers queries at our business intelligence layer, right? So this is a very simple, and effective architecture for most use cases. But at Uber scale, we ran into a few problems. So the first issue that we have is that, Uber is a pretty big company at this point, with a lot of users sending almost millions of queries every week. And at that scale, what we began to see was that a single cluster was not able to handle all the query traffic. So for those of you who have done an introductory course, on queueing theory, you will realize that basically, even though you could have all the query is processed through a single serving system. You will tend to see larger and larger queue wait times, as the number of queries pile up. And what this means in practice for end users, is that they are basically just seeing longer and longer query latencies. But even though the actual query execution time on Vertica itself, is probably less than a minute, their query sitting in the queue for a bunch of minutes, and that's the end user perceived latency. So this was a huge problem for us. The second problem we had was that the cluster becomes a single point of failure. Now Vertica can handle single node failures very gracefully, and it can probably also handle like two or three node failures depending on your cluster size and your application. But very soon, you will see that, when you basically have beyond a certain number of failures or nodes in maintenance, then your cluster will probably need to be restarted or you will start seeing some down times due to other issues. So another example of why you would have to have a downtime, is when you're upgrading software in your clusters. So, essentially we're a global company, and we have users all around the world, we really cannot afford to have downtime, even for one hour slot. So that turned out to be a big problem for us. And as I mentioned, we could have hardware issues. So we we might need to upgrade our machines, or we might need to replace storage or memory due to issues with the hardware in there, due to normal wear and tear, or due to abnormal issues. And so because of all of these things, having a single point of failure, having a single cluster was not really practical for us. So the next thing we did, was we set up multiple clusters, right? So we had a bunch of identities clusters, all of which have the same datasets. So then we would basically load data using ingestion pipelines from our data lake, onto each of these clusters. And then the business intelligence layer would be able to query any of these clusters. So this actually solved most of the issues that I pointed out in the previous slide. So we no longer had a single point of failure. Anytime we had to do version upgrades, we would just take off one cluster offline, upgrade the software on it. If we had node failures, we would probably just take out one cluster, if we had to, or we would just have some spare nodes, which would rotate into our production clusters and so forth. However, having multiple clusters, led to a new set of issues. So the first problem was that since we have multiple clusters, you would end up with inconsistent schema. So one of the things to understand about our platform, is that we are an infrastructure team. So we don't actually own or manage any of the data that is served on Vertica clusters. So we have dataset owners and publishers, who manage their own datasets. Now exposing multiple clusters to these dataset owners. Turns out, it's not a great idea, right? Because they are not really aware of, the importance of having consistency of schemas and datasets across different clusters. So over time, what we saw was that the schema for the same tables would basically get out of order, because they were all the updates are not consistently applied on all clusters. Or maybe they were just experimenting some new columns or some new tables in one cluster, but they forgot to delete it, whatever the case might be. We basically ended up in a situation where, we saw a lot of inconsistent schemas, even across some of our core tables in our different clusters. A second issue was, since we had ingestion pipelines that were ingesting data independently into all these clusters, these pipelines could fail independently as well. So what this meant is that if, for instance, the ingestion pipeline into cluster B failed, then the data there would be older than clusters A and C. So, when a query comes in from the BI layer, and if it happens to hit B, you would probably see different results, than you would if you went to a or C. And this was obviously not an ideal situation for our end users, because they would end up seeing slightly inconsistent, slightly different counts. But then that would lead to a bad situation for them where they would not able to fully trust the data that was, and the results and insights that were being returned by the SQL queries and Vertica systems. And then the third problem was, we had a lot of extra replication. So the 20/80 Rule, or maybe even the 90/10 Rule, applies to datasets on our clusters as well. So less than 10% of our datasets, for instance, in 90% of the queries, right? And so it doesn't really make sense for us to replicate all of our data on all the clusters. And so having this set up where we had to do that, was obviously very suboptimal for us. So then what we did, was we basically built some additional systems to solve these problems. So this brings us to our Vertica ecosystem that we have in production today. So on the ingestion side, we built a system called Vertica Data Manager, which basically manages all the ingestion into various clusters. So at this point, people who are managing datasets or dataset owners and publishers, they no longer have to be aware of individual clusters. They just set up their ingestion pipelines with an endpoint in Vertica Data Manager. And the Vertica Data Manager ensures that, all the schemas and data is consistent across all our clusters. And on the query side, we built a proxy layer. So what this ensures is that, when queries come in from the BI layer, the query was forwarded, smartly and with knowledge and data about which cluster up, which clusters are down, which clusters are available, which clusters are loaded, and so forth. So with these two layers of abstraction between our ingestion and our query, we were able to have a very consistent, almost single system view of our entire Vertica deployment. And the third bit, we had put in place, was the data manifest, which were the communication mechanism between ingestion and proxy. So the data manifest basically is a listing of, which tables are available on which clusters, which clusters are up to date, and so forth. So with this ecosystem in place, we were also able to solve the extra replication problem. So now we basically have some big clusters, where all the core tables, and all the tables, in fact, are served. So any query that hits 90%, less so tables, goes to the big clusters. And most of the queries which hit 10% heavily queried important tables, can also be served by many other small clusters, so much more efficient use of resources. So this basically is the view that we have today, of Vertica within Uber, so external to our team, folks, just have an endpoint, where they basically set up their ingestion jobs, and another endpoint where they can forward their Vertica SQL queries. And they are so to a proxy layer. So let's get a little more into details, about each of these layers. So, on the data management side, as I mentioned, we have two kinds of tables. So we have dimension tables. So these tables are updated every cycle, so the list of cities list of drivers, the list of users and so forth. So these change not so frequently, maybe once a day or so. And so we are able to, and since these datasets are not very big, we basically swap them out on every single cycle. Whereas the fact tables, so these are tables which have information about our trips or each orders and so forth. So these are partition. So we have one partition roughly per day, for the last couple of years, and then we have more of a hierarchical partitions set up for older data. So what we do is we load the partitions for the last three days on every cycle. The reason we do that, is because not all our data comes in at the same time. So we have updates for trips, going over the past two or three days, for instance, where people add ratings to their trips, or provide feedback for drivers and so forth. So we want to capture them all in the row corresponding to that particular trip. And so we upload partitions for the last few days to make sure we capture all those updates. And we also update older partitions, if for instance, records were deleted for retention purposes, or GDPR purposes, for instance, or other regulatory reasons. So we do this less frequently, but these are also updated if necessary. So there are endpoints which allow dataset owners to specify what partitions they want to update. And as I mentioned, data is typically managed using a hierarchical partitioning scheme. So in this way, we are able to make sure that, we take advantage of the data being clustered by day, so that we don't have to update all the data at once. So when we are recovering from an cluster event, like a version upgrade or software upgrade, or hardware fix or failure handling, or even when we are adding a new cluster to the system, the data manager takes care of updating the tables, and copying all the new partitions, making sure the schemas are all right. And then we update the data and schema consistency and make sure everything is up to date before we, add this cluster to our serving pool, and the proxy starts sending traffic to it. The second thing that the data manager provides is consistency. So the main thing we do here, is we do atomic updates of our tables and partitions for fact tables using a two-phase commit scheme. So what we do is we load all the new data in temp tables, in all the clusters in phase one. And then when all the clusters give us access signals, then we basically promote them to primary and set them as the main serving tables for incoming queries. We also optimize the load, using Vertica Data Copy. So what this means is earlier, in a parallel pipelines scheme, we had to ingest data individually from HDFS clusters into each of the Vertica clusters. That took a lot of HDFS bandwidth. But using this nice feature that Vertica provides called Vertica Data Copy, we just load it data into one cluster and then much more efficiently copy it, to the other clusters. So this has significantly reduced our ingestion overheads, and speed it up our load process. And as I mentioned as the second phase of the commit, all data is promoted at the same time. Finally, we make sure that all the data is up to date, by doing some checks around the number of rows and various other key signals for freshness and correctness, which we compare with the data in the data lake. So in terms of schema changes, VDM automatically applies these consistently across all the clusters. So first, what we do is we stage these changes to make sure that these are correct. So this catches errors that are trying to do, an incompatible update, like changing a column type or something like that. So we make sure that schema changes are validated. And then we apply them to all clusters atomically again for consistency. And provide a overall consistent view of our data to all our users. So on the proxy side, we have transparent support for, replicated clusters to all our users. So the way we handle that is, as I mentioned, the cluster to table mapping is maintained in the manifest database. And when we have an incoming query, the proxy is able to see which cluster has all the tables in that query, and route the query to the appropriate cluster based on the manifest information. Also the proxy is aware of the health of individual clusters. So if for some reason a cluster is down for maintenance or upgrades, the proxy is aware of this information. And it does the monitoring based on query response and execution times as well. And it uses this information to route queries to healthy clusters, and do some load balancing to ensure that we award hotspots on various clusters. So the key takeaways that I have from the stock, are primarily these. So we started off with single cluster mode on Vertica, and we ran into a bunch of issues around scaling and availability due to cluster downtime. We had then set up a bunch of replicated clusters to handle the scaling and availability issues. Then we run into issues around schema consistency, data staleness, and data replication. So we built an entire ecosystem around Vertica, with abstraction layers around data management and ingestion, and proxy. And with this setup, we were able to enforce consistency and improve storage utilization. So, hopefully this gives you all a brief idea of how we have been able to scale Vertica usage at Uber, and power some of our most business critical and important use cases. So as I mentioned at the beginning, I have a interesting and simple extra update for you. So an easy way in which you all can take advantage of many of the features that we have built into our ecosystem, is to use the Vertica Eon mode. So the Vertica Eon mode, allows you to set up multiple clusters with consistent data updates, and set them up at various different sizes to handle different query loads. And it automatically handles many of these issues that I mentioned in our ecosystem. So do check it out. We've also been, trying it out on DCP, and initial results look very, very promising. So thank you all for joining me on this talk today. I hope you guys learned something new. And hopefully you took away something that you can also apply to your systems. We have a few more time for some questions. So I'll pause for now and take any questions.

Published Date : Mar 30 2020

SUMMARY :

Any questions that we don't address, So the first issue that we have is that,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Girish BaligaPERSON

0.99+

UberORGANIZATION

0.99+

GirishPERSON

0.99+

10%QUANTITY

0.99+

one hourQUANTITY

0.99+

Sue LeClairePERSON

0.99+

90%QUANTITY

0.99+

AmazonORGANIZATION

0.99+

AWSORGANIZATION

0.99+

SuePERSON

0.99+

twoQUANTITY

0.99+

VerticaORGANIZATION

0.99+

DaraPERSON

0.99+

first issueQUANTITY

0.99+

less than a minuteQUANTITY

0.99+

MySQLTITLE

0.99+

FirstQUANTITY

0.99+

first problemQUANTITY

0.99+

third problemQUANTITY

0.99+

third bitQUANTITY

0.99+

less than 10%QUANTITY

0.99+

each platformQUANTITY

0.99+

secondQUANTITY

0.99+

one clusterQUANTITY

0.99+

oneQUANTITY

0.99+

second issueQUANTITY

0.99+

PythonTITLE

0.99+

todayDATE

0.99+

second phaseQUANTITY

0.99+

two kindsQUANTITY

0.99+

over 10,000 citiesQUANTITY

0.99+

over 70%QUANTITY

0.99+

each businessQUANTITY

0.99+

second thingQUANTITY

0.98+

second problemQUANTITY

0.98+

VerticaTITLE

0.98+

bothQUANTITY

0.98+

Vertica Data ManagerTITLE

0.98+

two-phaseQUANTITY

0.98+

firstQUANTITY

0.98+

90 percentileQUANTITY

0.98+

once a weekQUANTITY

0.98+

eachQUANTITY

0.98+

single pointQUANTITY

0.97+

SQLTITLE

0.97+

once a dayQUANTITY

0.97+

RedisTITLE

0.97+

one partitionQUANTITY

0.97+

under a minuteQUANTITY

0.97+

@ Uber ScaleORGANIZATION

0.96+

Murli Thirumale, Portworx & Satish Puranam, Ford | KubeCon + CloudNativeCon NA 2019


 

(upbeat music) >> Narrator: Live, from San Diego, California, it's theCUBE! Covering KubeCon, and CloudNativeCon. Brought to you by RedHat, the Cloud Native Computing Foundation, and its ecosystem partners. >> Welcome back, this is theCUBE's fourth year of covering KubeCon and CloudNativeCon. This is the North America show here in San Diego it's 2019, he is John Troyer, I am Stu Miniman, and happy to welcome to the program, first of all, I have Murli Thirumali, who is the co-founder and CEO of Portworx, and Murli, thank you so much for bringing one of your customers on the program, Satish Puranam, who is a Technical Specialist with Ford Motor Company. Gentlemen, thank you so much for joining us. >> Delighted to be here. >> All right, so Satish, we're going to start with you because, you know, the growth of this ecosystem has been phenomenal, there were End Users up on the mainstage, we've already had them, there's over, there's 129 now CNCF End User Participants there, but, you know, bring us in Ford, you know, we were getting ready for this, we're talking, there's so much change going in from, you know, of course, everybody talks about autonomous vehicles, and what there have, but, you know, technology has really embedded itself deeply into a company like Ford, so before we get into all the crew, just, bring us a little about into your world, what's happening, changing, and, you know, what your team does. >> Sure, in like uh, Ford generally has been in like a transformation journey for about the last two years now, that includes like, completely redoing our Data Centers, our Application Portfolio, as part of this monolithic journey, we started our journey with Cloud Foundry, we have been a huge favorite to Cloud Foundry shops for some time. And then, we also would like to start dabbling with like, Kubernetes and things, associated technologies primarily do for like, looking for like, data services, messaging services, lot of the stateful things, right? Cloud Native and like, Kubernetes, and I- Cloud Foundry, I am sorry, Did great wonders for us, for qualified graphs. So what do we do with like, stateful things? And that's what we started dabbling with Kubernetes and things like that. >> Yeah, Satish, if I could, I want to step back one second here, and say, you know, you do a transformation, consolidation, moving from monoliths to microservices, what was the business driver here, was it one day, some executive got up and said, you know, "hey this sounds really cool, go do it", or was there a specific driver in the business that now, your organization needs to respond to? >> I think the business drive is cost efficiency. Like, uh, there were, like, a lot of things that we would have not done, so there's a lot of technical debt we have to pay down, because of various fragmentation and various other things, so it's always about realizing business efficiencies, and most importantly, speed at which we deliver services to our customers internally, so that was the main driving force for our engaging in this transformation journey for like, about the last few years. >> Okay, Murli we'd love to have bring you to this conversation here. You obviously, agility is one of the things I hear most from customers, the driver of what new things. Infrastructure for the longest time, in many ways, it was like a boat anchor of what held us back. >> Murli: Yep. >> Especially you know, our friends in Networking and Storage, it is difficult to change and keep up with, with what's driving there, so bring us uh, bring us up to speed with Portworx and how you fit into Ford and more broadly. >> Yeah, just a quick introduction to Portworx, we've been around for about five years, now, right from the early days of containers and Kubernetes, and you know, we have quite a few customers now in Production, we have about 130 customers, 50 of the uh, the global 2k and so on, many, almost all those customers are in Production, deploying stat significant workloads. The interesting thing about Kubernetes in the last couple of years, especially, is that everybody recognizes it has won the war for orchestrating containers and applications, but the reality is, the customer still has to manage the whole stack, the stack of not just the app but the data itself underneath, and that's kind of the role of Portworx, Portworx is the platform for storage for Kubernetes, and we orchestrate all the underlying storage and the data applications, with that being said, I think one of the things that we've seen that Ford has kind of led the way in, and has been really amazing, is some of the many surprising things that people don't really know about Kubernetes, which has been happening now with customers like Ford for a while, one of them, for example, is just the use of Kubernetes in on-prem applications. Very few people really kind of, they think of Kubernetes as something that was born in the Cloud, and therefore, has kind of really only mushroomed in the Cloud, and you know, the, one of the key things about Kubernetes, and most of our customers are actually on-prem, and it to me is transforming the Data Center. The agility that Satish speaks about, is something that you don't just need because you are operating in the Cloud, you need that for all of your on-prem applications, too, and that's been one of the unique characteristics that we've seen from Ford. >> Yeah, and that's, I mean, you talked about your journey, Satish, you know, the pivotal folks really talk a lot about transformation and agility you know, no matter where your apps were sitting, I'm kind of curious in terms of the storage and the stateful- statefulness of the applications that your working with now, you know, what kind of a, if I looked at it, the diagram, what kind of a set-up would there be? So there's a Portworx layer underneath and beside Kubernetes that's managing some of the storage and some of the replication? Is it then, is the data sitting in a, you know, on a SAN somewhere, is it sitting in the Cloud, I mean, can you kind of describe what a typical application would look like? >> With your typical application, yes we draw storage, we've been drawing storage for the past several years from NetApp as being as the primary source of our data, and then we run on top of that, we run some kind of storage overlays, we dabble with quite a few technologies, including, uh, Rook, NetApp Trident, Uh, Loster, You know I'm like a, it was a journey A journey that we took us, to ultimately lead us to Portworx and we just didn't started with Portworx, but the toughest aspect has been the gravity that the stories bring along with it, and all the things that are, Cloud Native is great but Cloud Native has stayed somewhere and that has to be managed someplace, and we said "Hey, can we do that with Kubernetes?" Right? So, I think we have done a- I won't say an outstanding job but at least we've done a reasonably good job at actually at least wrapping our heads around it and we have quite a few workloads in production that are actually stateful, whether they are Base Systems, uh, there are also like Data Messaging Systems, many cards applications and all that stuff so that has been something that we have been working on for the past few years on our platforms at least. >> Yeah, well I wonder if you could expand a little bit on kind of the application suite you know, "What can we do? What can't we do?" Listen to the keynote this morning I definitely heard it was, if you look at a multi cluster environment, You know, you want to mirror and have the same things there. Well I can't just magically have all the data everywhere and data has gravity and the laws of physics do apply so I can't just automatically replicate terabytes from here to the Cloud or back so help us understand where we are. >> So, you know, one of the, uh, one of the things Satish told me yesterday which I loved was he saying, he said: "Stateful is almost easier than stateless now because of the fact that we have these extensions of Kubenetes." So, one of the things that's been very very impactful is that Kubernetes is now these extensions for managing you know, storage networking and so on, and in fact the way they do that is through an API that just an overlay, so we are an example of an overlay. And so think about it this way, if a customer about 60 percent of our customers are building a platform as their service, in many cases they don't even know what applications are going to be in there, so over our customer base we see the same alphabet soup over and over and over again. Guess what it is, Postgress, Cassandra's, all the databases Redis, right? You know, all of the messaging queues, right? Things like Kafta and uh, you know, Streaming Data, for example, Spark workloads. And so, one of the key things that is happening around with customers particularly on the enterprise side, like large enterprises, they are using all kinds of applications and they're all stateful. I mean they're very few enterprises that are not stateful and they're all running on some kind of a storage substrate that has virtualized the underlying storage. So we run on top of the underlying hardware, but then we're enabled to kind of work with all of the orchestration that Kubernetes provides but we're adding the orchestration of the Data infrastructure as well as the storage itself And I think that's one of the key things that's changed with Kubernetes in the last, I would say, two and a half years is, most people used to think of it as "in the cloud and stateless" but now it's "on-prem and stateful." >> So Satish, one of the things we've talked to customers is their journey of modernizing their applications, it's, there's things that you might build brand-new and are great here but, you know, I'm sure you have thousands of applications and-- >> Satish: Absolutely. >> You know, going from the old way to a brand new thing, there's lot of different ways to get there. Some of it you might need to just-- Where are you with the journey of getting things onto this platform layer that we're talking about? And what will that journey look like for Port? >> Net new apps, anything being new we're talking about writing and like Cloud Native, Twelve factor Apps, like, but anything new, I'm like, anything existing data services, messaging services, what we affectionately call as table stakes services, right? So, which are the Twelve Factor Apps rely on, we are targeting towards Kubernetes. The idea is, "are we there yet?" Probably "no" like We are getting there with along with our partners to put it on the platforms like Kubernetes, right? So, we are also doing a lot of automation orchestration on VMs itself. But the idea is heavy and heavier workloads are going to be lining on Kubernetes platforms, and there will be a lot of work in the upcoming years particularly 2020, where we will be concentrating more on those things and with the continuing growth would be on Twelve Factor, Net New, would be Twelve Factor, Net New, could be in Cloud Foundry, could be in Kubernetes. Time will tell, but uh, that's the guiding philosophy, so to speak, but uh, There's a lot that we have to learn in this journey right now. >> Well I was kind of curious about that Satish, we've talked about an alphabet soup, we've talked about a lot of different projects, and certainly here at KubeCon, the thing about the Cloud Native Computing Foundation is that not that they don't have opinions, but everybody has an opinon, there's lots of different components here, it's not one stack, it's a collection of things that could be put together in several different ways. So you've tried a bunch of different things with storage, I'm actually, I'm interested if there are, if there were kind of surprises or, you know, containerized activity is probably different than I/O activity and storage I/O is probably different than in a virtual machine, the storage itself has some different assumptions built into it, so like, do you have any advise for people? I'm interested in the storage case but also just in, you don't have to evaluate nerworking and security and compliance and a lot of different things. Like, how do you go about approaching this sort of evaluation in this trial; in this journey of when you have-- when you're facing an "alphabet soup" of options? >> I think uh, it all comes down to basic engineering, right? So, what I make, think about "what are your failure points?" I'm like, "could be servers failing, infrastructure, hardware failing" right? So, the basic tendance is that we try to introduce failure as early as possible, like, "what happens if you pull the wire?" and "what happens if the server failure, failure happens?" The question that always comes back is that "is there a way I can compose the same infrastructure so that I can spread it across a couple of failure domains?" I think that was the whole idea of when we started, is like, "can we decompose the problem such that we can actually take advantage of primitives that begged into Kebernetes?" The great thing with CSI, that we're just realizing, before that were all flex drivers, but, how do you actually organize storage in the back end that actually allows you to actually compose this thing on the front end using the Kubernetes primitives. I think that was the process we though. >> John: And CSI is a standard API, >> Correct. >> Yeah, storage API, yeah. >> Exactly. I mean that's what we are relying, we're hoping that it's going to help us with things like, uh, moving compute, uh, to the storage rather than moving storage to the compute. That's one of the evolving, thinking that we're working with. Portworx, we've been working with the community folks from work and a couple of other areas. It's, there's lot to be done here, like we're just in still early days I would say. >> Murli, want to make sure we get out there, Portworx had some updates for this week so what do you say to latest? >> Yeah, so, the updates actually relate to exactly to what Satish was talking about, you know, the idea of, so, container storage has kind of been on it's own journey right? In the early days that John remembers well, it was really providing storage per system, making that data available everywhere. It's then clearly moved to HA being having the High Availability say within the cluster and so on. So, but the data life cycle for the application that's been containerized extends well beyond that so we are making extensions to our own product that are kind of following that path. So, one of the things we launched a few months ago was disaster recovery, DR, which is very very specific to containers, so, container granular DR, so you can kind of you know, take a snapshot, not just of the data, but of the application state as well as the Kubernetes pods back and recover all three of them. At this KubeCon we're announcing two other things. One of them is backup, so our customers, as they make the journey through their app life cycle, inevitably they need to backup their data and we have, again, container granular backup, that will provide all of, by the way, on existing storage. We're not asking anybody to up change, there's underneath their hardware storage substructure. The last thing we're introducing is storage capacity management which is fully automated. You know one of the characteristics of Kubernetes is all of that is "get the person" "get the trouble to get out of the picture," right? The world is going to be automated. Kubernetes is one of the ways people are doing that. And what we have provided is the ability to auto-resize volumes, and auto-resize pods of storage and add more nodes automatically based on policy that is completely automated so that again, these applications, you know when the characteristics of containerized workloads, they aren't predictable; they go up and down and they grow very fast sometimes, and so all of that management, so autopilot, uh, you know, backup DR have now been added in addition to persistent in HA. >> Alright, so before I let you both go, uh, want to talk about 2020? >> So soon. >> Satish, I want to give you a wish. You talked about all the things you'd do the next couple of years, if you could get one thing more out if this ecosystem to make your lives easier for you and your team, you know? What would that be? >> I think standardization on more of these interfaces. Kerbenetes provides a great platform for everybody to interact equally. Uh, more things like CSI, CRI, stuff that's happening in the community. And more standardization will lead to actually, make my life and things and end prizes a lot more easier. Will like to see continue that happening, GPU space looks very interesting, um, so we'll see. That would be my wish at least. >> Alright so Murli, I'm not giving you a wish. You're going to tell me, what should we be looking for from Portworx in participation in, you know, in this community over the next year. >> I think one of the big changes that's happened, really, in the last couple of years that is really kind of achieving a hockey stick is that enterprises are recognizing that stateful apps are really, should be using Kubernetes and can use Kubernetes. So to me, what I predict is that I think, Kubernetes is going to move from not, from just managing applications, to actually managing infrastructure like storage. And so I, you know, my belief is that 2020 is the beginning of where Kubernetes becomes the control plane across the Data Center and Cloud. It's the new control plane. No, what Openstack was aspiring to be many years ago, and that it will be looking upwards to manage applications and downwards to manage infrastructure and, it's not just us who are doing that, folks like VMware with Project Pacific have kind of kind of indicated that that's the direction that we see. So I think it's roll is now much more than just an app orchestrator, it's really going to be the new control plane for infrastructure and apps in the enterprise and in the Cloud. >> Murli, Satish, thank you so much for sharing all the update. >> Thank you >> Pleasure to catch up with both of you >> Thanks. >> Northbound, Southbound, Multi Cloud, theCube is at all of these environments and all the shows. For John Trayer, I'm Stu Miniman as always, thank you for watching theCube.

Published Date : Nov 19 2019

SUMMARY :

Brought to you by RedHat, This is the North America show here in San Diego All right, so Satish, we're going to start with you messaging services, lot of the stateful things, right? that we would have not done, so there's a lot of You obviously, agility is one of the things I hear most and how you fit into Ford and more broadly. and the data applications, with that being said, and all the things that are, Cloud Native is great but and data has gravity and the laws of physics do apply because of the fact that we have Some of it you might need to just-- that's the guiding philosophy, so to speak, but uh, and certainly here at KubeCon, the thing about the So, the basic tendance is that we try to introduce failure that it's going to help us with things like, uh, So, one of the things we launched a few months ago was the next couple of years, if you could get one thing more stuff that's happening in the community. from Portworx in participation in, you know, kind of indicated that that's the direction that we see. for sharing all the update. thank you for watching theCube.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
John TroyerPERSON

0.99+

Satish PuranamPERSON

0.99+

John TrayerPERSON

0.99+

Stu MinimanPERSON

0.99+

Murli ThirumaliPERSON

0.99+

FordORGANIZATION

0.99+

MurliPERSON

0.99+

50QUANTITY

0.99+

JohnPERSON

0.99+

2019DATE

0.99+

SatishPERSON

0.99+

PortworxORGANIZATION

0.99+

OneQUANTITY

0.99+

Cloud Native Computing FoundationORGANIZATION

0.99+

Ford Motor CompanyORGANIZATION

0.99+

2020DATE

0.99+

San DiegoLOCATION

0.99+

San Diego, CaliforniaLOCATION

0.99+

RedHatORGANIZATION

0.99+

bothQUANTITY

0.99+

yesterdayDATE

0.99+

129QUANTITY

0.99+

two and a half yearsQUANTITY

0.99+

Project PacificORGANIZATION

0.99+

KubernetesTITLE

0.99+

KubeConEVENT

0.98+

Murli ThirumalePERSON

0.98+

North AmericaLOCATION

0.98+

fourth yearQUANTITY

0.98+

CloudNativeConEVENT

0.98+

about 60 percentQUANTITY

0.98+

2kQUANTITY

0.98+

about 130 customersQUANTITY

0.97+

Cloud NativeTITLE

0.97+

oneQUANTITY

0.97+

PostgressORGANIZATION

0.97+

one stackQUANTITY

0.97+

this weekDATE

0.97+

next yearDATE

0.96+

one secondQUANTITY

0.96+

Twelve factor AppsTITLE

0.95+

Net NewORGANIZATION

0.95+

VMwareORGANIZATION

0.94+

PortworxTITLE

0.94+

this morningDATE

0.93+

CNCFORGANIZATION

0.93+

Kalyan Ramanathan, Sumo Logic | Sumo Logic Illuminate 2019


 

>> Narrator: From Burlingame, California, it's theCUBE. Covering Sumo Logic Illuminate 2019. Brought to you by Sumo Logic. >> Hey, welcome back, everybody, Jeff Frick here with theCUBE. We're at Sumo Logic Illuminate 2019. It's at the Hyatt Regency San Francisco Airport. We're excited to be back. It's our second year, so third year of the show, and really, one of the key tenants of this whole event is the report. It's the fourth year of the report. It's The Continuous Intelligence Report, and here to tell us all about it is the VP of Product Marketing, Kalyan Ramanathan. He's, like I said, VP, Product Management of Sumo Logic. Great to see you again. >> All right, thank you, Jeff. >> What a beautiful report. >> Absolutely, I love the cover and I love the data in the report even more. >> Yeah, but you cheat, you cheat. >> How come? >> 'Cause it's not a survey. You guys actually take real data. >> Ah, that's exactly right, exactly right. >> No, I love them, let's jump into it. No, it's a pretty interesting fact, though, and it came out in the keynote that this is not a survey. Tell us how you get the data. >> Yeah, I mean, so as you already know, Sumo Logic is a continuous intelligence platform. And what we do is to help our customers manage the operations and security of the mission critical application. And the way we do that is by collecting machine data from our customers, and many of our customers, we have two thousand, our customers, they're all running modern applications in the cloud, and when we collect this machine data, we can grade insights into how are these customers building their applications, how are these customers running and securing their application, and that insight is what is reflected in this report. And so, you're exactly right, this is not a survey. This is data from our customers that we bring into our system and then what we do is really treat things once we get this data into our system. First and foremost, we completely anonymize this data. So, we don't-- >> I was going to say Let's make sure we have to get that out. >> Yes, absolutely, so we don't have any customer references in this data. Two, we genericize this data. So, we're not looking for anomalies. We are looking for broad patterns, broad trends that we can apply across all of our customers and all of these enterprises that are running modern mission critical applications in the cloud. And then three, we analyze ten weeks to Sunday. We look at these datas, we look at what stands out in terms of good sample sizes, and that's what we reflect in this report. >> Okay, and just to close a loop on that, are there some applications that you don't include? 'Cause they're just legacy applications that're running on the cloud that doesn't give you good information, or you're basically taking them all in? >> Yeah, it's a good point, I mean we collect all data and we collect all applications, so we don't opt-in applications or out applications for that matter because we don't care about it. But what we do look for is significant sample size because we want to make sure that we're not talking about onesie-twosie applications here or there. We're looking for applications that have significant eruption in the cloud and that's what gets reflected in this report. >> Okay, well, let's jump into it. We don't have time to go through the whole thing here now, but people can get it online. They can download their own version and go through it at their leisure. Biggest change from last year as the fourth year of the report. >> Yeah, I mean, look, there are three big insights that we see in this report. The first one is, while we continue to see AWS rule in the cloud and that's not surprising at all, we're starting to see pretty dramatic adoption of multi-cloud technologies. So, two years ago, we saw a smidgen of multi-cloud in this report. Now, we have seen almost a 50% growth year over year in terms of multi-cloud adoption amongst enterprises who are in the cloud, and that's a substantial jump albeit from a smaller baseline. >> Do you have visibility if those are new applications or are those existing ones that are migrating to different platforms? Are they splitting? Do you have any kind of visibility into that? >> Yeah, I mean, it's an interesting point, and part of this is very related to the growth of Kubernetes that we also see in this report. What ypu've seen is that, in AWS itself, Kubernetes adoption has gone up significantly, what's even more interesting is that, as you think about multi-cloud adoption, we see a lot of Kubernetes, Kubernetes as the platform that is driving this multi-cloud adoption. There is a very interesting chart in this report on page nine. Obviously, I think you guys can see this if they want to download the report. If you're looking at AWS only, we see one in five customers are adopting Kubernetes. If you're looking at AWS and GCP, Google Cloud Platform, we see almost 60% of our customers are adopting Kubernetes. Now, when you put in AWS-- >> One in five at AWS, 60% we got Google, so that means four out of five at GCP are using Kubernetes and bring that average up. >> And then, if you look at AWS, Azure, and GCP, now you're talking about the creme de la creme customers who want to adopt all three clouds, it's almost 80% adoption of Kubernetes, so what it tells you is that Kubernetes has almost become this new Linux in the cloud world. If I want to deploy my application across multiple clouds, guess what, Kubernetes is that platform that enables me to deploy my application and then port it and re-target it to any other cloud or, for that matter, even an on-prem environment. >> Now, I mean, you don't see motivation behind action, but I'm just curious how much of it is now that I have Kubernetes. I can do multi-cloud or I've been wanting to do multi-cloud, and now that I have Kubernetes, I have an avenue. >> Yeah, it started another question. What's the chicken and what's the egg right here? My general sense, and we've debated this endlessly in our company, our general sense has been that the initiative to go multi-cloud typically comes top down in an organization. It's usually the CIO or the CSO who says, you know what, we need to go multi-cloud. And there are various reasons to go multi-cloud, some of which you heard in our keynote today. It could be for more reliability, it could be for more choice that you may want, it could be because you don't want to get logged into any one cloud render, so that decision usually comes top down. But then, now, the engineering teams, the ops teams have to support that decision, and what these engineering teams and these ops teams have realized is that, if they deploy Kubernetes, they have a very good option available now to port their applications very easily across these various cloud platforms. So, Kubernetes, in some sense, is supporting the top down decision to go multi-cloud which is something that is shown in spades as a result of this report. >> So, another thing that jumped out at me, or is there another top trend you want to make sure we cover before we get in some of those specifics? >> I mean we can talk to-- >> Yeah, one of them, one of them that jumped out at me was Docker. The Docker adoption. So, Docker was the hottest thing since sliced bread about four years ago, and is the shade of Kubernetes, not that they're replacements for one another specifically, but it definitely put a little bit of appall in the buzz that was the Docker, yet here, the Docker utilization, Docker use is growing year over year. 30%! >> I'll be the first one to tell you that Docker adoption has not stalled at all. This is shown in the report. It's shown in customers that we talk to. I mean, everyone is down the path of containerizing their application. The value of Docker is indisputable. That I get better agility, that I get better portability with Docker cannot be questioned. Now, what is indeed happening is that everyone who is deploying Docker today is choosing a orchestration technology and that orchestration technology happens to be Kubernetes. Again, Kubernetes is the king of the hill. If I'm deploying Docker, I'm deploying Kubernetes along with it. >> Okay, another one that jumped out at me, which shouldn't be a big surprise, but I'm a huge fan of Andy Jassy, we do all the AWS shows, and one of always the shining moments is he throws up the slide, he's got the Customer slide. >> There you go. >> It's the Services slide which is, in like, 2.6 font across a 100-foot screen that fills Las Vegas, and yet, your guys' findings is that it's really: the top ten applications are the vast majority of the AWS offerings that are being consumed. >> Yep, not just that. It's that the top services in AWS are the infrastructure-as-a-service services. These are the core services that you need if you have to build an application in AWS. You need ECDO, I need Esri, I need identity access management. Otherwise, I can't even log into AWS. So, this again goes back to that first point that I was making was that multi-cloud adoption is top of mind for many, many customers right now. It's something that many enterprises think of, and so, if I want to indeed be able to port my application from AWS to any other environment, guess what I should be doing? I shouldn't be adopting every AWS service out there because if I frankly adopted all these AWS services, the tentacles of the cloud render are just so that I will not be able to port away from my cloud render to any other cloud service out there. So, to a certain extent, many of the data points that we have in this report support the story that enterprises are becoming more conscious of the cloud platform choices that they are making. They want to at least keep an option of adopting the second or the third cloud out there, and they're consciously, therefore choosing the services that they are building their applications with. >> So, another hot topic, right? Computer 101 is databases. We're just up the road from Oracle. Oracle OpenWorld's next week. A lot of verbal jabs between Oracle and some of the cloud providers on the databases, et cetera. So, what do the database findings come back as? >> I mean, look at the top four databases: Redis, MySQL, Postgres, Mongo. You know what's common across them? They're all open-source. They're all open-source database, so if you're building your application, find standard components that you can then build your application on, whether it's a community that you can then take and move to any other cloud that you want to. That's takeaway number one. Takeaway number two, look at where Oracle is in this report. I think they're the eighth database in the cloud. I actually talked to a few customers of ours today. >> Now, are you sampling from Oracle's cloud? Is that a dataset? >> No, this is-- >> Yes, right, okay. So, I thought I want to make sure. >> And, if AWS is almost the universe of cloud today, we can debate at some bids, but it is close enough, I'd say, it tells you where Oracle is in this cloud universe, so our friends at Redwood City may talk about cloud day in and day out, but it's very clear that they're not making much of intent in the cloud at this point. >> And then, is this the first year the rollup of the type of database that NoSQL exceeded relational database? >> No, I mean, we've been doing this for the last two years, and it's very clear that NoSQL is ahead of SQL in the cloud, and I think the way we think about it is primarily because, when you are re-architecting your applications in the cloud, the cloud gives you a timeline, it gives you an opportunity to reconsider how you build out your data layer, and many of our customers are saying NoSQL is the way to go. The scalability demands, the reliability demands, so if my application was such that I now have the opportunity to rethink and redo my data layer, and frankly, NoSQL is winning the game. >> Right, it's winning big time. Another big one: serverless, Lambda. Actually, I'm kind of surprised it took so long to get to Lambda 'cause we've been going to smaller atomic units of compute, store, and networking for so, so long, but it sounds like, looks like we're starting to hit some critical mass here. >> Yeah, I mean, look, Lambda's ready for primetime. I mean we have seen that tipping point out here. Almost one in three customers of ours are using Lambda in production environments. And then, if you cast a wider net, go beyond production and even look at dev tests, what we see is that almost 60% of Sumo Logic's customers, and if you look at 2,000 customers, that's a pretty big sample size. Almost 60% of enterprises are using Lambda in some way, shape, or form. So, I think it's not surprising that Lambda is getting used quite well in the enterprise. The question really is: what are these people doing with Lambda? What's the intent behind the use of Lambda? And that's where I think we have to do some more research. My general sense, and I think it's shared widely within Sumo Logic, is that Lambda's still at the edges of the application. It's not at the core of the application. People are not building your mission critical application on Lambda yet because I think that that paradigm of thinking about event-driven application is still a little foreign to many organizations, so I think it'll take a few more years for an entire application to be built on Lambda. >> But you would think, if it's variable demand applications, whether that's a marketing promotion around the Super Bowl or running the books at the end of the month, I guess it's easy enough to just fire up the servers versus doing a pure Lambda at this point in time, but it seems like a natural fit. >> If you're doing the utility type application and you want to start it and you want to kill it and not use it after an event has come and gone, absolutely, Lambda's the way to go. The economics of Lambda. Lambda absolutely makes sense. Having said that, I mean, if you're to build a true mission critical application that you're going to be keeping on for a while to come, I'm not seeing a lot of that in Lambda yet, but it's definitely getting there. I mean we have lots of customers who are building some serious stuff on Lambda. >> Well, a lot of great information. It's nice to have the longitudinal aspect as you do this year over year, and again, we're glad you're cheating 'cause you're getting good data. >> (chuckles) >> (laughs) You're not asking people questions. >> Yeah, I mean, I'd like to finish out by saying this is a report that Sumo Logic builds every year, not because we want to sell Sumo Logic. It's because we want to give back to our community. We want our community to build great apps. We want them to understand how their peers are building some amazing mission critical apps in the cloud and so, please download this report, learn from how your peers are doing things, and that's our only intent and goal from this report. >> Great, well, thanks for sharing the information and a great catch-up, nice event. >> All right, thank you very much, Jeff. >> All right, he's Kalyan, I'm Jeff. You're watching theCUBE. We're at Sumo Logic Illuminate 2019. Thanks for watching, we'll see you next time. (upbeat electronic music)

Published Date : Sep 12 2019

SUMMARY :

Brought to you by Sumo Logic. and really, one of the key tenants and I love the data in the report even more. 'Cause it's not a survey. and it came out in the keynote that this is not a survey. And the way we do that is by collecting Let's make sure we have to get that out. that we can apply across all of our customers that have significant eruption in the cloud as the fourth year of the report. that we see in this report. the growth of Kubernetes that we also see in this report. so that means four out of five at GCP and re-target it to any other cloud and now that I have Kubernetes, I have an avenue. it could be for more choice that you may want, and is the shade of Kubernetes, and that orchestration technology happens to be Kubernetes. and one of always the shining moments of the AWS offerings that are being consumed. These are the core services that you need and some of the cloud providers on the databases, et cetera. and move to any other cloud that you want to. So, I thought I want to make sure. much of intent in the cloud at this point. and many of our customers are saying NoSQL is the way to go. to get to Lambda 'cause we've been going and if you look at 2,000 customers, or running the books at the end of the month, and you want to start it and again, we're glad you're cheating You're not asking people questions. are building some amazing mission critical apps in the cloud and a great catch-up, nice event. Thanks for watching, we'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jeff FrickPERSON

0.99+

JeffPERSON

0.99+

AWSORGANIZATION

0.99+

OracleORGANIZATION

0.99+

Las VegasLOCATION

0.99+

Sumo LogicORGANIZATION

0.99+

Kalyan RamanathanPERSON

0.99+

Andy JassyPERSON

0.99+

secondQUANTITY

0.99+

KalyanPERSON

0.99+

30%QUANTITY

0.99+

last yearDATE

0.99+

60%QUANTITY

0.99+

GoogleORGANIZATION

0.99+

oneQUANTITY

0.99+

Super BowlEVENT

0.99+

NoSQLTITLE

0.99+

100-footQUANTITY

0.99+

fourth yearQUANTITY

0.99+

OneQUANTITY

0.99+

third yearQUANTITY

0.99+

threeQUANTITY

0.99+

2,000 customersQUANTITY

0.99+

FirstQUANTITY

0.99+

ten weeksQUANTITY

0.99+

fourQUANTITY

0.99+

Burlingame, CaliforniaLOCATION

0.99+

LambdaTITLE

0.99+

TwoQUANTITY

0.99+

next weekDATE

0.99+

fiveQUANTITY

0.99+

five customersQUANTITY

0.99+

three customersQUANTITY

0.98+

ECDOTITLE

0.98+

second yearQUANTITY

0.98+

first oneQUANTITY

0.98+

two years agoDATE

0.98+

two thousandQUANTITY

0.98+

EsriTITLE

0.98+

SundayDATE

0.98+

todayDATE

0.98+

first pointQUANTITY

0.98+

KubernetesORGANIZATION

0.97+

KubernetesTITLE

0.97+

MySQLTITLE

0.97+

third cloudQUANTITY

0.96+

almost 80%QUANTITY

0.96+

LinuxTITLE

0.96+

SQLTITLE

0.95+

Redwood CityLOCATION

0.95+

Breaking Analysis: Spending Data Shows Cloud Disrupting the Analytic Database Market


 

from the silicon angle media office in Boston Massachusetts it's the queue now here's your host David on tape hi everybody welcome to this special cube in size powered by ET our enterprise Technology Research our partner who's got this database to solve the spending data and what we're gonna do is a braking analysis on the analytic database market we're seeing that cloud and cloud players are disrupting that marketplace and that marketplace really traditionally has been known as the enterprise data warehouse market so Alex if you wouldn't mind bringing up the first slide I want to talk about some of the trends in the traditional EDW market I almost don't like to use that term anymore because it's sort of a pejorative but let's look at it's a very large market it's about twenty billion dollars today growing it you know high single digits low double digits it's expected to be in the 30 to 35 billion dollar size by mid next decade now historically this is dominated by teradata who started this market really back in the 1980s with the first appliance the first converged appliance or coal with Exadata you know IBM I'll talk about IBM a little bit they bought a company called mateesah back in the day and they've basically this month just basically killed the t's and killed the brand Microsoft has entered the fray and so it's it's been a fairly large market but I say it's failed to really live up to the promises that we heard about in the late 90s early parts of the 2000 namely that you were going to be able to get a 360 degree view of your data and you're gonna have this flexible easy access to the data you know the reality is data warehouses were really expensive they were slow you had to go through a few experts to to get data it took a long time I'll tell you I've done a lot of research on this space and when you talked to the the data warehouse practitioners they would tell you we always had to chase the chips anytime Intel would come out with a new chip we forced it in there because we just didn't have the performance to really run the analytics as we need to it's took so long one practitioner described it as a snake swallowing a basketball so you've got all those data which is the sort of metaphor for the basketball just really practitioners had a hard time standing up infrastructure and what happened as a spate of new players came into the marketplace these these MPP players trying to disrupt the market you had Vertica who was eventually purchased by HP and then they sold them to Micro Focus greenplum was buy bought by EMC and really you know company is de-emphasized greenplum Netezza 1.7 billion dollar acquisition by IBM IBM just this month month killed the brand they're kind of you know refactoring everything par Excel was interesting was it was a company based on an open-source platform that Amazon AWS did a one-time license with and created a redshift it ever actually put a lot of innovation redshift this is really doing well well show you some data on that we've also at the time saw a major shift toward unstructured data and read much much greater emphasis on analytics it coincided with Hadoop which also disrupted the market economics I often joked it the ROI of a dupe was reduction on investment and so you saw all these data lakes being built and of course they turned into the data swamps and you had dozens of companies come into the database space which used to be rather boring but Mike Amazon with dynamodb s AP with HANA data stacks Redis Mongo you know snowflake is another one that I'm going to talk about in detail today so you're starting to see the blurring of lines between relational and non relational and what was was what once thought of is no sequel became not only sequel sequel became the killer app for Hadoop and so at any rate you saw this new class of data stores emerging and snowflake was one of the more interesting and and I want to share some of that data with you some of the spending intentions so over the last several weeks and months we've shared spending intentions from ETR enterprise technology research they're a company that that the manages of the spending data and has a panel of about 4,500 end-users they go out and do spending in tension surveys periodically so Alex if you bring up this survey data I want to show you this so this is spending intentions and and what it shows is that the public cloud vendors in snowflake who really is a database as a service offering so cloud like are really leading the pack here so the sector that I'm showing is the enterprise data warehouse and I've added in the the analytics business intelligence and Big Data section so what this chart shows is the vendor on the left-hand side and then this bar chart has colors the the red is we're leaving the platform the gray is our spending will be flat so this is from the July survey expect to expectations for the second half of 2019 so gray is flat the the dark green is increase and the lime green is we are a new customer coming on to the platform so if you take the the greens and subtract out the red and there's two Reds the dark red is leaving the lighter red is spending less so if you subtract the Reds from the greens you get what's called a net score so the higher the net score the better so you can see here the net score of snowflake is 81% so that very very high you can also see AWS in Microsoft a very high and Google so the cloud vendors of which I would consider a snowflake at cloud vendor like at the cloud model all kicking butt now look at Oracle look at the the incumbents Oracle IBM and Tara data Oracle and IBM are in the single digits for a net score and the Terra data is in a negative 10% so that's obviously not a good sign for those guys so you're seeing share gains from the cloud company snowflake AWS Microsoft and Google at the expense of certainly of teradata but likely IBM and Oracle Oracle's little for animal they got Exadata and they're putting a lot of investments in there maybe talk about that a little bit more now you see on the right hand side this black says shared accounts so the N in this survey this July survey that ETR did is a thousand sixty eight so of a thousand sixty eight customers each er is asking them okay what's your spending going to be on enterprise data warehouse and analytics big data platforms and you can see the number of accounts out of that thousand sixty eight that are being cited so snowflake only had 52 and I'll show you some other data from from past surveys AWS 319 Microsoft the big you know whale here trillion dollar valuation 851 going down the line you see Oracle a number you know very large number and in Tara data and IBM pretty large as well certainly enough to get statistically valid results so takeaway here is snowflake you know very very strong and the other cloud vendors the hyper scale is AWS Microsoft and Google and their data stores doing very well in the marketplace and challenging the incumbents now the next slide that I want to show you is a time series for selected suppliers that can only show five on this chart but it's the spending intentions again in that EDW and analytics bi big data segment and it shows the spending intentions from January 17 survey all the way through July 19 so you can see the the period the periods that ETR takes this the snapshots and again the latest July survey is over a thousand n the other ones are very very large too so you can see here at the very top snowflake is that yellow line and they just showed up in the January 19 a survey and so you're seeing now actually you go back one yeah January 19 survey and then you see them in July you see the net score is the July next net score that I'm showing that's 35 that's the number of accounts out of the corpus of data that snowflake had in the survey back in January and now it's up to 52 you can see they lead the packet just in terms of the spending intention in terms of mentions AWS and Microsoft also up there very strong you see big gap down to Oracle and Terra data I didn't show I BM didn't show Google Google actually would be quite high to just around where Microsoft is but you can see the pressure that the cloud is placing on the incumbents so what are the incumbents going to do about it well certainly you're gonna see you know in the case of Oracle spending a lot of money trying to maybe rethink the the architecture refactor the architecture Oracle open worlds coming up shortly I'm sure you're gonna see a lot of new announcements around Exadata they're putting a lot of wood behind the the exadata arrow so you know we'll keep in touch with that and stay tuned but you can see again the big takeaways here is that cloud guys are really disrupting the traditional edw marketplace alright let's talk a little bit about snowflakes so I'm gonna highlight those guys and maybe give a little bit of inside baseball here but what you need to know about snowflakes so I've put some some points here just some quick points on the slide Alex if you want to bring that up very fast-growing cloud and SAS based data warehousing player growing that couple hundred percent annually their annual recurring revenue very high these guys are getting ready to do an IPO talk about that a little bit they were founded in 2012 and it kind of came out of stealth and hiding in 2014 after bringing Bob Moog Leon from Microsoft as the CEO it was really the background on these guys is they're three engineers from Oracle will probably bored out of their mind like you know what we got this great idea why should we give it to Oracle let's go pop out and start a company and that NIN's and as such they started a snowflake they really are disrupting the incumbents they've raised over 900 million dollars in venture and they've got almost a four billion dollar valuation last May they brought on Frank salute Minh and this is really a pivot point I think for the company and they're getting ready to do an IPO so and so let's talk a little bit about that in a moment but before we do that I want to bring up just this really simple picture of Alex if you if you'd bring this this slide up this block diagram it's like a kindergarten so that you know people like you know I can even understand it but basically the innovation around the snowflake architecture was that they they separated their claim is that they separated the storage from the compute and they've got this other layer called cloud services so let me talk about that for a minute snowflake fundamentally rethought the architecture of the data warehouse to really try to take advantage of the cloud so traditionally enterprise data warehouses are static you've got infrastructure that kind of dictates what you can do with the data warehouse and you got to predict you know your peak needs and you bring in a bunch of storage and compute and you say okay here's the infrastructure and this is what I got it's static if your workload grows or some new compliance regulation comes out or some new data set has to be analyzed well this is what you got you you got your infrastructure and yeah you can add to it in chunks of compute and storage together or you can forklift out and put in new infrastructure or you can chase more chips as I said it's that snake swallowing a basketball was not pretty so very static situation and you have to over provision whereas the cloud is all about you know pay buy the drink and it's about elasticity and on demand resources you got cheap storage and cheap compute and you can just pay for it as you use it so the innovation from snowflake was to separate the compute from storage so that you could independently scale those and decoupling those in a way that allowed you to sort of tune the knobs oh I need more compute dial it up I need more storage dial it up or dial it down and pay for only what you need now another nuance here is traditionally the computing and data warehousing happens on one cluster so you got contention for the resources of that cluster what snowflake does is you can spin up a warehouse on the fly you can size it up you can size it down based on the needs of the workload so that workload is what dictates the infrastructure also in snowflakes architecture you can access the same data from many many different houses so you got again that three layers that I'm showing you the storage the compute and the cloud services so let me go through some examples so you can really better understand this so you've got storage data you got customer data you got you know order data you got log files you might have parts data you know what's an inventory kind of thing and you want to build warehouses based on that data you might have marketing a warehouse you might have a sales warehouse you might have a finance warehouse maybe there's a supply chain warehouse so again by separating the compute from that sort of virtualized compute from the from the storage layer you can access any data leave the data where it is and I'll talk about this in more and bring the compute to the data so this is what in part the cloud layer does they've got security and governance they got data warehouse management in that cloud layer and and resource optimization but the key in in my opinion is this metadata management I think that's part of snowflakes secret sauce is the ability to leave data where it is and have the smarts and the algorithms to really efficiently bring the compute to the data so that you're not moving data around if you think about how traditional data warehouses work you put all the data into a central location so you can you know operate on it well that data movement takes a long long time it's very very complicated so that's part of the secret sauce is knowing what data lives where and efficiently bringing that compute to the data this dramatically improves performance it's a game changer and it's much much less expensive now when I come back to Frank's Luqman this is somebody that I've is a career that I've followed I've known had him on the cube of a number of times I first met Frank Sloot when he was at data domain he took that company took it public and then sold it originally NetApp made a bid for the company EMC Joe Tucci in the defensive play said no we're not gonna let Ned afgan it there was a little auction he ended up selling the company for I think two and a half billion dollars sloop and came in he helped clean up the the data protection business of EMC and then left did a stint as a VC and then took over service now when snoop and took over ServiceNow and a lot of people know this the ServiceNow is the the shiny toy on Wall Street today service that was a mess when saluteth took it over it's about 100 120 million dollar company he and his team took it to 1.2 billion dramatically increased the the valuation and one of the ways they did that was by thinking about the Tam and expanding that Tim that's part of a CEOs job as Tam expansion Steuben is also a great operational guy and he brought in an amazing team to do that I'll talk a little bit about that team effect uh well he just brought in Mike Scarpelli he was the CFO was the CFO of ServiceNow brought him in to run finance for snowflake so you've seen that playbook emerge you know be interesting Beth white was the CMO at data domain she was the CMO at ServiceNow helped take that company she's an amazing resource she kind of you know and in retirement she's young but she's kind of in retirement doing some advisory roles wonder if slooping will bring her back I wonder if Dan Magee who was ServiceNow is operational you know guru wonder if he'll come out of retirement how about Dave Schneider who runs the sales team at at ServiceNow well he you know be be lord over we'll see the kinds of things that Sluman looks for just in my view of observing his playbook over the years he looks for great product he looks for a big market he looks for disruption and he looks for off-the-chart ROI so his sales teams can go in and really make a strong business case to disrupt the existing legacy players so I one of the things I said that snoopin looks for is a large market so let's look at this market and this is the thing that people missed around ServiceNow and to credit Pat myself and David for in the back you know we saw the Tam potential of ServiceNow is to be many many tens of billions you know Gartner when they when ServiceNow first came out said hey helpdesk it's a small market couple billion dollars we saw the potential to transform not only IT operations but go beyond helpdesk change management at cetera IT Service Management into lines of business and we wrote a piece on wiki Vaughn back then it's showing the potential Tam and we think something similar could happen here so the market today let's call 20 billion growing to 30 Billy big first of all but a lot of players in here what if so one of the things that we see snowflake potentially being able to do with its architecture and its vision is able to bring enterprise search you know to the marketplace 80% of the data that's out there today sits behind firewalls it's not searchable by Google what if you could unlock that data and access it in query at anytime anywhere put the power in the hands of the line of business users to do that maybe think Google search for enterprises but with provenance and security and governance and compliance and the ability to run analytics for a line of business users it's think of it as citizens data analytics we think that tam could be 70 plus billion dollars so just think about that in terms of how this company might this company snowflake might go to market you by the time they do their IPO you know it could be they could be you know three four five hundred billion dollar company so we'll see we'll keep an eye on that now because the markets so big this is not like the ITSM the the market that ServiceNow was going after they crushed BMC HP was there but really not paying attention to it IBM had a product it had all these products that were old legacy products they weren't designed for the cloud and so you know ServiceNow was able to really crush that market and caught everybody by surprise and just really blew it out there's a similar dynamic here in that these guys are disrupting the legacy players with a cloud like model but at the same time so the Amazon with redshift so is Microsoft with its analytics platform you know teradata is trying to figure it out they you know they've got an inertia of a large install base but it's a big on-prem install base I think they struggle a little bit but their their advantages they've got customers locked in or go with exudate is very interesting Oracle has burned the boats and in gone to cloud first in Oracle mark my words is is reacting everything for the cloud now you can say Oh Oracle they're old school they're old guard that's fine but one of the things about Oracle and Larry Ellison they spend money on R&D they're very very heavy investor in Rd and and I think that you know you can see the exadata as it's actually been a very successful product they will react attacked exadata believe you me to to bring compute to the data they understand you can't just move all this the InfiniBand is not gonna solve their problem in terms of moving data around their architecture so you know watch Oracle you've got other competitors like Google who shows up well in the ETR survey so they got bigquery and BigTable and you got a you know a lot of other players here you know guys like data stacks are in there and you've got you've got Amazon with dynamo DB you've got couch base you've got all kinds of database players that are sort of blurring the lines as I said between sequel no sequel but the real takeaway here from the ETR data is you've got cloud again is winning it's driving the discussion and the spending discussion with an IT watch this company snowflake they're gonna do an IPO I guarantee it hopefully they will see if they'll get in before the booth before the market turns down but we've seen this play by Frank Sluman before and his team and and and the spending data shows that this company is hot you see them all over Silicon Valley you're seeing them show up in the in the spending data so we'll keep an eye on this it's an exciting market database market used to be kind of boring now it's red-hot so there you have it folks thanks for listening is a Dave Volante cube insights we'll see you next time

Published Date : Sep 6 2019

SUMMARY :

David for in the back you know we saw

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

January 19DATE

0.99+

Dave SchneiderPERSON

0.99+

2014DATE

0.99+

2012DATE

0.99+

Frank SlumanPERSON

0.99+

Mike ScarpelliPERSON

0.99+

Dan MageePERSON

0.99+

Frank SlootPERSON

0.99+

January 17DATE

0.99+

EMCORGANIZATION

0.99+

July 19DATE

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

ServiceNowORGANIZATION

0.99+

81%QUANTITY

0.99+

AWSORGANIZATION

0.99+

1.2 billionQUANTITY

0.99+

JulyDATE

0.99+

30QUANTITY

0.99+

OracleORGANIZATION

0.99+

52QUANTITY

0.99+

HPORGANIZATION

0.99+

Larry EllisonPERSON

0.99+

360 degreeQUANTITY

0.99+

GoogleORGANIZATION

0.99+

Joe TucciPERSON

0.99+

20 billionQUANTITY

0.99+

fiveQUANTITY

0.99+

JanuaryDATE

0.99+

PatPERSON

0.99+

HadoopTITLE

0.99+

Silicon ValleyLOCATION

0.99+

ExcelTITLE

0.99+

10%QUANTITY

0.99+

70 plus billion dollarsQUANTITY

0.99+

AmazonORGANIZATION

0.99+

35QUANTITY

0.99+

first slideQUANTITY

0.99+

Dave VolantePERSON

0.99+

about 4,500 end-usersQUANTITY

0.99+

over 900 million dollarsQUANTITY

0.99+

Boston MassachusettsLOCATION

0.99+

two and a half billion dollarsQUANTITY

0.98+

TaraORGANIZATION

0.98+

first applianceQUANTITY

0.98+

MikePERSON

0.98+

dozens of companiesQUANTITY

0.98+

over a thousandQUANTITY

0.98+

TimPERSON

0.98+

Reza Shafii, Red Hat | Red Hat Summit 2019


 

>> Announcer: Live from Boston, Massachusetts, it's theCUBE. Covering Red Hat Summit 2019. Brought to you by Red Hat. >> Good to have you back here on theCube we are live in Boston at the Convention Center here. Along with Stu Miniman, I'm John Walls and on theCUBE we're continuing our coverage of Red Hat Summit 2019 in Boston, as I said. Joined now by Reza Shafii, who is the VP of Platform Services at Red Hat. Former CoreOS guy >> That's right. >> Stu actually has his CoreOS socks on, >> He told me. >> Today, yeah, so he came dressed for the occasion. >> Shh, can't see those on camera, John. I can't be wearing vendor here. >> Don't show it to the camera. >> Well I just say they're cool! They're cool. Glad to have you with us, Reza. And first off, your impression, you have a big announcement, right, with OpenShift. OpenShift 4 being launched officially on the keynote stage today. That's some big news, right? >> It's a big deal, it's a big deal. The way I think about it is that it's really a culmination of the efforts that we planned out when we sat down between the CoreOS leadership team and the Red Hat leadership team, when the acquisition was closed. And we planned this out, I remember a meeting we had in the white board room. We planned this out. In terms of bringing the best of OpenShift and CoreOS technology together. And it's really great to see it out there on the keynote, and actually all demoed and working. >> And working, right? Key part. >> Reza, dig in for us a little bit here, because it's one thing to say okay, we got a white board and we put things together. You know, when I looked at both companies, at first both, CoreOS before the acquisition and Red Hat, I mean open source, absolutely as its core. I remember talking to the CoreOS team, I'm like, you guys are gonna build a whole bunch of really cool tools, but what's the business there? Do you guys think you're gonna be the next Red Hat? Come on. Well, now you're part of Red Hat. So, give us a little bit of the insight as to what it took to get from there to the announcements, CoreOS infused in many of the pieces that we heard announced this week. >> Yeah, so the way I like to think about it is that Red Hat's OpenShift's roots, it started with making sure that they create a really nice comfortable surface area for the deaf teams. The deaf teams can go in and start pushing the applications and it just ensures that it's running those applications in the right way. The CoreOS roots came from the operations perspective and the system administrator. We always looked at the world from the system administrator. Yes, you're right, CoreOS had a number of technologies they were working on, etcd, Rocket, clair. I used to joke that there's a constellation of open source services that we're working on, but where is the one product? And, towards the end, right before the acquisition, the one product I think was pretty clear is Tectonic, the Kubernetes software. Now, if you look at Tectonic, the key value difference was automated operations. The core tenants of what Alex Polvi and Brandon Philips said into the mindset of the company was we're outnumbered, the number of machines out there is going to be way more than we can handle, therefore we need to automate all operations. They started that on the operating system itself, with CoreOS, the namesake of the company. And then they brought that to Kubernetes. What you see with OpenShift is, OpenShift 4, you see us bringing that to, not only the Kubernetes core, that's the foundation of OpenShift 4, so all capabilities of running Kubernetes are automated with 20 plus operators now. But you see that apply to all the other value capabilities that are on top of OpenShift as well, and we're bringing that to ISV. I was walking around and a number of ISV's have their operators as the number one thing they're advertising. So you're seeing automated operations really take hold and with OpenShift 4 being a foundation for that. >> You talk about operations or operators, you have Operator Hub that was launched earlier this year, what was the driving force behind that? And then ultimately what are you trying to get out of that in terms of advancement and going forward here? >> Right, I think it means it's worked. Going back a little bit of history on this, the operator pattern was coined at CoreOS as a way to do things on a Kubernetes cluster to automate operations. The right way. You have to expose it as a proper API, you have to use a controller, so on and so forth. Then as the team started doing that we realized well there's a lot of demand for this pattern, we started documenting it, describing it better and so on. But then we realized there's a good case for a framework to help people build these automations. Therefore we announced the operator framework at Cubeacon. I think it was a year and a half ago. What happened then was interesting, suddenly we started seeing hundreds plus operators being built on the operator framework. But, it was hard because you could see five Redis operators, 10 MySQL operators. It was hard for our customers to know where can I find the right set of operators that have the right functionality and how do they compare to each other? OperatorHub.IO is a registry that we launched together with AWS, Google and Microsoft to solve for that problem. Now that we have a way to create operators easily and capture that automated operations, we have sort of created a pattern and a framework around it, where do you go to find the right set of operators. >> It's an interesting point because if you look in the container space, especially Kubernetes, it's like, okay well what's standardized, what works across all of these environments? We always worry, I've probably got some pain from previous projects and foundations as to well what's certified and what's not and how do we do that? So, did I see there's a certification now for operators and how do you balance that we need it to work everywhere, we don't wanna have it's Red Hat's building an open ecosystem not something that's limited to only this? >> Yes. So OperatorHub.IO is a community initiative. And, every operator you find on there should work on any Kubernetes. So in fact as part of the vetting process we make sure that that's the case. And then on the certification we launched today, actually, and you can see a number of, we have already 20 plus operators that are certified. This is where we take it a step further and we work with the vendors to make sure that it works on OpenShift. It's following a number of guidelines that we have, in terms of using, for example, Rail as the basis. They work with us to run the updates through security checks and so on. And that's just to give our enterprise customers more levels of guarantees and validation, if they would like to. >> So what are they getting out of that, out of the certification system? What, I guess, stability and certainty and all those kinds of things that I'm looking for, standardization of some kind, is that what's driving that? >> It's simple, at the end of the day they got three things. They get automated updates that are pushed through the OpenShift update mechanism. So if you are using the Redis one, for example, and it's certified, you're gonna be able to update the Redis operator through the same cluster administration mechanism, then you would apply it to the entire cluster itself. You see updates from Redis come in, you can put it through the same approval work so on, so on. The second is they get support. So they get first line of support from Red Hat. They can call Red Hat, our customers and actually we work with them on that. And the third is that they actually get that security vulnerability scans that we put them through to make sure that they pass certain checks. And actually one last one, they also get Rail as the basis of the operator, so, yup. >> Reza, help bring us into the customer point of view. What does all this mean to them, what are the big challenges, how do they modernize their applications and get more applications moving along this path? >> Yeah, in this case the operator customer is mainly the infrastructure administrators. It's important to point that out. The developers will get some benefit on that in that it's self service, so the provision, but there's other ways to do that as well. You can go to a Helm chart, deploy that Helm chart, you get that level of self service automated provisioning. To go ahead and configure for example, a charted MongoDB database on a Kubernetes cluster, you have to create something like 20 different objects. And then to update that to change the charts, you have to go and modify all those 20 different objects. Let's just stay at that level alone. An operator makes that before different parameters on a yaml file that you change. The operator takes that and applies all these configurations for you. So, it's all about simplifying the life of the infrastructure administrators. I truly believe that operators, human operators, infrastructure administrators are one of the least appreciated personas right now that we have out there. They're not the most important ones, but there is a lot of pain points and challenges that they have we're not really thinking about too much. And I think OpenShift goes a long way and operators go a long way to actually start thinking about their pain point as well. >> So what do you think their reaction was this morning when they're looking, first off, the general announcement, right? And then some of the demonstrations and all those things that are occurring? Is there, do you have or are you talking to customers? Are you getting the sense of relief or of anticipation or expectation? I mean, how would you characterize that? >> Think they're falling into a couple of different buckets. There's the customers we've talked to, for awhile now, that know this stuff, so this is not super new to them, but they're very happy to see it. There's one big automaker that's a customer of us and the main human operator was telling me awhile ago that he does not want any service on the cluster unless it has an operator, this is a year and a half ago. And he kept pushing me well I want a Kafka one and I want an Elasticsearch one, and you know. And we, CoreOS, were too small to try to build that ourselves. Obviously that's not, we can't maintain a Kafka operator and a CoreOS one. Now, he's able to go to our operator APP, he's gonna be able to get a Kafka operator that's maintained by Kafka experts. He's gonna be able to get a Redis operator that's maintained by Redis experts. So that bucket of customers are super happy. And then there's another one that's just starting to understand the power of all this. And I think they're just starting to kick the tires and play around with this. Hopefully they will get to the same point as the first bucket of customers, and be asking for everything to be operator based all the time. >> Convert the tire kickers, you're gonna be okay, right? >> That's right. >> Thank you for the time. >> Thank you. >> We appreciate that and continued success at Red Hat, and, once again, good to see you. >> Thank you, always a pleasure. >> You bet. Live, here on theCUBE, you're watching Red Hat Summit 2019. (upbeat music)

Published Date : May 8 2019

SUMMARY :

Brought to you by Red Hat. Good to have you back here on theCube I can't be wearing vendor here. Glad to have you with us, Reza. of the efforts that we planned out when we sat down And working, right? many of the pieces that we heard announced this week. is going to be way more than we can handle, Then as the team started doing that we realized and you can see a number of, we have already 20 plus It's simple, at the end of the day they got three things. What does all this mean to them, And then to update that to change the charts, and the main human operator was telling me awhile ago and, once again, good to see you. Live, here on theCUBE, you're watching Red Hat Summit 2019.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
BostonLOCATION

0.99+

Stu MinimanPERSON

0.99+

Red HatORGANIZATION

0.99+

Reza ShafiiPERSON

0.99+

MicrosoftORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Alex PolviPERSON

0.99+

John WallsPERSON

0.99+

CubeaconORGANIZATION

0.99+

20 plus operatorsQUANTITY

0.99+

TectonicORGANIZATION

0.99+

StuPERSON

0.99+

OpenShift 4TITLE

0.99+

JohnPERSON

0.99+

thirdQUANTITY

0.99+

20 different objectsQUANTITY

0.99+

fiveQUANTITY

0.99+

10QUANTITY

0.99+

both companiesQUANTITY

0.99+

RedisTITLE

0.99+

Boston, MassachusettsLOCATION

0.99+

CoreOSORGANIZATION

0.99+

Red Hat Summit 2019EVENT

0.99+

OpenShiftTITLE

0.99+

TodayDATE

0.99+

a year and a half agoDATE

0.99+

Brandon PhilipsPERSON

0.99+

todayDATE

0.99+

secondQUANTITY

0.98+

one productQUANTITY

0.98+

first bucketQUANTITY

0.98+

Convention CenterLOCATION

0.98+

three thingsQUANTITY

0.98+

CoreOSTITLE

0.98+

20 plus operatorsQUANTITY

0.98+

KubernetesTITLE

0.97+

RedisORGANIZATION

0.97+

hundreds plus operatorsQUANTITY

0.97+

this weekDATE

0.96+

earlier this yearDATE

0.96+

first lineQUANTITY

0.96+

oneQUANTITY

0.96+

KafkaTITLE

0.95+

OpenShiftORGANIZATION

0.94+

MongoDBTITLE

0.93+

one thingQUANTITY

0.92+

firstQUANTITY

0.91+

RezaPERSON

0.9+

Operator HubORGANIZATION

0.88+

bothQUANTITY

0.87+

ISVTITLE

0.86+

MySQLTITLE

0.85+

CoreOSCOMMERCIAL_ITEM

0.85+

OperatorHub.IOORGANIZATION

0.83+

this morningDATE

0.83+

KubernetesORGANIZATION

0.76+

Nima Badiey, Pivotal | Dell Boomi World 2018


 

(upbeat techno music) >> Live from Las Vegas, it's theCUBE. Covering Boomi World 2018. Brought to you by Dell Boomi. >> Good afternoon, welcome back to theCUBE's continuing coverage of Boomi World 2018 from Las Vegas. I'm Lisa Martin with John Furrier and we're welcoming back to theCUBE one of our alumni Nima Badiey, Head of Technology Ecosystems from Pivotal. Nima, welcome back. >> Thank you for having me back. >> So Pivotal, part of the Dell technologies part of the companies, >> Yeah. >> You guys IPOd recently. And I did read that of the first half 2018, eight of the 10 tech IPOs were powered by Boomi. >> Well, I don't know about that specific. I know that tech IPOs are making a big comeback. We did IPO on the 20th of April, so we've passed out six-month anniversary if you can say. But it's been a distinct privilege to be part of the overall Dell family of businesses. I think what you have in Michael as a leader, who, he has a specific vision, but he's left the independent operating units to work on their own, to find their path through that journey, and to help each other as brethren, as like sisters and brothers. And the fact that Pivotal is here supporting Boomi. That Boomi is within our conference of supporting our customers that we're working together really speaks volumes. I think if you take a look at it, a lot of things happened this week, right? So a couple weeks ago, IBM's acquiring RedHat, this morning VMWare's acquiring Heptio. That's a solid signal that the enterprise transformation and adoption of cloud native model is really taking off. So the new middleware is really all about the cloud native polyglock, multiglock environment. >> And what's interesting, I want to get your thoughts on this because first of all congratulations on the IP, some are saying Pivotal's never going to go public, and they did, you guys were spectacular, great success. But what's going on now is interesting. We're hearing here at this show, as other shows is, cloud scale and data are really at the center of this horizontally scalable cloud poly proposition. Okay great, you mention Kubernetes and Heptio and VM where, that's all great. The question that is how do you compete when ecosystems become the most important thing. You worked at VMware you're at Pivotal. Dell knows ecosystems. Boomi's got an ecosystem. Partners, which is also suppliers and integrators. >> Yeah. >> They integrate and also developers. This is a key competitive advantage. What's your take on that here? >> So I think you touched on the right point. You compete because of your ecosystem, not despite your ecosystem. We can't be completely hedgemonic like Microsoft or Cisco or Amazon can afford to be. And I don't think customers really want that. Customers actually want choice. They want the best options but from a variety of sources. And that's why one of the reasons that we not only invest Dell ecosystem but also in Pivotal's own ecosystem is to cultivate the right technologies that will help our customers on that journey. And our philosophy's always find the leaders in the quadrant. The Cadillac vendors, the Lexus vendors onboard them and the most important thing you can do is, to ensure a pristine customer experience. We're not measuring whether feature A from one partner is better than feature B from another partner. We really don't care. What we care about is we can hand wire and automate what would have been a very manual process for customers, so that, let's say Boomi with Cloud Foundry works perfectly out of the box. So the customers doesn't have to go through and hire consultants and additional external resources just to figure out how two pieces of software should work together, they just should. So when they make that buying decision they know that the day after that buying decision, everything's going to be installed and their developers and their app dev teams and their ops teams can be productive. So that's the power of the ecosystem. >> Can you talk about the relationship between Pivotal and Boomi, because Boomi's been born in the Cloud as start up. Acquired eight years ago. You're part of the Dell Technologies family. VMware's VMware, we know about VMware doing great. You guys doing great. Now Boomi's out there. So how do they factor into and what's the relationship you have with them and how does that work, how do you guys work together? >> Perfect question. So, in my primary role at Pivotal is to manage all of our partner ecosystems, specifically the technology partners. And what I look for are any force multipliers. Any essentially ISVs who can help us accomplish more together than we could on our own. Boomi's a classic example of that. What do they enable? So take your classic customer. Classic customer has, let's say, 100 applications in inventory that they have built, managed, and purchased procured off from shelf-to-shelf components. And roughly 20 or 30% are newish, green field applications, perfect for the cloud native transformation. Most 80% of them or 70% are going to be older, ground field applications that will have to be refactored. But there's always going to be that 15% towards the end that's legacy mainframe. It can't be changed, you cannot afford to modernize it, to restructure it, to refactor it. You're going to have to leave it alone, but you need it. Your inventory systems are there. >> These are critical systems, those people who think legacy as outdated, but they're actually just valued. >> No, they're critically valuable. >> Yes. >> We just cannot be modernized. >> Bingo. >> So a partner like Boomi will allow you to access the full breadth of those resources without having to change them. So I could potentially put Boomi in front of any number of older business applications and effectively modernize them by bridging those older legacy systems with the new systems that I want to build. So let's do an example. I am the Gap and I want to build a new version of our in-store procurement system that runs on my iPhone, that I can just point to a garment and it will automatically put it in my, ya know, check out box. How do I do that? Well I can build all the intelligence. And I can use AI and functions and I can build everything it's out of containers, that's great. But I still have to connect to the inventory system. Inventory system... >> Which is a database. All these systems are out there. >> Somewhere, something. And my developers don't know enough about the old legacy database to be able to use it. But if I put a restful interface using Boomi in front of it and a business connector that's not older XML or kind of inflexible, whatever, solo gateways. Then I have enabled my developer to actually build something that is real. That is customer focused. It is appropriate for that market without being hamstrung by my existing legacy infrastructure. And now my legacy infrastructure is not an anchor that's holding me back. >> You had mentioned force, me and Lisa talk about this all the time on theCUBE, where that scenario's totally legit and relevant because in the old version of IT you have to essentially build inventory management into the new app. You'd have to essentially kill the old to bring in the new. I think with containers and cloud native has shown is you can keep the old and sunset it if you want on your own time table or keep it there and make it productive. Make the data exposeble, but you can bring the cool relevant new stuff in. >> Yeah. >> I think that is what I see and we see from customers, like OK cool, I don't have to kill the old. I'll take care of it on my own timetable versus a complete switching cost analysis. Take down a production system. >> Exactly. >> Build something new, will it work. Ya know cross your fingers. Okay, again and this is a key IT different dynamic. >> It is and it's a realization that there are things you can move and those are immutable. They're simply just monolithic that will never move. And you're going to work within those confines. You can have the best of both worlds. You can maintain your legacy applications. They're still fine, they run most of your business. And still invent the new and explore new markets and new industries and new verticals. And just new capabilities all through and through without having to touch in your back end systems. Without having to bring the older vendors in and say can you please modernize your stuff because my business is dependent and I am going to lose that. I'm going to become the new Sears, I going to become the new Woolworth or whoever. Blockbuster that has missed an opportunity to vector into a new way of delivering their services. >> When you're having customer conversations, Nima, I'm curious, talking with enterprise organizations who have tons of data, all the systems including the legacy, which I'm glad that you brought up that that's not just old systems. There's a lot of business critical, mission critical application running on 'em. Where do you start that conversation with the large enterprise, who doesn't want to become a Blockbuster to your point, and going this is the suite of applications we have, where do we start? Talk to us about that customer journey that you help enable. >> That's great 'cause in most cases the customers already know exactly what they want. It's not the what that you have to have the conversation around, it's the how do I get there. I know what I want, I know what I want to be, I know what I want to design. And it's how do I transform my business fundamentally do an app transformation, enterprise transformation, digital transformation? Where do I begin? And so, ya know, our perspective at Pivotal is, ya know, we're diehard adopters of agile methodology. We truly, truly believe that you can be an agile development organization. We truly believe in Marc Andreessen's vision of software eating the world. Which let's unpack what that means. It just means that if you're going to survive the next 10 years you have to fundamentally become a software company, right? So look at all the companies we work with. Are you an insurance company or are you delivering an insurance product through software? Are you a bank or are you delivering banking product through software? Well, when was the last time you talked to a bank teller? Or the atm, most of your banking's done online. Your computer or your mobile device. Even my check cashing, I don't have to talk to anyone. It's wonderful. Ford Motor Company, do they bend sheet metal and put wheels on it or are they a software company? Well consider that your modern pickup truck has... >> They're an IOT company now. (laughing) (crosstalking) Manufacturing lines. >> That's what's crazy. You have a 150 million lines of code in your pickup truck. Your car, your pickup truck, your whatever is more software than it is anything else. >> But also data's key. I want to get your thoughts since this is super important Michael Dell brought up on the keynote today here at Boomi World was, okay the data's got to stay in the car. I don't need to have a latency issue of hey, I need to know nanosecond results. With data, cloud has become a great use case. With multicloud on the horizon, some people are going to throw data in multiple clouds and that's clear use case, and everyone can see the benefits of that. How do you guys look at this? 'Cause now data needs to be addressable across horizontal systems. You mentioned the Gap and the Gap example. >> That's great, so, one of the biggest trends we see in data is really event streaming. Is the idea that the ability to generate data far out exceeds the ability to consume it. So, what if we treated data as just a river? And I'm going to cast my line and only pick up what I want out of that stream. And this is where CAFCA and companies like Solice and any venturing networks and spring cloud functions and spring cloud data are really coming into play, is acknowledgement that yes we are not in a world where we can store all of the data all the time and figure out what to do with it after the fact. We need timely, and timely is within milliseconds, if not seconds. Action taken on an event or data even coming through. So why don't we modernize around, ya know, that type of data structure and data event and data horizon. So that's one of the trends we see. The second is that there is no one database to rule them all anymore. I can't get away with having oracle and that's my be all, end all. I now have my ESQL and SQL and Mongo and Cassandra and Redis and any other number of databases that are form, fit and function specific for a utility and they're perfect for that. I see graph databases, I see key value stores, I see distributed data warehouse. And so my options as a developer, as a user is really expanding, which means the total types of data components that I can use are also expanding exponentially. And that gives me a lot more flexibility on the types of products that I can build and the services that I can ultimately deliver. >> And that highlights micro services trend, because you have now a multitude of databases, it's not the one database rules them all. They'll be literally thousands of database on censors, so micro service has become the key element to connect all these systems. >> All of it together. And micro services really a higher level of abstraction. So we started with virtual machines and then we went to containers and then we went to functions and micro services. It's on an upward trend necessarily as it is an expansion. Into different ways of being able to do work. So some of my work products are going to be very, very small. They can afford to be ephemeral, but there may be many of them. How do I manage a cluster of millions of these potential work loads? Backing off I can have an ephemeral applications that run inside of containers or I can have ridged fixed applications that have to run inside a virtual machines. I'm going to have all of them. What I need is a platform that delivers all of this for me without me having to figure out how to hand wire these bits and pieces from various different either proprietary or open source kits just to make it work. I'm going to need a 60 to 100 or 200 person team just to maintain this very bespoke thing that I have developed. I'll just pull it off the shelf 'cause this is a solved problem. Right, Pivotal has already solved this problem. Other companies have already solved this problem. Let me start there and so now I'm here. I don't have to worry about all this left over plumbing. Now I can actually build on top of my business. The analogy I'd use is you don't bring furniture with you every time you check into a hotel. And we're telling customers every time you want to move to a different city just for business meeting or for work trip we're going to build you a house and you need to furnish it. Well, that's ridiculous. I'm going to check into a hotel and my expectation is I can check out of any other room and they'll all be the same, it doesn't really matter what floor I'm on, what room I'm in. But they'll have the same facilities, the same bed, the same, ya know, restroom facilities. That's what I want. That's what containers are. Eventually all the services surrounding that hotel room experience will be micro services. >> And we're the work load, the people. >> And we are the work load and we're the most important thing, we are the application, you're right. >> I love that. That's probably best analogy I've heard of containers. Nima, thanks so much for stopping by theCUBE, joining John and me today. And talking to us about what's going on with Pivotal and how you guys are really helping as part of Dell business dramatically transform. >> Been my pleasure. Thank you both. >> Thank you. >> Thank you. Thank you for watching theCUBE. I'm Lisa Martin with John Furrier. We are in Las Vegas at Boomi World '18. Stick around, John and I will be right back with our next guest. (light techno music)

Published Date : Nov 7 2018

SUMMARY :

Brought to you by Dell Boomi. back to theCUBE one of our alumni Nima Badiey, And I did read that of the first half 2018, That's a solid signal that the enterprise transformation The question that is how do you compete when ecosystems and also developers. and the most important thing you can do is, to ensure in the Cloud as start up. You're going to have to leave it alone, but you need it. those people who think legacy We just cannot that I can just point to a garment and it will automatically Which is a database. And my developers don't know enough about the old legacy because in the old version of IT you have to essentially like OK cool, I don't have to kill the old. Okay, again and this is a key IT different dynamic. It is and it's a realization that there are things you the legacy, which I'm glad that you brought up It's not the what that you have to have They're an IOT company now. You have a 150 million lines of code in your pickup truck. With multicloud on the horizon, some people are going to Is the idea that the ability to generate data far out so micro service has become the key element to connect applications that have to run inside a virtual machines. And we are the work load and we're the most important And talking to us about what's going on with Pivotal Thank you both. Thank you for watching theCUBE.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JohnPERSON

0.99+

Nima BadieyPERSON

0.99+

MicrosoftORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

Lisa MartinPERSON

0.99+

Michael DellPERSON

0.99+

CiscoORGANIZATION

0.99+

IBMORGANIZATION

0.99+

Marc AndreessenPERSON

0.99+

NimaPERSON

0.99+

BoomiPERSON

0.99+

CAFCAORGANIZATION

0.99+

Las VegasLOCATION

0.99+

SoliceORGANIZATION

0.99+

Ford Motor CompanyORGANIZATION

0.99+

LexusORGANIZATION

0.99+

six-monthQUANTITY

0.99+

MichaelPERSON

0.99+

two piecesQUANTITY

0.99+

DellORGANIZATION

0.99+

100 applicationsQUANTITY

0.99+

60QUANTITY

0.99+

15%QUANTITY

0.99+

20th of AprilDATE

0.99+

CadillacORGANIZATION

0.99+

iPhoneCOMMERCIAL_ITEM

0.99+

thousandsQUANTITY

0.99+

70%QUANTITY

0.99+

LisaPERSON

0.99+

30%QUANTITY

0.99+

PivotalORGANIZATION

0.99+

eight years agoDATE

0.99+

BoomiORGANIZATION

0.99+

one partnerQUANTITY

0.99+

SearsORGANIZATION

0.99+

150 million linesQUANTITY

0.99+

100QUANTITY

0.99+

feature BOTHER

0.99+

eightQUANTITY

0.99+

bothQUANTITY

0.99+

secondQUANTITY

0.99+

WoolworthORGANIZATION

0.99+

VMwareORGANIZATION

0.99+

feature AOTHER

0.99+

theCUBEORGANIZATION

0.98+

VMWareORGANIZATION

0.98+

John FurrierPERSON

0.98+

first half 2018DATE

0.98+

10 tech IPOsQUANTITY

0.98+

this weekDATE

0.98+

todayDATE

0.98+

Dell TechnologiesORGANIZATION

0.98+

oneQUANTITY

0.98+

SQLTITLE

0.97+

200 personQUANTITY

0.97+

ESQLTITLE

0.97+

80%QUANTITY

0.96+

Boomi World 2018EVENT

0.96+

both worldsQUANTITY

0.96+

millionsQUANTITY

0.95+

VMORGANIZATION

0.94+

Boomi World '18EVENT

0.92+

KubernetesORGANIZATION

0.92+

20QUANTITY

0.91+

HeptioORGANIZATION

0.88+

firstQUANTITY

0.87+

Boomi WorldORGANIZATION

0.86+

RedHatORGANIZATION

0.84+

couple weeks agoDATE

0.83+

Kalyan Ramanathan, Sumo Logic | Sumo Logic Illuminate 2018


 

>> From San Francisco, It's theCUBE. Covering Sumo Logic Illuminate 2018. Now here's Jeff Frick. >> Hey welcome back everybody. Jeff Frick here with theCUBE. We are at Sumo Logic Illuminate 2018: about 600 people. I think its three times as big as it was last year here at the Hyatt San Francisco Airport in Burlingame, and on of the big topics of today is the release of this new report. It's called The State of Modern Applications in DevOps Security, and to talk all about it and the results and kind of the process we are excited to have Kalyan Ramanathan, excuse me, VP of Product Marketing at Sumo Logic. Welcome. >> Thank you, Jeff. >> So you've been doing this report for a while, correct? >> Yeah, exactly, I think this is the third version of this report, and from what we know the first and only report that looks at how, you know, leading edge customers actually build, run, manage, and secure their applications in public cloud environments. >> Right, so just a little history for people that aren't familiar: Sumo launched in the cloud natively, right, and I think you guys launched on AWS. >> Absolutely. >> Way back when, I think, one of our very first AWS shows we went to in 2013, Summit San Francisco, I remember it well, we had you guys on, and so you guys have really grown along with AWS, but obviously you have expanded well beyond just simply inside of AWS. >> That's right. So, the company was founded in 2010, and we were one of the first big data services to run on AWS. I think our founders, you know, ran into one of the AWS architects who describe this new thing called a cloud, and they were completely smitten by it. They thought that this was the next new wave of how services are going to be delivered, so it just made a lot of sense to build this machine data analytics platform, that we were building, or that we were planning to build on AWS. >> Right. >> The scalability, the agility, the, ya know, the flexibility that AWS offered was exactly what our platform needed, and so this was a marriage made in heaven. But we can support applications that run just about anywhere. We obviously support applications running on AWS extremely well; that's our DNA. We get those applications, because we build and run those applications ourselves, but we also support Azure. We support GCP. We support hybrid environments. Many of our customers, ya know, are either, ya know, built in the cloud, and they know only cloud, but a few of them are also making the transition to the cloud, are migrating their applications to the cloud, and you know, we believe that we live in an age where flexibility is extremely important, and we support our customers where ever their applications are today. >> Right, so let's look at some of the findings, so. >> Absolutely. >> Just from a process point of view, you interviewed your customer data base, right? Your, your numbers here? >> Yeah, yeah, I think, yeah. We looked at our 16 hundred customers. >> Sixteen hundred customers, okay. >> And an important point to make out here is that we don't interview our customers. What we do is to, essentially, collect data from our customers, which is what we do when we are doing machine data analytics, we anonymize those data, and we represent as to what is happening in terms of these applications. How do our customers build these applications, you know, manage them, and secure them? >> Right. >> So this is not a >> It's the real data, though. >> It's real data. >> This is not, this is not what they think they know, and they're going to answer the survey. >> Absolutely. Exactly. Right. >> And all the survey biases that can come up. >> You are very right, I mean, you know, that's what makes this report unique, right? >> Right, right. >> It's the first report where we're actually reflecting what customers actually do. It's not a survey. It's not an aggregation of, you know, data from ten other sources. This is as close to truth as you can get in terms of running and building and securing applications. >> Right. >> In the cloud environment. >> So, I was happy to see that the data supports a number of the hypotheses that we derive at a lot of the shows. >> Absolutely. >> That we go through. You know, right of the top: Docker and the adoption of Kubernetes in orchestration is growing rapidly. >> Absolutely, I mean, ya know, everywhere you go you hear containers: container this, container that, so, ya know, we see a similar adoption. Docker has grown from 14 percent to about 28 percent in this, as we see in this report, but what's interesting is also the growth of Kubernetes and orchestration, right? If you were to ask anyone, even in this conference, you know, about orchestration, let's say two or three years ago, and even the word Kubernetes, ya know, I'm sure you'd have gotten blank stares. >> Right, right. >> Here we are, two years into Kubernetes becoming, you know, somewhat mainstream, and we are already starting to see 30 per cent adoption of orchestration within AWS, and out of that 30 per cent, we almost see fifteen per cent of those folks using Kubernetes as a native technology. AWS has just announced their own Kubernetes service. I am sure if, when we have this conversation next year, >> Right. >> Kubernetes, you know, will become a household name. You will see 30 per cent adoption of Kubernetes alone, >> Right. >> Ya know, in a report of this kind. >> Well it's funny: when we were at VMworld a couple weeks ago, and Kubernetes was both in Pat's, Pat Gelsinger's keynote. >> Uh-huh. >> As well as Sonjay's, you know, so it's just, it really shows how fast in this type of a world a new technology adoption can just be put into place. >> Yeah, I mean, if you bring the right capabilities, if you have the right support, which is what Kubernetes does, and, obviously, if you have the right backing, you know, in the form of Google, obviously, incubating this project and then, you know, promoting it as an open source standard, and everybody is now falling behind it. Ya know, we support it. We hear it from our customers, and, you know, the data also bears this out. >> Right, so what about on the database side? What did you find on the database side? >> Yeah, I mean, the database results are always interesting for us, right? You know, I think the most important thing that we learned is that, you know, as customers are building apps in their public cloud environments, they really have a choice, ya know? If you were to build an on-prem, you know, application, once upon a time, I mean, you are usually stuck with Oracle or, ya know, MySQL or SQL Silver or some of those standard database fares everyone has heard about. >> Right. >> But when you, now, go to the cloud, when you migrate to the cloud, or when you are, you know, incubating your applications in the cloud to begin with, you want to re-think your database layer. This is the core layer that powers your application, and there are lots more, ya know, opportunities to, and options out there. >> Right. >> So, what we are seeing is, one, the growth of no-SQL databases: they are way more scalable, ya know, they handle big data way better that, ya know, traditional SQL databases, so we're definitely seeing a growth of that, of no-SQL databases. >> Right. >> What's also interesting is that, ya know, is customers have the choice. They are looking at other forms of databases. Ya know, I could look at Redis, I could look at MongoDB, I could look at Posgres, and, right, I'm not stuck going back to, ya know, our favorite Oracle or SQL Silver anymore. >> Right. What strikes me is that the definition of the requirement has been flipped upside-down. Before it was, "What infrastructure do we have? What's available that IT can deliver to me? What do we have licenses for, and what can I build on top of?" Now the application has taken center stage, so now it's "This is what I want to do with my application. What is it that I need underneath the covers to deliver that capability?" So it really flipped the whole thing on its head. >> Ya know, this also goes back to that, you know, sort of the democratization of decisions where, you know, developers, now, can make these choices. You know, once upon a time, right, I mean, someone, a muckity-muck in your organization says Oracle is the way to go, and everybody follows suite, follows suit. That's not the case any more, right? >> Right. >> I mean, the engineer, they're a developer who is building their application, especially in the microservices world, they can make choices in terms of what is a data server that they may choose to build into that microservcie? And that doesn't have to be Oracle every time. It doesn't have to be SQL Silver every time. You know, if Redis makes sense, if MongoDB makes sense, let's go build that into our into our platform. >> Right, so, another one, you know, serverless is all the hot buzz, and clearly that is supported here with some of the data around Lambda adoption. >> Yeah, I mean, Lambda growth, you know, continues to astound us. We are seeing Lambda grow from twelve per cent two years ago, which is when we did our first report, to now, you know, almost 30 per cent, you know? So, imagine that, right: one in three enterprises today are using Lambda, and this is a technology that is very easy to use, but architecture-wise, you need to re-think how you are putting your applications together with Lambda, and we are starting to see, you know, some wide-spread Lambda adoption, you know, within enterprises. >> Right, but isn't that the ultimate goal, I mean, as we get closer and closer to, you know, atomic versions of store, compute, & networking, I mean, shouldn't it all, ultimately, get there. >> Yeah. >> I mean, there's requirements, and, you know, there's reality I don't deal, you know, luckily I don't have to go turn the stuff on and run it, but, you know, that is the vision, right? Atomic units of compute, atomic units of store, atomic units of network. >> Yeah, I mean, look, serverless is the ultimate Nirvana when it comes to the cloud, right? I mean, the notion of the cloud is that, you know, I have an application. It needs to run. I don't worry about the infrastructure, and to a certain extent, I don't even worry about the management. So, serverless and Lambda is the manifestation of that. >> Right. >> Right, and what we are starting to see is that many customers are, at least dabbling with Lambda. Now, I won't say that customers are building their core application with Lambda yet, because that requires a re-think of their application itself, but what we are starting to see is that Lambda is used in DevOps, Lambda is used in integration, Lambda is a glue-ware that sort of ties all of these applications together. >> Right. >> In fact, you know, this report that we put together, a lot of it has actually been put together on the back of Lambda. We use Lambda extensively to collect this kind of data, and create a report of this kind. >> (chuckles) That's great! Another piece I wanted to make sure that we talked about is really, kind of, the break-down of the clouds. >> Uh-huh. >> Obviously you guys have a huge percentage of your business is, you know, you ask customers, you guys were born in AWS. >> That's right. >> That seems pretty logical, but what's interesting is a lot of multi-cloud, so, you know, I don't know if you distinguish between multi-cloud, hyper-cloud, but at the end of the day, as I think Ramin talked about in the keynote, right, there's going to be different places for different workloads. >> Absolutely. You know, look, as you rightfully pointed out, we are born in AWS, so we have an affinity to AWS, and so AWS customers also have an affinity to Sumo Logic, so it's not wonder that a big swath of our customers are built in AWS. Now, having said that, what we are also seeing is actually an acceleration of our customers, you know, adopting more and more AWS. I mean, they are the leaders in the space. I mean, I think nobody can, nobody can question that statistics. What is interesting, though, is that we are starting to see increased adoption of multi-cloud. We saw about five per cent of our customers dabble in multi-cloud last year. We are at close to ten per cent this year. We are also seeing increased adoption of Azure. We had a, you know, about five per cent of our customers use Azure last year. We are starting to see almost, I should say about eight per cent of our customers used Azure last year. We saw, we're seeing about fifteen per cent of our customers use Azure this year. >> Right, right. >> Right, so Azure is a, you know, has definitely become a very credible second cloud alternative for many of our customers. >> Sure. >> Now, we do see interest in GCP. It's not translated into lots of GCP adoption in production environments yet, but we're definitely seeing that increased interest, and I'm sure, you know, when we put this report together next year, you'll see some very credible and statistically relevant GCP data in this report. >> Right, right, so, Kalyan, there's a lot here, and we could go on for (chuckles) and on and on. So, people can go to the website. They can download the report, but... It's so great, but what I like most about this report is you lay out the facts, right, you lay out your findings, people can question your data source or this or that, but you lay out your methodology, but then you have very specific instructions for the IT buyer about what they should consider, and I think that is so powerful, because I think from the position of an IT purchaser today, >> Right. >> They've got to just be getting creamed with, you know, like, with things we're talkin' about, like, with serverless and Lambda and security and DevOps and >> Right. >> And the pace of change for them keeps going faster, so where do they even begin when they're doin' this kind of analysis? It's not just putting it out for an RFP anymore, right? >> Yeah, I mean, that was the intent of this report, right? I mean, at the, you know, when we started this report our goal was to provide accurate, real-time information about, you know, where are these modern apps in the public cloud going? You know, our leading-edge customers, like Airbnb and the Twitters and the Salesforce and the Adobes, know how to do this well, but there is a huge swath of our community that is, in some sense, perplexed, right? I mean, they see this cloud adoption happening. They see this cloud wave coming. They have cut their teeth on, you know, data centers and applications in the data center. How do I make that transition to the cloud? How do I, you know, follow these cloud-first companies and learn from these companies? And, so, what we wanted to do was to collect this data, anonymously surface this data, and provide, you know, this insight to this community so that they can, you know, in some sense emulate, you know, these leading-edge companies and learn how to architect, build, run and secure their apps. >> Right, right, and I love this little, you know, kind of, the new stack, if you will, the architecture set-up. >> Right. >> Cake chart that you've done in the past. All right, great! So, a lot of, ton of information. I'll give you the last word as we're here at Illuminate, triple last year's numbers. A little bit about where you guys are goin' next. What's, kind of, top of your mind? >> You know, look, you know, Sumo Logic, as a company, you know, we are doing exceptionally well in this machine data analytics space. We are the only cloud-native machine data analytics vendor. We are where the market is going, right? We understand cloud; the apps are going to the cloud. We know how to manage these apps exceptionally well, but more importantly, you know, we think that it's also important and it behooves us to make sure that we take our developer community, our ops community, our security community along with us, and that's the intent of this report. >> Right. >> It's to not sell product, though we do want to sell it eventually. >> Yeah. >> But it's to provide you guys, actually I should say, provide the community with the right kinds of information so that, you know, they can do their jobs better. >> Right, right. >> That's the goal of Sumo Logic. It's all about, you know, empowering the people who power these modern apps, which is actually the theme of this event itself. >> Well, very good. Well, we'll leave it there, and thanks for taking a few minutes of your time. >> Thank you very much, Jeff. >> All right, he's Kalyan. I'm Jeff. You're watchin' theCUBE. We're at Sumo Logic Illuminate at San Francisco Hyatt Regency by the airport. See ya next time. (hip music)

Published Date : Sep 12 2018

SUMMARY :

It's theCUBE. and kind of the process we are excited to have that looks at how, you know, leading edge customers right, and I think you guys launched on AWS. and so you guys have really grown along with AWS, I think our founders, you know, ran into one of the and you know, we believe that we live in an age We looked at our 16 hundred customers. you know, manage them, and secure them? and they're going to answer the survey. Right. It's not an aggregation of, you know, a number of the hypotheses that we derive You know, right of the top: Docker and the adoption of Absolutely, I mean, ya know, everywhere you go you know, somewhat mainstream, and we are already Kubernetes, you know, will become a household name. Well it's funny: when we were at VMworld As well as Sonjay's, you know, so it's just, the right backing, you know, in the form of Google, is that, you know, as customers are building apps you know, incubating your applications So, what we are seeing is, one, the growth is customers have the choice. What strikes me is that the definition Ya know, this also goes back to that, you know, I mean, the engineer, they're a developer Right, so, another one, you know, serverless is and we are starting to see, you know, some wide-spread as we get closer and closer to, you know, I mean, there's requirements, and, you know, you know, I have an application. Right, and what we are starting to see is that In fact, you know, this report that we put together, is really, kind of, the break-down of the clouds. Obviously you guys have a huge percentage so, you know, I don't know if you distinguish We had a, you know, about five per cent Right, so Azure is a, you know, has definitely become and I'm sure, you know, when we put this report together is you lay out the facts, right, you lay out your findings, this insight to this community so that they can, you know, Right, right, and I love this little, you know, kind of, A little bit about where you guys are goin' next. You know, look, you know, Sumo Logic, as a company, It's to not sell product, though we do want so that, you know, they can do their jobs better. It's all about, you know, empowering the people and thanks for taking a few minutes of your time. San Francisco Hyatt Regency by the airport.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AWSORGANIZATION

0.99+

Pat GelsingerPERSON

0.99+

Jeff FrickPERSON

0.99+

JeffPERSON

0.99+

2010DATE

0.99+

Kalyan RamanathanPERSON

0.99+

14 percentQUANTITY

0.99+

Sumo LogicORGANIZATION

0.99+

KalyanPERSON

0.99+

last yearDATE

0.99+

30 per centQUANTITY

0.99+

twelve per centQUANTITY

0.99+

fifteen per centQUANTITY

0.99+

next yearDATE

0.99+

Sixteen hundred customersQUANTITY

0.99+

OracleORGANIZATION

0.99+

this yearDATE

0.99+

16 hundred customersQUANTITY

0.99+

GoogleORGANIZATION

0.99+

PatPERSON

0.99+

2013DATE

0.99+

BurlingameLOCATION

0.99+

RaminPERSON

0.99+

oneQUANTITY

0.99+

third versionQUANTITY

0.99+

firstQUANTITY

0.99+

first reportQUANTITY

0.99+

bothQUANTITY

0.98+

San FranciscoLOCATION

0.98+

AzureTITLE

0.98+

two years agoDATE

0.98+

LambdaTITLE

0.98+

ten other sourcesQUANTITY

0.98+

twoDATE

0.98+

three timesQUANTITY

0.98+

SQL SilverTITLE

0.98+

MySQLTITLE

0.98+

IlluminateORGANIZATION

0.98+

San Francisco Hyatt RegencyLOCATION

0.97+

about 600 peopleQUANTITY

0.97+

ten per centQUANTITY

0.97+

Sumo Logic IlluminateORGANIZATION

0.97+

two yearsQUANTITY

0.97+

about fifteen per centQUANTITY

0.96+

MongoDBTITLE

0.96+

about five per centQUANTITY

0.96+

AirbnbORGANIZATION

0.96+

three enterprisesQUANTITY

0.95+

todayDATE

0.95+

three years agoDATE

0.95+

KubernetesTITLE

0.95+