Image Title

Search Results for Each database:

Ed Walsh & Thomas Hazel | A New Database Architecture for Supercloud


 

(bright music) >> Hi, everybody, this is Dave Vellante, welcome back to Supercloud 2. Last August, at the first Supercloud event, we invited the broader community to help further define Supercloud, we assessed its viability, and identified the critical elements and deployment models of the concept. The objectives here at Supercloud too are, first of all, to continue to tighten and test the concept, the second is, we want to get real world input from practitioners on the problems that they're facing and the viability of Supercloud in terms of applying it to their business. So on the program, we got companies like Walmart, Sachs, Western Union, Ionis Pharmaceuticals, NASDAQ, and others. And the third thing that we want to do is we want to drill into the intersection of cloud and data to project what the future looks like in the context of Supercloud. So in this segment, we want to explore the concept of data architectures and what's going to be required for Supercloud. And I'm pleased to welcome one of our Supercloud sponsors, ChaosSearch, Ed Walsh is the CEO of the company, with Thomas Hazel, who's the Founder, CTO, and Chief Scientist. Guys, good to see you again, thanks for coming into our Marlborough studio. >> Always great. >> Great to be here. >> Okay, so there's a little debate, I'm going to put you right in the spot. (Ed chuckling) A little debate going on in the community started by Bob Muglia, a former CEO of Snowflake, and he was at Microsoft for a long time, and he looked at the Supercloud definition, said, "I think you need to tighten it up a little bit." So, here's what he came up with. He said, "A Supercloud is a platform that provides a programmatically consistent set of services hosted on heterogeneous cloud providers." So he's calling it a platform, not an architecture, which was kind of interesting. And so presumably the platform owner is going to be responsible for the architecture, but Dr. Nelu Mihai, who's a computer scientist behind the Cloud of Clouds Project, he chimed in and responded with the following. He said, "Cloud is a programming paradigm supporting the entire lifecycle of applications with data and logic natively distributed. Supercloud is an open architecture that integrates heterogeneous clouds in an agnostic manner." So, Ed, words matter. Is this an architecture or is it a platform? >> Put us on the spot. So, I'm sure you have concepts, I would say it's an architectural or design principle. Listen, I look at Supercloud as a mega trend, just like cloud, just like data analytics. And some companies are using the principle, design principles, to literally get dramatically ahead of everyone else. I mean, things you couldn't possibly do if you didn't use cloud principles, right? So I think it's a Supercloud effect, you're able to do things you're not able to. So I think it's more a design principle, but if you do it right, you get dramatic effect as far as customer value. >> So the conversation that we were having with Muglia, and Tristan Handy of dbt Labs, was, I'll set it up as the following, and, Thomas, would love to get your thoughts, if you have a CRM, think about applications today, it's all about forms and codifying business processes, you type a bunch of stuff into Salesforce, and all the salespeople do it, and this machine generates a forecast. What if you have this new type of data app that pulls data from the transaction system, the e-commerce, the supply chain, the partner ecosystem, et cetera, and then, without humans, actually comes up with a plan. That's their vision. And Muglia was saying, in order to do that, you need to rethink data architectures and database architectures specifically, you need to get down to the level of how the data is stored on the disc. What are your thoughts on that? Well, first of all, I'm going to cop out, I think it's actually both. I do think it's a design principle, I think it's not open technology, but open APIs, open access, and you can build a platform on that design principle architecture. Now, I'm a database person, I love solving the database problems. >> I'm waited for you to launch into this. >> Yeah, so I mean, you know, Snowflake is a database, right? It's a distributed database. And we wanted to crack those codes, because, multi-region, multi-cloud, customers wanted access to their data, and their data is in a variety of forms, all these services that you're talked about. And so what I saw as a core principle was cloud object storage, everyone streams their data to cloud object storage. From there we said, well, how about we rethink database architecture, rethink file format, so that we can take each one of these services and bring them together, whether distributively or centrally, such that customers can access and get answers, whether it's operational data, whether it's business data, AKA search, or SQL, complex distributed joins. But we had to rethink the architecture. I like to say we're not a first generation, or a second, we're a third generation distributed database on pure, pure cloud storage, no caching, no SSDs. Why? Because all that availability, the cost of time, is a struggle, and cloud object storage, we think, is the answer. >> So when you're saying no caching, so when I think about how companies are solving some, you know, pretty hairy problems, take MySQL Heatwave, everybody thought Oracle was going to just forget about MySQL, well, they come out with Heatwave. And the way they solve problems, and you see their benchmarks against Amazon, "Oh, we crush everybody," is they put it all in memory. So you said no caching? You're not getting performance through caching? How is that true, and how are you getting performance? >> Well, so five, six years ago, right? When you realize that cloud object storage is going to be everywhere, and it's going to be a core foundational, if you will, fabric, what would you do? Well, a lot of times the second generation say, "We'll take it out of cloud storage, put in SSDs or something, and put into cache." And that adds a lot of time, adds a lot of costs. But I said, what if, what if we could actually make the first read hot, the first read distributed joins and searching? And so what we went out to do was said, we can't cache, because that's adds time, that adds cost. We have to make cloud object storage high performance, like it feels like a caching SSD. That's where our patents are, that's where our technology is, and we've spent many years working towards this. So, to me, if you can crack that code, a lot of these issues we're talking about, multi-region, multicloud, different services, everybody wants to send their data to the data lake, but then they move it out, we said, "Keep it right there." >> You nailed it, the data gravity. So, Bob's right, the data's coming in, and you need to get the data from everywhere, but you need an environment that you can deal with all that different schema, all the different type of technology, but also at scale. Bob's right, you cannot use memory or SSDs to cache that, that doesn't scale, it doesn't scale cost effectively. But if you could, and what you did, is you made object storage, S3 first, but object storage, the only persistence by doing that. And then we get performance, we should talk about it, it's literally, you know, hundreds of terabytes of queries, and it's done in seconds, it's done without memory caching. We have concepts of caching, but the only caching, the only persistence, is actually when we're doing caching, we're just keeping another side-eye track of things on the S3 itself. So we're using, actually, the object storage to be a database, which is kind of where Bob was saying, we agree, but that's what you started at, people thought you were crazy. >> And maybe make it live. Don't think of it as archival or temporary space, make it live, real time streaming, operational data. What we do is make it smart, we see the data coming in, we uniquely index it such that you can get your use cases, that are search, observability, security, or backend operational. But we don't have to have this, I dunno, static, fixed, siloed type of architecture technologies that were traditionally built prior to Supercloud thinking. >> And you don't have to move everything, essentially, you can do it wherever the data lands, whatever cloud across the globe, you're able to bring it together, you get the cost effectiveness, because the only persistence is the cheapest storage persistent layer you can buy. But the key thing is you cracked the code. >> We had to crack the code, right? That was the key thing. >> That's where the plans are. >> And then once you do that, then everything else gets easier to scale, your architecture, across regions, across cloud. >> Now, it's a general purpose database, as Bob was saying, but we use that database to solve a particular issue, which is around operational data, right? So, we agree with Bob's. >> Interesting. So this brings me to this concept of data, Jimata Gan is one of our speakers, you know, we talk about data fabric, which is a NetApp, originally NetApp concept, Gartner's kind of co-opted it. But so, the basic concept is, data lives everywhere, whether it's an S3 bucket, or a SQL database, or a data lake, it's just a node on the data mesh. So in your view, how does this fit in with Supercloud? Ed, you've said that you've built, essentially, an enabler for that, for the data mesh, I think you're an enabler for the Supercloud-like principles. This is a big, chewy opportunity, and it requires, you know, a team approach. There's got to be an ecosystem, there's not going to be one Supercloud to rule them all, so where does the ecosystem fit into the discussion, and where do you fit into the ecosystem? >> Right, so we agree completely, there's not one Supercloud in effect, but we use Supercloud principles to build our platform, and then, you know, the ecosystem's going to be built on leveraging what everyone else's secret powers are, right? So our power, our superpower, based upon what we built is, we deal with, if you're having any scale, or cost effective scale issues, with data, machine generated data, like business observability or security data, we are your force multiplier, we will take that in singularly, just let it, simply put it in your object storage wherever it sits, and we give you uniformity access to that using OpenAPI access, SQL, or you know, Elasticsearch API. So, that's what we do, that's our superpower. So I'll play it into data mesh, that's a perfect, we are a node on a data mesh, but I'll play it in the soup about how, the ecosystem, we see it kind of playing, and we talked about it in just in the last couple days, how we see this kind of possibly. Short term, our superpowers, we deal with this data that's coming at these environments, people, customers, building out observability or security environments, or vendors that are selling their own Supercloud, I do observability, the Datadogs of the world, dot dot dot, the Splunks of the world, dot dot dot, and security. So what we do is we fit in naturally. What we do is a cost effective scale, just land it anywhere in the world, we deal with ingest, and it's a cost effective, an order of magnitude, or two or three order magnitudes more cost effective. Allows them, their customers are asking them to do the impossible, "Give me fast monitoring alerting. I want it snappy, but I want it to keep two years of data, (laughs) and I want it cost effective." It doesn't work. They're good at the fast monitoring alerting, we're good at the long-term retention. And yet there's some gray area between those two, but one to one is actually cheaper, so we would partner. So the first ecosystem plays, who wants to have the ability to, really, all the data's in those same environments, the security observability players, they can literally, just through API, drag our data into their point to grab. We can make it seamless for customers. Right now, we make it helpful to customers. Your Datadog, we make a button, easy go from Datadog to us for logs, save you money. Same thing with Grafana. But you can also look at ecosystem, those same vendors, it used to be a year ago it was, you know, its all about how can you grow, like it's growth at all costs, now it's about cogs. So literally we can go an environment, you supply what your customer wants, but we can help with cogs. And one-on one in a partnership is better than you trying to build on your own. >> Thomas, you were saying you make the first read fast, so you think about Snowflake. Everybody wants to talk about Snowflake and Databricks. So, Snowflake, great, but you got to get the data in there. All right, so that's, can you help with that problem? >> I mean we want simple in, right? And if you have to have structure in, you're not simple. So the idea that you have a simple in, data lake, schema read type philosophy, but schema right type performance. And so what I wanted to do, what we have done, is have that simple lake, and stream that data real time, and those access points of Search or SQL, to go after whatever business case you need, security observability, warehouse integration. But the key thing is, how do I make that click, click, click answer, and do it quickly? And so what we want to do is, that first read has to be fast. Why? 'Cause then you're going to do all this siloing, layers, complexity. If your first read's not fast, you're at a disadvantage, particularly in cost. And nobody says I want less data, but everyone has to, whether they say we're going to shorten the window, we're going to use AI to choose, but in a security moment, when you don't have that answer, you're in trouble. And that's why we are this service, this Supercloud service, if you will, providing access, well-known search, well-known SQL type access, that if you just have one access point, you're at a disadvantage. >> We actually talked about Snowflake and BigQuery, and a different platform, Data Bricks. That's kind of where we see the phase two of ecosystem. One is easy, the low-hanging fruit is observability and security firms. But the next one is, what we do, our super power is dealing with this messy data that schema is changing like night and day. Pipelines are tough, and it's changing all the time, but you want these things fast, and it's big data around the world. That's the next point, just use us alongside, or inside, one of their platforms, and now we get the best of both worlds. Our superpower is keeping this messy data as a streaming, okay, not a batch thing, allow you to do that. So, that's the second one. And then to be honest, the third one, which plays you to Supercloud, it also plays perfectly in the data mesh, is if you really go to the ultimate thing, what we have done is made object storage, S3, GCS, and blob storage, we made it a database. Put, get, complex query with big joins. You know, so back to your original thing, and Muglia teed it up perfectly, we've done that. Now imagine if that's an ecosystem, who would want that? If it's, again, it's uniform available across all the regions, across all the clouds, and it's right next to where you are building a service, or a client's trying, that's where the ecosystem, I think people are going to use Superclouds for their superpowers. We're really good at this, allows that short term. I think the Snowflakes and the Data Bricks are the medium term, you know? And then I think eventually gets to, hey, listen if you can make object storage fast, you can just go after it with simple SQL queries, or elastic. Who would want that? I think that's where people are going to leverage it. It's not going to be one Supercloud, and we leverage the super clouds. >> Our viewpoint is smart object storage can be programmable, and so we agree with Bob, but we're not saying do it here, do it here. This core, fundamental layer across regions, across clouds, that everyone has? Simple in. Right now, it's hard to get data in for access for analysis. So we said, simply, we'll automate the entire process, give you API access across regions, across clouds. And again, how do you do a distributed join that's fast? How do you do a distributed join that doesn't cost you an arm or a leg? And how do you do it at scale? And that's where we've been focused. >> So prior, the cloud object store was a niche. >> Yeah. >> S3 obviously changed that. How standard is, essentially, object store across the different cloud platforms? Is that a problem for you? Is that an easy thing to solve? >> Well, let's talk about it. I mean we've fundamentally, yeah we've extracted it, but fundamentally, cloud object storage, put, get, and list. That's why it's so scalable, 'cause it doesn't have all these other components. That complexity is where we have moved up, and provide direct analytical API access. So because of its simplicity, and costs, and security, and reliability, it can scale naturally. I mean, really, distributed object storage is easy, it's put-get anywhere, now what we've done is we put a layer of intelligence, you know, call it smart object storage, where access is simple. So whether it's multi-region, do a query across, or multicloud, do a query across, or hunting, searching. >> We've had clients doing Amazon and Google, we have some Azure, but we see Amazon and Google more, and it's a consistent service across all of them. Just literally put your data in the bucket of choice, or folder of choice, click a couple buttons, literally click that to say "that's hot," and after that, it's hot, you can see it. But we're not moving data, the data gravity issue, that's the other. That it's already natively flowing to these pools of object storage across different regions and clouds. We don't move it, we index it right there, we're spinning up stateless compute, back to the Supercloud concept. But now that allows us to do all these other things, right? >> And it's no longer just cheap and deep object storage. Right? >> Yeah, we make it the same, like you have an analytic platform regardless of where you're at, you don't have to worry about that. Yeah, we deal with that, we deal with a stateless compute coming up -- >> And make it programmable. Be able to say, "I want this bucket to provide these answers." Right, that's really the hope, the vision. And the complexity to build the entire stack, and then connect them together, we said, the fabric is cloud storage, we just provide the intelligence on top. >> Let's bring it back to the customers, and one of the things we're exploring in Supercloud too is, you know, is Supercloud a solution looking for a problem? Is a multicloud really a problem? I mean, you hear, you know, a lot of the vendor marketing says, "Oh, it's a disaster, because it's all different across the clouds." And I talked to a lot of customers even as part of Supercloud too, they're like, "Well, I solved that problem by just going mono cloud." Well, but then you're not able to take advantage of a lot of the capabilities and the primitives that, you know, like Google's data, or you like Microsoft's simplicity, their RPA, whatever it is. So what are customers telling you, what are their near term problems that they're trying to solve today, and how are they thinking about the future? >> Listen, it's a real problem. I think it started, I think this is a a mega trend, just like cloud. Just, cloud data, and I always add, analytics, are the mega trends. If you're looking at those, if you're not considering using the Supercloud principles, in other words, leveraging what I have, abstracting it out, and getting the most out of that, and then build value on top, I think you're not going to be able to keep up, In fact, no way you're going to keep up with this data volume. It's a geometric challenge, and you're trying to do linear things. So clients aren't necessarily asking, hey, for Supercloud, but they're really saying, I need to have a better mechanism to simplify this and get value across it, and how do you abstract that out to do that? And that's where they're obviously, our conversations are more amazed what we're able to do, and what they're able to do with our platform, because if you think of what we've done, the S3, or GCS, or object storage, is they can't imagine the ingest, they can't imagine how easy, time to glass, one minute, no matter where it lands in the world, querying this in seconds for hundreds of terabytes squared. People are amazed, but that's kind of, so they're not asking for that, but they are amazed. And then when you start talking on it, if you're an enterprise person, you're building a big cloud data platform, or doing data or analytics, if you're not trying to leverage the public clouds, and somehow leverage all of them, and then build on top, then I think you're missing it. So they might not be asking for it, but they're doing it. >> And they're looking for a lens, you mentioned all these different services, how do I bring those together quickly? You know, our viewpoint, our service, is I have all these streams of data, create a lens where they want to go after it via search, go after via SQL, bring them together instantly, no e-tailing out, no define this table, put into this database. We said, let's have a service that creates a lens across all these streams, and then make those connections. I want to take my CRM with my Google AdWords, and maybe my Salesforce, how do I do analysis? Maybe I want to hunt first, maybe I want to join, maybe I want to add another stream to it. And so our viewpoint is, it's so natural to get into these lake platforms and then provide lenses to get that access. >> And they don't want it separate, they don't want something different here, and different there. They want it basically -- >> So this is our industry, right? If something new comes out, remember virtualization came out, "Oh my God, this is so great, it's going to solve all these problems." And all of a sudden it just got to be this big, more complex thing. Same thing with cloud, you know? It started out with S3, and then EC2, and now hundreds and hundreds of different services. So, it's a complex matter for a lot of people, and this creates problems for customers, especially when you got divisions that are using different clouds, and you're saying that the solution, or a solution for the part of the problem, is to really allow the data to stay in place on S3, use that standard, super simple, but then give it what, Ed, you've called superpower a couple of times, to make it fast, make it inexpensive, and allow you to do that across clouds. >> Yeah, yeah. >> I'll give you guys the last word on that. >> No, listen, I think, we think Supercloud allows you to do a lot more. And for us, data, everyone says more data, more problems, more budget issue, everyone knows more data is better, and we show you how to do it cost effectively at scale. And we couldn't have done it without the design principles of we're leveraging the Supercloud to get capabilities, and because we use super, just the object storage, we're able to get these capabilities of ingest, scale, cost effectiveness, and then we built on top of this. In the end, a database is a data platform that allows you to go after everything distributed, and to get one platform for analytics, no matter where it lands, that's where we think the Supercloud concepts are perfect, that's where our clients are seeing it, and we're kind of excited about it. >> Yeah a third generation database, Supercloud database, however we want to phrase it, and make it simple, but provide the value, and make it instant. >> Guys, thanks so much for coming into the studio today, I really thank you for your support of theCUBE, and theCUBE community, it allows us to provide events like this and free content. I really appreciate it. >> Oh, thank you. >> Thank you. >> All right, this is Dave Vellante for John Furrier in theCUBE community, thanks for being with us today. You're watching Supercloud 2, keep it right there for more thought provoking discussions around the future of cloud and data. (bright music)

Published Date : Feb 17 2023

SUMMARY :

And the third thing that we want to do I'm going to put you right but if you do it right, So the conversation that we were having I like to say we're not a and you see their So, to me, if you can crack that code, and you need to get the you can get your use cases, But the key thing is you cracked the code. We had to crack the code, right? And then once you do that, So, we agree with Bob's. and where do you fit into the ecosystem? and we give you uniformity access to that so you think about Snowflake. So the idea that you have are the medium term, you know? and so we agree with Bob, So prior, the cloud that an easy thing to solve? you know, call it smart object storage, and after that, it's hot, you can see it. And it's no longer just you don't have to worry about And the complexity to and one of the things we're and how do you abstract it's so natural to get and different there. and allow you to do that across clouds. I'll give you guys and we show you how to do it but provide the value, I really thank you for around the future of cloud and data.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
WalmartORGANIZATION

0.99+

Dave VellantePERSON

0.99+

NASDAQORGANIZATION

0.99+

Bob MugliaPERSON

0.99+

ThomasPERSON

0.99+

Thomas HazelPERSON

0.99+

Ionis PharmaceuticalsORGANIZATION

0.99+

Western UnionORGANIZATION

0.99+

Ed WalshPERSON

0.99+

BobPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Nelu MihaiPERSON

0.99+

SachsORGANIZATION

0.99+

Tristan HandyPERSON

0.99+

twoQUANTITY

0.99+

AmazonORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

two yearsQUANTITY

0.99+

Supercloud 2TITLE

0.99+

firstQUANTITY

0.99+

Last AugustDATE

0.99+

threeQUANTITY

0.99+

OracleORGANIZATION

0.99+

SnowflakeORGANIZATION

0.99+

bothQUANTITY

0.99+

dbt LabsORGANIZATION

0.99+

John FurrierPERSON

0.99+

EdPERSON

0.99+

GartnerORGANIZATION

0.99+

Jimata GanPERSON

0.99+

third oneQUANTITY

0.99+

one minuteQUANTITY

0.99+

secondQUANTITY

0.99+

first generationQUANTITY

0.99+

third generationQUANTITY

0.99+

GrafanaORGANIZATION

0.99+

second generationQUANTITY

0.99+

second oneQUANTITY

0.99+

hundreds of terabytesQUANTITY

0.98+

SQLTITLE

0.98+

fiveDATE

0.98+

oneQUANTITY

0.98+

DatabricksORGANIZATION

0.98+

a year agoDATE

0.98+

ChaosSearchORGANIZATION

0.98+

MugliaPERSON

0.98+

MySQLTITLE

0.98+

both worldsQUANTITY

0.98+

third thingQUANTITY

0.97+

MarlboroughLOCATION

0.97+

theCUBEORGANIZATION

0.97+

todayDATE

0.97+

SupercloudORGANIZATION

0.97+

ElasticsearchTITLE

0.96+

NetAppTITLE

0.96+

DatadogORGANIZATION

0.96+

OneQUANTITY

0.96+

EC2TITLE

0.96+

each oneQUANTITY

0.96+

S3TITLE

0.96+

one platformQUANTITY

0.95+

Supercloud 2EVENT

0.95+

first readQUANTITY

0.95+

six years agoDATE

0.95+

Analyst Power Panel: Future of Database Platforms


 

(upbeat music) >> Once a staid and boring business dominated by IBM, Oracle, and at the time newcomer Microsoft, along with a handful of wannabes, the database business has exploded in the past decade and has become a staple of financial excellence, customer experience, analytic advantage, competitive strategy, growth initiatives, visualizations, not to mention compliance, security, privacy and dozens of other important use cases and initiatives. And on the vendor's side of the house, we've seen the rapid ascendancy of cloud databases. Most notably from Snowflake, whose massive raises leading up to its IPO in late 2020 sparked a spate of interest and VC investment in the separation of compute and storage and all that elastic resource stuff in the cloud. The company joined AWS, Azure and Google to popularize cloud databases, which have become a linchpin of competitive strategies for technology suppliers. And if I get you to put your data in my database and in my cloud, and I keep innovating, I'm going to build a moat and achieve a hugely attractive lifetime customer value in a really amazing marginal economics dynamic that is going to fund my future. And I'll be able to sell other adjacent services, not just compute and storage, but machine learning and inference and training and all kinds of stuff, dozens of lucrative cloud offerings. Meanwhile, the database leader, Oracle has invested massive amounts of money to maintain its lead. It's building on its position as the king of mission critical workloads and making typical Oracle like claims against the competition. Most were recently just yesterday with another announcement around MySQL HeatWave. An extension of MySQL that is compatible with on-premises MySQLs and is setting new standards in price performance. We're seeing a dramatic divergence in strategies across the database spectrum. On the far left, we see Amazon with more than a dozen database offerings each with its own API and primitives. AWS is taking a right tool for the right job approach, often building on open source platforms and creating services that it offers to customers to solve very specific problems for developers. And on the other side of the line, we see Oracle, which is taking the Swiss Army Knife approach, converging database functionality, enabling analytic and transactional workloads to run in the same data store, eliminating the need to ETL, at the same time adding capabilities into its platform like automation and machine learning. Welcome to this database Power Panel. My name is Dave Vellante, and I'm so excited to bring together some of the most respected industry analyst in the community. Today we're going to assess what's happening in the market. We're going to dig into the competitive landscape and explore the future of database and database platforms and decode what it means to customers. Let me take a moment to welcome our guest analyst today. Matt Kimball is a vice president and principal analysts at Moor Insights and Strategy, Matt. He knows products, he knows industry, he's got real world IT expertise, and he's got all the angles 25 plus years of experience in all kinds of great background. Matt, welcome. Thanks very much for coming on theCUBE. Holgar Mueller, friend of theCUBE, vice president and principal analyst at Constellation Research in depth knowledge on applications, application development, knows developers. He's worked at SAP and Oracle. And then Bob Evans is Chief Content Officer and co-founder of the Acceleration Economy, founder and principle of Cloud Wars. Covers all kinds of industry topics and great insights. He's got awesome videos, these three minute hits. If you haven't seen 'em, checking them out, knows cloud companies, his Cloud Wars minutes are fantastic. And then of course, Marc Staimer is the founder of Dragon Slayer Research. A frequent contributor and guest analyst at Wikibon. He's got a wide ranging knowledge across IT products, knows technology really well, can go deep. And then of course, Ron Westfall, Senior Analyst and Director Research Director at Futurum Research, great all around product trends knowledge. Can take, you know, technical dives and really understands competitive angles, knows Redshift, Snowflake, and many others. Gents, thanks so much for taking the time to join us in theCube today. It's great to have you on, good to see you. >> Good to be here, thanks for having us. >> Thanks, Dave. >> All right, let's start with an around the horn and briefly, if each of you would describe, you know, anything I missed in your areas of expertise and then you answer the following question, how would you describe the state of the database, state of platform market today? Matt Kimball, please start. >> Oh, I hate going first, but that it's okay. How would I describe the world today? I would just in one sentence, I would say, I'm glad I'm not in IT anymore, right? So, you know, it is a complex and dangerous world out there. And I don't envy IT folks I'd have to support, you know, these modernization and transformation efforts that are going on within the enterprise. It used to be, you mentioned it, Dave, you would argue about IBM versus Oracle versus this newcomer in the database space called Microsoft. And don't forget Sybase back in the day, but you know, now it's not just, which SQL vendor am I going to go with? It's all of these different, divergent data types that have to be taken, they have to be merged together, synthesized. And somehow I have to do that cleanly and use this to drive strategic decisions for my business. That is not easy. So, you know, you have to look at it from the perspective of the business user. It's great for them because as a DevOps person, or as an analyst, I have so much flexibility and I have this thing called the cloud now where I can go get services immediately. As an IT person or a DBA, I am calling up prevention hotlines 24 hours a day, because I don't know how I'm going to be able to support the business. And as an Oracle or as an Oracle or a Microsoft or some of the cloud providers and cloud databases out there, I'm licking my chops because, you know, my market is expanding and expanding every day. >> Great, thank you for that, Matt. Holgar, how do you see the world these days? You always have a good perspective on things, share with us. >> Well, I think it's the best time to be in IT, I'm not sure what Matt is talking about. (laughing) It's easier than ever, right? The direction is going to cloud. Kubernetes has won, Google has the best AI for now, right? So things are easier than ever before. You made commitments for five plus years on hardware, networking and so on premise, and I got gray hair about worrying it was the wrong decision. No, just kidding. But you kind of both sides, just to be controversial, make it interesting, right. So yeah, no, I think the interesting thing specifically with databases, right? We have this big suite versus best of breed, right? Obviously innovation, like you mentioned with Snowflake and others happening in the cloud, the cloud vendors server, where to save of their databases. And then we have one of the few survivors of the old guard as Evans likes to call them is Oracle who's doing well, both their traditional database. And now, which is really interesting, remarkable from that because Oracle it was always the power of one, have one database, add more to it, make it what I call the universal database. And now this new HeatWave offering is coming and MySQL open source side. So they're getting the second (indistinct) right? So it's interesting that older players, traditional players who still are in the market are diversifying their offerings. Something we don't see so much from the traditional tools from Oracle on the Microsoft side or the IBM side these days. >> Great, thank you Holgar. Bob Evans, you've covered this business for a while. You've worked at, you know, a number of different outlets and companies and you cover the competition, how do you see things? >> Dave, you know, the other angle to look at this from is from the customer side, right? You got now CEOs who are any sort of business across all sorts of industries, and they understand that their future success is going to be dependent on their ability to become a digital company, to understand data, to use it the right way. So as you outline Dave, I think in your intro there, it is a fantastic time to be in the database business. And I think we've got a lot of new buyers and influencers coming in. They don't know all this history about IBM and Microsoft and Oracle and you know, whoever else. So I think they're going to take a long, hard look, Dave, at some of these results and who is able to help these companies not serve up the best technology, but who's going to be able to help their business move into the digital future. So it's a fascinating time now from every perspective. >> Great points, Bob. I mean, digital transformation has gone from buzzword to imperative. Mr. Staimer, how do you see things? >> I see things a little bit differently than my peers here in that I see the database market being segmented. There's all the different kinds of databases that people are looking at for different kinds of data, and then there is databases in the cloud. And so database as cloud service, I view very differently than databases because the traditional way of implementing a database is changing and it's changing rapidly. So one of the premises that you stated earlier on was that you viewed Oracle as a database company. I don't view Oracle as a database company anymore. I view Oracle as a cloud company that happens to have a significant expertise and specialty in databases, and they still sell database software in the traditional way, but ultimately they're a cloud company. So database cloud services from my point of view is a very distinct market from databases. >> Okay, well, you gave us some good meat on the bone to talk about that. Last but not least-- >> Dave did Marc, just say Oracle's a cloud company? >> Yeah. (laughing) Take away the database, it would be interesting to have that discussion, but let's let Ron jump in here. Ron, give us your take. >> That's a great segue. I think it's truly the era of the cloud database, that's something that's rising. And the key trends that come with it include for example, elastic scaling. That is the ability to scale on demand, to right size workloads according to customer requirements. And also I think it's going to increase the prioritization for high availability. That is the player who can provide the highest availability is going to have, I think, a great deal of success in this emerging market. And also I anticipate that there will be more consolidation across platforms in order to enable cost savings for customers, and that's something that's always going to be important. And I think we'll see more of that over the horizon. And then finally security, security will be more important than ever. We've seen a spike (indistinct), we certainly have seen geopolitical originated cybersecurity concerns. And as a result, I see database security becoming all the more important. >> Great, thank you. Okay, let me share some data with you guys. I'm going to throw this at you and see what you think. We have this awesome data partner called Enterprise Technology Research, ETR. They do these quarterly surveys and each period with dozens of industry segments, they track clients spending, customer spending. And this is the database, data warehouse sector okay so it's taxonomy, so it's not perfect, but it's a big kind of chunk. They essentially ask customers within a category and buy a specific vendor, you're spending more or less on the platform? And then they subtract the lesses from the mores and they derive a metric called net score. It's like NPS, it's a measure of spending velocity. It's more complicated and granular than that, but that's the basis and that's the vertical axis. The horizontal axis is what they call market share, it's not like IDC market share, it's just pervasiveness in the data set. And so there are a couple of things that stand out here and that we can use as reference point. The first is the momentum of Snowflake. They've been off the charts for many, many, for over two years now, anything above that dotted red line, that 40%, is considered by ETR to be highly elevated and Snowflake's even way above that. And I think it's probably not sustainable. We're going to see in the next April survey, next month from those guys, when it comes out. And then you see AWS and Microsoft, they're really pervasive on the horizontal axis and highly elevated, Google falls behind them. And then you got a number of well funded players. You got Cockroach Labs, Mongo, Redis, MariaDB, which of course is a fork on MySQL started almost as protest at Oracle when they acquired Sun and they got MySQL and you can see the number of others. Now Oracle who's the leading database player, despite what Marc Staimer says, we know, (laughs) and they're a cloud player (laughing) who happens to be a leading database player. They dominate in the mission critical space, we know that they're the king of that sector, but you can see here that they're kind of legacy, right? They've been around a long time, they get a big install base. So they don't have the spending momentum on the vertical axis. Now remember this is, just really this doesn't capture spending levels, so that understates Oracle but nonetheless. So it's not a complete picture like SAP for instance is not in here, no Hana. I think people are actually buying it, but it doesn't show up here, (laughs) but it does give an indication of momentum and presence. So Bob Evans, I'm going to start with you. You've commented on many of these companies, you know, what does this data tell you? >> Yeah, you know, Dave, I think all these compilations of things like that are interesting, and that folks at ETR do some good work, but I think as you said, it's a snapshot sort of a two-dimensional thing of a rapidly changing, three dimensional world. You know, the incidents at which some of these companies are mentioned versus the volume that happens. I think it's, you know, with Oracle and I'm not going to declare my religious affiliation, either as cloud company or database company, you know, they're all of those things and more, and I think some of our old language of how we classify companies is just not relevant anymore. But I want to ask too something in here, the autonomous database from Oracle, nobody else has done that. So either Oracle is crazy, they've tried out a technology that nobody other than them is interested in, or they're onto something that nobody else can match. So to me, Dave, within Oracle, trying to identify how they're doing there, I would watch autonomous database growth too, because right, it's either going to be a big plan and it breaks through, or it's going to be caught behind. And the Snowflake phenomenon as you mentioned, that is a rare, rare bird who comes up and can grow 100% at a billion dollar revenue level like that. So now they've had a chance to come in, scare the crap out of everybody, rock the market with something totally new, the data cloud. Will the bigger companies be able to catch up and offer a compelling alternative, or is Snowflake going to continue to be this outlier. It's a fascinating time. >> Really, interesting points there. Holgar, I want to ask you, I mean, I've talked to certainly I'm sure you guys have too, the founders of Snowflake that came out of Oracle and they actually, they don't apologize. They say, "Hey, we not going to do all that complicated stuff that Oracle does, we were trying to keep it real simple." But at the same time, you know, they don't do sophisticated workload management. They don't do complex joints. They're kind of relying on the ecosystems. So when you look at the data like this and the various momentums, and we talked about the diverging strategies, what does this say to you? >> Well, it is a great point. And I think Snowflake is an example how the cloud can turbo charge a well understood concept in this case, the data warehouse, right? You move that and you find steroids and you see like for some players who've been big in data warehouse, like Sentara Data, as an example, here in San Diego, what could have been for them right in that part. The interesting thing, the problem though is the cloud hides a lot of complexity too, which you can scale really well as you attract lots of customers to go there. And you don't have to build things like what Bob said, right? One of the fascinating things, right, nobody's answering Oracle on the autonomous database. I don't think is that they cannot, they just have different priorities or the database is not such a priority. I would dare to say that it's for IBM and Microsoft right now at the moment. And the cloud vendors, you just hide that right through scripts and through scale because you support thousands of customers and you can deal with a little more complexity, right? It's not against them. Whereas if you have to run it yourself, very different story, right? You want to have the autonomous parts, you want to have the powerful tools to do things. >> Thank you. And so Matt, I want to go to you, you've set up front, you know, it's just complicated if you're in IT, it's a complicated situation and you've been on the customer side. And if you're a buyer, it's obviously, it's like Holgar said, "Cloud's supposed to make this stuff easier, but the simpler it gets the more complicated gets." So where do you place your bets? Or I guess more importantly, how do you decide where to place your bets? >> Yeah, it's a good question. And to what Bob and Holgar said, you know, the around autonomous database, I think, you know, part of, as I, you know, play kind of armchair psychologist, if you will, corporate psychologists, I look at what Oracle is doing and, you know, databases where they've made their mark and it's kind of, that's their strong position, right? So it makes sense if you're making an entry into this cloud and you really want to kind of build momentum, you go with what you're good at, right? So that's kind of the strength of Oracle. Let's put a lot of focus on that. They do a lot more than database, don't get me wrong, but you know, I'm going to short my strength and then kind of pivot from there. With regards to, you know, what IT looks at and what I would look at you know as an IT director or somebody who is, you know, trying to consume services from these different cloud providers. First and foremost, I go with what I know, right? Let's not forget IT is a conservative group. And when we look at, you know, all the different permutations of database types out there, SQL, NoSQL, all the different types of NoSQL, those are largely being deployed by business users that are looking for agility or businesses that are looking for agility. You know, the reason why MongoDB is so popular is because of DevOps, right? It's a great platform to develop on and that's where it kind of gained its traction. But as an IT person, I want to go with what I know, where my muscle memory is, and that's my first position. And so as I evaluate different cloud service providers and cloud databases, I look for, you know, what I know and what I've invested in and where my muscle memory is. Is there enough there and do I have enough belief that that company or that service is going to be able to take me to, you know, where I see my organization in five years from a data management perspective, from a business perspective, are they going to be there? And if they are, then I'm a little bit more willing to make that investment, but it is, you know, if I'm kind of going in this blind or if I'm cloud native, you know, that's where the Snowflakes of the world become very attractive to me. >> Thank you. So Marc, I asked Andy Jackson in theCube one time, you have all these, you know, data stores and different APIs and primitives and you know, very granular, what's the strategy there? And he said, "Hey, that allows us as the market changes, it allows us to be more flexible. If we start building abstractions layers, it's harder for us." I think also it was not a good time to market advantage, but let me ask you, I described earlier on that spectrum from AWS to Oracle. We just saw yesterday, Oracle announced, I think the third major enhancement in like 15 months to MySQL HeatWave, what do you make of that announcement? How do you think it impacts the competitive landscape, particularly as it relates to, you know, converging transaction and analytics, eliminating ELT, I know you have some thoughts on this. >> So let me back up for a second and defend my cloud statement about Oracle for a moment. (laughing) AWS did a great job in developing the cloud market in general and everything in the cloud market. I mean, I give them lots of kudos on that. And a lot of what they did is they took open source software and they rent it to people who use their cloud. So I give 'em lots of credit, they dominate the market. Oracle was late to the cloud market. In fact, they actually poo-pooed it initially, if you look at some of Larry Ellison's statements, they said, "Oh, it's never going to take off." And then they did 180 turn, and they said, "Oh, we're going to embrace the cloud." And they really have, but when you're late to a market, you've got to be compelling. And this ties into the announcement yesterday, but let's deal with this compelling. To be compelling from a user point of view, you got to be twice as fast, offer twice as much functionality, at half the cost. That's generally what compelling is that you're going to capture market share from the leaders who established the market. It's very difficult to capture market share in a new market for yourself. And you're right. I mean, Bob was correct on this and Holgar and Matt in which you look at Oracle, and they did a great job of leveraging their database to move into this market, give 'em lots of kudos for that too. But yesterday they announced, as you said, the third innovation release and the pace is just amazing of what they're doing on these releases on HeatWave that ties together initially MySQL with an integrated builtin analytics engine, so a data warehouse built in. And then they added automation with autopilot, and now they've added machine learning to it, and it's all in the same service. It's not something you can buy and put on your premise unless you buy their cloud customers stuff. But generally it's a cloud offering, so it's compellingly better as far as the integration. You don't buy multiple services, you buy one and it's lower cost than any of the other services, but more importantly, it's faster, which again, give 'em credit for, they have more integration of a product. They can tie things together in a way that nobody else does. There's no additional services, ETL services like Glue and AWS. So from that perspective, they're getting better performance, fewer services, lower cost. Hmm, they're aiming at the compelling side again. So from a customer point of view it's compelling. Matt, you wanted to say something there. >> Yeah, I want to kind of, on what you just said there Marc, and this is something I've found really interesting, you know. The traditional way that you look at software and, you know, purchasing software and IT is, you look at either best of breed solutions and you have to work on the backend to integrate them all and make them all work well. And generally, you know, the big hit against the, you know, we have one integrated offering is that, you lose capability or you lose depth of features, right. And to what you were saying, you know, that's the thing I found interesting about what Oracle is doing is they're building in depth as they kind of, you know, build that service. It's not like you're losing a lot of capabilities, because you're going to one integrated service versus having to use A versus B versus C, and I love that idea. >> You're right. Yeah, not only you're not losing, but you're gaining functionality that you can't get by integrating a lot of these. I mean, I can take Snowflake and integrate it in with machine learning, but I also have to integrate in with a transactional database. So I've got to have connectors between all of this, which means I'm adding time. And what it comes down to at the end of the day is expertise, effort, time, and cost. And so what I see the difference from the Oracle announcements is they're aiming at reducing all of that by increasing performance as well. Correct me if I'm wrong on that but that's what I saw at the announcement yesterday. >> You know, Marc, one thing though Marc, it's funny you say that because I started out saying, you know, I'm glad I'm not 19 anymore. And the reason is because of exactly what you said, it's almost like there's a pseudo level of witchcraft that's required to support the modern data environment right in the enterprise. And I need simpler faster, better. That's what I need, you know, I am no longer wearing pocket protectors. I have turned from, you know, break, fix kind of person, to you know, business consultant. And I need that point and click simplicity, but I can't sacrifice, you know, a depth of features of functionality on the backend as I play that consultancy role. >> So, Ron, I want to bring in Ron, you know, it's funny. So Matt, you mentioned Mongo, I often and say, if Oracle mentions you, you're on the map. We saw them yesterday Ron, (laughing) they hammered RedShifts auto ML, they took swipes at Snowflake, a little bit of BigQuery. What were your thoughts on that? Do you agree with what these guys are saying in terms of HeatWaves capabilities? >> Yes, Dave, I think that's an excellent question. And fundamentally I do agree. And the question is why, and I think it's important to know that all of the Oracle data is backed by the fact that they're using benchmarks. For example, all of the ML and all of the TPC benchmarks, including all the scripts, all the configs and all the detail are posted on GitHub. So anybody can look at these results and they're fully transparent and replicate themselves. If you don't agree with this data, then by all means challenge it. And we have not really seen that in all of the new updates in HeatWave over the last 15 months. And as a result, when it comes to these, you know, fundamentals in looking at the competitive landscape, which I think gives validity to outcomes such as Oracle being able to deliver 4.8 times better price performance than Redshift. As well as for example, 14.4 better price performance than Snowflake, and also 12.9 better price performance than BigQuery. And so that is, you know, looking at the quantitative side of things. But again, I think, you know, to Marc's point and to Matt's point, there are also qualitative aspects that clearly differentiate the Oracle proposition, from my perspective. For example now the MySQL HeatWave ML capabilities are native, they're built in, and they also support things such as completion criteria. And as a result, that enables them to show that hey, when you're using Redshift ML for example, you're having to also use their SageMaker tool and it's running on a meter. And so, you know, nobody really wants to be running on a meter when, you know, executing these incredibly complex tasks. And likewise, when it comes to Snowflake, they have to use a third party capability. They don't have the built in, it's not native. So the user, to the point that he's having to spend more time and it increases complexity to use auto ML capabilities across the Snowflake platform. And also, I think it also applies to other important features such as data sampling, for example, with the HeatWave ML, it's intelligent sampling that's being implemented. Whereas in contrast, we're seeing Redshift using random sampling. And again, Snowflake, you're having to use a third party library in order to achieve the same capabilities. So I think the differentiation is crystal clear. I think it definitely is refreshing. It's showing that this is where true value can be assigned. And if you don't agree with it, by all means challenge the data. >> Yeah, I want to come to the benchmarks in a minute. By the way, you know, the gentleman who's the Oracle's architect, he did a great job on the call yesterday explaining what you have to do. I thought that was quite impressive. But Bob, I know you follow the financials pretty closely and on the earnings call earlier this month, Ellison said that, "We're going to see HeatWave on AWS." And the skeptic in me said, oh, they must not be getting people to come to OCI. And then they, you remember this chart they showed yesterday that showed the growth of HeatWave on OCI. But of course there was no data on there, it was just sort of, you know, lines up and to the right. So what do you guys think of that? (Marc laughs) Does it signal Bob, desperation by Oracle that they can't get traction on OCI, or is it just really a smart tame expansion move? What do you think? >> Yeah, Dave, that's a great question. You know, along the way there, and you know, just inside of that was something that said Ellison said on earnings call that spoke to a different sort of philosophy or mindset, almost Marc, where he said, "We're going to make this multicloud," right? With a lot of their other cloud stuff, if you wanted to use any of Oracle's cloud software, you had to use Oracle's infrastructure, OCI, there was no other way out of it. But this one, but I thought it was a classic Ellison line. He said, "Well, we're making this available on AWS. We're making this available, you know, on Snowflake because we're going after those users. And once they see what can be done here." So he's looking at it, I guess you could say, it's a concession to customers because they want multi-cloud. The other way to look at it, it's a hunting expedition and it's one of those uniquely I think Oracle ways. He said up front, right, he doesn't say, "Well, there's a big market, there's a lot for everybody, we just want on our slice." Said, "No, we are going after Amazon, we're going after Redshift, we're going after Aurora. We're going after these users of Snowflake and so on." And I think it's really fairly refreshing these days to hear somebody say that, because now if I'm a buyer, I can look at that and say, you know, to Marc's point, "Do they measure up, do they crack that threshold ceiling? Or is this just going to be more pain than a few dollars savings is worth?" But you look at those numbers that Ron pointed out and that we all saw in that chart. I've never seen Dave, anything like that. In a substantive market, a new player coming in here, and being able to establish differences that are four, seven, eight, 10, 12 times better than competition. And as new buyers look at that, they're going to say, "What the hell are we doing paying, you know, five times more to get a poor result? What's going on here?" So I think this is going to rattle people and force a harder, closer look at what these alternatives are. >> I wonder if the guy, thank you. Let's just skip ahead of the benchmarks guys, bring up the next slide, let's skip ahead a little bit here, which talks to the benchmarks and the benchmarking if we can. You know, David Floyer, the sort of semiretired, you know, Wikibon analyst said, "Dave, this is going to force Amazon and others, Snowflake," he said, "To rethink actually how they architect databases." And this is kind of a compilation of some of the data that they shared. They went after Redshift mostly, (laughs) but also, you know, as I say, Snowflake, BigQuery. And, like I said, you can always tell which companies are doing well, 'cause Oracle will come after you, but they're on the radar here. (laughing) Holgar should we take this stuff seriously? I mean, or is it, you know, a grain salt? What are your thoughts here? >> I think you have to take it seriously. I mean, that's a great question, great point on that. Because like Ron said, "If there's a flaw in a benchmark, we know this database traditionally, right?" If anybody came up that, everybody will be, "Oh, you put the wrong benchmark, it wasn't audited right, let us do it again," and so on. We don't see this happening, right? So kudos to Oracle to be aggressive, differentiated, and seem to having impeccable benchmarks. But what we really see, I think in my view is that the classic and we can talk about this in 100 years, right? Is the suite versus best of breed, right? And the key question of the suite, because the suite's always slower, right? No matter at which level of the stack, you have the suite, then the best of breed that will come up with something new, use a cloud, put the data warehouse on steroids and so on. The important thing is that you have to assess as a buyer what is the speed of my suite vendor. And that's what you guys mentioned before as well, right? Marc said that and so on, "Like, this is a third release in one year of the HeatWave team, right?" So everybody in the database open source Marc, and there's so many MySQL spinoffs to certain point is put on shine on the speed of (indistinct) team, putting out fundamental changes. And the beauty of that is right, is so inherent to the Oracle value proposition. Larry's vision of building the IBM of the 21st century, right from the Silicon, from the chip all the way across the seven stacks to the click of the user. And that what makes the database what Rob was saying, "Tied to the OCI infrastructure," because designed for that, it runs uniquely better for that, that's why we see the cross connect to Microsoft. HeatWave so it's different, right? Because HeatWave runs on cheap hardware, right? Which is the breadth and butter 886 scale of any cloud provider, right? So Oracle probably needs it to scale OCI in a different category, not the expensive side, but also allow us to do what we said before, the multicloud capability, which ultimately CIOs really want, because data gravity is real, you want to operate where that is. If you have a fast, innovative offering, which gives you more functionality and the R and D speed is really impressive for the space, puts away bad results, then it's a good bet to look at. >> Yeah, so you're saying, that we versus best of breed. I just want to sort of play back then Marc a comment. That suite versus best of breed, there's always been that trade off. If I understand you Holgar you're saying that somehow Oracle has magically cut through that trade off and they're giving you the best of both. >> It's the developing velocity, right? The provision of important features, which matter to buyers of the suite vendor, eclipses the best of breed vendor, then the best of breed vendor is in the hell of a potential job. >> Yeah, go ahead Marc. >> Yeah and I want to add on what Holgar just said there. I mean the worst job in the data center is data movement, moving the data sucks. I don't care who you are, nobody likes it. You never get any kudos for doing it well, and you always get the ah craps, when things go wrong. So it's in- >> In the data center Marc all the time across data centers, across cloud. That's where the bleeding comes. >> It's right, you get beat up all the time. So nobody likes to move data, ever. So what you're looking at with what they announce with HeatWave and what I love about HeatWave is it doesn't matter when you started with it, you get all the additional features they announce it's part of the service, all the time. But they don't have to move any of the data. You want to analyze the data that's in your transactional, MySQL database, it's there. You want to do machine learning models, it's there, there's no data movement. The data movement is the key thing, and they just eliminate that, in so many ways. And the other thing I wanted to talk about is on the benchmarks. As great as those benchmarks are, they're really conservative 'cause they're underestimating the cost of that data movement. The ETLs, the other services, everything's left out. It's just comparing HeatWave, MySQL cloud service with HeatWave versus Redshift, not Redshift and Aurora and Glue, Redshift and Redshift ML and SageMaker, it's just Redshift. >> Yeah, so what you're saying is what Oracle's doing is saying, "Okay, we're going to run MySQL HeatWave benchmarks on analytics against Redshift, and then we're going to run 'em in transaction against Aurora." >> Right. >> But if you really had to look at what you would have to do with the ETL, you'd have to buy two different data stores and all the infrastructure around that, and that goes away so. >> Due to the nature of the competition, they're running narrow best of breed benchmarks. There is no suite level benchmark (Dave laughs) because they created something new. >> Well that's you're the earlier point they're beating best of breed with a suite. So that's, I guess to Floyer's earlier point, "That's going to shake things up." But I want to come back to Bob Evans, 'cause I want to tap your Cloud Wars mojo before we wrap. And line up the horses, you got AWS, you got Microsoft, Google and Oracle. Now they all own their own cloud. Snowflake, Mongo, Couchbase, Redis, Cockroach by the way they're all doing very well. They run in the cloud as do many others. I think you guys all saw the Andreessen, you know, commentary from Sarah Wang and company, to talk about the cost of goods sold impact of cloud. So owning your own cloud has to be an advantage because other guys like Snowflake have to pay cloud vendors and negotiate down versus having the whole enchilada, Safra Catz's dream. Bob, how do you think this is going to impact the market long term? >> Well, Dave, that's a great question about, you know, how this is all going to play out. If I could mention three things, one, Frank Slootman has done a fantastic job with Snowflake. Really good company before he got there, but since he's been there, the growth mindset, the discipline, the rigor and the phenomenon of what Snowflake has done has forced all these bigger companies to really accelerate what they're doing. And again, it's an example of how this intense competition makes all the different cloud vendors better and it provides enormous value to customers. Second thing I wanted to mention here was look at the Adam Selipsky effect at AWS, took over in the middle of May, and in Q2, Q3, Q4, AWS's growth rate accelerated. And in each of those three quotas, they grew faster than Microsoft's cloud, which has not happened in two or three years, so they're closing the gap on Microsoft. The third thing, Dave, in this, you know, incredibly intense competitive nature here, look at Larry Ellison, right? He's got his, you know, the product that for the last two or three years, he said, "It's going to help determine the future of the company, autonomous database." You would think he's the last person in the world who's going to bring in, you know, in some ways another database to think about there, but he has put, you know, his whole effort and energy behind this. The investments Oracle's made, he's riding this horse really hard. So it's not just a technology achievement, but it's also an investment priority for Oracle going forward. And I think it's going to form a lot of how they position themselves to this new breed of buyer with a new type of need and expectations from IT. So I just think the next two or three years are going to be fantastic for people who are lucky enough to get to do the sorts of things that we do. >> You know, it's a great point you made about AWS. Back in 2018 Q3, they were doing about 7.4 billion a quarter and they were growing in the mid forties. They dropped down to like 29% Q4, 2020, I'm looking at the data now. They popped back up last quarter, last reported quarter to 40%, that is 17.8 billion, so they more doubled and they accelerated their growth rate. (laughs) So maybe that pretends, people are concerned about Snowflake right now decelerating growth. You know, maybe that's going to be different. By the way, I think Snowflake has a different strategy, the whole data cloud thing, data sharing. They're not trying to necessarily take Oracle head on, which is going to make this next 10 years, really interesting. All right, we got to go, last question. 30 seconds or less, what can we expect from the future of data platforms? Matt, please start. >> I have to go first again? You're killing me, Dave. (laughing) In the next few years, I think you're going to see the major players continue to meet customers where they are, right. Every organization, every environment is, you know, kind of, we use these words bespoke in Snowflake, pardon the pun, but Snowflakes, right. But you know, they're all opinionated and unique and what's great as an IT person is, you know, there is a service for me regardless of where I am on my journey, in my data management journey. I think you're going to continue to see with regards specifically to Oracle, I think you're going to see the company continue along this path of being all things to all people, if you will, or all organizations without sacrificing, you know, kind of richness of features and sacrificing who they are, right. Look, they are the data kings, right? I mean, they've been a database leader for an awful long time. I don't see that going away any time soon and I love the innovative spirit they've brought in with HeatWave. >> All right, great thank you. Okay, 30 seconds, Holgar go. >> Yeah, I mean, the interesting thing that we see is really that trend to autonomous as Oracle calls or self-driving software, right? So the database will have to do more things than just store the data and support the DVA. It will have to show it can wide insights, the whole upside, it will be able to show to one machine learning. We haven't really talked about that. How in just exciting what kind of use case we can get of machine learning running real time on data as it changes, right? So, which is part of the E5 announcement, right? So we'll see more of that self-driving nature in the database space. And because you said we can promote it, right. Check out my report about HeatWave latest release where I post in oracle.com. >> Great, thank you for that. And Bob Evans, please. You're great at quick hits, hit us. >> Dave, thanks. I really enjoyed getting to hear everybody's opinion here today and I think what's going to happen too. I think there's a new generation of buyers, a new set of CXO influencers in here. And I think what Oracle's done with this, MySQL HeatWave, those benchmarks that Ron talked about so eloquently here that is going to become something that forces other companies, not just try to get incrementally better. I think we're going to see a massive new wave of innovation to try to play catch up. So I really take my hat off to Oracle's achievement from going to, push everybody to be better. >> Excellent. Marc Staimer, what do you say? >> Sure, I'm going to leverage off of something Matt said earlier, "Those companies that are going to develop faster, cheaper, simpler products that are going to solve customer problems, IT problems are the ones that are going to succeed, or the ones who are going to grow. The one who are just focused on the technology are going to fall by the wayside." So those who can solve more problems, do it more elegantly and do it for less money are going to do great. So Oracle's going down that path today, Snowflake's going down that path. They're trying to do more integration with third party, but as a result, aiming at that simpler, faster, cheaper mentality is where you're going to continue to see this market go. >> Amen brother Marc. >> Thank you, Ron Westfall, we'll give you the last word, bring us home. >> Well, thank you. And I'm loving it. I see a wave of innovation across the entire cloud database ecosystem and Oracle is fueling it. We are seeing it, with the native integration of auto ML capabilities, elastic scaling, lower entry price points, et cetera. And this is just going to be great news for buyers, but also developers and increased use of open APIs. And so I think that is really the key takeaways. Just we're going to see a lot of great innovation on the horizon here. >> Guys, fantastic insights, one of the best power panel as I've ever done. Love to have you back. Thanks so much for coming on today. >> Great job, Dave, thank you. >> All right, and thank you for watching. This is Dave Vellante for theCube and we'll see you next time. (soft music)

Published Date : Mar 31 2022

SUMMARY :

and co-founder of the and then you answer And don't forget Sybase back in the day, the world these days? and others happening in the cloud, and you cover the competition, and Oracle and you know, whoever else. Mr. Staimer, how do you see things? in that I see the database some good meat on the bone Take away the database, That is the ability to scale on demand, and they got MySQL and you I think it's, you know, and the various momentums, and Microsoft right now at the moment. So where do you place your bets? And to what Bob and Holgar said, you know, and you know, very granular, and everything in the cloud market. And to what you were saying, you know, functionality that you can't get to you know, business consultant. you know, it's funny. and all of the TPC benchmarks, By the way, you know, and you know, just inside of that was of some of the data that they shared. the stack, you have the suite, and they're giving you the best of both. of the suite vendor, and you always get the ah In the data center Marc all the time And the other thing I wanted to talk about and then we're going to run 'em and all the infrastructure around that, Due to the nature of the competition, I think you guys all saw the Andreessen, And I think it's going to form I'm looking at the data now. and I love the innovative All right, great thank you. and support the DVA. Great, thank you for that. And I think what Oracle's done Marc Staimer, what do you say? or the ones who are going to grow. we'll give you the last And this is just going to Love to have you back. and we'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
David FloyerPERSON

0.99+

Dave VellantePERSON

0.99+

Ron WestfallPERSON

0.99+

DavePERSON

0.99+

Marc StaimerPERSON

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

MarcPERSON

0.99+

EllisonPERSON

0.99+

Bob EvansPERSON

0.99+

OracleORGANIZATION

0.99+

MattPERSON

0.99+

Holgar MuellerPERSON

0.99+

AWSORGANIZATION

0.99+

Frank SlootmanPERSON

0.99+

RonPERSON

0.99+

StaimerPERSON

0.99+

Andy JacksonPERSON

0.99+

BobPERSON

0.99+

Matt KimballPERSON

0.99+

GoogleORGANIZATION

0.99+

100%QUANTITY

0.99+

Sarah WangPERSON

0.99+

San DiegoLOCATION

0.99+

AmazonORGANIZATION

0.99+

RobPERSON

0.99+

Breaking Analysis: Chasing Snowflake in Database Boomtown


 

(upbeat music) >> From theCUBE studios in Palo Alto, in Boston bringing you data-driven insights from theCUBE and ETR. This is braking analysis with Dave Vellante. >> Database is the heart of enterprise computing. The market is both exploding and it's evolving. The major force is transforming the space include Cloud and data, of course, but also new workloads, advanced memory and IO capabilities, new processor types, a massive push towards simplicity, new data sharing and governance models, and a spate of venture investment. Snowflake stands out as the gold standard for operational excellence and go to market execution. The company has attracted the attention of customers, investors, and competitors and everyone from entrenched players to upstarts once in the act. Hello everyone and welcome to this week's Wikibon CUBE Insights powered by ETR. In this breaking analysis, we'll share our most current thinking on the database marketplace and dig into Snowflake's execution. Some of its challenges and we'll take a look at how others are making moves to solve customer problems and try to get a piece of the growing database pie. Let's look at some of the factors that are driving market momentum. First, customers want lower license costs. They want simplicity. They want to avoid database sprawl. They want to run anywhere and manage new data types. These needs often are divergent and they pull vendors and technologies in different direction. It's really hard for any one platform to accommodate every customer need. The market is large and it's growing. Gardner has it at around 60 to 65 billion with a CAGR of somewhere around 20% over the next five years. But the market, as we know it is being redefined. Traditionally, databases have served two broad use cases, OLTP or transactions and reporting like data warehouses. But a diversity of workloads and new architectures and innovations have given rise to a number of new types of databases to accommodate all these diverse customer needs. Many billions have been spent over the last several years in venture money and it continues to pour in. Let me just give you some examples. Snowflake prior to its IPO, raised around 1.4 billion. Redis Labs has raised more than 1/2 billion dollars so far, Cockroach Labs, more than 350 million, Couchbase, 250 million, SingleStore formerly MemSQL, 238 million, Yellowbrick Data, 173 million. And if you stretch the definition of database a little bit to including low-code or no-code, Airtable has raised more than 600 million. And that's by no means a complete list. Now, why is all this investment happening? Well, in a large part, it's due to the TAM. The TAM is huge and it's growing and it's being redefined. Just how big is this market? Let's take a look at a chart that we've shown previously. We use this chart to Snowflakes TAM, and it focuses mainly on the analytics piece, but we'll use it here to really underscore the market potential. So the actual database TAM is larger than this, we think. Cloud and Cloud-native technologies have changed the way we think about databases. Virtually 100% of the database players that they're are in the market have pivoted to a Cloud first strategy. And many like Snowflake, they're pretty dogmatic and have a Cloud only strategy. Databases has historically been very difficult to manage, they're really sensitive to latency. So that means they require a lot of tuning. Cloud allows you to throw virtually infinite resources on demand and attack performance problems and scale very quickly, minimizing the complexity and tuning nuances. This idea, this layer of data as a service we think of it as a staple of digital transformation. Is this layer that's forming to support things like data sharing across ecosystems and the ability to build data products or data services. It's a fundamental value proposition of Snowflake and one of the most important aspects of its offering. Snowflake tracks a metric called edges, which are external connections in its data Cloud. And it claims that 15% of its total shared connections are edges and that's growing at 33% quarter on quarter. This notion of data sharing is changing the way people think about data. We use terms like data as an asset. This is the language of the 2010s. We don't share our assets with others, do we? No, we protect them, we secure or them, we even hide them. But we absolutely don't want to share those assets but we do want to share our data. I had a conversation recently with Forrester analyst, Michelle Goetz. And we both agreed we're going to scrub data as an asset from our phrasiology. Increasingly, people are looking at sharing as a way to create, as I said, data products or data services, which can be monetized. This is an underpinning of Zhamak Dehghani's concept of a data mesh, make data discoverable, shareable and securely governed so that we can build data products and data services that can be monetized. This is where the TAM just explodes and the market is redefining. And we think is in the hundreds of billions of dollars. Let's talk a little bit about the diversity of offerings in the marketplace. Again, databases used to be either transactional or analytic. The bottom lines and top lines. And this chart here describe those two but the types of databases, you can see the middle of mushrooms, just looking at this list, blockchain is of course a specialized type of database and it's also finding its way into other database platforms. Oracle is notable here. Document databases that support JSON and graph data stores that assist in visualizing data, inference from multiple different sources. That's is one of the ways in which adtech has taken off and been so effective. Key Value stores, log databases that are purpose-built, machine learning to enhance insights, spatial databases to help build the next generation of products, the next automobile, streaming databases to manage real time data flows and time series databases. We might've missed a few, let us know if you think we have, but this is a kind of pretty comprehensive list that is somewhat mind boggling when you think about it. And these unique requirements, they've spawned tons of innovation and companies. Here's a small subset on this logo slide. And this is by no means an exhaustive list, but you have these companies here which have been around forever like Oracle and IBM and Teradata and Microsoft, these are the kind of the tier one relational databases that have matured over the years. And they've got properties like atomicity, consistency, isolation, durability, what's known as ACID properties, ACID compliance. Some others that you may or may not be familiar with, Yellowbrick Data, we talked about them earlier. It's going after the best price, performance and analytics and optimizing to take advantage of both hybrid installations and the latest hardware innovations. SingleStore, as I said, formerly known as MemSQL is a very high end analytics and transaction database, supports mixed workloads, extremely high speeds. We're talking about trillions of rows per second that could be ingested in query. Couchbase with hybrid transactions and analytics, Redis Labs, open source, no SQL doing very well, as is Cockroach with distributed SQL, MariaDB with its managed MySQL, Mongo and document database has a lot of momentum, EDB, which supports open source Postgres. And if you stretch the definition a bit, Splunk, for log database, why not? ChaosSearch, really interesting startup that leaves data in S-3 and is going after simplifying the ELK stack, New Relic, they have a purpose-built database for application performance management and we probably could have even put Workday in the mix as it developed a specialized database for its apps. Of course, we can't forget about SAP with how not trying to pry customers off of Oracle. And then the big three Cloud players, AWS, Microsoft and Google with extremely large portfolios of database offerings. The spectrum of products in this space is very wide, with you've got AWS, which I think we're up to like 16 database offerings, all the way to Oracle, which has like one database to do everything not withstanding MySQL because it owns MySQL got that through the Sun Acquisition. And it recently, it made some innovations there around the heat wave announcement. But essentially Oracle is investing to make its database, Oracle database run any workload. While AWS takes the approach of the right tool for the right job and really focuses on the primitives for each database. A lot of ways to skin a cat in this enormous and strategic market. So let's take a look at the spending data for the names that make it into the ETR survey. Not everybody we just mentioned will be represented because they may not have quite the market presence of the ends in the survey, but ETR that capture a pretty nice mix of players. So this chart here, it's one of the favorite views that we like to share quite often. It shows the database players across the 1500 respondents in the ETR survey this past quarter and it measures their net score. That's spending momentum and is shown on the vertical axis and market share, which is the pervasiveness in the data set is on the horizontal axis. The Snowflake is notable because it's been hovering around 80% net score since the survey started picking them up. Anything above 40%, that red line there, is considered by us to be elevated. Microsoft and AWS, they also stand out because they have both market presence and they have spending velocity with their platforms. Oracle is very large but it doesn't have the spending momentum in the survey because nearly 30% of Oracle installations are spending less, whereas only 22% are spending more. Now as a caution, this survey doesn't measure dollar spent and Oracle will be skewed toward the big customers with big budgets. So you got to consider that caveat when evaluating this data. IBM is in a similar position although its market share is not keeping up with Oracle's. Google, they've got great tech especially with BigQuery and it has elevated momentum. So not a bad spot to be in although I'm sure it would like to be closer to AWS and Microsoft on the horizontal axis, so it's got some work to do there. And some of the others we mentioned earlier, like MemSQL, Couchbase. As shown MemSQL here, they're now SingleStore. Couchbase, Reddis, Mongo, MariaDB, all very solid scores on the vertical axis. Cloudera just announced that it was selling to private equity and that will hopefully give it some time to invest in this platform and get off the quarterly shot clock. MapR was acquired by HPE and it's part of HPE's Ezmeral platform, their data platform which doesn't yet have the market presence in the survey. Now, something that is interesting in looking at in Snowflakes earnings last quarter, is this laser focused on large customers. This is a hallmark of Frank Slootman and Mike Scarpelli who I know they don't have a playbook but they certainly know how to go whale hunting. So this chart isolates the data that we just showed you to the global 1000. Note that both AWS and Snowflake go up higher on the X-axis meaning large customers are spending at a faster rate for these two companies. The previous chart had an end of 161 for Snowflake, and a 77% net score. This chart shows the global 1000, in the end there for Snowflake is 48 accounts and the net score jumps to 85%. We're not going to show it here but when you isolate the ETR data, nice you can just cut it, when you isolate it on the fortune 1000, the end for Snowflake goes to 59 accounts in the data set and Snowflake jumps another 100 basis points in net score. When you cut the data by the fortune 500, the Snowflake N goes to 40 accounts and the net score jumps another 200 basis points to 88%. And when you isolate on the fortune 100 accounts is only 18 there but it's still 18, their net score jumps to 89%, almost 90%. So it's very strong confirmation that there's a proportional relationship between larger accounts and spending momentum in the ETR data set. So Snowflakes large account strategy appears to be working. And because we think Snowflake is sticky, this probably is a good sign for the future. Now we've been talking about net score, it's a key measure in the ETR data set, so we'd like to just quickly remind you what that is and use Snowflake as an example. This wheel chart shows the components of net score, that lime green is new adoptions. 29% of the customers in the ETR dataset that are new to Snowflake. That's pretty impressive. 50% of the customers are spending more, that's the forest green, 20% are flat, that's the gray, and only 1%, the pink, are spending less. And 0% zero or replacing Snowflake, no defections. What you do here to get net scores, you subtract the red from the green and you get a net score of 78%. Which is pretty sick and has been sick as in good sick and has been steady for many, many quarters. So that's how the net score methodology works. And remember, it typically takes Snowflake customers many months like six to nine months to start consuming it's services at the contracted rate. So those 29% new adoptions, they're not going to kick into high gear until next year, so that bodes well for future revenue. Now, it's worth taking a quick snapshot at Snowflakes most recent quarter, there's plenty of stuff out there that you can you can google and get a summary but let's just do a quick rundown. The company's product revenue run rate is now at 856 million they'll surpass $1 billion on a run rate basis this year. The growth is off the charts very high net revenue retention. We've explained that before with Snowflakes consumption pricing model, they have to account for retention differently than what a SaaS company. Snowflake added 27 net new $1 million accounts in the quarter and claims to have more than a hundred now. It also is just getting its act together overseas. Slootman says he's personally going to spend more time in Europe, given his belief, that the market is huge and they can disrupt it and of course he's from the continent. He was born there and lived there and gross margins expanded, do in a large part to renegotiation of its Cloud costs. Welcome back to that in a moment. Snowflake it's also moving from a product led growth company to one that's more focused on core industries. Interestingly media and entertainment is one of the largest along with financial services and it's several others. To me, this is really interesting because Disney's example that Snowflake often puts in front of its customers as a reference. And it seems to me to be a perfect example of using data and analytics to both target customers and also build so-called data products through data sharing. Snowflake has to grow its ecosystem to live up to its lofty expectations and indications are that large SIS are leaning in big time. Deloitte cross the $100 million in deal flow in the quarter. And the balance sheet's looking good. Thank you very much with $5 billion in cash. The snarks are going to focus on the losses, but this is all about growth. This is a growth story. It's about customer acquisition, it's about adoption, it's about loyalty and it's about lifetime value. Now, as I said at the IPO, and I always say this to young people, don't buy a stock at the IPO. There's probably almost always going to be better buying opportunities ahead. I'm not always right about that, but I often am. Here's a chart of Snowflake's performance since IPO. And I have to say, it's held up pretty well. It's trading above its first day close and as predicted there were better opportunities than day one but if you have to make a call from here. I mean, don't take my stock advice, do your research. Snowflake they're priced to perfection. So any disappointment is going to be met with selling. You saw that the day after they beat their earnings last quarter because their guidance in revenue growth,. Wasn't in the triple digits, it sort of moderated down to the 80% range. And they pointed, they pointed to a new storage compression feature that will lower customer costs and consequently, it's going to lower their revenue. I swear, I think that that before earnings calls, Scarpelli sits back he's okay, what kind of creative way can I introduce the dampen enthusiasm for the guidance. Now I'm not saying lower storage costs will translate into lower revenue for a period of time. But look at dropping storage prices, customers are always going to buy more, that's the way the storage market works. And stuff like did allude to that in all fairness. Let me introduce something that people in Silicon Valley are talking about, and that is the Cloud paradox for SaaS companies. And what is that? I was a clubhouse room with Martin Casado of Andreessen when I first heard about this. He wrote an article with Sarah Wang, calling it to question the merits of SaaS companies sticking with Cloud at scale. Now the basic premise is that for startups in early stages of growth, the Cloud is a no brainer for SaaS companies, but at scale, the cost of Cloud, the Cloud bill approaches 50% of the cost of revenue, it becomes an albatross that stifles operating leverage. Their conclusion ended up saying that as much as perhaps as much as the back of the napkin, they admitted that, but perhaps as much as 1/2 a trillion dollars in market cap is being vacuumed away by the hyperscalers that could go to the SaaS providers as cost savings from repatriation. And that Cloud repatriation is an inevitable path for large SaaS companies at scale. I was particularly interested in this as I had recently put on a post on the Cloud repatriation myth. I think in this instance, there's some merit to their conclusions. But I don't think it necessarily bleeds into traditional enterprise settings. But for SaaS companies, maybe service now has it right running their own data centers or maybe a hybrid approach to hedge bets and save money down the road is prudent. What caught my attention in reading through some of the Snowflake docs, like the S-1 in its most recent 10-K were comments regarding long-term purchase commitments and non-cancelable contracts with Cloud companies. And the companies S-1, for example, there was disclosure of $247 million in purchase commitments over a five plus year period. And the company's latest 10-K report, that same line item jumped to 1.8 billion. Now Snowflake is clearly managing these costs as it alluded to when its earnings call. But one has to wonder, at some point, will Snowflake follow the example of say Dropbox which Andreessen used in his blog and start managing its own IT? Or will it stick with the Cloud and negotiate hard? Snowflake certainly has the leverage. It has to be one of Amazon's best partners and customers even though it competes aggressively with Redshift but on the earnings call, CFO Scarpelli said, that Snowflake was working on a new chip technology to dramatically increase performance. What the heck does that mean? Is this Snowflake is not becoming a hardware company? So I going to have to dig into that a little bit and find out what that it means. I'm guessing, it means that it's taking advantage of ARM-based processes like graviton, which many ISVs ar allowing their software to run on that lower cost platform. Or maybe there's some deep dark in the weeds secret going on inside Snowflake, but I doubt it. We're going to leave all that for there for now and keep following this trend. So it's clear just in summary that Snowflake they're the pace setter in this new exciting world of data but there's plenty of room for others. And they still have a lot to prove. For instance, one customer in ETR, CTO round table express skepticism that Snowflake will live up to its hype because its success is going to lead to more competition from well-established established players. This is a common theme you hear it all the time. It's pretty easy to reach that conclusion. But my guess is this the exact type of narrative that fuels Slootman and sucked him back into this game of Thrones. That's it for now, everybody. Remember, these episodes they're all available as podcasts, wherever you listen. All you got to do is search braking analysis podcast and please subscribe to series. Check out ETR his website at etr.plus. We also publish a full report every week on wikinbon.com and siliconangle.com. You can get in touch with me, Email is David.vellante@siliconangle.com. You can DM me at DVelante on Twitter or comment on our LinkedIn posts. This is Dave Vellante for theCUBE Insights powered by ETR. Have a great week everybody, be well and we'll see you next time. (upbeat music)

Published Date : Jun 5 2021

SUMMARY :

This is braking analysis and the net score jumps to 85%.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Michelle GoetzPERSON

0.99+

AWSORGANIZATION

0.99+

Mike ScarpelliPERSON

0.99+

Dave VellantePERSON

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

Sarah WangPERSON

0.99+

AmazonORGANIZATION

0.99+

50%QUANTITY

0.99+

Palo AltoLOCATION

0.99+

AndreessenPERSON

0.99+

EuropeLOCATION

0.99+

40 accountsQUANTITY

0.99+

$1 billionQUANTITY

0.99+

Frank SlootmanPERSON

0.99+

SlootmanPERSON

0.99+

OracleORGANIZATION

0.99+

Redis LabsORGANIZATION

0.99+

ScarpelliPERSON

0.99+

TAMORGANIZATION

0.99+

sixQUANTITY

0.99+

33%QUANTITY

0.99+

$5 billionQUANTITY

0.99+

80%QUANTITY

0.99+

GoogleORGANIZATION

0.99+

1.8 billionQUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

59 accountsQUANTITY

0.99+

Cockroach LabsORGANIZATION

0.99+

DisneyORGANIZATION

0.99+

TeradataORGANIZATION

0.99+

18QUANTITY

0.99+

77%QUANTITY

0.99+

85%QUANTITY

0.99+

29%QUANTITY

0.99+

20%QUANTITY

0.99+

BostonLOCATION

0.99+

78%QUANTITY

0.99+

Martin CasadoPERSON

0.99+

48 accountsQUANTITY

0.99+

856 millionQUANTITY

0.99+

1500 respondentsQUANTITY

0.99+

nine monthsQUANTITY

0.99+

Zhamak DehghaniPERSON

0.99+

0%QUANTITY

0.99+

wikinbon.comOTHER

0.99+

88%QUANTITY

0.99+

twoQUANTITY

0.99+

$100 millionQUANTITY

0.99+

89%QUANTITY

0.99+

AirtableORGANIZATION

0.99+

next yearDATE

0.99+

SnowflakeORGANIZATION

0.99+

two companiesQUANTITY

0.99+

DeloitteORGANIZATION

0.99+

200 basis pointsQUANTITY

0.99+

FirstQUANTITY

0.99+

HPEORGANIZATION

0.99+

15%QUANTITY

0.99+

more than 600 millionQUANTITY

0.99+

last quarterDATE

0.99+

161QUANTITY

0.99+

David.vellante@siliconangle.comOTHER

0.99+

$247 millionQUANTITY

0.99+

27 netQUANTITY

0.99+

2010sDATE

0.99+

siliconangle.comOTHER

0.99+

ForresterORGANIZATION

0.99+

MemSQLTITLE

0.99+

Yellowbrick DataORGANIZATION

0.99+

more than 1/2 billion dollarsQUANTITY

0.99+

DropboxORGANIZATION

0.99+

MySQLTITLE

0.99+

BigQueryTITLE

0.99+

Thenu Kittappa, Anand Akela & Tajeshwar Singh | Introducing a New Era in Database Management


 

>>from around the globe. It's the Cube with digital coverage of a new era and database management brought to you by Nutanix. >>Welcome back. I'm still minimum and we're covering Nutanix Is New Era database launch Of course, we had to do instead of conversation with Monica Ambala talking about era to Dato and to dig into it a little bit further. We have some new tennis guests as well as what? One of their close partners. So going across the channel, first of all, happy to welcome to the program. Uh, the new kid UPA she is the gsc strategy and go to market with Nutanix sitting in the middle chair we have on and Akila whose product marketing leader with Nutanix and then from HCL happy to welcome to the program Tasing who is the senior vice president with HCL Technologies. I mentioned all three of you. Thank you so much for joining us. >>Glad to be here, >>right? Uh they knew What? Why don't we start with you? You handle the relationship between Nutanix and HCL. As I said, some exciting announcements database services help us understand how Ah partner like HCL takes the technology and what will help bring it to market. >>Let me start by thanking used to for this opportunity. Head Seal is a very significant partner for Nutanix and we've had this partnership for a long time now. It's one of our long standing partnership. Over the five years we've closed over 100 accounts across all three theaters. Trained professionals both on the Nutanix side on the outside, on built a 3 60 relationships so we can deliver the best experience around solutions to our partners. In the very recent announcement, we're looking to build a database as a service offering. With that CL we want Thio leverage are intelligent technology that allows us to simplify off and increase operating efficiency. Andi Couple it with head seals ability to offer world class services on it. It's a scale to reach the go to market needs needed right. We're very confident that the solution is going to drive significant incremental business for both our companies. >>Excellent taste. We would love to hear from your standpoint. What is it that excites you? We we know HCL knows the data space real well. So I think you've got some customers that air looking to take advantage of some of these new offerings. >>Yeah, So if you look at where the focus has been so far, most of the focus is on taking applications to cloud and moving them from VM two probably containers one of the most. Uh, I won't say, uh, neglected, but the space that needs to change now is the entire database space on. If you look at how customers are managing databases today, they have taken hardware on a KPIX model. They have the operating system and the database licenses on L. A model from the E. M s on. Then they have, ah, teams which are siloed depending upon the database technology that is there in the environment and managing that I think that whole model is has to change, enabling customers to transform Onda accelerate the digital transformation journey on. That is where our offering off database as a service ises very unique because it offers a full stack off services which includes right from hardware and all the way to operations on a completely utility model powered by the Nutanix era. >>Yeah, on it might make sense if you could give us a little bit of a broader context for your users. Some of the data that you have around this offering, >>yeah, you know, attend effect. All the solution, our joint solutions. Our customers, uh, they are trying to deliver the best individual experience, right? That's at the heart of it. What they're trying to do, I'll give you a couple of customer examples. For example, Arbil Bank in India. You know, they deployed their database solutions and applications, and Nutanix got 16 fasters application response. That means like they used to take 180 seconds. Uh, Thio logging into the application. And now it's, uh, 20 seconds, 36 times faster. Another example I could give. I can give many examples, but when this one is really interesting, Delaware Valley community held, you know, at the time of Kobe they went remote. They started working from home and they had medical systems applications. EMR electronic medical record applications and used to take even before they were working from home, is take like 171 seconds to log into medical systems before they could, you know, talk to their patients and look at their, you know, health results and everything and that from 171 seconds, it went to 19 seconds. So these are some of the values that customers seeing when it comes to delivering the individual experience to their customers. >>Yeah, absolutely. We've seen police stage go ahead. >>Yeah, and I just had to What men? Who said that? It's also the ability tohave self service with dynamic provisioning capability that really brings the value toe the to the I T teams and to the application teams who are consuming these services. So we have cases where customers were waiting for about a week, 10 days for the environments to be provisioned to them. And now it's a matter of seconds or minutes where they can have a full fledged environments leading to develop a productivity. And that also really adds the whole acceleration that we just spoke about. >>Yeah, we we've absolutely seen such a transformation in database for the longest time. It was, you know, a database. It didn't change too much. That's what everything run on Now there's a lot of flexibility. Open source is a big piece of what's going on there. I'd like to come back to you and you know, they know. I know you're gonna want to chime in here. You know, HCL doesn't just, you know, take this off the shelf and, you know, resell it, help us understand. You know what is unique about the offering that that HCL brings market? Uh, with with >>Nutanix. Right. So one is that we have standardized reference architectures, which really x ray the time to consume the offering. We're not building anything from from from ground up. Three Nutanix is also part off our velocity framework, which helps customers deploy software defined infrastructure as the as a foundation element for their for their private cloud. Now, what is unique is also the ability toe not only provide operations on different databases that are there in the environment on a completely utility model, but also help customers, you know, move to cloud and adopt the database clouded of databases and then manage the whole show seamlessly using using the BP platform and that really, you know, if you look at the trend that is there, there's a short term impact on the long term impact off transformation. In the short term, there's hardly an industry which is not touched by by covert on most of our customers are either looking at cost or initiatives or are looking at ah platform, which will help them in a weight or find new business model to to sail through. In the long term, we strongly believe that the customers will be in a hybrid, multi cloud world where they will still have the heritage environments. The article and the Sequels on a lot off cloud native data business will also start coming into picture. How do you manage is also seamlessly is what will be the next challenge for for most of the customers. And that's where we come in, along with Nutanix, to solve the problem. >>Well, very simply put right, we have different categories of customers. One off them refers to buy the ingredients and make their own meal on some really large customers, and global customers prefer to buy the meal and pay for it on on as consumer basis. What that seal does is take era, which simplifies a lot of the database operations, puts it into a full stack solution and gives the customer the full stack solution. Everything from assessing that environment to deploying, to making sure that the designers I accurate and then of course, the day and through they do through and, uh, uh, environment, right. So literally the customer can Now I'll offload any off their data center, our database management and operation to hit cl from my perspective on do rest assured, run their projects toe, etc. Also, excel becomes their extended arm, the beauty off. It is also like working with dead C. Elgar now able to offer the entire solution on a pay as you go model or pay as you use model, which is very relevant to the existing times where everybody is trying to cut their Catholics costs and and optimized on the utilization. >>Well, great. Great to hear about that. You've mentioned that this partnership has been for many years, so I know you've got plenty of joint customers. Anything specifically could share about these new offerings on. And I know you've got a lot of the customer stories there. Maybe you could start would look love, freedom. The rest of you, >>Thio, I'll start what? You know, Like I talked about a couple of customers. But recently I'm really excited about. And this is something that to be a announcing today as well. Ah, study that we did with Forrester called Forrester T I study, which is what it means total economic impact study. And what they do is that they topped with customers, uh, interviewed them, four of them. And based on their experience, uh, you know what? They observe what kind of benefit they got, what challenges they had, what was cause they built an economic model. And based on that economic model, they found that customers were rolled all off them were able to get their payback within six months. So Bala talked about it earlier that, you know, like all the great experience, all the great value that we offer, but at a very, very good cost. So the six less than six months payback was used and the r y for the three years period and again, this is ah, model based on four enterprises was 2 91 100% almost like three times mawr. So whatever they invested, I think on an average day the cost was 2.3 million and the benefit was nine million or so so huge value customers have observed already. And with this new launch, I believe that it will just go to the next level. All the things about provisioning copy data saving that the stories All of that adds to the R Y that I'm talking about and our joint customers with SCL or otherwise, who are customers who are running their applications, their business critical applications on you can X Platform managed by era an era is built out off a bunch off best practices that over time that we have done. I talked about custom performance earlier, and a lot of the performance comes from fine tuning. You do that like a lot of tea tuning and to get to the right kind of performance. Uh, era comes with that, those best practices. So when your provisioning an application, you know, it gives you you don't have to do all that tuning. So that's the value customers are experiencing. And I'm really excited about the joint customers what they could experience and benefit out off the new expanded solution. >>Great Tiger. Any other customer examples that you'd like to share? >>Well, we got a lot of go ahead page, >>but it's okay. >>No, I was just saying that we've had a lot of success with Head cl across the board anywhere from data center organization Thio v. D. I. We had a very large manufacturing company in America where we partner together. They have a huge number of sub brands. We partnered together to go evaluate that environment and then also even that is a B infrastructure with databases. It's a relatively new offering we're announcing today. But we're leveraging the expertise that SCL has in the market, uh, to go to go deeper into that market with cl eso. I will leave it to page to give us the NCL examples. >>So one thing that is happening is the very definition off infrastructure and infrastructure operation itself is changing. So a couple of years ago, for many of our customers, it was about operating system management, hardware management, network management and all the use. Uh, the concept that you're going back to customer is about platform operations. That means everything to do with application operations. Downward is going to be done by one integrated unit. Now, with Nutanix, we can we can really bring a lot of change, and we're bringing a lot of change in our in the operations model for for lot off a large customers where earlier you had siloed teams around Compute network storage, offering system databases both at the Level two and level three, and you had a level one, which was basically command center. Now, we're saying is that with the artificial intelligence and machine learning driven OBS, you can practically eliminate the need for command center on the level two layer because the platform enables you toe be multi skilled. You need not have siloed engineers looking after databases separately on and operating system separately. You can have the same sort of people who are cross train, multi skilled, looking at the entire state. On at level three. You may want to keep people who are deep into databases as a separate team, then from people who are managing the Nutanix platform, which is a combination off compute storage and and and and the SCN. So that's the change that we're bringing. A lot of our customers were going about infrastructure, platform modernization, Azaz, the public cloud or hybrid clubs. >>Well, I think you're really articulated well, that modernization journey we've seen so many companies going through. The thing I've been saying with Nutanix for years is modernize the platform, then you can modernize everything that runs on top of it. All the applications on, of course, did databases a major piece of this on. And that brings up a point I want to get your take on. We haven't talked about developers, you know, the DEV ops trend. Something we've seen, you know, huge growth for for a number of years. So what >>does this >>mean from developers? This something that you know, mostly the infrastructure team's gonna handle. Or how do you bridge that gap to the people that really are? You know, building and building and building the APS. >>Yeah. And in this digital world, you know the cycle time from idea to production. Everyone is trying to reduce that. What that means is that things are moving left. People are trying to develop and test early in the life cycle when it is easy to find a problem and easy to and cheaper to fix. Right. So for that, you need a your application environment, your application and database available to test and develop in, uh, you know, like in volume. And that's where databases the service era helps developers and develops professionals to provision in the whole infrastructure for testing and involvement in hundreds and thousands of them at the same time without, you know, worrying about the storage back back and how much story it is consuming. So it is. It helps developers to to really expedite their development and testing left lifecycle ultimately resulting in excellent and unique experience. >>Yeah, absolutely way no. Of just moving faster. Being able to respond to the business so critically important. Uh, they know Tasia wanna let you have the final word Talk about the partnership and what we should expect, you know, in the coming months and quarters. >>So, uh, I'll go first. And then we can come in, uh, a salon and Nutanix you to share the same values where we believe that we need to provide a very innovative platform for our customers to accelerate their digital transformation journey. No matter what it is right, we share common values and way have a 3 60 degree relationship. It started way back in 2015 and we have come a long way since then. A C also does engineering services for for Nutanix, and we have closed about 850 r plus people who has prayed and 35 on Nutanix Solutions. Providing manage services to our customers on Nutanix is also part off our software defined infrastructure portfolio on we're taking it to our customers as part of our entire infrastructure platform modernization that, I suppose talk about earlier three recent announcement off Nutanix clusters running on AWS. I think it's a significant announcement and it will provide a lot off options to our customers. And as an S, I, uh, you know, we are able to bring a lot of value to our customers. We're looking at adopting cloud the database as a service offering. I think we're very excited about it. I I think we have about 300 plus customers, and many of them are still stuck with the way they are managing databases the old way. And we can bring in a lot of value to those customers, whether it is about reducing cars or increasing agility or helping them modern ice, The platform one ended up hybrid multi club >>business critical lapse are growing, are still growing, and data is pretty much gold in these scenarios, right? It's it's doubling every two years, if not more with every transaction being remote today with zeal. We actually look forward to addressing that market and optimizing the environment for our customers. Both of our companies believe in partnership crossed and the customer first mindset. And when you have that belief, trust comes with delivering the best experience to our customers. So we're looking forward to this partnership and you're looking forward to growing our joint revenue and modernizing our customers platforms with this often? >>Well, I wanna thank all three of you for for sharing the exciting news. Absolutely. It looks like a strong partnership. Lots of potential there for the future. So thank you so much for joining us. Thank you for >>having thank you. Mhm. >>All right, when I think the audience were watching this lot with Nutanix, the new era in database management personally, a big thank you to the Nutanix community has been a pleasure being able to host these interviews with Nutanix for for many years. So I'm still minimum and thank you as always for watching the Cube

Published Date : Oct 6 2020

SUMMARY :

coverage of a new era and database management brought to you by Nutanix. and go to market with Nutanix sitting in the middle chair we have on and Ah partner like HCL takes the technology and what will help bring it to the solution is going to drive significant incremental business for both our companies. What is it that excites you? most of the focus is on taking applications to cloud and moving them from VM two probably containers Some of the data that you have around this offering, before they could, you know, talk to their patients and look at their, Yeah, absolutely. And that also really adds the whole acceleration that we just spoke about. I'd like to come back to you and you know, and that really, you know, if you look at the trend that is there, there's a short term impact C. Elgar now able to offer the entire solution on a pay as you go model Maybe you could start would look love, of that adds to the R Y that I'm talking about and our joint customers with SCL Any other customer examples that you'd like to share? to go to go deeper into that market with cl eso. both at the Level two and level three, and you had a level one, which was basically command center. We haven't talked about developers, you know, the DEV ops trend. This something that you know, mostly the infrastructure team's gonna handle. at the same time without, you know, worrying about the storage back and what we should expect, you know, in the coming months and quarters. And as an S, I, uh, you know, we are able to bring a lot of value to our customers. Both of our companies believe in partnership crossed and the customer first mindset. So thank you so much for joining having thank you. So I'm still minimum and thank you as always

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NutanixORGANIZATION

0.99+

HCLORGANIZATION

0.99+

Monica AmbalaPERSON

0.99+

ForresterORGANIZATION

0.99+

10 daysQUANTITY

0.99+

2.3 millionQUANTITY

0.99+

AmericaLOCATION

0.99+

Arbil BankORGANIZATION

0.99+

nine millionQUANTITY

0.99+

IndiaLOCATION

0.99+

36 timesQUANTITY

0.99+

20 secondsQUANTITY

0.99+

35QUANTITY

0.99+

19 secondsQUANTITY

0.99+

BothQUANTITY

0.99+

171 secondsQUANTITY

0.99+

180 secondsQUANTITY

0.99+

AWSORGANIZATION

0.99+

171 secondsQUANTITY

0.99+

Tajeshwar SinghPERSON

0.99+

hundredsQUANTITY

0.99+

three yearsQUANTITY

0.99+

2015DATE

0.99+

todayDATE

0.99+

BalaPERSON

0.99+

Anand AkelaPERSON

0.99+

HCL TechnologiesORGANIZATION

0.99+

bothQUANTITY

0.99+

about 300 plus customersQUANTITY

0.98+

oneQUANTITY

0.98+

Delaware ValleyLOCATION

0.98+

3 60 relationshipsQUANTITY

0.98+

16 fastersQUANTITY

0.98+

six monthsQUANTITY

0.98+

threeQUANTITY

0.97+

AzazORGANIZATION

0.97+

over 100 accountsQUANTITY

0.97+

AkilaPERSON

0.97+

ThioPERSON

0.96+

KobeORGANIZATION

0.96+

C. ElgarPERSON

0.96+

OneQUANTITY

0.96+

firstQUANTITY

0.96+

2 91 100%QUANTITY

0.96+

fourQUANTITY

0.95+

NutanixTITLE

0.95+

six less than six monthsQUANTITY

0.94+

SCLORGANIZATION

0.94+

excelTITLE

0.94+

one thingQUANTITY

0.93+

AndiPERSON

0.93+

about a weekQUANTITY

0.92+

one integrated unitQUANTITY

0.91+

couple of years agoDATE

0.91+

level threeQUANTITY

0.9+

3 60 degreeQUANTITY

0.9+

five yearsQUANTITY

0.9+

CatholicsORGANIZATION

0.9+

level oneQUANTITY

0.87+

about 850 r plus peopleQUANTITY

0.86+

HCLTITLE

0.8+

every two yearsQUANTITY

0.78+

three timesQUANTITY

0.77+

level two layerQUANTITY

0.76+

Monica Kumar & Bala Kuchibhotla, Nutanix | Introducing a New Era in Database Management


 

>> Narrator: From around the globe. It's theCUBE with digital coverage of A New Era In Database Management. Brought to you by Nutanix. >> Hi, I'm Stu Miniman. And welcome to this special presentation with Nutanix. We're talking about A New Era In Database Management. To help us dig into it, first of all, I have the Senior Vice President and General Manager of Nutanix Era Databases and Business Critical Applications, that is Bala Kuchibhotla. And one of our other CUBE alongs, Monica Kumar. Who's an SVP also with Nutanix. Bala, Monica, thank you so much for joining us. >> Thank you, thank you so... >> Great to be here. All right, so first of all, Bala a new Era. We, have a little bit of a punj. You've got me with some punjs there. Of course we know that the database for Nutanix solution is Era. So, we always like to bring out the news first. Why don't you tell us, what does this mean? What is Nutanix announcing today? >> Awesome. Thank you, Stu. Yeah, so today's a very big day for us. I'm super excited to inform all of us and our audience that we are announcing the Eratory dot two GA bits for customers to enjoy it. Some customers can download and start playing with it. So what's new with Nutanix Eratory dot two? As you knows 1.0 is a single cluster solution meaning the customers have to have a Nutanix cluster and then have around the same cluster to enjoy the databases. But with Eratory dot two, it becomes multi-cluster solution. It's not just a multi-cluster solution, but customers can enjoy database across clusters, That means that they can have their Always On Availability Groups SQL servers, their Postgres servers across Nutanix clusters. That means that they can spread across Azure Availability Zones. Now, the most interesting point of this is, it's not just across clusters, customers can place these clusters in the cloud. That is AWS. You can have Nutanix cluster in the AWS cluster and then the primary production clusters maybe on the Nutanix and primary enterprise cloud kind of stuff, that's number one. Number two, we have extended our data management capabilities, data management platform capabilities, and what we call them as global time mission. Global time mission with a data access management. Like racing river, that you need to harness the racing river by constructing a dam and then harness it for multipurpose either irrigation projects or hydroelectric project kind of stuff. You need to kind of do the similar things for your data in company, enterprise company. You need to make sure that the right persons get the right amount of data, so that you don't kind of give all production data to everyone in the company. At the same time, they also need the accessible, with one click they can get the database, the data they want. So that's the data access management. Imagine a QA person only gets the sanitized snapshots or sanitize database backups for them to create the copies. And then we are extending our database engine portfolios too to introduce SAP HANA to the thing. As you know, that we support Oracle today, Postgres, MalSQL, Mariadb SQL server. I'm excited to inform that we are introducing SAP HANA. Our customers can do one click sandbox creation into an environment for SAP HANA predown intense platform. And lastly, I'm super excited to inform that we are becoming a Postgres vendor. We are willing to give 24 by seven, 365 day support but Postgres database engine, that's kind of a provision through Nutanix setup platform. So this way the customers can enjoy the engine, platform, service all together in one single shot with a single 180 company that they can call and get the support they want. I'm super duper excited that this is going to make the customers a truly multicloud multi cluster data management platform. Thank you. >> Yeah. And I'll just add to that too. It's fantastic that we are now offering this new capability. I just want to kind of remind our audience that Nutanix for many years has been providing the foundation the infrastructure software, where you can run all these multiple workloads including databases today. And what we're doing with Era is fantastic because now they are giving our customers the ability to take that database that they run on top of Nutanix to provide that as a service now. So now are talking to a whole different organization here. It's database administrations, it's administrators, it's teams that run databases, it teams that care about data and providing access to data and organizations. >> Well, first of all, congratulations, I've taught for a couple of years to the teams at Nutanix especially some of the people working on PostgreSQL really exciting stuff and you've both seen really the unlocking of database. It used to be ,we talked about, I have one database it's kind of the one that everything runs on. Now, customers they have more databases. You talked about that flexibility is then, where we run it. We'd love to hear, maybe Monica we start with you. You talk about the customers, what does this really mean for them? Because one of our most mission critical applications we talk about, we're not just throwing our databases or what. I don't wake up in the morning and say, Oh let me move it to this cloud and put it in this data center. This needs to be reliable. I need to have access to the data. I need to be able to work with it. So, what does this really mean? And what does it unlock for your customers? >> Yes absolutely, I love to talk about this topic. I mean, if you think about databases, they are means to an end. And in this case, the end is being able to mine insights from the data and then make meaningful decisions based on that. So when we talk to customers, it's really clear that data has not become one of the most valuable assets that an organization owns. Well, of course, in addition to the employees that are part of the organization and our customers. Data is one of the most important assets. But most organizations, the challenges they face is a lot of data gets collected. And in fact, we've heard numbers thrown around for many years like, almost 80% of world's data has been created in the last like three or four years. And data is doubling every two years in terms of volume. Well guess what? Data gets collected. It sits there and organizations are struggling to get access to it with the right performance, the right security and regulation compliance, the reliability, availability, by persona, developers need certain access, analysts needs different access line of businesses need different access. So what we see is organizations are struggling in getting access to data at the right time by the right person on the team and when they need it. And I think that's where database as a service is critical. It's not just about having the database software which is of course important but how you know not make that service available to your stakeholders, to developers to lines of business within the SLAs that they demand. So is it instantly? How quickly can you make it available? How quickly can you use have access to data and do something meaningful with it? And mind the insights for smarter business? And then the one thing I'd like to add is that's where IT and business really come together. That's the glue. If you think about it today, what is the blue between an IT Organization and a business organization? It's the data. And that's where they're really coming together to say how can we together deliver the right service? So you, the business owner can deliver the right outcome for our business. >> That's very true. Maybe I'll just add a couple of comments there. What we're trying to do is we are trying to bring the cloud experience, the RDS-like experience to the enterprise cloud and then hybrid cloud. So the customers will now have a choice of cloud. They don't need to be locked in a particular cloud, at the same time enjoy the true cloud utility experience. We help customers create clouds, database clouds either by themselves if that's big enough to manage the cloud themselves or they can partner with a GSIs like Wipro, WorkHCL and then create a completely managed database service kind of stuff. So, this brings this cloud neutrality, portability for customers and give them the choice and their terms, Stu. >> Well Bala, absolutely we've seen a huge growth in managed services as you've said, maybe bring us inside a little bit. What is free up customers? What we've said for so long that back when HCI first started, it was some of the storage administrators might bristle because you were taking things away from them. It was like, no, we're going to free you up to do other things that as Monica said, deliver more business value not mapping LUNs and doing that. How about from the DBA standpoint? What are some of those repetitive, undifferentiated heavy lifting that we're going to take away from them so that they can focus on the business value. >> Yep. Thank you Stu. So think about this. We all do copy paste operations in laptops. Something of that sort happens in data center at a much larger scale. Meaning that the same kind of copy paste operation happens to databases and petabytes and terabytes of scale. Hundreds of petabytes. It has become the most dreaded complex, long running error prone operation. Why should it be that way? Why should the DBS spend all this mundane tasks and then get busy for every cloning operation? It's a two day job for me, every backup job. It's like a hobby job for provisioning takes like three days. We can take this undifferentiated heavy lifting by this and then let the DBS focus on designing the cloud for them. Looking for the database tuning, design data modeling, ML aspects of the data kind of stuff. So we are freeing up the database Ops people, in a way that they can design the database cloud, and make sure that they are energy focused on high valid things and more towards the business center kind of stuff. >> Yeah. And you know automation is really important. You were talking about is automating mundane grunt work. Like IT spends 80% of its time in maintaining systems. So then where is the time for innovation. So if we can automate stuff that's repetitive, stuff that the machine can do, the software can do, why not? And I think that's what our database as a service often does. And I would add this, the big thing our database as a service does really is provide IT organizations and DV organizations a way to manage heterogeneous databases too. It's not like, here's my environment for Postgres. Here's my environment for my SQL. Here's my environment for Oracle. Here's my environment for SQL server. Now with a single offering, a single tool you can manage your heterogeneous environment across different clouds. On premises cloud, or in a public cloud environment. So I think that's the beauty we are talking about with Nutanix's Era. Is a truly, truly gives organizations that single environment to manage heterogeneous databases, apply the same automation and the ease of management across all these different environments. >> Yeah. I'll just add one comment to that. A true managed PaaS obviously customers in like a single shop go to public cloud, just click through and then they get the database and point. And then if someone is managing the database for them. But if you look at the enterprise data centers, they need to bring that enterprise GalNets and structure to these databases. It's not like anyone can do anything to any or these databases. So we are kind of getting the best of both, the needed enterprise GalNets by these enterprise people at the same time bringing the convenience for the application teams and developers they want to consume these databases like utility. So bringing the cloud experience, bringing the enterprise GalNets. At same time, I'm super confident we can cut down the cost. So that is what Nutanix Era is all about across all the clouds, including the enterprise cloud. >> Well, Bala being simpler and being less expensive are one of the original promises of the cloud that don't necessarily always come out there. So, that's super important. One of the other things, you talk about these hybrid environments. It's not just studied, in the public cloud want to understand these environments, if I'm in the public cloud, can I still leverage some of the services that are in the public cloud? So, if I want to run some analytics, if I want to use some of the phenomenal services that are coming out every day. Is that something that can be done in this environment? >> Yeah, beautiful. Thank you Stu. So we are seeing customers who two categories. There is a public cloud customer, completely born in public cloud cloud, native services. They realize that for every database that maintaining five or seven different copies and the management of these copies is prohibited just because every copy is a faulty copy in the public cloud. Meaning you take a backup snapshot and restore it. Your meter like New York taxi, it starts with running for your EBS   and that you are looking at it kind of stuff. So they can leverage Nutanix clusters and then have a highly efficient cloning capability so that they can cut down some of these costs for these secondary environments that I talk about. What we call is copy data management, that's one kind of use case. The other kind of customers that we are seeing who's cloud is a phenomenon. There's no way that people have to move to cloud. That's the something at a C level mandate that happens. These customers are enjoying their database experience on our enterprise cloud. But when they try to go to these big hyperscalers, they are seeing the disconnect that they're not able to enjoy some of the things that they are seeing on the enterprise cloud with us. So this transition, they are talking to us. Can you get this kind of functionality with Nutanix platform onto some of these big hyperscalers? So there are kind of customers moving both sides, some customers that are public cloud they're time to enjoy our facilities other than copy data management and Nutanix. Customers that are on-prem but they have a mandate to good public cloud ,with our hybrid cloud strategy. They get to enjoy the same kind of convenience that they are seeing it on enterprise and bringing the same kind of governance that they used to do it. so that maybe see customers. Yeah. >> Yeah. Monica, I want to go back to something you talked about customers dealing with that heterogeneous environment that they have reminds me of a lot of the themes that we talked about at nutanix.next because customers have they have multiple clouds they're using, requires different skillsets, different tooling. It's that simplicity layer that Nutanix has been working to deliver since day one. What are you from your customers? How are they doing with this? And especially in the database world. What are some of those challenges that they're really facing that we're looking to help solve with the solution today. >> Yeah. I mean, if you think about it, what customers at least in our experience, what they want or what they're looking for is this modern cloud platform that can really work across multiple cloud environments. Cause people don't want to change running, let's say an Oracle database you're on-prem on a certain stack and then using a whole different stack to run Oracle database in the cloud. What they want is the same exact foundation. So be so they can be, for sure have the right performance. Availability, reliability, the applications don't have to be rewritten on top of Oracle database. They want to preserve all of that, but they want the flexibility to be able to run that cloud platform wherever they choose to. So that's one. So that's choosing the right and modernizing and choosing the right cloud platform is definitely very important to our customers, but you nailed it on the head Stu. It's been about how do you manage it? How do you operate it on a daily basis? And that's where our customers are struggling with multiple types of tools out there, custom tool for every single environment. And that's what they don't want. They want to be able to manage, simply across multiple environments using the same tools and skillsets. And again, and I'm going to beat the same drum, but that's when Nutanix shines. That's a design principle is. It's the exact same technology foundation that you provide to customers to run any applications. In this case it happens to be databases. Exact same foundation you can use to run databases on-prem in the cloud. And then on top of that using Era boom! Simple management, simple operations, simple provisioning simple copy data management, simple patching, all of that becomes easy using just a single framework to manage and operate. And I will tell you this, when we talk to customers, what is it that DBS and database teams are struggling with? They're struggling with SLS and performance on scalability, that's one, number two they're struggling with keeping it up and running and fulfilling the demands of the stakeholders because they cannot keep up with how many databases they need to keep provisioning and patching and updating. So at Nutanix now we are actually solving both those problems with the platform. We are solving the problem of a very specific SLA that we can deliver in any cloud. And with Era, you're solving the issue of that operational complexity. We're making it really easy. So again, IT stakeholders DBS can fulfill the demands of the business stakeholders and really help them monetize the data. >> Yeah. I'll just add on with one concrete examples too. Like we have a big financial customer, they want to run Postgres. They are looking at the public cloud. Can we do a manage services kind of stuff, but you look at this, that the cost difference between a Postgres and your company infrastructure versus managed services almost like $3X to $4X dollars. Now, with Nutanix platform and Era, we were able to show that they can do at much reduced cost, manage their best service experience including their DBA cost are including the cloud administration cost. Like we added the infrastructure picture. We added the people who are going to manage the cloud, internal cloud and then intern experience being, plus plus of what they can see to public cloud. That's what makes the big difference. And this is what data sovereignty, data control, compliance and infrastructure governance, all these things coupled with cloud experiences, what customers really see the value of Era and the enterprise cloud and with an extension to the public cloud, with our hybrid cloud strategy. if they want to move this workload to public cloud they can do it. So, today with AWS clusters and tomorrow with our Azure clusters. So that gives them that kind of insurance not getting locked in by a big hyperscaler, but at same time enjoy the cloud experience. That's what big customers are looking for. >> Alright Bala, all the things you laid out here, what's the availability of Era rotically dot two? >> Era rotically dot two is actually available today. The customers can enjoy download the bits. We already have bunches of beta customers who are trying it out with the recall big telco companies are financial companies, and even big companies that manage big pensions kind of stuff. Let's talk about that kind of stuff. People are looking to us. In fact, there are customers who are looking for, when is this available for Azure cluster so that we can move some of our workloads to and manage the databases in Azure classes. So it is available and I'm looking forward to great feedback from our customers. And I'm hoping that it will solve some of their major critical problems. And in the process they get the best of Nutanix. >> Monica, last question I have for you. This doesn't seem like it's necessarily the same traditional infrastructure go to market for a solution like this. If I think back to, people think of HCI it was like, Oh! well, it was kind of a new box. We know Nutanix is a software company. More of what you do today is subscription based. So, maybe if you could talk a little bit to just how Nutanix goes to market with a solution like this. >> Yeah. And you know what, maybe people don't realize it but I'm hoping a lot of people do that. Nutanix is not just an infrastructure company anymore. In the last many years we've developed a full cloud platform in not only do we offer the infrastructure services with hyperconverged infrastructure which is now really the foundation. It's the hybrid cloud infrastructure. As you know, Stu, we talked to you a month ago and we talked about the evolution of XCI to really becoming the hybrid cloud infrastructure. But in addition to that, we also offer other data center services on storage DR Networking. We also offer DevOps services with application provisioning automation, application orchestration and then of course, database services that we talking about today and we offer desktop services. So Nutanix has really evolved in the last few years to a complete cloud platform really focusing on the application and workloads that run on top of the infrastructure stack. So not just the infrastructure layer but how can we be the best platform to run your databases? Your end is the computing workloads, your analytics applications your enterprise applications, cloud native applications. So that's what this is. And databases is one of our most successful workloads that's that runs a Nutanix very well because of the way the infrastructure software is architected. Because it's really great to scale high performance because again our superior architecture. And now with Era, it's a tool, it's all in one. Now it's also about really simplifying the management of databases and delivering them speedily and with agility to drive innovation in the organizations. >> Yep. Thank you Monica. Thank you. I I'll just add a couple of lines of comments into that. DTM for databases as erotically dots two, is going to be a challenge. And historically we are seen as an infrastructure company but the beauty of databases is so and to send to the infrastructure, the storage. So the language slightly becomes easy. And in fact, this holistic way of looking at solving the problem at the solution level rather than infrastructure helps us to go to a different kind of buyer, different kinds of decision maker, and we are learning. And I can tell you confidently the kind of progress that we have seen for in one enough year, the kind of customers that we are winning. And we are proving that we can bring a big difference to them. Though there is a challenge of DTM speaking the language of database, but the sheer nature of cloud platform the way they are a hundred hyperscale work. That's the kind of language that we take. You can run your solution. And here is how you can cut down your database backup time from hours to less than minute. Here's how you can cut down your patching from 16 hours to less than one hour. It is how you can cut down your provisioning time from multiple weeks to let them like matter of minutes. That holistic way of approaching it coupled with the power of the platform, really making the big difference for us. And I usually tell every time I meet, can you give us an opportunity to cut down your database cost, the PC vote, total cost of operations by close to 50%? That gets them excited that lets then move lean in and say, how do you plan to do it? And then we go about how do we do it? And we do a deep dive and PC people and all of it. So I'm excited. I think this is going to be a big play for Nutanix. We're going to make big difference. >> Absolutely well, Bala, congratulations to the team. Monica, both of you thank you so much for joining, really excited for all the announcements. >> Thank you so much. >> Thank you >> Stay with us. We're going to dig in a little bit more with one more interview for this product launch of the New Era and Database Management from Nutanix. I'm Stu Minimam as always, thank you for watching theCUBE. (cool music)

Published Date : Oct 6 2020

SUMMARY :

Narrator: From around the globe. I have the Senior Vice that the database for the customers have to our customers the ability I have one database it's kind of the one of the most valuable assets So the customers will now How about from the DBA standpoint? Meaning that the same kind of stuff that the machine can do, So bringing the cloud experience, of the services that are and the management of these of a lot of the themes that we talked about at nutanix.next demands of the stakeholders of Era and the enterprise And in the process they the same traditional of the way the infrastructure the kind of customers that we are winning. really excited for all the announcements. the New Era and Database

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MonicaPERSON

0.99+

NutanixORGANIZATION

0.99+

Monica KumarPERSON

0.99+

fiveQUANTITY

0.99+

$3XQUANTITY

0.99+

AWSORGANIZATION

0.99+

DBSORGANIZATION

0.99+

two dayQUANTITY

0.99+

$4XQUANTITY

0.99+

Stu MinimanPERSON

0.99+

PostgresORGANIZATION

0.99+

oneQUANTITY

0.99+

80%QUANTITY

0.99+

threeQUANTITY

0.99+

24QUANTITY

0.99+

16 hoursQUANTITY

0.99+

New YorkLOCATION

0.99+

three daysQUANTITY

0.99+

Bala KuchibhotlaPERSON

0.99+

less than one hourQUANTITY

0.99+

BalaPERSON

0.99+

bothQUANTITY

0.99+

four yearsQUANTITY

0.99+

OracleORGANIZATION

0.99+

tomorrowDATE

0.99+

SAP HANATITLE

0.99+

365 dayQUANTITY

0.99+

Stu MinimamPERSON

0.99+

StuPERSON

0.99+

Hundreds of petabytesQUANTITY

0.99+

both sidesQUANTITY

0.98+

AzureTITLE

0.98+

a month agoDATE

0.98+

todayDATE

0.98+

SQLTITLE

0.98+

two categoriesQUANTITY

0.98+

OneQUANTITY

0.98+

SAP HANATITLE

0.98+

HCIORGANIZATION

0.98+

single clusterQUANTITY

0.97+

seven different copiesQUANTITY

0.96+

single shopQUANTITY

0.96+

almost 80%QUANTITY

0.96+

nutanix.nextORGANIZATION

0.96+

sevenQUANTITY

0.96+

single frameworkQUANTITY

0.95+

one kindQUANTITY

0.94+

Eratory dot twoTITLE

0.94+

every two yearsQUANTITY

0.93+

UNLIST TILL 4/2 - The Road to Autonomous Database Management: How Domo is Delivering SLAs for Less


 

hello everybody and thank you for joining us today at the virtual Vertica BBC 2020 today's breakout session is entitled the road to autonomous database management how Domo is delivering SLA for less my name is su LeClair I'm the director of marketing at Vertica and I'll be your host for this webinar joining me is Ben white senior database engineer at Domo but before we begin I want to encourage you to submit questions or comments during the virtual session you don't have to wait just type your question or comment in the question box below the slides and click Submit there will be a Q&A session at the end of the presentation we'll answer as many questions as we're able to during that time any questions that we aren't able to address or drew our best to answer them offline alternatively you can visit vertical forums to post your questions there after the session our engineering team is planning to join the forum to keep the conversation going also as a reminder you can maximize your screen by clicking the double arrow button in the lower right corner of the slide and yes this virtual session is being recorded and will be available to view on demand this week we'll send you notification as soon as it's ready now let's get started then over to you greetings everyone and welcome to our virtual Vertica Big Data conference 2020 had we been in Boston the song you would have heard playing in the intro would have been Boogie Nights by heatwaves if you've never heard of it it's a great song to fully appreciate that song the way I do you have to believe that I am a genuine database whisperer then you have to picture me at 3 a.m. on my laptop tailing a vertical log getting myself all psyched up now as cool as they may sound 3 a.m. boogie nights are not sustainable they don't scale in fact today's discussion is really all about how Domo engineers the end of 3 a.m. boogie nights again well I am Ben white senior database engineer at Domo and as we heard the topic today the road to autonomous database management how Domo is delivering SLA for less the title is a mouthful in retrospect I probably could have come up with something snazzy er but it is I think honest for me the most honest word in that title is Road when I hear that word it evokes for me thoughts of the journey and how important it is to just enjoy it when you truly embrace the journey often you look up and wonder how did we get here where are we and of course what's next right now I don't intend to come across this too deep so I'll submit there's nothing particularly prescient and simply noticing the elephant in the room when it comes to database economy my opinion is then merely and perhaps more accurately my observation the office context imagine a place where thousands and thousands of users submit millions of ad-hoc queries every hour now imagine someone promised all these users that we could deliver bi leverage at cloud scale in record time I know what many of you should be thinking who in the world would do such a thing of course that news was well received and after the cheers from executives and business analysts everywhere and chance of Keep Calm and query on finally started to subside someone that turns an ass that's possible we can do that right except this is no imaginary place this is a very real challenge we face the demo through imaginative engineering demo continues to redefine what's possible the beautiful minds at Domo truly embrace the database engineering paradigm that one size does not fit all that little philosophical nugget is one I would pick up while reading the white papers and books of some guy named stone breaker so to understand how I and by extension Domo came to truly value analytic database administration look no further than that philosophy and what embracing it would mean it meant really that while others were engineering skyscrapers we would endeavor to build Datta neighborhoods with a diverse kapala G of database configuration this is where our journey at Domo really gets under way without any purposeful intent to define our destination not necessarily thinking about database as a service or anything like that we had planned this ecosystem of clusters capable of efficiently performing varied workloads we achieve this with custom configurations for node count resource pool configuration parameters etc but it also meant concerning ourselves with the unattended consequences of our ambition the impact of increased DDL activities on the catalog system overhead in general what would be the management requirements of an ever-evolving infrastructure we would be introducing multiple points of failure what are the advantages the disadvantages those types of discussions and considerations really help to define what would be the basic characteristics of our system the database itself needed to be trivial redundant potentially ephemeral customizable and above all scalable and we'll get more into that later with this knowledge of what we were getting into automation would have to be an integral part of development one might even say automation will become the first point of interest on our journey now using popular DevOps tools like saltstack terraform ServiceNow everything would be automated I mean it discluded everything from larger multi-step tasks like database designs database cluster creation and reboots to smaller routine tasks like license updates move-out and projection refreshes all of this cool automation certainly made it easier for us to respond to problems within the ecosystem these methods alone still if our database administration reactionary and reacting to an unpredictable stream of slow query complaints is not a good way to manage a database in fact that's exactly how three a.m. Boogie Nights happen and again I understand there was a certain appeal to them but ultimately managing that level of instability is not sustainable earlier I mentioned an elephant in the room which brings us to the second point of interest on our road to autonomy analytics more specifically analytic database administration why our analytics so important not just in this case but generally speaking I mean we have a whole conference set up to discuss it domo itself is self-service analytics the answer is curiosity analytics is the method in which we feed the insatiable human curiosity and that really is the impetus for analytic database administration analytics is also the part of the road I like to think of as a bridge the bridge if you will from automation to autonomy and with that in mind I say to you my fellow engineers developers administrators that as conductors of the symphony of data we call analytics we have proven to be capable producers of analytic capacity you take pride in that and rightfully so the challenge now is to become more conscientious consumers in some way shape or form many of you already employ some level of analytics to inform your decisions far too often we are using data that would be categorized as nagging perhaps you're monitoring slow queries in the management console better still maybe you consult the workflows analyzing how about a logging and alerting system like sumo logic if you're lucky you do have demo where you monitor and alert on query metrics like this all examples of analytics that help inform our decisions being a Domo the incorporation of analytics into database administration is very organic in other words pretty much company mandated as a company that provides BI leverage a cloud scale it makes sense that we would want to use our own product could be better at the business of doma adoption of stretches across the entire company and everyone uses demo to deliver insights into the hands of the people that need it when they need it most so it should come as no surprise that we have from the very beginning use our own product to make informed decisions as it relates to the application back engine in engineering we call it our internal system demo for Domo Domo for Domo in its current iteration uses a rules-based engine with elements through machine learning to identify and eliminate conditions that cause slow query performance pulling data from a number of sources including our own we could identify all sorts of issues like global query performance actual query count success rate for instance as a function of query count and of course environment timeout errors this was a foundation right this recognition that we should be using analytics to be better conductors of curiosity these types of real-time alerts were a legitimate step in the right direction for the engineering team though we saw ourselves in an interesting position as far as demo for demo we started exploring the dynamics of using the platform to not only monitor an alert of course but to also triage and remediate just how much economy could we give the application what were the pros and cons of that Trust is a big part of that equation trust in the decision-making process trust that we can mitigate any negative impacts and Trust in the very data itself still much of the data comes from systems that interacted directly and in some cases in directly with the database by its very nature much of the data was past tense and limited you know things that had already happened without any reference or correlation to the condition the mayor to those events fortunately the vertical platform holds a tremendous amount of information about the transaction it had performed its configurations the characteristics of its objects like tables projections containers resource pools etc this treasure trove of metadata is collected in the vertical system tables and the appropriately named data collector tables as a version 9 3 there are over 190 tables that define the system tables while the data collector is the collection of 215 components a rich collection can be found in the vertical system tables these tables provide a robust stable set of views that let you monitor information about your system resources background processes workload and performance allowing you to more efficiently profile diagnose and correlate historical data such as low streams query profiles to pool mover operations and more here you see a simple query to retrieve the names and descriptions of the system tables and an example of some of the tables you'll find the system tables are divided into two schemas the catalog schema contains information about persistent objects and the monitor schema tracks transient system States most of the tables you find there can be grouped into the following areas system information system resources background processes and workload and performance the Vertica data collector extends system table functionality by gathering and retaining aggregating information about your database collecting the data collector mixes information available in system table a moment ago I show you how you get a list of the system tables in their description but here we see how to get that information for the data collector tables with data from the data collecting tables in the system tables we now have enough data to analyze that we would describe as conditional or leading data that will allow us to be proactive in our system management this is a big deal for Domo and particularly Domo for demo because from here we took the critical next step where we analyze this data for conditions we know or suspect lead to poor performance and then we can suggest the recommended remediation really for the first time we were using conditional data to be proactive in a database management in record time we track many of the same conditions the Vertica support analyzes via scrutinize like tables with too many production or non partition fact tables which can negatively affect query performance and life in vertical in viral suggests if the table has a data a time step column you recommend the partitioning by the month we also can track catalog sizes percentage of total memory and alert thresholds and trigger remediations requests per hour is a very important metric in determining when a trigger are scaling solution tracking memory usage over time allows us to adjust resource pool parameters to achieve the optimal performance for the workload of course the workload analyzer is a great example of analytic database administration I mean from here one can easily see the logical next step where we were able to execute these recommendations manually or automatically be of some configuration parameter now when I started preparing for this discussion this slide made a lot of sense as far as the logical next iteration for the workload analyzing now I left it in because together with the next slide it really illustrates how firmly Vertica has its finger on the pulse of the database engineering community in 10 that OS management console tada we have the updated work lies will load analyzer we've added a column to show tuning commands the management console allows the user to select to run certain recommendations currently tuning commands that are louder and alive statistics but you can see where this is going for us using Domo with our vertical connector we were able to then pull the metadata from all of our clusters we constantly analyze that data for any number of known conditions we build these recommendations into script that we can then execute immediately the actions or we can save it to a later time for manual execution and as you would expect those actions are triggered by thresholds that we can set from the moment nyan mode was released to beta our team began working on a serviceable auto-scaling solution the elastic nature of AI mode separated store that compute clearly lent itself to our ecosystems requirement for scalability in building our system we worked hard to overcome many of the obstacles they came with the more rigid architecture of enterprise mode but with the introduction is CRM mode we now have a practical way of giving our ecosystem at Domo the architectural elasticity our model requires using analytics we can now scale our environment to match demand what we've built is a system that scales without adding management overhead or our necessary cost all the while maintaining optimal performance well we're really this is just our journey up to now and which begs the question what's next for us we expand the use of Domo for Domo within our own application stack maybe more importantly we continue to build logic into the tools we have by bringing machine learning and artificial intelligence to our analysis and decision making really do to further illustrate those priorities we announced the support for Amazon sage maker autopilot at our demo collusive conference just a couple of weeks ago for vertical the future must include in database economy the enhanced capabilities in the new management console to me are clear nod to that future in fact with a streamline and lightweight database design process all the pieces should be in place versions deliver economists database management itself we'll see well I would like to thank you for listening and now of course we will have a Q&A session hopefully very robust thank you [Applause]

Published Date : Mar 31 2020

SUMMARY :

conductors of the symphony of data we

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
BostonLOCATION

0.99+

VerticaORGANIZATION

0.99+

thousandsQUANTITY

0.99+

DomoORGANIZATION

0.99+

3 a.m.DATE

0.99+

AmazonORGANIZATION

0.99+

todayDATE

0.99+

first timeQUANTITY

0.98+

this weekDATE

0.97+

over 190 tablesQUANTITY

0.97+

two schemasQUANTITY

0.96+

second pointQUANTITY

0.96+

215 componentsQUANTITY

0.96+

first pointQUANTITY

0.96+

three a.m.DATE

0.96+

Boogie NightsTITLE

0.96+

millions of ad-hoc queriesQUANTITY

0.94+

DomoTITLE

0.93+

Vertica Big Data conference 2020EVENT

0.93+

Ben whitePERSON

0.93+

10QUANTITY

0.91+

thousands of usersQUANTITY

0.9+

one sizeQUANTITY

0.89+

saltstackTITLE

0.88+

4/2DATE

0.86+

a couple of weeks agoDATE

0.84+

DattaORGANIZATION

0.82+

end of 3 a.m.DATE

0.8+

Boogie NightsEVENT

0.78+

double arrowQUANTITY

0.78+

every hourQUANTITY

0.74+

ServiceNowTITLE

0.72+

DevOpsTITLE

0.72+

Database ManagementTITLE

0.69+

su LeClairPERSON

0.68+

many questionsQUANTITY

0.63+

SLATITLE

0.62+

The RoadTITLE

0.58+

Vertica BBCORGANIZATION

0.56+

2020EVENT

0.55+

database managementTITLE

0.52+

Domo DomoTITLE

0.46+

version 9 3OTHER

0.44+

UNLIST TILL 4/2 - Vertica Database Designer - Today and Tomorrow


 

>> Jeff: Hello everybody and thank you for joining us today for the Virtual VERTICA BDC 2020. Today's breakout session has been titled, "VERTICA Database Designer Today and Tomorrow." I'm Jeff Healey, Product VERTICA Marketing, I'll be your host for this breakout session. Joining me today is Yuanzhe Bei, Senior Technical Manager from VERTICA Engineering. But before we begin, (clearing throat) I encourage you to submit questions or comments during the virtual session. You don't have to wait, just type your question or comment in the question box below the slides and click Submit. As always, there will be a Q&A session at the end of the presentation. We'll answer as many questions, as we're able to during that time, any questions we don't address, we'll do our best to answer them offline. Alternatively, visit VERTICA forums at forum.vertica.com to post your questions there after the session. Our engineering team is planning to join the forums, to keep the conversation going. Also, a reminder that you can maximize your screen by clicking the double arrow button at the lower right corner of the slides. And yes, this virtual session is being recorded and will be available to view on demand this week. We will send you a notification as soon as it's ready. Now let's get started. Over to you Yuanzhe. >> Yuanzhe: Thanks Jeff. Hi everyone, my name is Yuanzhe Bei, I'm a Senior Technical Manager at VERTICA Server RND Group. I run the query optimizer, catalog and the disaggregated engine team. Very glad to be here today, to talk about, the "VERTICA Database Designer Today and Tomorrow". This presentation will be organized as the following; I will first refresh some knowledge about, VERTICA fundamentals such as Tables and Projections, which will bring to the question, "What is Database Designer?" and "Why we need this tool?". Then I will take you through a deep dive, into a Database Designer or we call DBD, and see how DBD's internals works, after that I'll show you some exciting DBD improvements, we have planned for 10.0 release and lastly, I will share with you, some DBD future roadmap we planned next. As most of you should already know, VERTICA is built on a columnar architecture. That means, data is stored column wise. Here we can see a very simple example, of table with four columns, and the many of you may also know, table in VERTICA is a virtual concept. It's just a logical representation of data, which means user can write SQL query, to reference the table names and column, just like other relational database management system, but the actual physical storage of data, is called Projection. A Projection can reference a subset, or all of the columns all to its anchor table, and must be sorted by at least one column. Each table need at least one C for projection which reference all the columns to the table. If you load data to a table with no projection, and automated, auto production will be created, which will be arbitrarily assorted by, the first couple of columns in the table. As you can imagine, even though such other production, can be used to answer any query, the performance is not optimized in most cases. A common practice in VERTICA, is to create multiple projections, contain difference step of column, and sorted in different ways on the same table. When query is sent to the server, the optimizer will pick the projection, that can answer the query in the most efficient way. For example, here you can say, let's say you have a query, that select columns B, D, C and sorted by B and D, the third projection will be ideal, because the data is already sorted, so you can save the sorting costs while executing the query. Basically when you choose the design of the projection, you need to consider four things. First and foremost, of course the sort order. The data already sorted in the right way, can benefit quite a lot of the query actually, like Ordered by, Group By, Analytics, Merge, Join, Predicates and so on. The select column group is also important, because the projection must contain, all the columns referenced by your workflow query. Even missing one column in the projection, this projection cannot be used for a particular query. In addition, VERTICA is the distributed database, and allow projection to be segmented, based on the hash of a set of columns, which is beneficial if the segmentation merged, the join keys or group keys. And finally encoding of each per columns is also part of the design, because the data is sorted in different way, may completely change the optimal encoding for each column. This example only show the benefit of the first two, but you can imagine the rest too are also important. But even for that, it doesn't sound that hard, right? Well I hope you change your mind already when you see this, at least I do. These machine generated queries, really beats me. It will probably take an experienced DBA hours, to figure out which projection can be benefit these queries, not even mentioning there could be hundreds of such queries, in the regular work logs in the real world. So what can we do? That's why we need DBD. DBD is a tool integrated in the VERTICA server, that it can help DBA to perform an access, on their work log query, tabled schema and data, and then automatically figure out, the most optimized projection design for their workload. In addition, DBD also a sophisticated tool, that can take customize by a user, by sending a lot of parameters objectives and so on. And lastly, DBD has access to the optimizer, so DB knows what kind of attribute, the projection need to have, in order to have the optimizer to benefit from them. DBD has been there for years, and I'm sure there are plenty of materials available online, to show you how DBD can be used in different scenarios, whether to achieve the query optimize, or load optimize, whether it's the comprehensive design, or the incremental design, whether it's a dumping deployment script, and manual deployment later, or let the DBD do the order deployment for you, and the many other options. I'm not planning to talk about this today, instead, I will take the opportunity today, to open this black box DBD, and show you what exactly hide inside. DBD is a complex tool and I have tried my best to summarize the DBD design process into seven steps; Extract, Permute, Prune, Build, Score, Identify and Encode. What do they mean? Don't worry, I will show you step by step. The first step is Extract. Extract Interesting Columns. In this step, DBD pass the design queries, and figure out the operations that can be benefited, by the potential projection design, and extract the corresponding columns, as interesting columns. So Predicates, Group By, Order By, Joint Condition, and analytics are all interesting Column to the DBD. As you can see this three simple sample queries, DBD can extract the interest in column sets on the right. Some of these column sets are unordered. For example, the green one for Group By a1 and b1, the DBD extracts the interesting column set, and put them in the own orders set, because either data sorted by a1 first or b1 first, can benefit from this Group By operation. Some of the other sets are ordered, and the best example is here, order by clause a2 and b2, and obviously you cannot sort it by b2 and then a2. These interesting columns set will be used as if, to extend to actual projection sort order candidates. The next step is Permute, once DBD extract all the C's, it will enumerate sort order using C, and how does DBD do that? I'm starting with a very simple example. So here you can see DBD can enumerate two sort orders, by extending d1 with the unordered set a1, b1, and the derived at two sort order candidates, d1, a1, b1, and d1, b1, a1. This sort order can benefit queries with predicate on d1, and also benefit queries by Group By a1, b1, when a1, sorry when d1 is constant. So with the same idea, DBD will try to extend other States with each other, and populate more sort order permutations. You can imagine that how many of them, there could be many of them, these candidates, based on how many queries you have in the design and that can be handled of the sort order candidates. That comes to the third step, which is Pruning. This step is to limit the candidates sort order, so that the design won't be running forever. DBD uses very simple capping mechanism. It sorts all the, sort all the candidates, are ranked by length, and only a certain number of the sort order, with longest length, will be moved forward to the next step. And now we have all the sort orders candidate, that we want to try, but whether this sort order candidate, will be actually be benefit from the optimizer, DBD need to ask the optiizer. So this step before that happens, this step has to build those projection candidate, in the catalog. So this step will build, will generates the projection DBL's, surround the sort order, and create this projection in the catalog. These projections won't be loaded with real data, because that takes a lot of time, instead, DBD will copy over the statistic, on existing projections, to this projection candidates, so that the optimizer can use them. The next step is Score. Scoring with optimizer. Now projection candidates are built in the catalog. DBD can send a work log queries to optimizer, to generate a query plan. And then optimizer will return the query plan, DBD will go through the query plan, and investigate whether, there are certain benefits being achieved. The benefits list have been growing over time, when optimizer add more optimizations. Let's say in this case because the projection candidates, can be sorted by the b1 and a1, it is eligible for Group By Pipe benefit. Each benefit has a preset score. The overall benefit score of all design queries, will be aggregated and then recorded, for each projection candidate. We are almost there. Now we have all the total benefit score, for the projection candidates, we derived on the work log queries. Now the job is easy. You can just pick the sort order with the highest score as the winner. Here we have the winner d1, b1 and a1. Sometimes you need to find more winners, because the chosen winner may only benefit a subset, of the work log query you provided to the DBD. So in order to have the rest of the queries, to be also benefit, you need more projections. So in this case, DBD will go to the next iteration, and let's say in this case find to another winner, d1, c1, to benefit the work log queries, that cannot be benefit by d1, b1 and a1. The number of iterations and thus the winner outcome, DBD really depends on the design objective that uses that. It can be load optimized, which means that only one, super projection winner will be selected, or query optimized, where DBD try to create as many projections, to cover most of the work log queries, or somewhat balance an objective in the middle. The last step is to decide encoding, for each projection columns, for the projection winners. Because the data are sorted differently, the encoding benefits, can be very different from the existing projection. So choose the right projection encoding design, will save the disk footprint a significant factor. So it's worth the effort, to find out the best thing encoding. DBD picks the encoding, based on the actual sampling the data, and measure the storage footprint. For example, in this case, the projection winner has three columns, and say each column has a few encoding options. DBD will write the sample data in the way this projection is sorted, and then you can see with different encoding, the disk footprint is different. DBD will then compare the disk footprint of each, of different options for each column, and pick the best encoding options, based on the one that has the smallest storage footprint. Nothing magical here, but it just works pretty well. And basic that how DBD internal works, of course, I think we've heard it quite a lot. For example, I didn't mention how the DBD handles segmentation, but the idea is similar to analyze the sort order. But I hope this section gave you some basic idea, about DBD for today. So now let's talk about tomorrow. And here comes the exciting part. In version 10.0, we significantly improve the DBD in many ways. In this talk I will highlight four issues in old DBD and describe how the 10.0 version new DBD, will address those issues. The first issue is that a DBD API is too complex. In most situations, what user really want is very simple. My queries were slow yesterday, with the new or different projection can help speed it up? However, to answer a simple question like this using DBD, user will be very likely to have the documentation open on the side, because they have to go through it's whole complex flow, from creating a projection, run the design, get outputs and then create a design in the end. And that's not there yet, for each step, there are several functions user need to call in order. So adding these up, user need to write the quite long script with dozens of functions, it's just too complicated, and most of you may find it annoying. They either manually tune the projection to themselves, or simply live with the performance and come back, when it gets really slow again, and of course in most situations, they never come back to use the DBD. In 10.0 VERTICA support the new simplified API, to run DBD easily. There will be just one function designer_single_run and one argument, the interval that you think, your query was slow. In this case, user complained about it yesterday. So what does this user to need to do, is just specify one day, as argument and run it. The user don't need to provide anything else, because the DBD will look up his query or history, within that time window and automatically populate design, run design and export the projection design, and the clean up, no user intervention needed. No need to have the documentation on the side and carefully write a script, and a debug, just one function call. That's it. Very simple. So that must be pretty impressive, right? So now here comes to another issue. To fully utilize this single round function, users are encouraged to run DBD on the production cluster. However, in fact, VERTICA used to not recommend, to run a design on a production cluster. One of the reasons issue, is that DBD picks massive locks, both table locks and catalog locks, which will badly interfere the running workload, on a production cluster. As of 10.0, we eliminated all the table and ten catalog locks from DBD. Yes, we eliminate 100% of them, simple improvement, clear win. The third issue, which user may not be aware of, is that DBD writes intermediate result. into real VERTICA tables, the real DBD have to do that is, DBD is the background task. So the intermediate results, some user needs to monitor it, the progress of the DBD in concurrent session. For complex design, the intermediate result can be quite massive, and as a result, many lost files will be created, and written to the disk, and we should both stress, the catalog, and that the disk can slow down the design. For ER mode, it's even worse because, the table are shared on communal storage. So writing to the regular table, means that it has to upload the data, to the communal storage, which is even more expensive and disruptive. In 10.0, we significantly restructure the intermediate results buffer, and make this shared in memory data structure. Monitoring queries will go directly look up, in memory data structure, and go through the system table, and return the results. No Intermediate Results files will be written anymore. Another expensive lubidge of local disk for DBD is encoding design, as I mentioned earlier in the deep dive, to determine which encoding works the best for the new projection design, there's no magic way, but the DBD need to actually write down, the sample data to the disk, using the different encoding options, and to find out which ones have the smallest footprint, or pick it as the best choice. These written sample data will be useless after this, and it will be wiped out right away, and you can imagine this is a huge waste of the system resource. In 10.0 we improve this process. So instead of writing, the different encoded data on the disk, and then read the file size, DBD aggregate the data block size on-the-fly. The data block will not be written to the disk, so the overall encoding and design is more efficient and non-disruptive. Of course, this is just about the start. The reason why we put a significant amount of the resource on the improving the DBD in 10.0, is because the VERTICA DBD, as essential component of the out of box performance design campaign. To simply illustrate the timeline, we are now on the second step, where we significantly reduced, the running overhead of the DBD, so that user will no longer fear, to run DBD on their production cluster. Please be noted that as of 10.0, we haven't really started changing, how DBD design algorithm works, so that what we have discussed in the deep dive today, still holds. For the next phase of DBD, we will briefly make the design process smarter, and this will include better enumeration mechanism, so that the pruning is more intelligence rather than brutal, then that will result in better design quality, and also faster design. The longer term is to make DBD to achieve the automation. What entail automation and what I really mean is that, instead of having user to decide when to use DBD, until their query is slow, VERTICA have to know, detect this event, and have have DBD run automatically for users, and suggest the better projections design, if the existing projection is not good enough. Of course, there will be a lot of work that need to be done, before we can actually fully achieve the automation. But we are working on that. At the end of day, what the user really wants, is the fast database, right? And thank you for listening to my presentation. so I hope you find it useful. Now let's get ready for the Q&A.

Published Date : Mar 30 2020

SUMMARY :

at the end of the presentation. and the many of you may also know,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

Yuanzhe BeiPERSON

0.99+

Jeff HealeyPERSON

0.99+

100%QUANTITY

0.99+

forum.vertica.comOTHER

0.99+

one dayQUANTITY

0.99+

second stepQUANTITY

0.99+

third stepQUANTITY

0.99+

tomorrowDATE

0.99+

third issueQUANTITY

0.99+

todayDATE

0.99+

FirstQUANTITY

0.99+

yesterdayDATE

0.99+

Each benefitQUANTITY

0.99+

TodayDATE

0.99+

third projectionQUANTITY

0.99+

OneQUANTITY

0.99+

b2OTHER

0.99+

each columnQUANTITY

0.99+

first issueQUANTITY

0.99+

one columnQUANTITY

0.99+

three columnsQUANTITY

0.99+

VERTICA EngineeringORGANIZATION

0.99+

YuanzhePERSON

0.99+

each stepQUANTITY

0.98+

Each tableQUANTITY

0.98+

first stepQUANTITY

0.98+

DBDTITLE

0.98+

DBDORGANIZATION

0.98+

seven stepsQUANTITY

0.98+

DBLORGANIZATION

0.98+

eachQUANTITY

0.98+

one argumentQUANTITY

0.98+

VERTICATITLE

0.98+

each projectionQUANTITY

0.97+

first twoQUANTITY

0.97+

firstQUANTITY

0.97+

this weekDATE

0.97+

hundredsQUANTITY

0.97+

one functionQUANTITY

0.97+

clause a2OTHER

0.97+

oneQUANTITY

0.97+

each per columnsQUANTITY

0.96+

TomorrowDATE

0.96+

bothQUANTITY

0.96+

four issuesQUANTITY

0.95+

VERTICAORGANIZATION

0.95+

b1OTHER

0.95+

single roundQUANTITY

0.94+

4/2DATE

0.94+

first couple of columnsQUANTITY

0.92+

VERTICA Database Designer Today and TomorrowTITLE

0.91+

VerticaORGANIZATION

0.91+

10.0QUANTITY

0.89+

one function callQUANTITY

0.89+

a1OTHER

0.89+

four thingsQUANTITY

0.88+

c1OTHER

0.87+

two sort orderQUANTITY

0.85+

Breaking Analysis: Spending Data Shows Cloud Disrupting the Analytic Database Market


 

from the silicon angle media office in Boston Massachusetts it's the queue now here's your host David on tape hi everybody welcome to this special cube in size powered by ET our enterprise Technology Research our partner who's got this database to solve the spending data and what we're gonna do is a braking analysis on the analytic database market we're seeing that cloud and cloud players are disrupting that marketplace and that marketplace really traditionally has been known as the enterprise data warehouse market so Alex if you wouldn't mind bringing up the first slide I want to talk about some of the trends in the traditional EDW market I almost don't like to use that term anymore because it's sort of a pejorative but let's look at it's a very large market it's about twenty billion dollars today growing it you know high single digits low double digits it's expected to be in the 30 to 35 billion dollar size by mid next decade now historically this is dominated by teradata who started this market really back in the 1980s with the first appliance the first converged appliance or coal with Exadata you know IBM I'll talk about IBM a little bit they bought a company called mateesah back in the day and they've basically this month just basically killed the t's and killed the brand Microsoft has entered the fray and so it's it's been a fairly large market but I say it's failed to really live up to the promises that we heard about in the late 90s early parts of the 2000 namely that you were going to be able to get a 360 degree view of your data and you're gonna have this flexible easy access to the data you know the reality is data warehouses were really expensive they were slow you had to go through a few experts to to get data it took a long time I'll tell you I've done a lot of research on this space and when you talked to the the data warehouse practitioners they would tell you we always had to chase the chips anytime Intel would come out with a new chip we forced it in there because we just didn't have the performance to really run the analytics as we need to it's took so long one practitioner described it as a snake swallowing a basketball so you've got all those data which is the sort of metaphor for the basketball just really practitioners had a hard time standing up infrastructure and what happened as a spate of new players came into the marketplace these these MPP players trying to disrupt the market you had Vertica who was eventually purchased by HP and then they sold them to Micro Focus greenplum was buy bought by EMC and really you know company is de-emphasized greenplum Netezza 1.7 billion dollar acquisition by IBM IBM just this month month killed the brand they're kind of you know refactoring everything par Excel was interesting was it was a company based on an open-source platform that Amazon AWS did a one-time license with and created a redshift it ever actually put a lot of innovation redshift this is really doing well well show you some data on that we've also at the time saw a major shift toward unstructured data and read much much greater emphasis on analytics it coincided with Hadoop which also disrupted the market economics I often joked it the ROI of a dupe was reduction on investment and so you saw all these data lakes being built and of course they turned into the data swamps and you had dozens of companies come into the database space which used to be rather boring but Mike Amazon with dynamodb s AP with HANA data stacks Redis Mongo you know snowflake is another one that I'm going to talk about in detail today so you're starting to see the blurring of lines between relational and non relational and what was was what once thought of is no sequel became not only sequel sequel became the killer app for Hadoop and so at any rate you saw this new class of data stores emerging and snowflake was one of the more interesting and and I want to share some of that data with you some of the spending intentions so over the last several weeks and months we've shared spending intentions from ETR enterprise technology research they're a company that that the manages of the spending data and has a panel of about 4,500 end-users they go out and do spending in tension surveys periodically so Alex if you bring up this survey data I want to show you this so this is spending intentions and and what it shows is that the public cloud vendors in snowflake who really is a database as a service offering so cloud like are really leading the pack here so the sector that I'm showing is the enterprise data warehouse and I've added in the the analytics business intelligence and Big Data section so what this chart shows is the vendor on the left-hand side and then this bar chart has colors the the red is we're leaving the platform the gray is our spending will be flat so this is from the July survey expect to expectations for the second half of 2019 so gray is flat the the dark green is increase and the lime green is we are a new customer coming on to the platform so if you take the the greens and subtract out the red and there's two Reds the dark red is leaving the lighter red is spending less so if you subtract the Reds from the greens you get what's called a net score so the higher the net score the better so you can see here the net score of snowflake is 81% so that very very high you can also see AWS in Microsoft a very high and Google so the cloud vendors of which I would consider a snowflake at cloud vendor like at the cloud model all kicking butt now look at Oracle look at the the incumbents Oracle IBM and Tara data Oracle and IBM are in the single digits for a net score and the Terra data is in a negative 10% so that's obviously not a good sign for those guys so you're seeing share gains from the cloud company snowflake AWS Microsoft and Google at the expense of certainly of teradata but likely IBM and Oracle Oracle's little for animal they got Exadata and they're putting a lot of investments in there maybe talk about that a little bit more now you see on the right hand side this black says shared accounts so the N in this survey this July survey that ETR did is a thousand sixty eight so of a thousand sixty eight customers each er is asking them okay what's your spending going to be on enterprise data warehouse and analytics big data platforms and you can see the number of accounts out of that thousand sixty eight that are being cited so snowflake only had 52 and I'll show you some other data from from past surveys AWS 319 Microsoft the big you know whale here trillion dollar valuation 851 going down the line you see Oracle a number you know very large number and in Tara data and IBM pretty large as well certainly enough to get statistically valid results so takeaway here is snowflake you know very very strong and the other cloud vendors the hyper scale is AWS Microsoft and Google and their data stores doing very well in the marketplace and challenging the incumbents now the next slide that I want to show you is a time series for selected suppliers that can only show five on this chart but it's the spending intentions again in that EDW and analytics bi big data segment and it shows the spending intentions from January 17 survey all the way through July 19 so you can see the the period the periods that ETR takes this the snapshots and again the latest July survey is over a thousand n the other ones are very very large too so you can see here at the very top snowflake is that yellow line and they just showed up in the January 19 a survey and so you're seeing now actually you go back one yeah January 19 survey and then you see them in July you see the net score is the July next net score that I'm showing that's 35 that's the number of accounts out of the corpus of data that snowflake had in the survey back in January and now it's up to 52 you can see they lead the packet just in terms of the spending intention in terms of mentions AWS and Microsoft also up there very strong you see big gap down to Oracle and Terra data I didn't show I BM didn't show Google Google actually would be quite high to just around where Microsoft is but you can see the pressure that the cloud is placing on the incumbents so what are the incumbents going to do about it well certainly you're gonna see you know in the case of Oracle spending a lot of money trying to maybe rethink the the architecture refactor the architecture Oracle open worlds coming up shortly I'm sure you're gonna see a lot of new announcements around Exadata they're putting a lot of wood behind the the exadata arrow so you know we'll keep in touch with that and stay tuned but you can see again the big takeaways here is that cloud guys are really disrupting the traditional edw marketplace alright let's talk a little bit about snowflakes so I'm gonna highlight those guys and maybe give a little bit of inside baseball here but what you need to know about snowflakes so I've put some some points here just some quick points on the slide Alex if you want to bring that up very fast-growing cloud and SAS based data warehousing player growing that couple hundred percent annually their annual recurring revenue very high these guys are getting ready to do an IPO talk about that a little bit they were founded in 2012 and it kind of came out of stealth and hiding in 2014 after bringing Bob Moog Leon from Microsoft as the CEO it was really the background on these guys is they're three engineers from Oracle will probably bored out of their mind like you know what we got this great idea why should we give it to Oracle let's go pop out and start a company and that NIN's and as such they started a snowflake they really are disrupting the incumbents they've raised over 900 million dollars in venture and they've got almost a four billion dollar valuation last May they brought on Frank salute Minh and this is really a pivot point I think for the company and they're getting ready to do an IPO so and so let's talk a little bit about that in a moment but before we do that I want to bring up just this really simple picture of Alex if you if you'd bring this this slide up this block diagram it's like a kindergarten so that you know people like you know I can even understand it but basically the innovation around the snowflake architecture was that they they separated their claim is that they separated the storage from the compute and they've got this other layer called cloud services so let me talk about that for a minute snowflake fundamentally rethought the architecture of the data warehouse to really try to take advantage of the cloud so traditionally enterprise data warehouses are static you've got infrastructure that kind of dictates what you can do with the data warehouse and you got to predict you know your peak needs and you bring in a bunch of storage and compute and you say okay here's the infrastructure and this is what I got it's static if your workload grows or some new compliance regulation comes out or some new data set has to be analyzed well this is what you got you you got your infrastructure and yeah you can add to it in chunks of compute and storage together or you can forklift out and put in new infrastructure or you can chase more chips as I said it's that snake swallowing a basketball was not pretty so very static situation and you have to over provision whereas the cloud is all about you know pay buy the drink and it's about elasticity and on demand resources you got cheap storage and cheap compute and you can just pay for it as you use it so the innovation from snowflake was to separate the compute from storage so that you could independently scale those and decoupling those in a way that allowed you to sort of tune the knobs oh I need more compute dial it up I need more storage dial it up or dial it down and pay for only what you need now another nuance here is traditionally the computing and data warehousing happens on one cluster so you got contention for the resources of that cluster what snowflake does is you can spin up a warehouse on the fly you can size it up you can size it down based on the needs of the workload so that workload is what dictates the infrastructure also in snowflakes architecture you can access the same data from many many different houses so you got again that three layers that I'm showing you the storage the compute and the cloud services so let me go through some examples so you can really better understand this so you've got storage data you got customer data you got you know order data you got log files you might have parts data you know what's an inventory kind of thing and you want to build warehouses based on that data you might have marketing a warehouse you might have a sales warehouse you might have a finance warehouse maybe there's a supply chain warehouse so again by separating the compute from that sort of virtualized compute from the from the storage layer you can access any data leave the data where it is and I'll talk about this in more and bring the compute to the data so this is what in part the cloud layer does they've got security and governance they got data warehouse management in that cloud layer and and resource optimization but the key in in my opinion is this metadata management I think that's part of snowflakes secret sauce is the ability to leave data where it is and have the smarts and the algorithms to really efficiently bring the compute to the data so that you're not moving data around if you think about how traditional data warehouses work you put all the data into a central location so you can you know operate on it well that data movement takes a long long time it's very very complicated so that's part of the secret sauce is knowing what data lives where and efficiently bringing that compute to the data this dramatically improves performance it's a game changer and it's much much less expensive now when I come back to Frank's Luqman this is somebody that I've is a career that I've followed I've known had him on the cube of a number of times I first met Frank Sloot when he was at data domain he took that company took it public and then sold it originally NetApp made a bid for the company EMC Joe Tucci in the defensive play said no we're not gonna let Ned afgan it there was a little auction he ended up selling the company for I think two and a half billion dollars sloop and came in he helped clean up the the data protection business of EMC and then left did a stint as a VC and then took over service now when snoop and took over ServiceNow and a lot of people know this the ServiceNow is the the shiny toy on Wall Street today service that was a mess when saluteth took it over it's about 100 120 million dollar company he and his team took it to 1.2 billion dramatically increased the the valuation and one of the ways they did that was by thinking about the Tam and expanding that Tim that's part of a CEOs job as Tam expansion Steuben is also a great operational guy and he brought in an amazing team to do that I'll talk a little bit about that team effect uh well he just brought in Mike Scarpelli he was the CFO was the CFO of ServiceNow brought him in to run finance for snowflake so you've seen that playbook emerge you know be interesting Beth white was the CMO at data domain she was the CMO at ServiceNow helped take that company she's an amazing resource she kind of you know and in retirement she's young but she's kind of in retirement doing some advisory roles wonder if slooping will bring her back I wonder if Dan Magee who was ServiceNow is operational you know guru wonder if he'll come out of retirement how about Dave Schneider who runs the sales team at at ServiceNow well he you know be be lord over we'll see the kinds of things that Sluman looks for just in my view of observing his playbook over the years he looks for great product he looks for a big market he looks for disruption and he looks for off-the-chart ROI so his sales teams can go in and really make a strong business case to disrupt the existing legacy players so I one of the things I said that snoopin looks for is a large market so let's look at this market and this is the thing that people missed around ServiceNow and to credit Pat myself and David for in the back you know we saw the Tam potential of ServiceNow is to be many many tens of billions you know Gartner when they when ServiceNow first came out said hey helpdesk it's a small market couple billion dollars we saw the potential to transform not only IT operations but go beyond helpdesk change management at cetera IT Service Management into lines of business and we wrote a piece on wiki Vaughn back then it's showing the potential Tam and we think something similar could happen here so the market today let's call 20 billion growing to 30 Billy big first of all but a lot of players in here what if so one of the things that we see snowflake potentially being able to do with its architecture and its vision is able to bring enterprise search you know to the marketplace 80% of the data that's out there today sits behind firewalls it's not searchable by Google what if you could unlock that data and access it in query at anytime anywhere put the power in the hands of the line of business users to do that maybe think Google search for enterprises but with provenance and security and governance and compliance and the ability to run analytics for a line of business users it's think of it as citizens data analytics we think that tam could be 70 plus billion dollars so just think about that in terms of how this company might this company snowflake might go to market you by the time they do their IPO you know it could be they could be you know three four five hundred billion dollar company so we'll see we'll keep an eye on that now because the markets so big this is not like the ITSM the the market that ServiceNow was going after they crushed BMC HP was there but really not paying attention to it IBM had a product it had all these products that were old legacy products they weren't designed for the cloud and so you know ServiceNow was able to really crush that market and caught everybody by surprise and just really blew it out there's a similar dynamic here in that these guys are disrupting the legacy players with a cloud like model but at the same time so the Amazon with redshift so is Microsoft with its analytics platform you know teradata is trying to figure it out they you know they've got an inertia of a large install base but it's a big on-prem install base I think they struggle a little bit but their their advantages they've got customers locked in or go with exudate is very interesting Oracle has burned the boats and in gone to cloud first in Oracle mark my words is is reacting everything for the cloud now you can say Oh Oracle they're old school they're old guard that's fine but one of the things about Oracle and Larry Ellison they spend money on R&D they're very very heavy investor in Rd and and I think that you know you can see the exadata as it's actually been a very successful product they will react attacked exadata believe you me to to bring compute to the data they understand you can't just move all this the InfiniBand is not gonna solve their problem in terms of moving data around their architecture so you know watch Oracle you've got other competitors like Google who shows up well in the ETR survey so they got bigquery and BigTable and you got a you know a lot of other players here you know guys like data stacks are in there and you've got you've got Amazon with dynamo DB you've got couch base you've got all kinds of database players that are sort of blurring the lines as I said between sequel no sequel but the real takeaway here from the ETR data is you've got cloud again is winning it's driving the discussion and the spending discussion with an IT watch this company snowflake they're gonna do an IPO I guarantee it hopefully they will see if they'll get in before the booth before the market turns down but we've seen this play by Frank Sluman before and his team and and and the spending data shows that this company is hot you see them all over Silicon Valley you're seeing them show up in the in the spending data so we'll keep an eye on this it's an exciting market database market used to be kind of boring now it's red-hot so there you have it folks thanks for listening is a Dave Volante cube insights we'll see you next time

Published Date : Sep 6 2019

SUMMARY :

David for in the back you know we saw

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

January 19DATE

0.99+

Dave SchneiderPERSON

0.99+

2014DATE

0.99+

2012DATE

0.99+

Frank SlumanPERSON

0.99+

Mike ScarpelliPERSON

0.99+

Dan MageePERSON

0.99+

Frank SlootPERSON

0.99+

January 17DATE

0.99+

EMCORGANIZATION

0.99+

July 19DATE

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

ServiceNowORGANIZATION

0.99+

81%QUANTITY

0.99+

AWSORGANIZATION

0.99+

1.2 billionQUANTITY

0.99+

JulyDATE

0.99+

30QUANTITY

0.99+

OracleORGANIZATION

0.99+

52QUANTITY

0.99+

HPORGANIZATION

0.99+

Larry EllisonPERSON

0.99+

360 degreeQUANTITY

0.99+

GoogleORGANIZATION

0.99+

Joe TucciPERSON

0.99+

20 billionQUANTITY

0.99+

fiveQUANTITY

0.99+

JanuaryDATE

0.99+

PatPERSON

0.99+

HadoopTITLE

0.99+

Silicon ValleyLOCATION

0.99+

ExcelTITLE

0.99+

10%QUANTITY

0.99+

70 plus billion dollarsQUANTITY

0.99+

AmazonORGANIZATION

0.99+

35QUANTITY

0.99+

first slideQUANTITY

0.99+

Dave VolantePERSON

0.99+

about 4,500 end-usersQUANTITY

0.99+

over 900 million dollarsQUANTITY

0.99+

Boston MassachusettsLOCATION

0.99+

two and a half billion dollarsQUANTITY

0.98+

TaraORGANIZATION

0.98+

first applianceQUANTITY

0.98+

MikePERSON

0.98+

dozens of companiesQUANTITY

0.98+

over a thousandQUANTITY

0.98+

TimPERSON

0.98+

Jay Marshall, Neural Magic | AWS Startup Showcase S3E1


 

(upbeat music) >> Hello, everyone, and welcome to theCUBE's presentation of the "AWS Startup Showcase." This is season three, episode one. The focus of this episode is AI/ML: Top Startups Building Foundational Models, Infrastructure, and AI. It's great topics, super-relevant, and it's part of our ongoing coverage of startups in the AWS ecosystem. I'm your host, John Furrier, with theCUBE. Today, we're excited to be joined by Jay Marshall, VP of Business Development at Neural Magic. Jay, thanks for coming on theCUBE. >> Hey, John, thanks so much. Thanks for having us. >> We had a great CUBE conversation with you guys. This is very much about the company focuses. It's a feature presentation for the "Startup Showcase," and the machine learning at scale is the topic, but in general, it's more, (laughs) and we should call it "Machine Learning and AI: How to Get Started," because everybody is retooling their business. Companies that aren't retooling their business right now with AI first will be out of business, in my opinion. You're seeing massive shift. This is really truly the beginning of the next-gen machine learning AI trend. It's really seeing ChatGPT. Everyone sees that. That went mainstream. But this is just the beginning. This is scratching the surface of this next-generation AI with machine learning powering it, and with all the goodness of cloud, cloud scale, and how horizontally scalable it is. The resources are there. You got the Edge. Everything's perfect for AI 'cause data infrastructure's exploding in value. AI is just the applications. This is a super topic, so what do you guys see in this general area of opportunities right now in the headlines? And I'm sure you guys' phone must be ringing off the hook, metaphorically speaking, or emails and meetings and Zooms. What's going on over there at Neural Magic? >> No, absolutely, and you pretty much nailed most of it. I think that, you know, my background, we've seen for the last 20-plus years. Even just getting enterprise applications kind of built and delivered at scale, obviously, amazing things with AWS and the cloud to help accelerate that. And we just kind of figured out in the last five or so years how to do that productively and efficiently, kind of from an operations perspective. Got development and operations teams. We even came up with DevOps, right? But now, we kind of have this new kind of persona and new workload that developers have to talk to, and then it has to be deployed on those ITOps solutions. And so you pretty much nailed it. Folks are saying, "Well, how do I do this?" These big, generational models or foundational models, as we're calling them, they're great, but enterprises want to do that with their data, on their infrastructure, at scale, at the edge. So for us, yeah, we're helping enterprises accelerate that through optimizing models and then delivering them at scale in a more cost-effective fashion. >> Yeah, and I think one of the things, the benefits of OpenAI we saw, was not only is it open source, then you got also other models that are more proprietary, is that it shows the world that this is really happening, right? It's a whole nother level, and there's also new landscape kind of maps coming out. You got the generative AI, and you got the foundational models, large LLMs. Where do you guys fit into the landscape? Because you guys are in the middle of this. How do you talk to customers when they say, "I'm going down this road. I need help. I'm going to stand this up." This new AI infrastructure and applications, where do you guys fit in the landscape? >> Right, and really, the answer is both. I think today, when it comes to a lot of what for some folks would still be considered kind of cutting edge around computer vision and natural language processing, a lot of our optimization tools and our runtime are based around most of the common computer vision and natural language processing models. So your YOLOs, your BERTs, you know, your DistilBERTs and what have you, so we work to help optimize those, again, who've gotten great performance and great value for customers trying to get those into production. But when you get into the LLMs, and you mentioned some of the open source components there, our research teams have kind of been right in the trenches with those. So kind of the GPT open source equivalent being OPT, being able to actually take, you know, a multi-$100 billion parameter model and sparsify that or optimize that down, shaving away a ton of parameters, and being able to run it on smaller infrastructure. So I think the evolution here, you know, all this stuff came out in the last six months in terms of being turned loose into the wild, but we're staying in the trenches with folks so that we can help optimize those as well and not require, again, the heavy compute, the heavy cost, the heavy power consumption as those models evolve as well. So we're staying right in with everybody while they're being built, but trying to get folks into production today with things that help with business value today. >> Jay, I really appreciate you coming on theCUBE, and before we came on camera, you said you just were on a customer call. I know you got a lot of activity. What specific things are you helping enterprises solve? What kind of problems? Take us through the spectrum from the beginning, people jumping in the deep end of the pool, some people kind of coming in, starting out slow. What are the scale? Can you scope the kind of use cases and problems that are emerging that people are calling you for? >> Absolutely, so I think if I break it down to kind of, like, your startup, or I maybe call 'em AI native to kind of steal from cloud native years ago, that group, it's pretty much, you know, part and parcel for how that group already runs. So if you have a data science team and an ML engineering team, you're building models, you're training models, you're deploying models. You're seeing firsthand the expense of starting to try to do that at scale. So it's really just a pure operational efficiency play. They kind of speak natively to our tools, which we're doing in the open source. So it's really helping, again, with the optimization of the models they've built, and then, again, giving them an alternative to expensive proprietary hardware accelerators to have to run them. Now, on the enterprise side, it varies, right? You have some kind of AI native folks there that already have these teams, but you also have kind of, like, AI curious, right? Like, they want to do it, but they don't really know where to start, and so for there, we actually have an open source toolkit that can help you get into this optimization, and then again, that runtime, that inferencing runtime, purpose-built for CPUs. It allows you to not have to worry, again, about do I have a hardware accelerator available? How do I integrate that into my application stack? If I don't already know how to build this into my infrastructure, does my ITOps teams, do they know how to do this, and what does that runway look like? How do I cost for this? How do I plan for this? When it's just x86 compute, we've been doing that for a while, right? So it obviously still requires more, but at least it's a little bit more predictable. >> It's funny you mentioned AI native. You know, born in the cloud was a phrase that was out there. Now, you have startups that are born in AI companies. So I think you have this kind of cloud kind of vibe going on. You have lift and shift was a big discussion. Then you had cloud native, kind of in the cloud, kind of making it all work. Is there a existing set of things? People will throw on this hat, and then what's the difference between AI native and kind of providing it to existing stuff? 'Cause we're a lot of people take some of these tools and apply it to either existing stuff almost, and it's not really a lift and shift, but it's kind of like bolting on AI to something else, and then starting with AI first or native AI. >> Absolutely. It's a- >> How would you- >> It's a great question. I think that probably, where I'd probably pull back to kind of allow kind of retail-type scenarios where, you know, for five, seven, nine years or more even, a lot of these folks already have data science teams, you know? I mean, they've been doing this for quite some time. The difference is the introduction of these neural networks and deep learning, right? Those kinds of models are just a little bit of a paradigm shift. So, you know, I obviously was trying to be fun with the term AI native, but I think it's more folks that kind of came up in that neural network world, so it's a little bit more second nature, whereas I think for maybe some traditional data scientists starting to get into neural networks, you have the complexity there and the training overhead, and a lot of the aspects of getting a model finely tuned and hyperparameterization and all of these aspects of it. It just adds a layer of complexity that they're just not as used to dealing with. And so our goal is to help make that easy, and then of course, make it easier to run anywhere that you have just kind of standard infrastructure. >> Well, the other point I'd bring out, and I'd love to get your reaction to, is not only is that a neural network team, people who have been focused on that, but also, if you look at some of the DataOps lately, AIOps markets, a lot of data engineering, a lot of scale, folks who have been kind of, like, in that data tsunami cloud world are seeing, they kind of been in this, right? They're, like, been experiencing that. >> No doubt. I think it's funny the data lake concept, right? And you got data oceans now. Like, the metaphors just keep growing on us, but where it is valuable in terms of trying to shift the mindset, I've always kind of been a fan of some of the naming shift. I know with AWS, they always talk about purpose-built databases. And I always liked that because, you know, you don't have one database that can do everything. Even ones that say they can, like, you still have to do implementation detail differences. So sitting back and saying, "What is my use case, and then which database will I use it for?" I think it's kind of similar here. And when you're building those data teams, if you don't have folks that are doing data engineering, kind of that data harvesting, free processing, you got to do all that before a model's even going to care about it. So yeah, it's definitely a central piece of this as well, and again, whether or not you're going to be AI negative as you're making your way to kind of, you know, on that journey, you know, data's definitely a huge component of it. >> Yeah, you would have loved our Supercloud event we had. Talk about naming and, you know, around data meshes was talked about a lot. You're starting to see the control plane layers of data. I think that was the beginning of what I saw as that data infrastructure shift, to be horizontally scalable. So I have to ask you, with Neural Magic, when your customers and the people that are prospects for you guys, they're probably asking a lot of questions because I think the general thing that we see is, "How do I get started? Which GPU do I use?" I mean, there's a lot of things that are kind of, I won't say technical or targeted towards people who are living in that world, but, like, as the mainstream enterprises come in, they're going to need a playbook. What do you guys see, what do you guys offer your clients when they come in, and what do you recommend? >> Absolutely, and I think where we hook in specifically tends to be on the training side. So again, I've built a model. Now, I want to really optimize that model. And then on the runtime side when you want to deploy it, you know, we run that optimized model. And so that's where we're able to provide. We even have a labs offering in terms of being able to pair up our engineering teams with a customer's engineering teams, and we can actually help with most of that pipeline. So even if it is something where you have a dataset and you want some help in picking a model, you want some help training it, you want some help deploying that, we can actually help there as well. You know, there's also a great partner ecosystem out there, like a lot of folks even in the "Startup Showcase" here, that extend beyond into kind of your earlier comment around data engineering or downstream ITOps or the all-up MLOps umbrella. So we can absolutely engage with our labs, and then, of course, you know, again, partners, which are always kind of key to this. So you are spot on. I think what's happened with the kind of this, they talk about a hockey stick. This is almost like a flat wall now with the rate of innovation right now in this space. And so we do have a lot of folks wanting to go straight from curious to native. And so that's definitely where the partner ecosystem comes in so hard 'cause there just isn't anybody or any teams out there that, I literally do from, "Here's my blank database, and I want an API that does all the stuff," right? Like, that's a big chunk, but we can definitely help with the model to delivery piece. >> Well, you guys are obviously a featured company in this space. Talk about the expertise. A lot of companies are like, I won't say faking it till they make it. You can't really fake security. You can't really fake AI, right? So there's going to be a learning curve. They'll be a few startups who'll come out of the gate early. You guys are one of 'em. Talk about what you guys have as expertise as a company, why you're successful, and what problems do you solve for customers? >> No, appreciate that. Yeah, we actually, we love to tell the story of our founder, Nir Shavit. So he's a 20-year professor at MIT. Actually, he was doing a lot of work on kind of multicore processing before there were even physical multicores, and actually even did a stint in computational neurobiology in the 2010s, and the impetus for this whole technology, has a great talk on YouTube about it, where he talks about the fact that his work there, he kind of realized that the way neural networks encode and how they're executed by kind of ramming data layer by layer through these kind of HPC-style platforms, actually was not analogous to how the human brain actually works. So we're on one side, we're building neural networks, and we're trying to emulate neurons. We're not really executing them that way. So our team, which one of the co-founders, also an ex-MIT, that was kind of the birth of why can't we leverage this super-performance CPU platform, which has those really fat, fast caches attached to each core, and actually start to find a way to break that model down in a way that I can execute things in parallel, not having to do them sequentially? So it is a lot of amazing, like, talks and stuff that show kind of the magic, if you will, a part of the pun of Neural Magic, but that's kind of the foundational layer of all the engineering that we do here. And in terms of how we're able to bring it to reality for customers, I'll give one customer quote where it's a large retailer, and it's a people-counting application. So a very common application. And that customer's actually been able to show literally double the amount of cameras being run with the same amount of compute. So for a one-to-one perspective, two-to-one, business leaders usually like that math, right? So we're able to show pure cost savings, but even performance-wise, you know, we have some of the common models like your ResNets and your YOLOs, where we can actually even perform better than hardware-accelerated solutions. So we're trying to do, I need to just dumb it down to better, faster, cheaper, but from a commodity perspective, that's where we're accelerating. >> That's not a bad business model. Make things easier to use, faster, and reduce the steps it takes to do stuff. So, you know, that's always going to be a good market. Now, you guys have DeepSparse, which we've talked about on our CUBE conversation prior to this interview, delivers ML models through the software so the hardware allows for a decoupling, right? >> Yep. >> Which is going to drive probably a cost advantage. Also, it's also probably from a deployment standpoint it must be easier. Can you share the benefits? Is it a cost side? Is it more of a deployment? What are the benefits of the DeepSparse when you guys decouple the software from the hardware on the ML models? >> No you actually, you hit 'em both 'cause that really is primarily the value. Because ultimately, again, we're so early. And I came from this world in a prior life where I'm doing Java development, WebSphere, WebLogic, Tomcat open source, right? When we were trying to do innovation, we had innovation buckets, 'cause everybody wanted to be on the web and have their app and a browser, right? We got all the money we needed to build something and show, hey, look at the thing on the web, right? But when you had to get in production, that was the challenge. So to what you're speaking to here, in this situation, we're able to show we're just a Python package. So whether you just install it on the operating system itself, or we also have a containerized version you can drop on any container orchestration platform, so ECS or EKS on AWS. And so you get all the auto-scaling features. So when you think about that kind of a world where you have everything from real-time inferencing to kind of after hours batch processing inferencing, the fact that you can auto scale that hardware up and down and it's CPU based, so you're paying by the minute instead of maybe paying by the hour at a lower cost shelf, it does everything from pure cost to, again, I can have my standard IT team say, "Hey, here's the Kubernetes in the container," and it just runs on the infrastructure we're already managing. So yeah, operational, cost and again, and many times even performance. (audio warbles) CPUs if I want to. >> Yeah, so that's easier on the deployment too. And you don't have this kind of, you know, blank check kind of situation where you don't know what's on the backend on the cost side. >> Exactly. >> And you control the actual hardware and you can manage that supply chain. >> And keep in mind, exactly. Because the other thing that sometimes gets lost in the conversation, depending on where a customer is, some of these workloads, like, you know, you and I remember a world where even like the roundtrip to the cloud and back was a problem for folks, right? We're used to extremely low latency. And some of these workloads absolutely also adhere to that. But there's some workloads where the latency isn't as important. And we actually even provide the tuning. Now, if we're giving you five milliseconds of latency and you don't need that, you can tune that back. So less CPU, lower cost. Now, throughput and other things come into play. But that's the kind of configurability and flexibility we give for operations. >> All right, so why should I call you if I'm a customer or prospect Neural Magic, what problem do I have or when do I know I need you guys? When do I call you in and what does my environment look like? When do I know? What are some of the signals that would tell me that I need Neural Magic? >> No, absolutely. So I think in general, any neural network, you know, the process I mentioned before called sparcification, it's, you know, an optimization process that we specialize in. Any neural network, you know, can be sparcified. So I think if it's a deep-learning neural network type model. If you're trying to get AI into production, you have cost concerns even performance-wise. I certainly hate to be too generic and say, "Hey, we'll talk to everybody." But really in this world right now, if it's a neural network, it's something where you're trying to get into production, you know, we are definitely offering, you know, kind of an at-scale performant deployable solution for deep learning models. >> So neural network you would define as what? Just devices that are connected that need to know about each other? What's the state-of-the-art current definition of neural network for customers that may think they have a neural network or might not know they have a neural network architecture? What is that definition for neural network? >> That's a great question. So basically, machine learning models that fall under this kind of category, you hear about transformers a lot, or I mentioned about YOLO, the YOLO family of computer vision models, or natural language processing models like BERT. If you have a data science team or even developers, some even regular, I used to call myself a nine to five developer 'cause I worked in the enterprise, right? So like, hey, we found a new open source framework, you know, I used to use Spring back in the day and I had to go figure it out. There's developers that are pulling these models down and they're figuring out how to get 'em into production, okay? So I think all of those kinds of situations, you know, if it's a machine learning model of the deep learning variety that's, you know, really specifically where we shine. >> Okay, so let me pretend I'm a customer for a minute. I have all these videos, like all these transcripts, I have all these people that we've interviewed, CUBE alumnis, and I say to my team, "Let's AI-ify, sparcify theCUBE." >> Yep. >> What do I do? I mean, do I just like, my developers got to get involved and they're going to be like, "Well, how do I upload it to the cloud? Do I use a GPU?" So there's a thought process. And I think a lot of companies are going through that example of let's get on this AI, how can it help our business? >> Absolutely. >> What does that progression look like? Take me through that example. I mean, I made up theCUBE example up, but we do have a lot of data. We have large data models and we have people and connect to the internet and so we kind of seem like there's a neural network. I think every company might have a neural network in place. >> Well, and I was going to say, I think in general, you all probably do represent even the standard enterprise more than most. 'Cause even the enterprise is going to have a ton of video content, a ton of text content. So I think it's a great example. So I think that that kind of sea or I'll even go ahead and use that term data lake again, of data that you have, you're probably going to want to be setting up kind of machine learning pipelines that are going to be doing all of the pre-processing from kind of the raw data to kind of prepare it into the format that say a YOLO would actually use or let's say BERT for natural language processing. So you have all these transcripts, right? So we would do a pre-processing path where we would create that into the file format that BERT, the machine learning model would know how to train off of. So that's kind of all the pre-processing steps. And then for training itself, we actually enable what's called sparse transfer learning. So that's transfer learning is a very popular method of doing training with existing models. So we would be able to retrain that BERT model with your transcript data that we have now done the pre-processing with to get it into the proper format. And now we have a BERT natural language processing model that's been trained on your data. And now we can deploy that onto DeepSparse runtime so that now you can ask that model whatever questions, or I should say pass, you're not going to ask it those kinds of questions ChatGPT, although we can do that too. But you're going to pass text through the BERT model and it's going to give you answers back. It could be things like sentiment analysis or text classification. You just call the model, and now when you pass text through it, you get the answers better, faster or cheaper. I'll use that reference again. >> Okay, we can create a CUBE bot to give us questions on the fly from the the AI bot, you know, from our previous guests. >> Well, and I will tell you using that as an example. So I had mentioned OPT before, kind of the open source version of ChatGPT. So, you know, typically that requires multiple GPUs to run. So our research team, I may have mentioned earlier, we've been able to sparcify that over 50% already and run it on only a single GPU. And so in that situation, you could train OPT with that corpus of data and do exactly what you say. Actually we could use Alexa, we could use Alexa to actually respond back with voice. How about that? We'll do an API call and we'll actually have an interactive Alexa-enabled bot. >> Okay, we're going to be a customer, let's put it on the list. But this is a great example of what you guys call software delivered AI, a topic we chatted about on theCUBE conversation. This really means this is a developer opportunity. This really is the convergence of the data growth, the restructuring, how data is going to be horizontally scalable, meets developers. So this is an AI developer model going on right now, which is kind of unique. >> It is, John, I will tell you what's interesting. And again, folks don't always think of it this way, you know, the AI magical goodness is now getting pushed in the middle where the developers and IT are operating. And so it again, that paradigm, although for some folks seem obvious, again, if you've been around for 20 years, that whole all that plumbing is a thing, right? And so what we basically help with is when you deploy the DeepSparse runtime, we have a very rich API footprint. And so the developers can call the API, ITOps can run it, or to your point, it's developer friendly enough that you could actually deploy our off-the-shelf models. We have something called the SparseZoo where we actually publish pre-optimized or pre-sparcified models. And so developers could literally grab those right off the shelf with the training they've already had and just put 'em right into their applications and deploy them as containers. So yeah, we enable that for sure as well. >> It's interesting, DevOps was infrastructure as code and we had a last season, a series on data as code, which we kind of coined. This is data as code. This is a whole nother level of opportunity where developers just want to have programmable data and apps with AI. This is a whole new- >> Absolutely. >> Well, absolutely great, great stuff. Our news team at SiliconANGLE and theCUBE said you guys had a little bit of a launch announcement you wanted to make here on the "AWS Startup Showcase." So Jay, you have something that you want to launch here? >> Yes, and thank you John for teeing me up. So I'm going to try to put this in like, you know, the vein of like an AWS, like main stage keynote launch, okay? So we're going to try this out. So, you know, a lot of our product has obviously been built on top of x86. I've been sharing that the past 15 minutes or so. And with that, you know, we're seeing a lot of acceleration for folks wanting to run on commodity infrastructure. But we've had customers and prospects and partners tell us that, you know, ARM and all of its kind of variance are very compelling, both cost performance-wise and also obviously with Edge. And wanted to know if there was anything we could do from a runtime perspective with ARM. And so we got the work and, you know, it's a hard problem to solve 'cause the instructions set for ARM is very different than the instruction set for x86, and our deep tensor column technology has to be able to work with that lower level instruction spec. But working really hard, the engineering team's been at it and we are happy to announce here at the "AWS Startup Showcase," that DeepSparse inference now has, or inference runtime now has support for AWS Graviton instances. So it's no longer just x86, it is also ARM and that obviously also opens up the door to Edge and further out the stack so that optimize once run anywhere, we're not going to open up. So it is an early access. So if you go to neuralmagic.com/graviton, you can sign up for early access, but we're excited to now get into the ARM side of the fence as well on top of Graviton. >> That's awesome. Our news team is going to jump on that news. We'll get it right up. We get a little scoop here on the "Startup Showcase." Jay Marshall, great job. That really highlights the flexibility that you guys have when you decouple the software from the hardware. And again, we're seeing open source driving a lot more in AI ops now with with machine learning and AI. So to me, that makes a lot of sense. And congratulations on that announcement. Final minute or so we have left, give a summary of what you guys are all about. Put a plug in for the company, what you guys are looking to do. I'm sure you're probably hiring like crazy. Take the last few minutes to give a plug for the company and give a summary. >> No, I appreciate that so much. So yeah, joining us out neuralmagic.com, you know, part of what we didn't spend a lot of time here, our optimization tools, we are doing all of that in the open source. It's called SparseML and I mentioned SparseZoo briefly. So we really want the data scientists community and ML engineering community to join us out there. And again, the DeepSparse runtime, it's actually free to use for trial purposes and for personal use. So you can actually run all this on your own laptop or on an AWS instance of your choice. We are now live in the AWS marketplace. So push button, deploy, come try us out and reach out to us on neuralmagic.com. And again, sign up for the Graviton early access. >> All right, Jay Marshall, Vice President of Business Development Neural Magic here, talking about performant, cost effective machine learning at scale. This is season three, episode one, focusing on foundational models as far as building data infrastructure and AI, AI native. I'm John Furrier with theCUBE. Thanks for watching. (bright upbeat music)

Published Date : Mar 9 2023

SUMMARY :

of the "AWS Startup Showcase." Thanks for having us. and the machine learning and the cloud to help accelerate that. and you got the foundational So kind of the GPT open deep end of the pool, that group, it's pretty much, you know, So I think you have this kind It's a- and a lot of the aspects of and I'd love to get your reaction to, And I always liked that because, you know, that are prospects for you guys, and you want some help in picking a model, Talk about what you guys have that show kind of the magic, if you will, and reduce the steps it takes to do stuff. when you guys decouple the the fact that you can auto And you don't have this kind of, you know, the actual hardware and you and you don't need that, neural network, you know, of situations, you know, CUBE alumnis, and I say to my team, and they're going to be like, and connect to the internet and it's going to give you answers back. you know, from our previous guests. and do exactly what you say. of what you guys call enough that you could actually and we had a last season, that you want to launch here? And so we got the work and, you know, flexibility that you guys have So you can actually run Vice President of Business

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JayPERSON

0.99+

Jay MarshallPERSON

0.99+

John FurrierPERSON

0.99+

JohnPERSON

0.99+

AWSORGANIZATION

0.99+

fiveQUANTITY

0.99+

Nir ShavitPERSON

0.99+

20-yearQUANTITY

0.99+

AlexaTITLE

0.99+

2010sDATE

0.99+

sevenQUANTITY

0.99+

PythonTITLE

0.99+

MITORGANIZATION

0.99+

each coreQUANTITY

0.99+

Neural MagicORGANIZATION

0.99+

JavaTITLE

0.99+

YouTubeORGANIZATION

0.99+

TodayDATE

0.99+

nine yearsQUANTITY

0.98+

bothQUANTITY

0.98+

BERTTITLE

0.98+

theCUBEORGANIZATION

0.98+

ChatGPTTITLE

0.98+

20 yearsQUANTITY

0.98+

over 50%QUANTITY

0.97+

second natureQUANTITY

0.96+

todayDATE

0.96+

ARMORGANIZATION

0.96+

oneQUANTITY

0.95+

DeepSparseTITLE

0.94+

neuralmagic.com/gravitonOTHER

0.94+

SiliconANGLEORGANIZATION

0.94+

WebSphereTITLE

0.94+

nineQUANTITY

0.94+

firstQUANTITY

0.93+

Startup ShowcaseEVENT

0.93+

five millisecondsQUANTITY

0.92+

AWS Startup ShowcaseEVENT

0.91+

twoQUANTITY

0.9+

YOLOORGANIZATION

0.89+

CUBEORGANIZATION

0.88+

OPTTITLE

0.88+

last six monthsDATE

0.88+

season threeQUANTITY

0.86+

doubleQUANTITY

0.86+

one customerQUANTITY

0.86+

SupercloudEVENT

0.86+

one sideQUANTITY

0.85+

VicePERSON

0.85+

x86OTHER

0.83+

AI/ML: Top Startups Building Foundational ModelsTITLE

0.82+

ECSTITLE

0.81+

$100 billionQUANTITY

0.81+

DevOpsTITLE

0.81+

WebLogicTITLE

0.8+

EKSTITLE

0.8+

a minuteQUANTITY

0.8+

neuralmagic.comOTHER

0.79+

Daren Brabham & Erik Bradley | What the Spending Data Tells us About Supercloud


 

(gentle synth music) (music ends) >> Welcome back to Supercloud 2, an open industry collaboration between technologists, consultants, analysts, and of course practitioners to help shape the future of cloud. At this event, one of the key areas we're exploring is the intersection of cloud and data. And how building value on top of hyperscale clouds and across clouds is evolving, a concept of course we call "Supercloud". And we're pleased to welcome our friends from Enterprise Technology research, Erik Bradley and Darren Brabham. Guys, thanks for joining us, great to see you. we love to bring the data into these conversations. >> Thank you for having us, Dave, I appreciate it. >> Yeah, thanks. >> You bet. And so, let me do the setup on what is Supercloud. It's a concept that we've floated, Before re:Invent 2021, based on the idea that cloud infrastructure is becoming ubiquitous, incredibly powerful, but there's a lack of standards across the big three clouds. That creates friction. So we defined over the period of time, you know, better part of a year, a set of essential elements, deployment models for so-called supercloud, which create this common experience for specific cloud services that, of course, again, span multiple clouds and even on-premise data. So Erik, with that as background, I wonder if you could add your general thoughts on the term supercloud, maybe play proxy for the CIO community, 'cause you do these round tables, you talk to these guys all the time, you gather a lot of amazing information from senior IT DMs that compliment your survey. So what are your thoughts on the term and the concept? >> Yeah, sure. I'll even go back to last year when you and I did our predictions panel, right? And we threw it out there. And to your point, you know, there's some haters. Anytime you throw out a new term, "Is it marketing buzz? Is it worth it? Why are you even doing it?" But you know, from my own perspective, and then also speaking to the IT DMs that we interview on a regular basis, this is just a natural evolution. It's something that's inevitable in enterprise tech, right? The internet was not built for what it has become. It was never intended to be the underlying infrastructure of our daily lives and work. The cloud also was not built to be what it's become. But where we're at now is, we have to figure out what the cloud is and what it needs to be to be scalable, resilient, secure, and have the governance wrapped around it. And to me that's what supercloud is. It's a way to define operantly, what the next generation, the continued iteration and evolution of the cloud and what its needs to be. And that's what the supercloud means to me. And what depends, if you want to call it metacloud, supercloud, it doesn't matter. The point is that we're trying to define the next layer, the next future of work, which is inevitable in enterprise tech. Now, from the IT DM perspective, I have two interesting call outs. One is from basically a senior developer IT architecture and DevSecOps who says he uses the term all the time. And the reason he uses the term, is that because multi-cloud has a stigma attached to it, when he is talking to his business executives. (David chuckles) the stigma is because it's complex and it's expensive. So he switched to supercloud to better explain to his business executives and his CFO and his CIO what he's trying to do. And we can get into more later about what it means to him. But the inverse of that, of course, is a good CSO friend of mine for a very large enterprise says the concern with Supercloud is the reduction of complexity. And I'll explain, he believes anything that takes the requirement of specific expertise out of the equation, even a little bit, as a CSO worries him. So as you said, David, always two sides to the coin, but I do believe supercloud is a relevant term, and it is necessary because the cloud is continuing to be defined. >> You know, that's really interesting too, 'cause you know, Darren, we use Snowflake a lot as an example, sort of early supercloud, and you think from a security standpoint, we've always pushed Amazon and, "Are you ever going to kind of abstract the complexity away from all these primitives?" and their position has always been, "Look, if we produce these primitives, and offer these primitives, we we can move as the market moves. When you abstract, then it becomes harder to peel the layers." But Darren, from a data standpoint, like I say, we use Snowflake a lot. I think of like Tim Burners-Lee when Web 2.0 came out, he said, "Well this is what the internet was always supposed to be." So in a way, you know, supercloud is maybe what multi-cloud was supposed to be. But I mean, you think about data sharing, Darren, across clouds, it's always been a challenge. Snowflake always, you know, obviously trying to solve that problem, as are others. But what are your thoughts on the concept? >> Yeah, I think the concept fits, right? It is reflective of, it's a paradigm shift, right? Things, as a pendulum have swung back and forth between needing to piece together a bunch of different tools that have specific unique use cases and they're best in breed in what they do. And then focusing on the duct tape that holds 'em all together and all the engineering complexity and skill, it shifted from that end of the pendulum all the way back to, "Let's streamline this, let's simplify it. Maybe we have budget crunches and we need to consolidate tools or eliminate tools." And so then you kind of see this back and forth over time. And with data and analytics for instance, a lot of organizations were trying to bring the data closer to the business. That's where we saw self-service analytics coming in. And tools like Snowflake, what they did was they helped point to different databases, they helped unify data, and organize it in a single place that was, you know, in a sense neutral, away from a single cloud vendor or a single database, and allowed the business to kind of be more flexible in how it brought stuff together and provided it out to the business units. So Snowflake was an example of one of those times where we pulled back from the granular, multiple points of the spear, back to a simple way to do things. And I think Snowflake has continued to kind of keep that mantle to a degree, and we see other tools trying to do that, but that's all it is. It's a paradigm shift back to this kind of meta abstraction layer that kind of simplifies what is the reality, that you need a complex multi-use case, multi-region way of doing business. And it sort of reflects the reality of that. >> And you know, to me it's a spectrum. As part of Supercloud 2, we're talking to a number of of practitioners, Ionis Pharmaceuticals, US West, we got Walmart. And it's a spectrum, right? In some cases the practitioner's saying, "You know, the way I solve multi-cloud complexity is mono-cloud, I just do one cloud." (laughs) Others like Walmart are saying, "Hey, you know, we actually are building an abstraction layer ourselves, take advantage of it." So my general question to both of you is, is this a concept, is the lack of standards across clouds, you know, really a problem, you know, or is supercloud a solution looking for a problem? Or do you hear from practitioners that "No, this is really an issue, we have to bring together a set of standards to sort of unify our cloud estates." >> Allow me to answer that at a higher level, and then we're going to hand it over to Dr. Brabham because he is a little bit more detailed on the realtime streaming analytics use cases, which I think is where we're going to get to. But to answer that question, it really depends on the size and the complexity of your business. At the very large enterprise, Dave, Yes, a hundred percent. This needs to happen. There is complexity, there is not only complexity in the compute and actually deploying the applications, but the governance and the security around them. But for lower end or, you know, business use cases, and for smaller businesses, it's a little less necessary. You certainly don't need to have all of these. Some of the things that come into mind from the interviews that Darren and I have done are, you know, financial services, if you're doing real-time trading, anything that has real-time data metrics involved in your transactions, is going to be necessary. And another use case that we hear about is in online travel agencies. So I think it is very relevant, the complexity does need to be solved, and I'll allow Darren to explain a little bit more about how that's used from an analytics perspective. >> Yeah, go for it. >> Yeah, exactly. I mean, I think any modern, you know, multinational company that's going to have a footprint in the US and Europe, in China, or works in different areas like manufacturing, where you're probably going to have on-prem instances that will stay on-prem forever, for various performance reasons. You have these complicated governance and security and regulatory issues. So inherently, I think, large multinational companies and or companies that are in certain areas like finance or in, you know, online e-commerce, or things that need real-time data, they inherently are going to have a very complex environment that's going to need to be managed in some kind of cleaner way. You know, they're looking for one door to open, one pane of glass to look at, one thing to do to manage these multi points. And, streaming's a good example of that. I mean, not every organization has a real-time streaming use case, and may not ever, but a lot of organizations do, a lot of industries do. And so there's this need to use, you know, they want to use open-source tools, they want to use Apache Kafka for instance. They want to use different megacloud vendors offerings, like Google Pub/Sub or you know, Amazon Kinesis Firehose. They have all these different pieces they want to use for different use cases at different stages of maturity or proof of concept, you name it. They're going to have to have this complexity. And I think that's why we're seeing this need, to have sort of this supercloud concept, to juggle all this, to wrangle all of it. 'Cause the reality is, it's complex and you have to simplify it somehow. >> Great, thanks you guys. All right, let's bring up the graphic, and take a look. Anybody who follows the breaking analysis, which is co-branded with ETR Cube Insights powered by ETR, knows we like to bring data to the table. ETR does amazing survey work every quarter, 1200 plus 1500 practitioners that that answer a number of questions. The vertical axis here is net score, which is ETR's proprietary methodology, which is a measure of spending momentum, spending velocity. And the horizontal axis here is overlap, but it's the presence pervasiveness, and the dataset, the ends, that table insert on the bottom right shows you how the dots are plotted, the net score and then the ends in the survey. And what we've done is we've plotted a bunch of the so-called supercloud suspects, let's start in the upper right, the cloud platforms. Without these hyperscale clouds, you can't have a supercloud. And as always, Azure and AWS, up and to the right, it's amazing we're talking about, you know, 80 plus billion dollar company in AWS. Azure's business is, if you just look at the IaaS is in the 50 billion range, I mean it's just amazing to me the net scores here. Anything above 40% we consider highly elevated. And you got Azure and you got Snowflake, Databricks, HashiCorp, we'll get to them. And you got AWS, you know, right up there at that size, it's quite amazing. With really big ends as well, you know, 700 plus ends in the survey. So, you know, kind of half the survey actually has these platforms. So my question to you guys is, what are you seeing in terms of cloud adoption within the big three cloud players? I wonder if you could could comment, maybe Erik, you could start. >> Yeah, sure. Now we're talking data, now I'm happy. So yeah, we'll get into some of it. Right now, the January, 2023 TSIS is approaching 1500 survey respondents. One caveat, it's not closed yet, it will close on Friday, but with an end that big we are over statistically significant. We also recently did a cloud survey, and there's a couple of key points on that I want to get into before we get into individual vendors. What we're seeing here, is that annual spend on cloud infrastructure is expected to grow at almost a 70% CAGR over the next three years. The percentage of those workloads for cloud infrastructure are expected to grow over 70% as three years as well. And as you mentioned, Azure and AWS are still dominant. However, we're seeing some share shift spreading around a little bit. Now to get into the individual vendors you mentioned about, yes, Azure is still number one, AWS is number two. What we're seeing, which is incredibly interesting, CloudFlare is number three. It's actually beating GCP. That's the first time we've seen it. What I do want to state, is this is on net score only, which is our measure of spending intentions. When you talk about actual pervasion in the enterprise, it's not even close. But from a spending velocity intention point of view, CloudFlare is now number three above GCP, and even Salesforce is creeping up to be at GCPs level. So what we're seeing here, is a continued domination by Azure and AWS, but some of these other players that maybe might fit into your moniker. And I definitely want to talk about CloudFlare more in a bit, but I'm going to stop there. But what we're seeing is some of these other players that fit into your Supercloud moniker, are starting to creep up, Dave. >> Yeah, I just want to clarify. So as you also know, we track IaaS and PaaS revenue and we try to extract, so AWS reports in its quarterly earnings, you know, they're just IaaS and PaaS, they don't have a SaaS play, a little bit maybe, whereas Microsoft and Google include their applications and so we extract those out and if you do that, AWS is bigger, but in the surveys, you know, customers, they see cloud, SaaS to them as cloud. So that's one of the reasons why you see, you know, Microsoft as larger in pervasion. If you bring up that survey again, Alex, the survey results, you see them further to the right and they have higher spending momentum, which is consistent with what you see in the earnings calls. Now, interesting about CloudFlare because the CEO of CloudFlare actually, and CloudFlare itself uses the term supercloud basically saying, "Hey, we're building a new type of internet." So what are your thoughts? Do you have additional information on CloudFlare, Erik that you want to share? I mean, you've seen them pop up. I mean this is a really interesting company that is pretty forward thinking and vocal about how it's disrupting the industry. >> Sure, we've been tracking 'em for a long time, and even from the disruption of just a traditional CDN where they took down Akamai and what they're doing. But for me, the definition of a true supercloud provider can't just be one instance. You have to have multiple. So it's not just the cloud, it's networking aspect on top of it, it's also security. And to me, CloudFlare is the only one that has all of it. That they actually have the ability to offer all of those things. Whereas you look at some of the other names, they're still piggybacking on the infrastructure or platform as a service of the hyperscalers. CloudFlare does not need to, they actually have the cloud, the networking, and the security all themselves. So to me that lends credibility to their own internal usage of that moniker Supercloud. And also, again, just what we're seeing right here that their net score is now creeping above AGCP really does state it. And then just one real last thing, one of the other things we do in our surveys is we track adoption and replacement reasoning. And when you look at Cloudflare's adoption rate, which is extremely high, it's based on technical capabilities, the breadth of their feature set, it's also based on what we call the ability to avoid stack alignment. So those are again, really supporting reasons that makes CloudFlare a top candidate for your moniker of supercloud. >> And they've also announced an object store (chuckles) and a database. So, you know, that's going to be, it takes a while as you well know, to get database adoption going, but you know, they're ambitious and going for it. All right, let's bring the chart back up, and I want to focus Darren in on the ecosystem now, and really, we've identified Snowflake and Databricks, it's always fun to talk about those guys, and there are a number of other, you know, data platforms out there, but we use those too as really proxies for leaders. We got a bunch of the backup guys, the data protection folks, Rubric, Cohesity, and Veeam. They're sort of in a cluster, although Rubric, you know, ahead of those guys in terms of spending momentum. And then VMware, Tanzu and Red Hat as sort of the cross cloud platform. But I want to focus, Darren, on the data piece of it. We're seeing a lot of activity around data sharing, governed data sharing. Databricks is using Delta Sharing as their sort of place, Snowflakes is sort of this walled garden like the app store. What are your thoughts on, you know, in the context of Supercloud, cross cloud capabilities for the data platforms? >> Yeah, good question. You know, I think Databricks is an interesting player because they sort of have made some interesting moves, with their Data Lakehouse technology. So they're trying to kind of complicate, or not complicate, they're trying to take away the complications of, you know, the downsides of data warehousing and data lakes, and trying to find that middle ground, where you have the benefits of a managed, governed, you know, data warehouse environment, but you have sort of the lower cost, you know, capability of a data lake. And so, you know, Databricks has become really attractive, especially by data scientists, right? We've been tracking them in the AI machine learning sector for quite some time here at ETR, attractive for a data scientist because it looks and acts like a lake, but can have some managed capabilities like a warehouse. So it's kind of the best of both worlds. So in some ways I think you've seen sort of a data science driver for the adoption of Databricks that has now become a little bit more mainstream across the business. Snowflake, maybe the other direction, you know, it's a cloud data warehouse that you know, is starting to expand its capabilities and add on new things like Streamlit is a good example in the analytics space, with apps. So you see these tools starting to branch and creep out a bit, but they offer that sort of neutrality, right? We heard one IT decision maker we recently interviewed that referred to Snowflake and Databricks as the quote unquote Switzerland of what they do. And so there's this desirability from an organization to find these tools that can solve the complex multi-headed use-case of data and analytics, which every business unit needs in different ways. And figure out a way to do that, an elegant way that's governed and centrally managed, that federated kind of best of both worlds that you get by bringing the data close to the business while having a central governed instance. So these tools are incredibly powerful and I think there's only going to be room for growth, for those two especially. I think they're going to expand and do different things and maybe, you know, join forces with others and a lot of the power of what they do well is trying to define these connections and find these partnerships with other vendors, and try to be seen as the nice add-on to your existing environment that plays nicely with everyone. So I think that's where those two tools are going, but they certainly fit this sort of label of, you know, trying to be that supercloud neutral, you know, layer that unites everything. >> Yeah, and if you bring the graphic back up, please, there's obviously big data plays in each of the cloud platforms, you know, Microsoft, big database player, AWS is, you know, 11, 12, 15, data stores. And of course, you know, BigQuery and other, you know, data platforms within Google. But you know, I'm not sure the big cloud guys are going to go hard after so-called supercloud, cross-cloud services. Although, we see Oracle getting in bed with Microsoft and Azure, with a database service that is cross-cloud, certainly Google with Anthos and you know, you never say never with with AWS. I guess what I would say guys, and I'll I'll leave you with this is that, you know, just like all players today are cloud players, I feel like anybody in the business or most companies are going to be so-called supercloud players. In other words, they're going to have a cross-cloud strategy, they're going to try to build connections if they're coming from on-prem like a Dell or an HPE, you know, or Pure or you know, many of these other companies, Cohesity is another one. They're going to try to connect to their on-premise states, of course, and create a consistent experience. It's natural that they're going to have sort of some consistency across clouds. You know, the big question is, what's that spectrum look like? I think on the one hand you're going to have some, you know, maybe some rudimentary, you know, instances of supercloud or maybe they just run on the individual clouds versus where Snowflake and others and even beyond that are trying to go with a single global instance, basically building out what I would think of as their own cloud, and importantly their own ecosystem. I'll give you guys the last thought. Maybe you could each give us, you know, closing thoughts. Maybe Darren, you could start and Erik, you could bring us home on just this entire topic, the future of cloud and data. >> Yeah, I mean I think, you know, two points to make on that is, this question of these, I guess what we'll call legacy on-prem players. These, mega vendors that have been around a long time, have big on-prem footprints and a lot of people have them for that reason. I think it's foolish to assume that a company, especially a large, mature, multinational company that's been around a long time, it's foolish to think that they can just uproot and leave on-premises entirely full scale. There will almost always be an on-prem footprint from any company that was not, you know, natively born in the cloud after 2010, right? I just don't think that's reasonable anytime soon. I think there's some industries that need on-prem, things like, you know, industrial manufacturing and so on. So I don't think on-prem is going away, and I think vendors that are going to, you know, go very cloud forward, very big on the cloud, if they neglect having at least decent connectors to on-prem legacy vendors, they're going to miss out. So I think that's something that these players need to keep in mind is that they continue to reach back to some of these players that have big footprints on-prem, and make sure that those integrations are seamless and work well, or else their customers will always have a multi-cloud or hybrid experience. And then I think a second point here about the future is, you know, we talk about the three big, you know, cloud providers, the Google, Microsoft, AWS as sort of the opposite of, or different from this new supercloud paradigm that's emerging. But I want to kind of point out that, they will always try to make a play to become that and I think, you know, we'll certainly see someone like Microsoft trying to expand their licensing and expand how they play in order to become that super cloud provider for folks. So also don't want to downplay them. I think you're going to see those three big players continue to move, and take over what players like CloudFlare are doing and try to, you know, cut them off before they get too big. So, keep an eye on them as well. >> Great points, I mean, I think you're right, the first point, if you're Dell, HPE, Cisco, IBM, your strategy should be to make your on-premise state as cloud-like as possible and you know, make those differences as minimal as possible. And you know, if you're a customer, then the business case is going to be low for you to move off of that. And I think you're right. I think the cloud guys, if this is a real problem, the cloud guys are going to play in there, and they're going to make some money at it. Erik, bring us home please. >> Yeah, I'm going to revert back to our data and this on the macro side. So to kind of support this concept of a supercloud right now, you know Dave, you and I know, we check overall spending and what we're seeing right now is total year spent is expected to only be 4.6%. We ended 2022 at 5% even though it began at almost eight and a half. So this is clearly declining and in that environment, we're seeing the top two strategies to reduce spend are actually vendor consolidation with 36% of our respondents saying they're actively seeking a way to reduce their number of vendors, and consolidate into one. That's obviously supporting a supercloud type of play. Number two is reducing excess cloud resources. So when I look at both of those combined, with a drop in the overall spending reduction, I think you're on the right thread here, Dave. You know, the overall macro view that we're seeing in the data supports this happening. And if I can real quick, couple of names we did not touch on that I do think deserve to be in this conversation, one is HashiCorp. HashiCorp is the number one player in our infrastructure sector, with a 56% net score. It does multiple things within infrastructure and it is completely agnostic to your environment. And if we're also speaking about something that's just a singular feature, we would look at Rubric for data, backup, storage, recovery. They're not going to offer you your full cloud or your networking of course, but if you are looking for your backup, recovery, and storage Rubric, also number one in that sector with a 53% net score. Two other names that deserve to be in this conversation as we watch it move and evolve. >> Great, thank you for bringing that up. Yeah, we had both of those guys in the chart and I failed to focus in on HashiCorp. And clearly a Supercloud enabler. All right guys, we got to go. Thank you so much for joining us, appreciate it. Let's keep this conversation going. >> Always enjoy talking to you Dave, thanks. >> Yeah, thanks for having us. >> All right, keep it right there for more content from Supercloud 2. This is Dave Valente for John Ferg and the entire Cube team. We'll be right back. (gentle synth music) (music fades)

Published Date : Feb 17 2023

SUMMARY :

is the intersection of cloud and data. Thank you for having period of time, you know, and evolution of the cloud So in a way, you know, supercloud the data closer to the business. So my general question to both of you is, the complexity does need to be And so there's this need to use, you know, So my question to you guys is, And as you mentioned, Azure but in the surveys, you know, customers, the ability to offer and there are a number of other, you know, and maybe, you know, join forces each of the cloud platforms, you know, the three big, you know, And you know, if you're a customer, you and I know, we check overall spending and I failed to focus in on HashiCorp. to you Dave, thanks. Ferg and the entire Cube team.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
IBMORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

ErikPERSON

0.99+

DellORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

John FergPERSON

0.99+

DavePERSON

0.99+

WalmartORGANIZATION

0.99+

Erik BradleyPERSON

0.99+

DavidPERSON

0.99+

AWSORGANIZATION

0.99+

Dave ValentePERSON

0.99+

January, 2023DATE

0.99+

ChinaLOCATION

0.99+

USLOCATION

0.99+

HPEORGANIZATION

0.99+

50 billionQUANTITY

0.99+

Ionis PharmaceuticalsORGANIZATION

0.99+

Darren BrabhamPERSON

0.99+

56%QUANTITY

0.99+

4.6%QUANTITY

0.99+

EuropeLOCATION

0.99+

OracleORGANIZATION

0.99+

53%QUANTITY

0.99+

36%QUANTITY

0.99+

TanzuORGANIZATION

0.99+

DarrenPERSON

0.99+

1200QUANTITY

0.99+

Red HatORGANIZATION

0.99+

VMwareORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

FridayDATE

0.99+

RubricORGANIZATION

0.99+

last yearDATE

0.99+

two sidesQUANTITY

0.99+

DatabricksORGANIZATION

0.99+

5%QUANTITY

0.99+

CohesityORGANIZATION

0.99+

two toolsQUANTITY

0.99+

VeeamORGANIZATION

0.99+

CloudFlareTITLE

0.99+

twoQUANTITY

0.99+

bothQUANTITY

0.99+

2022DATE

0.99+

OneQUANTITY

0.99+

Daren BrabhamPERSON

0.99+

three yearsQUANTITY

0.99+

TSISORGANIZATION

0.99+

BrabhamPERSON

0.99+

CloudFlareORGANIZATION

0.99+

1500 survey respondentsQUANTITY

0.99+

second pointQUANTITY

0.99+

first pointQUANTITY

0.98+

SnowflakeTITLE

0.98+

oneQUANTITY

0.98+

SupercloudORGANIZATION

0.98+

ETRORGANIZATION

0.98+

SnowflakeORGANIZATION

0.98+

AkamaiORGANIZATION

0.98+

Juan Loaiza, Oracle | Building the Mission Critical Supercloud


 

(upbeat music) >> Welcome back to Supercloud two where we're gathering a number of industry luminaries to discuss the future of cloud services. And we'll be focusing on various real world practitioners today, their challenges, their opportunities with an emphasis on data, self-service infrastructure and how organizations are evolving their data and cloud strategies to prepare for that next era of digital innovation. And we really believe that support for multiple cloud estates is a first step of any Supercloud. And in that regard Oracle surprise some folks with its Azure collaboration the Oracle database and exit database services. And to discuss the challenges of developing a mission critical Supercloud we welcome Juan Loaiza, who's the executive vice president of Mission Critical Database Technologies at Oracle. Juan, you're many time CUBE alums so welcome back to the show. Great to see you. >> Great to see you, and happy to be here with you. >> Yeah, thank you. So a lot of people felt that Oracle was resistant to multicloud strategies and preferred to really have everything run just on the Oracle cloud infrastructure, OCI and maybe that was a misperception maybe you guys were misunderstood or maybe you had to change your heart. Take us through the decision to support multiple cloud platforms >> Now we've supported multiple cloud platforms for many years, so I think that was probably a misperception. Oracle database, we partnered up with Amazon very early on in their cloud when they had kind of the the first cloud out there. And we had Oracle database running on their cloud. We have backup, we have a lot of stuff running. So, yeah, part of the philosophy of Oracle has always been we partner with every platform. We're very open we started with SQL and APIs. As we develop new technologies we push them into the SQL standard. So that's always been part of the ecosystem at Oracle. That's how we think we get an advantage by being more open. I think if we try to create this isolated little world it actually hurts us and hurts customers. So for us it's a win-win to be open across the clouds. >> So Supercloud is this concept that we put forth to describe a platform or some people think it's an architecture if you have an opinion, and I'd love to hear it but it provides a programmatically consistent set of services that hosted on heterogeneous cloud providers. And so we look at the Oracle database service for Azure as fitting within this definition. In your view, is this accurate? >> Yeah, I would broaden it. I'd see a little bit more than that. We just think that services should be available from everywhere, right? So, I mean, it's a little bit like if you go back to the pre-internet world, there was things like AOL and CompuServe and those were kind of islands. And if you were on AOL, you really didn't have access to anything on CompuServe and vice versa. And the cloud world has evolved a little bit like that. And we just think that's the wrong model. They shouldn't these clouds are part of the world and they need to be interconnected like all the rest of the world. It's been a long time with telephones internet, everything, everything's interconnected. Everything should work seamlessly together. So that's how we believe if you're running in one cloud and you're running let's say an application, one cloud you want to use a service from another cloud should be completely simple to do that. It shouldn't be, I can only use what's in AOL or CompuServe or whatever else. It should not be isolated. >> Well, we got a long way to go before that Nirvana exists but one example is the Oracle database service with Azure. So what exactly does that service provide? I'm interested in how consistent the service experience is across clouds. Did you create a purpose-built PaaS layer to achieve this common experience? Or is it off the shelf Terraform? Is there unique value in the PaaS layer? Let's dig into some of those questions. I know I just threw six at you. >> Yeah, I mean, so what this is, is what we're trying to do is very simple. Which is, for example, starting with the Oracle database we want to make that seamless to use from anywhere you're running. Whether it's on-prem, on some other cloud, anywhere else you should be able to seamlessly use the Oracle database and it should look like the internet. There's no friction. There's not a lot of hoops you got to jump just because you're trying to use a database that isn't local to you. So it's pretty straightforward. And in terms of things like Azure, it's not easy to do because all these clouds have a lot of kind of very unique technologies. So what we've done is at Oracle is we've said, "Okay we're going to make Oracle database look exactly like if it was running on Azure." That means we'll use the Azure security systems, the identity management systems, the networking, there's things like monitoring and management. So we'll push all these technologies. For example, when we have monitoring event or we have alerts we'll push those into the Azure console. So as a user, it looks to you exactly as if that Oracle database was running inside Azure. Also, the networking is a big challenge across these clouds. So we've basically made that whole thing seamless. So we create the super high bandwidth network between Azure and Oracle. We make sure that's extremely low latency, under two milliseconds round trip. It's all within the local metro region. So it's very fast, very high bandwidth, very low latency. And we take care establishing the links and making sure that it's secure and all that kind of stuff. So at a high level, it looks to you like the database is--even the look and feel of the screens. It's the Azure colors, it's the Azure buttons it's the Azure layout of the screens so it looks like you're running there and we take care of all the technical details underlying that which there's a lot which has taken a lot of work to make it work seamlessly. >> In the magic of that abstraction. Juan, does it happen at the PaaS layer? Could you take us inside that a little bit? Is there intelligence in there that helps you deal with latency or are there any kind of purpose-built functions for this service? >> You could think of it as... I mean it happens at a lot of different layers. It happens at the identity management layer, it happens at the networking layer, it happens at the database layer, it happens at the monitoring layer, at the management layer. So all those things have been integrated. So it's not one thing that you just go and do. You have to integrate all these different services together. You can access files in Azure from the Oracle database. Again, that's completely seamless. You, it's just like if it was local to our cloud you get your Azure files in your kind of S3 equivalent. So yeah, the, it's not one thing. There's a whole lot of pieces to the ecosystem. And what we've done is we've worked on each piece separately to make sure that it's completely seamless and transparent so you don't have to think about it, it just works. >> So you kind of answered my next question which is one of the technical hurdles. It sounds like the technical hurdles are that integration across the entire stack. That's the sort of architecture that you've built. What was the catalyst for this service? >> Yeah, the catalyst is just fulfilling our vision of an open cloud world. It's really like I said, Oracle, from the very beginning has been believed in open standards. Customers should be able to have choice customers should be able to use whatever they want from wherever they want. And we saw that, you know in the new world of cloud that had broken down everybody had their own authentication system management system, monitoring system networking system, configuration system. And it became very difficult. There was a lot of friction to using services across cloud. So we said, "Well, okay we can fix that." It's work, it's significant amount of work but we know how to do it and let's just go do it and make it easy for customers. >> So given Oracle is really your main focus is on mission critical workloads. You talked about this low latency network, I mean but you still have physical distances, so how are you managing that latency? What's the experience been for customers across Azure and OCI? >> Yeah, so it, it's a good point. I mean, latency can be an issue. So the good thing about clouds is we have a lot of cloud data centers. We have dozens and dozens of cloud data centers around the world. And Azure has dozens and dozens of cloud data centers. And in most cases, they're in the same metro region because there's kind of natural metro regions within each country that you want to put your cloud data centers in. So most of our data centers are actually very close to the Azure data centers. There's the kind of northern Virginia, there's London, there's Tokyo I mean, there's natural places where everybody puts their data centers Seoul et cetera. And so that's the real key. So that allows us to put a very high bandwidth and low latency network. The real problems with latency come when you're trying to go along physical distance. If you're trying to connect, you know across the Pacific or you know across the country or something like that, then you can get in trouble with latency within the same metro region. It's extremely fast. It tends to be around one, you know the highest two millisecond that's roundtrip through all the routers and connections and gateways and everything else. With everything taken into consideration, what we guarantee is it's always less than two millisecond which is a very low latency time. So that tends to not be a problem because it's extremely low latency. >> I was going to ask you less than two milliseconds. So, earlier in the program we had Jack Greenfield who runs architecture for Walmart, and he was explaining what we call their Supercloud, and it's runs across Azure, GCP, and they're on-prem. They have this thing called the triplet model. So my question to you is, are you in situations where you guaranteeing that less than two milliseconds do you have situations where you're bringing, you know Exadata Cloud, a customer on-prem to achieve that? Or is this just across clouds? >> Yeah, in this case, we're talking public cloud data center to public cloud data center. >> Oh okay. >> So add your public cloud data center to Oracle Public Cloud data center. They're in the same metro region. We set up the connections, we do all the technology to make it seamless. And from a customer point of view they don't really see the network. Also, remember that SQL is actually designed to have very low bandwidth and latency requirements. So it is a language. So you don't go to the database and say do this one little thing for me. You send it a SQL statement that can actually access lots of data while in the database. So the real latency requirement of a SQL database is within the database. So I need to access all that data fast. So I need very fast access to storage very fast access across node. That's what exit data gives you. But you send one request and that request can do a huge amount of work and then return one answer. And that's kind of the design point of SQL. So SQL is inherently low bandwidth requirements, it was used back in the eighties when we used to have 10 megabit networks and the the biggest companies in the world ran back then. So right now we're talking over hundred hundreds of gigabits. So it's really not much of a challenge. When you're designed to run on 10 megabit to say, okay I'm going to give you 10,000 times what you were designed for it's really, it's a pretty low hurdle jump. >> What about the deployment models? How do you handle this? Is it a single global instance across clouds or do you sort of instantiate in each you got exudate in Azure and exudates in OCI? What's the deployment model look like? >> It's pretty straightforward. So customer decides where they want to run their application and database. So there's natural places where people go. If you're in Tokyo, you're going to choose the local Tokyo data centers for both, you know Microsoft and Oracle. If you're in London, you're going to do that. If you're in California you're going to choose maybe San Jose, something like that. So a customer just chooses. We both have data centers in that metro region. So they create their service on Azure and then they go to our console which looks just like an Azure console and say all right create me a database. And then we choose the closest Oracle data center which is generally a few miles away, and then it it all gets created. So from a customer point of view, it's very straightforward. >> I'm always in awe about how simple you make things sound. All right what about security? You talked a little bit before about identity access how you sort of abstracting the Azure capabilities away so that you've simplified it for your customers but are there any other specific security things that you need to do? How much did you have to abstract the underlying primitives of Azure or OCI to present that common experience to customers? >> Yeah, so there's really two big things. One is the identity management. Like my name is X on Azure and I have this set of privileges. Oracle has its own identity management system, right? So what we didn't want is that you have to kind of like bridge these things yourself. It's a giant pain to do that. So we actually what we call federate across these identity managements. So you put your credentials into Azure and then they automatically get to use the exact same credentials and identity in the Oracle cloud. So again, you don't have to think about it, it just works. And then the second part is that the whole bridging the network. So within a cloud you generally have virtual network that's private to your company. And so at Oracle, we bridge the private network that you created in, for example, Azure to the private network that we create for you in Oracle. So it is still a private network without you having to do a whole bunch of work. So it's just like if you were in your own data center other people can't get into your network. So it's secured at the network level, it's secured at the identity management, and encryption level. And again we did a lot of work to make that seamless for customers and they don't have to worry about it because we did the work. That's really as simple as it gets. >> That's what's Supercloud's supposed to be all about. Alright, we were talking earlier about sort of the misperception around multicloud, your view of Open I think, which is you run the Oracle database, wherever the customer wants to run it. So you got this database service across OCI and Azure customers today, they run Oracle database in AWS. You got heat wave, MySQL, heat wave that you announced on AWS, Google touts a bare metal offering where you can run Oracle on GCP. Do you see a day when you extend an OCI Azure like situation across multiple clouds? Would that bring benefits to customers or will the world of database generally remain largely fenced with maybe a few exceptions like what you're doing with OCI and Azure? I'm particularly interested in your thoughts on egress fees as maybe one of the reasons that there is a barrier to this happening and why maybe these stove pipes, exist today and in the future. What are your thoughts on that? >> Yeah, we're very open to working with everyone else out there. Like I said, we've always been, big believers in customers should have choice and you should be able to run wherever you want. So that's been kind of a founding principle of Oracle. We have the Azure, we did a partnership with them, we're open to doing other partnerships and you're going to see other things coming down the pipe on the topic of egress. Yeah, the large egress fees, it's pretty obvious what goes on with that. Various vendors like to have large egress fees because they want to keep things kind of locked into their cloud. So it's not a very customer friendly thing to do. And I think everybody recognizes that it's really trying to kind of course or put a lot of friction on moving data out of a particular cloud. And that's not what we do. We have very, very low egress fees. So we don't really do that and we don't think anybody else should do that. But I think customers at the end of the day, will win that battle. They're going to have to go back to their vendor and say, well I have choice in clouds and if you're going to impose these limits on me, maybe I'll make a different choice. So that's ultimately how these things get resolved. >> So do you think other cloud providers are going to take a page out of what you're doing with Azure and provide similar solutions? >> Yeah, well I think customers want, I mean, I've talked to a lot of customers, this is what they want, right? I mean, there's really no doubt no customer wants to be locked into a single ecosystem. There's nobody out there that wants that. And as the competition, when they start seeing an open ecosystem evolving they're going to be like, okay, I'd rather go there than the closed ecosystem, and that's going to put pressure on the closed ecosystems. So that's the nature of competition. That's what ultimately will tip the balance on these things. >> So Juan, even though you have this capability of distributing a workload across multiple clouds as in our Supercloud premise it's still something that's relatively new. It's a big decision that maybe many people might consider somewhat of a risk. So I'm curious who's driving the decisions for your initial customers? What do they want to get out of it? What's the decision point there? >> Yeah, I mean, this is generally driven by customers that want a specific technology in a cloud. I think the risk, I haven't seen a lot of people worry too much about the risk. Everybody involved in this is a very well known, very reputable firm. I mean, Oracle's been around for 40 years. We run most of the world's largest companies. I think customers understand we're not going to build a solution that's going to put their technology and their business at risk. And the same thing with Azure and others. So I don't see customers too worried about this is a risky move because it's really not. And you know, everybody understands networking at the end the day networking works. I mean, how does the internet work? It's a known quantity. It's not like it's some brand new invention. What we're really doing is breaking down the barriers to interconnecting things. Automating 'em, making 'em easy. So there's not a whole lot of risk here for customers. And like I said, every single customer in the world loves an open ecosystem. It's just not a question. If you go to a customer would you rather put your technology or your business to run on a closed ecosystem or an open system? It's kind of not even worth asking a question. It's a no-brainer. >> All right, so we got to go. My last question. What do you think of the term "Supercloud"? You think it'll stick? >> We'll see. There's a lot of terms out there and it's always fun to see which terms stick. It's a cool term. I like it, but the decision makers are actually the public, what sticks and what doesn't. It's very hard to predict. >> Yeah well, it's been a lot of fun having you on, Juan. Really appreciate your time and always good to see you. >> All right, Dave, thanks a lot. It's always fun to talk to you. >> You bet. All right, keep it right there. More Supercloud two content from theCUBE Community Dave Vellante for John Furrier. We'll be right back. (upbeat music)

Published Date : Jan 12 2023

SUMMARY :

and cloud strategies to prepare happy to be here with you. just on the Oracle cloud of the ecosystem at Oracle. and I'd love to hear it And the cloud world has Or is it off the shelf Terraform? So at a high level, it looks to you Juan, does it happen at the PaaS layer? it happens at the database layer, So you kind of And we saw that, you know What's the experience been for customers across the Pacific or you know So my question to you is, to public cloud data center. So the real latency requirement and then they go to our console the Azure capabilities away So it's secured at the network level, So you got this database We have the Azure, we did So that's the nature of competition. What's the decision point there? down the barriers to the term "Supercloud"? and it's always fun to and always good to see you. It's always fun to talk to you. Vellante for John Furrier.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MicrosoftORGANIZATION

0.99+

OracleORGANIZATION

0.99+

DavePERSON

0.99+

WalmartORGANIZATION

0.99+

Juan LoaizaPERSON

0.99+

AmazonORGANIZATION

0.99+

San JoseLOCATION

0.99+

CaliforniaLOCATION

0.99+

Dave VellantePERSON

0.99+

TokyoLOCATION

0.99+

JuanPERSON

0.99+

LondonLOCATION

0.99+

sixQUANTITY

0.99+

10,000 timesQUANTITY

0.99+

Jack GreenfieldPERSON

0.99+

GoogleORGANIZATION

0.99+

second partQUANTITY

0.99+

AWSORGANIZATION

0.99+

less than two millisecondQUANTITY

0.99+

less than two millisecondsQUANTITY

0.99+

OneQUANTITY

0.99+

SQLTITLE

0.99+

10 megabitQUANTITY

0.99+

bothQUANTITY

0.99+

AOLORGANIZATION

0.98+

each pieceQUANTITY

0.98+

MySQLTITLE

0.98+

first cloudQUANTITY

0.98+

singleQUANTITY

0.98+

each countryQUANTITY

0.98+

John FurrierPERSON

0.98+

two big thingsQUANTITY

0.98+

under two millisecondsQUANTITY

0.98+

oneQUANTITY

0.98+

northern VirginiaLOCATION

0.98+

CompuServeORGANIZATION

0.97+

first stepQUANTITY

0.97+

Mission Critical Database TechnologiesORGANIZATION

0.97+

one requestQUANTITY

0.97+

SeoulLOCATION

0.97+

AzureTITLE

0.97+

eachQUANTITY

0.97+

two millisecondQUANTITY

0.97+

AzureORGANIZATION

0.96+

one cloudQUANTITY

0.95+

one thingQUANTITY

0.95+

cloud data centersQUANTITY

0.95+

one answerQUANTITY

0.95+

SupercloudORGANIZATION

0.94+

Nikesh Arora, Palo Alto Networks | Palo Alto Networks Ignite22


 

Upbeat music plays >> Voice Over: TheCUBE presents Ignite 22, brought to you by Palo Alto Networks. >> Good morning everyone. Welcome to theCUBE. Lisa Martin here with Dave Vellante. We are live at Palo Alto Networks Ignite. This is the 10th annual Ignite. There's about 3,000 people here, excited to really see where this powerhouse organization is taking security. Dave, it's great to be here. Our first time covering Ignite. People are ready to be back. They.. and security is top. It's a board level conversation. >> It is the other Ignite, I like to call it cuz of course there's another big company has a conference name Ignite, so I'm really excited to be here. Palo Alto Networks, a company we've covered for a number of years, as we just wrote in our recent breaking analysis, we've called them the gold standard but it's not just our opinion, we've backed it up with data. The company's on track. We think to do close to 7 billion in revenue by 2023. That's double it's 2020 revenue. You can measure it with execution, market cap M and A prowess. I'm super excited to have the CEO here. >> We have the CEO here, Nikesh Arora joins us from Palo Alto Networks. Nikesh, great to have you on theCube. Thank you for joining us. >> Well thank you very much for having me Lisa and Dave >> Lisa: It was great to see your keynote this morning. You said that, you know fundamentally security is a data problem. Well these days every company has to be a data company. Grocery stores, gas stations, car dealers. How is Palo Alto networks making customers, these data companies, more secure? >> Well Lisa, you know, (coughs) I've only done cybersecurity for about four, four and a half years so when I came to the industry I was amazed to see how security is so reactive as opposed to proactive. We should be able to stop bad threats, right? as they're happening. But I think a lot of threats get through because we don't have the right infrastructure and the right tooling and right products in there. So I think we've been working hard for the last four and a half years to turn it around so we can have consistent data flow across an enterprise and then mine that data for threats and anomalous behavior and try and protect our customers. >> You know the problem, I wrote this, this weekend, the problem in cybersecurity is well understood, you put up that Optiv graph and it's like 8,000 companies >> Yes >> and I think you mentioned your keynote on average, you know 30 to 40 tools, maybe 50, at least 20, >> Yes. >> from the folks that I talked to. So, okay, great, but actually solving that problem is not trivial. To be a consolidator, I mean, everybody wants to consolidate tools. So in your three to four years and security as you well know, it's, you can't fake security. It's a really, really challenging topic. So when you joined Palo Alto Networks and you heard that strategy, I know you guys have been thinking about this for some time, what did you see as the challenges to actually executing on that and how is it that you've been able to sort of get through that knot hole. >> So Dave, you know, it's interesting if you look at the history of cybersecurity, I call them the flavor of the decade, a flare, you know a new threat vector gets created, very large market gets created, a solution comes through, people flock, you get four or five companies will chase that opportunity, and then they become leaders in that space whether it's firewalls or endpoints or identity. And then people stick to their swim lane. The problem is that's a very product centric approach to security. It's not a customer-centric approach. The customer wants a more secure enterprise. They don't want to solve 20 different solutions.. problems with 20 different point solutions. But that's kind of how the industry's grown up, and it's been impossible for a large security company in one category, to actually have a substantive presence in the next category. Now what we've been able to do in the last four and a half years is, you know, from our firewall base we had resources, we had intellectual capability from a security perspective and we had cash. So we used that to pay off our technical debt. We acquired a bunch of companies, we created capability. In the last three years, four years we've created three incremental businesses which are all on track to hit a billion dollars the next 12 to 18 months. >> Yeah, so it's interesting on Twitter last night we had a little conversation about acquirers and who was a good, who was not so good. It was, there was Oracle, they came up actually very high, they'd done pretty, pretty good Job, VMware was on the list, IBM, Cisco, ServiceNow. And if you look at IBM and Cisco's strategy, they tend to be very services heavy, >> Mm >> right? How is it that you have been able to, you mentioned get rid of your technical debt, you invested in that. I wonder if you could, was it the, the Cloud, even though a lot of the Cloud was your own Cloud, was that a difference in terms of your ability to integrate? Because so many companies have tried it in the past. Oracle I think has done a good job, but it took 'em 10 to 12 years, you know, to, to get there. What was the sort of secret sauce? Is it culture, is it just great engineering? >> Dave it's a.. thank you for that. I think, look, it's, it's a mix of everything. First and foremost, you know, there are certain categories we didn't play in so there was nothing to integrate. We built a capability in a category in automation. We didn't have a product, we acquired a company. It's a net new capability in instant response. We didn't have a capability. It was net new capability. So there was, there was, other than integrating culturally and into the organization into our core to market processes there was no technical integration needed. Most of our technical integration was needed in our Cloud platform, which we bought five or six companies, we integrated then we just bought one recently called cyber security as well, which is going to get integrated in the Cloud platform. >> Dave: Yeah. >> And the thing is like, the Cloud platform is net new in the industry. We.. nobody's created a Cloud security platform yet, so we're working hard to create it because we don't want to replicate the mistakes of the past, that were made in enterprise security, in Cloud security. So it's a combination of cultural integration it's a combination of technical integration. The two things we do differently I think, than most people in the industry is look, we have no pride of, you know of innovations. Like, if somebody else has done it, we respect it and we'll acquire it, but we always want to acquire number one or number two in their category. I don't want number three or four. There's three or four for a reason and there still leaves one or two out there to compete with. So we've always acquired one or two, one. And the second thing, which is as important is most of these companies are in the early stage of development. So it's very important for the founding team to be around. So we spend a lot of time making sure they stick around. We actually make our people work for them. My principle is, listen, if they beat us in the open market with all our resources and our people, then they deserve to run this as opposed to us. So most of our new product categories are run by founders of companies required. >> So a little bit of Jack Welch, a little bit of Franks Lubens is a, you know always deference to the founders. But go ahead Lisa. >> Speaking of cultural transformation, you were mentioning your keynote this morning, there's been a significant workforce transformation at Palo Alto Networks. >> Yeah >> Talk a little bit about that, cause that's a big challenge, for many organizations to achieve. Sounds like you've done it pretty well. >> Well you know, my old boss, Eric Schmidt, used to say, 'revenue solves all known problems'. Which kind of, you know, it is a part joking, part true, but you know as Dave mentioned, we've doubled or two and a half time the revenues in the last four and a half years. That allows you to grow, that allows you to increase headcount. So we've gone from four and a half thousand people to 14,000 people. Good news is that's 9,500 people are net new to the company. So you can hire a whole new set of people who have new skills, new capabilities and there's some attrition four and a half thousand, some part of that turns over in four and a half years, so we effectively have 80% net new people, and the people we have, who are there from before, are amazing because they've built a phenomenal firewall business. So it's kind of been right sized across the board. It's very hard to do this if you're not growing. So you got to focus on growing. >> Dave: It's like winning in sports. So speaking of firewalls, I got to ask you does self-driving cars need brakes? So if I got a shout out to my friend Zeus Cararvela so like that's his line about why you need firewalls, right? >> Nikesh: Yes. >> I mean you mentioned it in your keynote today. You said it's the number one question that you get. >> and I don't get it why P industry observers don't go back and say that's, this is ridiculous. The network traffic is doubling or tripling. (clears throat) In fact, I gave an interesting example. We shut down our data centers, as I said, we are all on Google Cloud and Amazon Cloud and then, you know our internal team comes in, we'd want a bigger firewall. I'm like, why do you want a bigger firewall? We shut down our data centers as well. The traffic coming in and out of our campus is doubled. We need a bigger firewall. So you still need a firewall even if you're in the Cloud. >> So I'm going to come back to >> Nikesh: (coughs) >> the M and A strategy. My question is, can you be both best of breed and develop a comprehensive suite number.. part one and part one A of that is do you even have to, because generally sweets win out over best of breed. But what, how do you, how do you respond? >> Well, you know, this is this age old debate and people get trapped in that, I think in my mind, and let me try and expand the analogy which I tried to do up in my keynote. You know, let's assume that Oracle, Microsoft, Dynamics and Salesforce did not exist, okay? And you were running a large company of 50,000 people and your job was to manage the customer process which easier to understand than security. And I said, okay, guess what? I have a quoting system and a lead system but the lead system doesn't talk to my coding system. So I get leads, but I don't know who those customers. And I write codes for a whole new set of customers and I have a customer database. Then when they come as purchase orders, I have a new database with all the customers who've bought something from me, and then when I go get them licensing I have a new database and when I go have customer support, I have a fifth database and there are customers in all five databases. You'll say Nikesh you're crazy, you should have one customer database, otherwise you're never going to be able to make this work. But security is the same problem. >> Dave: Mm I should.. I need consistency in data from suit to nuts. If it's in Cloud, if you're writing code, I need to understand the security flaws before they go into deployment, before they go into production. We for somehow ridiculously have bought security like IT. Now the difference between IT and security is, IT is required to talk to each other, so a Dell server and HP server work very similarly but a Palo Alto firewall and a Checkpoint firewall Fortnight firewall work formally differently. And then how that transitions into endpoints is a whole different ball game. So you need consistency in data, as Lisa was saying earlier, it's a data problem. You need consistency as you traverse to the enterprise. And that's why that's the number one need. Now, when you say best of breed, (coughs) best of breed, if it's fine, if it's a specific problem that you're trying to solve. But if you're trying to make sure that's the data flow that happens, you need both best of breed, you know, technology that stops things and need integration on data. So what we are trying to do is we're trying to give people best to breed solutions in the categories they want because otherwise they won't buy us. But we're also trying to make sure we stitch the data. >> But that definition of best of breed is a little bit of nuance than different in security is what I'm hearing because that consistency >> Nikesh: (coughs) Yes, >> across products. What about across Cloud? You mentioned Google and Amazon. >> Yeah so that's great question. >> Dave: Are you building the security super Cloud, I call it, above the Cloud? >> It's, it's not, it's, less so a super Cloud, It's more like Switzerland and I used to work at Google for 10 years, not a secret. And we used to sell advertising and we decided to go into pub into display ads or publishing, right. Now we had no publishing platform so we had to be good at everybody else's publishing platform >> Dave: Mm >> but we never were able to search ads for everybody else because we only focus on our own platform. So part of it is when the Cloud guys they're busy solving security for their Cloud. Google is not doing anything about Amazon Cloud or Microsoft Cloud, Microsoft's Azure, right? AWS is not doing anything about Google Cloud or Azure. So what we do is we don't have a Cloud. Our job in providing Cloud securities, be Switzerland make sure it works consistently across every Cloud. Now if you try to replicate what we offer Prisma Cloud, by using AWS, Azure and GCP, you'd have to first of all, have three panes of glass for all three of them. But even within them they have four panes of glass for the capabilities we offer. So you could end up with 12 different interfaces to manage a development process, we give you one. Now you tell me which is better. >> Dave: Sounds like a super Cloud to me Lisa (laughing) >> He's big on super Cloud >> Uber Cloud, there you >> Hey I like that, Uber Cloud. Well, so I want to understand Nikesh, what's realistic. You mentioned in your keynote Dave, brought it up that the average organization has 30 to 50 tools, security tools. >> Nikesh: Yes, yes >> On their network. What is realistic for from a consolidation perspective where Palo Alto can come in and say, let me make this consistent and simple for you. >> Well, I'll give you your own example, right? (clears throat) We're probably sub 10 substantively, right? There may be small things here and there we do. But on a substantive protecting the enterprise perspective you be should be down to eight or 10 vendors, and that is not perfect but it's a lot better than 50, >> Lisa: Right? >> because don't forget 50 tools means you have to have capability to understand what those 50 tools are doing. You have to have the capability to upgrade them on a constant basis, learn about their new capabilities. And I just can't imagine why customers have two sets of firewalls right. Now you got to learn both the files on how to deploy both them. That's silly because that's why we need 7 million more people. You need people to understand, so all these tools, who work for companies. If you had less tools, we need less people. >> Do you think, you know I wrote about this as well, that the security industry is anomalous and that the leader has, you know, single digit, low single digit >> Yes >> market shares. Do you think that you can change that? >> Well, you know, when I started that was exactly the observation I had Dave, which you highlighted in your article. We were the largest by revenue, by small margin. And we were one and half percent of the industry. Now we're closer to three, three to four percent and we're still at, you know, like you said, going to be around $7 billion. So I see a path for us to double from here and then double from there, and hopefully as we keep doubling and some point in time, you know, I'd like to get to double digits to start with. >> One of the things that I think has to happen is this has to grow dramatically, the ecosystem. I wonder if you could talk about the ecosystem and your strategy there. >> Well, you know, it's a matter of perspective. I think we have to get more penetrated in our largest customers. So we have, you know, 1800 of the top 2000 customers in the world are Palo Alto customers. But we're not fully penetrated with all our capabilities and the same customers set, so yes the ecosystem needs to grow, but the pandemic has taught us the ecosystem can grow wherever they are without having to come to Vegas. Which I don't think is a bad thing to be honest. So the ecosystem is growing. You are seeing new players come to the ecosystem. Five years ago you didn't see a lot of systems integrators and security. You didn't see security offshoots of telecom companies. You didn't see the Optivs, the WWTs, the (indistinct) of the world (coughs) make a concerted shift towards consolidation or services and all that is happening >> Dave: Mm >> as we speak today in the audience you will find people from Google, Amazon Microsoft are sitting in the audience. People from telecom companies are sitting in the audience. These people weren't there five years ago. So you are seeing >> Dave: Mm >> the ecosystem's adapting. They're, they want to be front and center of solving the customer's problem around security and they want to consolidate capability, they need. They don't want to go work with a hundred vendors because you know, it's like, it's hard. >> And the global system integrators are key. I always say they like to eat at the trough and there's a lot of money in security. >> Yes. >> Dave: (laughs) >> Well speaking of the ecosystem, you had Thomas Curry and Google Cloud CEO in your fireside chat in the keynote. Talk a little bit about how Google Cloud plus Palo Alto Networks, the Zero Trust Partnership and what it's enable customers to achieve. >> Lisa, that's a great question. (clears his throat) Thank you for bringing it up. Look, you know the, one of the most fundamental shifts that is happening is obviously the shift to the Cloud. Now when that shift fully, sort of, takes shape you will realize if your network has changed and you're delivering everything to the Cloud you need to go figure out how to bring the traffic to the Cloud. You don't have to bring it back to your data center you can bring it straight to the Cloud. So in that context, you know we use Google Cloud and Amazon Cloud, to be able to carry our traffic. We're going from a product company to a services company in addition, right? Cuz when we go from firewalls to SASE we're not carrying your traffic. When we carry our traffic, we need to make sure we have underlying capability which is world class. We think GCP and AWS and Azure run some of the biggest and best networks in the world. So our partnership with Google is such that we use their public Cloud, we sit on top of their Cloud, they give us increased enhanced functionality so that our customers SASE traffic gets delivered in priority anywhere in the world. They give us tooling to make sure that there's high reliability. So you know, we partner, they have Beyond Corp which is their version of Zero Trust which allows you to take unmanaged devices with browsers. We have SASE, which allows you to have managed devices. So the combination gives our collective customers the ability for Zero Trust. >> Do you feel like there has to be more collaboration within the ecosystem, the security, you know, landscape even amongst competitors? I mean I think about Google acquires Mandiant. You guys have Unit 42. Should and will, like, Wendy Whitmore and maybe they already are, Kevin Mandia talk more and share more data. If security's a data problem is all this data >> Nikesh: Yeah look I think the industry shares threat data, both in private organizations as well as public and private context, so that's not a problem. You know the challenge with too much collaboration in security is you never know. Like you know, the moment you start sharing your stuff at third parties, you go out of Secure Zone. >> Lisa: Mm >> Our biggest challenge is, you know, I can't trust a third party competitor partner product. I have to treat it with as much suspicion as anything else out there because the only way I can deliver Zero Trust is to not trust anything. So collaboration in Zero Trust are a bit of odds with each other. >> Sounds like another problem you can solve >> (laughs) >> Nikesh last question for you. >> Yes >> Favorite customer or example that you think really articulates the value of what Palo Alto was delivering? >> Look you know, it's a great question, Lisa. I had this seminal conversation with a customer and I explained all those things we were talking about and the customer said to me, great, okay so what do I need to do? I said, fun, you got to trust me because you know, we are on a journey, because in the past, customers have had to take the onus on themselves of integrating everything because they weren't sure a small startup will be independent, be bought by another cybersecurity company or a large cybersecurity company won't get gobbled up and split into pieces by private equity because every one of the cybersecurity companies have had a shelf life. So you know, our aspiration is to be the evergreen cybersecurity company. We will always be around and we will always tackle innovation and be on the front line. So the customer understood what we're doing. Over the last three years we've been working on a transformation journey with them. We're trying to bring them, or we have brought them along the path of Zero Trust and we're trying to work with them to deliver this notion of reducing their meantime to remediate from days to minutes. Now that's an outcome based approach that's a partnership based approach and we'd like, love to have more and more customers of that kind. I think we weren't ready to be honest as a company four and a half years ago, but I think today we're ready. Hence my keynote was called The Perfect Storm. I think we're at the right time in the industry with the right capabilities and the right ecosystem to be able to deliver what the industry needs. >> The perfect storm, partners, customers, investors, employees. Nikesh, it's been such a pleasure having you on theCUBE. Thank you for coming to talk to Dave and me right after your keynote. We appreciate that and we look forward to two days of great coverage from your executives, your customers, and your partners. Thank you. >> Well, thank you for having me, Lisa and Dave and thank you >> Dave: Pleasure >> for what you guys do for our industry. >> Our pleasure. For Nikesh Arora and Dave Vellante, I'm Lisa Martin, you're watching theCUBE live at MGM Grand Hotel in Las Vegas, Palo Alto Ignite 22. Stick around Dave and I will be joined by our next guest in just a minute. (cheerful music plays out)

Published Date : Dec 13 2022

SUMMARY :

brought to you by Palo Alto Networks. Dave, it's great to be here. I like to call it cuz Nikesh, great to have you on theCube. You said that, you know and the right tooling and and you heard that strategy, So Dave, you know, it's interesting And if you look at IBM How is it that you have been able to, First and foremost, you know, of, you know of innovations. Lubens is a, you know you were mentioning your for many organizations to achieve. and the people we have, So speaking of firewalls, I got to ask you I mean you mentioned and then, you know our that is do you even have to, Well, you know, this So you need consistency in data, and Amazon. so that's great question. and we decided to go process, we give you one. that the average organization and simple for you. Well, I'll give you You have to have the Do you think that you can change that? and some point in time, you know, I wonder if you could So we have, you know, 1800 in the audience you will find because you know, it's like, it's hard. And the global system and Google Cloud CEO in your So in that context, you security, you know, landscape Like you know, the moment I have to treat it with as much suspicion for you. and the customer said to me, great, okay Thank you for coming Arora and Dave Vellante,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
CiscoORGANIZATION

0.99+

DavePERSON

0.99+

IBMORGANIZATION

0.99+

Dave VellantePERSON

0.99+

LisaPERSON

0.99+

Lisa MartinPERSON

0.99+

fiveQUANTITY

0.99+

Eric SchmidtPERSON

0.99+

GoogleORGANIZATION

0.99+

DellORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

Kevin MandiaPERSON

0.99+

OracleORGANIZATION

0.99+

30QUANTITY

0.99+

Palo Alto NetworksORGANIZATION

0.99+

AWSORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

80%QUANTITY

0.99+

HPORGANIZATION

0.99+

2020DATE

0.99+

threeQUANTITY

0.99+

10 yearsQUANTITY

0.99+

10QUANTITY

0.99+

oneQUANTITY

0.99+

9,500 peopleQUANTITY

0.99+

2023DATE

0.99+

six companiesQUANTITY

0.99+

50QUANTITY

0.99+

VegasLOCATION

0.99+

four and a half yearsQUANTITY

0.99+

14,000 peopleQUANTITY

0.99+

twoQUANTITY

0.99+

Wendy WhitmorePERSON

0.99+

50,000 peopleQUANTITY

0.99+

Jack WelchPERSON

0.99+

10 vendorsQUANTITY

0.99+

Five years agoDATE

0.99+

Thomas CurryPERSON

0.99+

fourQUANTITY

0.99+

50 toolsQUANTITY

0.99+

1800QUANTITY

0.99+

Zero TrustORGANIZATION

0.99+

SalesforceORGANIZATION

0.99+

12 different interfacesQUANTITY

0.99+

Holger Mueller, Constellation Research | AWS re:Invent 2022


 

(upbeat music) >> Hey, everyone, welcome back to Las Vegas, "theCube" is on our fourth day of covering AWS re:Invent, live from the Venetian Expo Center. This week has been amazing. We've created a ton of content, as you know, 'cause you've been watching. But, there's been north of 55,000 people here, hundreds of thousands online. We've had amazing conversations across the AWS ecosystem. Lisa Martin, Paul Gillan. Paul, what's your, kind of, take on day four of the conference? It's still highly packed. >> Oh, there's lots of people here. (laughs) >> Yep. Unusual for the final day of a conference. I think Werner Vogels, if I'm pronouncing it right kicked things off today when he talked about asymmetry and how the world is, you know, asymmetric. We build symmetric software, because it's convenient to do so, but asymmetric software actually scales and evolves much better. And I think that that was a conversation starter for a lot of what people are talking about here today, which is how the cloud changes the way we think about building software. >> Absolutely does. >> Our next guest, Holger Mueller, that's one of his key areas of focus. And Holger, welcome, thanks for joining us on the "theCube". >> Thanks for having me. >> What did you take away from the keynote this morning? >> Well, how do you feel on the final day of the marathon, right? We're like 23, 24 miles. Hit the ball yesterday, right? >> We are going strong Holger. And, of course, >> Yeah. >> you guys, we can either talk about business transformation with cloud or the World Cup. >> Or we can do both. >> The World Cup, hands down. World Cup. (Lisa laughs) Germany's out, I'm unbiased now. They just got eliminated. >> Spain is out now. >> What will the U.S. do against Netherlands tomorrow? >> They're going to win. What's your forecast? U.S. will win? >> They're going to win 2 to 1. >> What do you say, 2:1? >> I'm optimistic, but realistic. >> 3? >> I think Netherlands. >> Netherlands will win? >> 2 to nothing. >> Okay, I'll vote for the U.S.. >> Okay, okay >> 3:1 for the U.S.. >> Be optimistic. >> Root for the U.S.. >> Okay, I like that. >> Hope for the best wherever you work. >> Tomorrow you'll see how much soccer experts we are. >> If your prediction was right. (laughs) >> (laughs) Ja, ja. Or yours was right, right, so. Cool, no, but the event, I think the event is great to have 50,000 people. Biggest event of the year again, right? Not yet the 70,000 we had in 2019. But it's great to have the energy. I've never seen the show floor going all the way down like this, right? >> I haven't either. >> I've never seen that. I think it's a record. Often vendors get the space here and they have the keynote area, and the entertainment area, >> Yeah. >> and the food area, and then there's an exposition, right? This is packed. >> It's packed. >> Maybe it'll pay off. >> You don't see the big empty booths that you often see. >> Oh no. >> Exactly, exactly. You know, the white spaces and so on. >> No. >> Right. >> Which is a good thing. >> There's lots of energy, which is great. And today's, of course, the developer day, like you said before, right now Vogels' a rockstar in the developer community, right. Revered visionary on what has been built, right? And he's becoming a little professorial is my feeling, right. He had these moments before too, when it was justifying how AWS moved off the Oracle database about the importance of data warehouses and structures and why DynamoDB is better and so on. But, he had a large part of this too, and this coming right across the keynotes, right? Adam Selipsky talking about Antarctica, right? Scott against almonds and what went wrong. He didn't tell us, by the way, which often the tech winners forget. Scott banked on technology. He had motorized sleds, which failed after three miles. So, that's not the story to tell the technology. Let everything down. Everybody went back to ponies and horses and dogs. >> Maybe goes back to these asynchronous behavior. >> Yeah. >> The way of nature. >> And, yesterday, Swami talking about the bridges, right? The root bridges, right? >> Right. >> So, how could Werner pick up with his video at the beginning. >> Yeah. >> And then talk about space and other things? So I think it's important to educate about event-based architecture, right? And we see this massive transformation. Modern software has to be event based, right? Because, that's how things work and we didn't think like this before. I see this massive transformation in my other research area in other platforms about the HR space, where payrolls are being rebuilt completely. And payroll used to be one of the three peaks of ERP, right? You would size your ERP machine before the cloud to financial close, to run the payroll, and to do an MRP manufacturing run if you're manufacturing. God forbid you run those three at the same time. Your machine wouldn't be able to do that, right? So it was like start the engine, start the boosters, we are running payroll. And now the modern payroll designs like you see from ADP or from Ceridian, they're taking every payroll relevant event. You check in time wise, right? You go overtime, you take a day of vacation and right away they trigger and run the payroll, so it's up to date for you, up to date for you, which, in this economy, is super important, because we have more gig workers, we have more contractors, we have employees who are leaving suddenly, right? The great resignation, which is happening. So, from that perspective, it's the modern way of building software. So it's great to see Werner showing that. The dirty little secrets though is that is more efficient software for the cloud platform vendor too. Takes less resources, gets less committed things, so it's a much more scalable architecture. You can move the events, you can work asynchronously much better. And the biggest showcase, right? What's the biggest transactional showcase for an eventually consistent asynchronous transactional application? I know it's a mouthful, but we at Amazon, AWS, Amazon, right? You buy something on Amazon they tell you it's going to come tomorrow. >> Yep. >> They don't know it's going to come tomorrow by that time, because it's not transactionally consistent, right? We're just making every ERP vendor, who lives in transactional work, having nightmares of course, (Lisa laughs) but for them it's like, yes we have the delivery to promise, a promise to do that, right? But they come back to you and say, "Sorry, we couldn't make it, delivery didn't work and so on. It's going to be a new date. We are out of the product.", right? So these kind of event base asynchronous things are more and more what's going to scale around the world. It's going to be efficient for everybody, it's going to be better customer experience, better employee experience, ultimately better user experience, it's going to be better for the enterprise to build, but we have to learn to build it. So big announcement was to build our environment to build better eventful applications from today. >> Talk about... This is the first re:Invent... Well, actually, I'm sorry, it's the second re:Invent under Adam Selipsky. >> Right. Adam Selipsky, yep. >> But his first year. >> Right >> We're hearing a lot of momentum. What's your takeaway with what he delivered with the direction Amazon is going, their vision? >> Ja, I think compared to the Jassy times, right, we didn't see the hockey stick slide, right? With a number of innovations and releases. That was done in 2019 too, right? So I think it's a more pedestrian pace, which, ultimately, is good for everybody, because it means that when software vendors go slower, they do less width, but more depth. >> Yeah. >> And depth is what customers need. So Amazon's building more on the depth side, which is good news. I also think, and that's not official, right, but Adam Selipsky came from Tableau, right? >> Yeah. So he is a BI analytics guy. So it's no surprise we have three data lake offerings, right? Security data lake, we have a healthcare data lake and we have a supply chain data lake, right? Where all, again, the epigonos mentioned them I was like, "Oh, my god, Amazon's coming to supply chain.", but it's actually data lakes, which is an interesting part. But, I think it's not a surprise that someone who comes heavily out of the analytics BI world, it's off ringside, if I was pitching internally to him maybe I'd do something which he's is familiar with and I think that's what we see in the major announcement of his keynote on Tuesday. >> I mean, speaking of analytics, one of the big announcements early on was Amazon is trying to bridge the gap between Aurora. >> Yep. >> And Redshift. >> Right. >> And setting up for continuous pipelines, continuous integration. >> Right. >> Seems to be a trend that is common to all database players. I mean, Oracle is doing the same thing. SAP is doing the same thing. MariaDB. Do you see the distinction between transactional and analytical databases going away? >> It's coming together, right? Certainly coming together, from that perspective, but there's a fundamental different starting point, right? And with the big idea part, right? The universal database, which does everything for you in one system, whereas the suite of specialized databases, right? Oracle is in the classic Oracle database in the universal database camp. On the other side you have Amazon, which built a database. This is one of the first few Amazon re:Invents. It's my 10th where there was no new database announced. Right? >> No. >> So it was always add another one specially- >> I think they have enough. >> It's a great approach. They have enough, right? So it's a great approach to build something quick, which Amazon is all about. It's not so great when customers want to leverage things. And, ultimately, which I think with Selipsky, AWS is waking up to the enterprise saying, "I have all this different database and what is in them matters to me." >> Yeah. >> "So how can I get this better?" So no surprise between the two most popular database, Aurora and RDS. They're bring together the data with some out of the box parts. I think it's kind of, like, silly when Swami's saying, "Hey, no ETL.". (chuckles) Right? >> Yeah. >> There shouldn't be an ETL from the same vendor, right? There should be data pipes from that perspective anyway. So it looks like, on the overall value proposition database side, AWS is moving closer to the universal database on the Oracle side, right? Because, if you lift, of course, the universal database, under the hood, you see, well, there's different database there, different part there, you do something there, you have to configure stuff, which is also the case but it's one part of it, right, so. >> With that shift, talk about the value that's going to be in it for customers regardless of industry. >> Well, the value for customers is great, because when software vendors, or platform vendors, go in depth, you get more functionality, you get more maturity you get easier ways of setting up the whole things. You get ways of maintaining things. And you, ultimately, get lower TCO to build them, which is super important for enterprise. Because, here, this is the developer cloud, right? Developers love AWS. Developers are scarce, expensive. Might not be want to work for you, right? So developer velocity getting more done with same amount of developers, getting less done, less developers getting more done, is super crucial, super important. So this is all good news for enterprise banking on AWS and then providing them more efficiency, more automation, out of the box. >> Some of your customer conversations this week, talk to us about some of the feedback. What's the common denominator amongst customers right now? >> Customers are excited. First of all, like, first event, again in person, large, right? >> Yeah. >> People can travel, people meet each other, meet in person. They have a good handle around the complexity, which used to be a huge challenge in the past, because people say, "Do I do this?" I know so many CXOs saying, "Yeah, I want to build, say, something in IoT with AWS. The first reference built it like this, the next reference built it completely different. The third one built it completely different again. So now I'm doubting if my team has the skills to build things successfully, because will they be smart enough, like your teams, because there's no repetitiveness and that repetitiveness is going to be very important for AWS to come up with some higher packaging and version numbers.", right? But customers like that message. They like that things are working better together. They're not missing the big announcement, right? One of the traditional things of AWS would be, and they made it even proud, as a system, Jassy was saying, "If we look at the IT spend and we see something which is, like, high margin for us and not served well and we announced something there, right?" So Quick Start, Workspaces, where all liaisons where AWS went after traditional IT spend and had an offering. We haven't had this in 2019, we don't have them in 2020. Last year and didn't have it now. So something is changing on the AWS side. It's a little bit too early to figure out what, but they're not chewing off as many big things as they used in the past. >> Right. >> Yep. >> Did you get the sense that... Keith Townsend, from "The CTO Advisor", was on earlier. >> Yep. >> And he said he's been to many re:Invents, as you have, and he said that he got the sense that this is Amazon's chance to do a victory lap, as he called it. That this is a way for Amazon to reinforce the leadership cloud. >> Ja. >> And really, kind of, establish that nobody can come close to them, nobody can compete with them. >> You don't think that- >> I don't think that's at all... I mean, love Keith, he's a great guy, but I don't think that's the mindset at all, right? So, I mean, Jassy was always saying, "It's still the morning of the day in the cloud.", right? They're far away from being done. They're obsessed over being right. They do more work with the analysts. We think we got something right. And I like the passion, from that perspective. So I think Amazon's far from being complacent and the area, which is the biggest bit, right, the biggest. The only thing where Amazon truly has floundered, always floundered, is the AI space, right? So, 2018, Werner Vogels was doing more technical stuff that "Oh, this is all about linear regression.", right? And Amazon didn't start to put algorithms on silicon, right? And they have a three four trail and they didn't announce anything new here, behind Google who's been doing this for much, much longer than TPU platform, so. >> But they have now. >> They're keen aware. >> Yep. >> They now have three, or they own two of their own hardware platforms for AI. >> Right. >> They support the Intel platform. They seem to be catching up in that area. >> It's very hard to catch up on hardware, right? Because, there's release cycles, right? And just the volume that, just talking about the largest models that we have right now, to do with the language models, and Google is just doing a side note of saying, "Oh, we supported 50 less or 30 less, not little spoken languages, which I've never even heard of, because they're under banked and under supported and here's the language model, right? And I think it's all about little bit the organizational DNA of a company. I'm a strong believer in that. And, you have to remember AWS comes from the retail side, right? >> Yeah. >> Their roll out of data centers follows their retail strategy. Open secret, right? But, the same thing as the scale of the AI is very very different than if you take a look over at Google where it makes sense of the internet, right? The scale right away >> Right. >> is a solution, which is a good solution for some of the DNA of AWS. Also, Microsoft Azure is good. There has no chance to even get off the ship of that at Google, right? And these leaders with Google and it's not getting smaller, right? We didn't hear anything. I mean so much focused on data. Why do they focus so much on data? Because, data is the first step for AI. If AWS was doing a victory lap, data would've been done. They would own data, right? They would have a competitor to BigQuery Omni from the Google side to get data from the different clouds. There's crickets on that topic, right? So I think they know that they're catching up on the AI side, but it's really, really hard. It's not like in software where you can't acquire someone they could acquire in video. >> Not at Core Donovan. >> Might play a game, but that's not a good idea, right? So you can't, there's no shortcuts on the hardware side. As much as I'm a software guy and love software and don't like hardware, it's always a pain, right? There's no shortcuts there and there's nothing, which I think, has a new Artanium instance, of course, certainly, but they're not catching up. The distance is the same, yep. >> One of the things is funny, one of our guests, I think it was Tuesday, it was, it was right after Adam's keynote. >> Sure. >> Said that Adam Selipsky stood up on stage and talked about data for 52 minutes. >> Yeah. Right. >> It was timed, 52 minutes. >> Right. >> Huge emphasis on that. One of the things that Adam said to John Furrier when they were able to sit down >> Yeah >> a week or so ago at an event preview, was that CIOs and CEOs are not coming to Adam to talk about technology. They want to talk about transformation. They want to talk about business transformation. >> Sure, yes, yes. >> Talk to me in our last couple of minutes about what CEOs and CIOs are coming to you saying, "Holger, help us figure this out. We have to transform the business." >> Right. So we advise, I'm going quote our friends at Gartner, once the type A company. So we'll use technology aggressively, right? So take everything in the audience with a grain of salt, followers are the laggards, and so on. So for them, it's really the cusp of doing AI, right? Getting that data together. It has to be in the cloud. We live in the air of infinite computing. The cloud makes computing infinite, both from a storage, from a compute perspective, from an AI perspective, and then define new business models and create new best practices on top of that. Because, in the past, everything was fine out on premise, right? We talked about the (indistinct) size. Now in the cloud, it's just the business model to say, "Do I want to have a little more AI? Do I want a to run a little more? Will it give me the insight in the business?". So, that's the transformation that is happening, really. So, bringing your data together, this live conversation data, but not for bringing the data together. There's often the big win for the business for the first time to see the data. AWS is banking on that. The supply chain product, as an example. So many disparate systems, bring them them together. Big win for the business. But, the win for the business, ultimately, is when you change the paradigm from the user showing up to do something, to software doing stuff for us, right? >> Right. >> We have too much in this operator paradigm. If the user doesn't show up, doesn't find the click, doesn't find where to go, nothing happens. It can't be done in the 21st century, right? Software has to look over your shoulder. >> Good point. >> Understand one for you, autonomous self-driving systems. That's what CXOs, who're future looking, will be talked to come to AWS and all the other cloud vendors. >> Got it, last question for you. We're making a sizzle reel on Instagram. >> Yeah. >> If you had, like, a phrase, like, or a 30 second pitch that would describe re:Invent 2022 in the direction the company's going. What would that elevator pitch say? >> 30 second pitch? >> Yeah. >> All right, just timing. AWS is doing well. It's providing more depth, less breadth. Making things work together. It's catching up in some areas, has some interesting offerings, like the healthcare offering, the security data lake offering, which might change some things in the industry. It's staying the course and it's going strong. >> Ah, beautifully said, Holger. Thank you so much for joining Paul and me. >> Might have been too short. I don't know. (laughs) >> About 10 seconds left over. >> It was perfect, absolutely perfect. >> Thanks for having me. >> Perfect sizzle reel. >> Appreciate it. >> We appreciate your insights, what you're seeing this week, and the direction the company is going. We can't wait to see what happens in the next year. And, yeah. >> Thanks for having me. >> And of course, we've been on so many times. We know we're going to have you back. (laughs) >> Looking forward to it, thank you. >> All right, for Holger Mueller and Paul Gillan, I'm Lisa Martin. You're watching "theCube", the leader in live enterprise and emerging tech coverage. (upbeat music)

Published Date : Dec 1 2022

SUMMARY :

across the AWS ecosystem. of people here. and how the world is, And Holger, welcome, on the final day of the marathon, right? And, of course, or the World Cup. They just got eliminated. What will the U.S. do They're going to win. Hope for the best experts we are. was right. Biggest event of the year again, right? and the entertainment area, and the food area, the big empty booths You know, the white spaces in the developer community, right. Maybe goes back to So, how could Werner pick up and run the payroll, the enterprise to build, This is the first re:Invent... Right. a lot of momentum. compared to the Jassy times, right, more on the depth side, in the major announcement one of the big announcements early on And setting up for I mean, Oracle is doing the same thing. This is one of the first to build something quick, So no surprise between the So it looks like, on the overall talk about the value Well, the value for customers is great, What's the common denominator First of all, like, So something is changing on the AWS side. Did you get the sense that... and he said that he got the sense that can come close to them, And I like the passion, or they own two of their own the Intel platform. and here's the language model, right? But, the same thing as the scale of the AI from the Google side to get The distance is the same, yep. One of the things is funny, Said that Adam Selipsky Yeah. One of the things that are not coming to Adam coming to you saying, for the first time to see the data. It can't be done in the come to AWS and all the We're making a sizzle reel on Instagram. 2022 in the direction It's staying the course Paul and me. I don't know. It was perfect, and the direction the company is going. And of course, we've the leader in live enterprise

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
PaulPERSON

0.99+

HolgerPERSON

0.99+

AdamPERSON

0.99+

ScottPERSON

0.99+

Adam SelipskyPERSON

0.99+

Lisa MartinPERSON

0.99+

AmazonORGANIZATION

0.99+

JassyPERSON

0.99+

KeithPERSON

0.99+

GartnerORGANIZATION

0.99+

Paul GillanPERSON

0.99+

23QUANTITY

0.99+

AWSORGANIZATION

0.99+

twoQUANTITY

0.99+

2019DATE

0.99+

TuesdayDATE

0.99+

2020DATE

0.99+

Las VegasLOCATION

0.99+

Last yearDATE

0.99+

GoogleORGANIZATION

0.99+

Holger MuellerPERSON

0.99+

Keith TownsendPERSON

0.99+

Werner VogelsPERSON

0.99+

OracleORGANIZATION

0.99+

WernerPERSON

0.99+

21st centuryDATE

0.99+

52 minutesQUANTITY

0.99+

threeQUANTITY

0.99+

yesterdayDATE

0.99+

2018DATE

0.99+

Holger MuellerPERSON

0.99+

10thQUANTITY

0.99+

firstQUANTITY

0.99+

TomorrowDATE

0.99+

NetherlandsORGANIZATION

0.99+

U.S.ORGANIZATION

0.99+

50QUANTITY

0.99+

tomorrowDATE

0.99+

LisaPERSON

0.99+

first timeQUANTITY

0.99+

50,000 peopleQUANTITY

0.99+

John FurrierPERSON

0.99+

AntarcticaLOCATION

0.99+

MicrosoftORGANIZATION

0.99+

third oneQUANTITY

0.99+

2QUANTITY

0.99+

Wes Barnes, Pfizer and Jon Harrison, Accenture | AWS Executive Summit 2022


 

(mellow music) >> Oh, welcome back to theCUBE. We continue our coverage here at AWS re:Invent 22. We're in the Venetian in Las Vegas, and this place is hopping. I'm tell you what. It is a nearly standing room only that exhibit floor is jam packed, and it's been great to be along for the ride here on Accenture's sponsorship at the Executive summit as well. We'll talk about Pfizer today, you know them quite well, one of the largest biopharmaceutical companies in the world but their tech footprint is impressive, to say the least. And to talk more about that is Wes Barnes, senior Director of Pfizer's Digital Hosting Solutions. Wes, good to see you, sir. >> Good to meet you, John. >> And Jon Harrison, the North American lead for Infrastructure and engineering at Accenture. Jon, good to see you as well. >> Good to see you as well. >> Thanks for joining us. >> Happy to be here. >> Alright, so let's jump in. Pfizer, we make drugs, right? >> Pharmaceuticals. >> Yes. >> Among the most preeminent, as I said biopharms in the world. But your tech capabilities and your tech focus as we were talking about earlier, has changed dramatically in the 18 years that you've been there. >> Yep. >> Now, talk about that evolution a little bit to where you were and what you have to be now. >> Yeah, yeah. It's interesting. When I started at Pfizer, IT was an enabling function. It was akin to HR or our facilities function. And over the past couple years, it's dramatically changed. Where Digital now is really at the center of everything we do across Pfizer. You know it really is a core strategic element of our business. >> Yeah. And those elements that you were talking about, just in terms of whether it's research, whether it's your patients, I don't want to go through the laundry lists the litany of things, but the touch points with data and what you need it to do for you in terms of you know, computations, what you, the list is long. It's pretty impressive. >> Yeah, yeah, for sure yeah. >> I mean, shed some light on that for us. >> We cannot release a medicine without the use of technology. And if you think about research now, a huge component of our research is computational chemistry. Manufacturing medicines now is a practice in using data and analytics and predictive machine learning and analytics capabilities to help us determine how to best you know, apply the capabilities to deliver the outcomes that we need. The way in which we connect with patients and payers now is wholly digital. So it's an entirely different way of operating than it was 10 years ago. >> And the past three years, pretty remarkable in many respects, to say the least, I would think, I mean, John, you've seen what Pfizer's been up to, talk about maybe just this, the recent past and all that has happened and what they've been able to do. >> Yeah, I mean, what is so exciting to me about working with a company like Pfizer and working in life sciences more broadly is the impact that they make on patients around the world world, right? I mean, think about those past three years and Pfizer stepped up and met the moment for all of us, right? And as we talk a little bit about the role that we played together with Pfizer with AWS in their journey to the cloud, it's so motivating for myself personally it's so motivating for every single person on the team that we ask to spend nights and weekends migrating things to the cloud, creating new capabilities, knowing that at the end of the day, the work that they're doing is making the world a healthier place. >> Yeah, we talk so much about modernization now, right? And it's, but it kind of means different things to different people depending on where you're coming into the game, right? If you've been smart and been planning all along then this is not a dramatic shift in some cases though, for others it is. Right? >> Yeah. >> Traumatic in some cases for some people. >> For sure. >> For Pfizer, I mean talk about how do you see modernization and what does it mean to your operations? >> Following our success of the COVID program of 2021, I mean it became evident to us that, you know we needed to maintain a new pace of innovation and in fact try to find ways to accelerate that pace of innovation. And as I said earlier everything we do at Pfizer is centered around digital. But despite that, and despite 10 years of consolidating infrastructure and moving towards modern technology, last year, only 10% of Pfizer's infrastructure was in the native public cloud. So we had a problem to solve. In fact, I remember, you know, we had to build up our clinical systems to support the volume of work that we were doing for COVID-19 vaccine. We were rolling things into our data center to build up the capacity to achieve what we needed to achieve. Moving to the public cloud became more imperative to try to achieve the scale and the modern capabilities that we need. >> And so where did you come into play here with this? Because obviously as a partner you're right alongside for the ride but you saw these inherent challenges that they had and how did Accenture answer the bell there? >> Well, so look, I mean we saw Pfizer react to the pandemic. We saw them seize the moment. We talked together about how IT needed to move quicker and quicker towards the cloud to unlock capabilities that would serve Pfizer's business well into the future. And together we laid out some pretty ambitious goals. I mean, really moving at a velocity in a pace that I think for both Accenture and AWS surpassed the velocity and pace that we've done anywhere else. >> Yeah, right, yeah. >> So we've set out on an ambitious plan together. You know, I was kind of reflecting about some of the successes, what went well what didn't in preparation for re:Invent. And you know, many of the folks that'll listen to this will remember the old days of moving data centers when you'd have a war room you'd have a conference bridge open the whole time. Someone would be running around the tile floor in the data center, do a task, call back up to the bridge and say, what do I do next, right? Then when I think about what we did together at Pfizer in moving towards the public cloud, I mean, we had weekends most weekends where we were running a wave with 10,000 plus discrete activities. >> Yeah. >> Wow. >> Right, so that old model doesn't scale. >> Right. >> And we really anchored, >> You have a very crowded data center with a lot of people running into each other. >> You'd have a whole lot of people running around. But we really anchored to an Accenture capability that we call myNav Migrate. I know you guys have talked about it here before so I won't go into that. But what we found is that we approached this problem of velocity not as a technical problem to solve for but as a loading and optimization problem of resources. Right, thought about it just a little bit different way and made sure that we could programmatically control command and control of the program in a way that people didn't have to wait around all Saturday afternoon to be notified that their next activity was ready, right? They could go out, they could live their day and they could get a notification from the platform that says, hey it's about your turn. Right, they could claim it they could do it, they could finish it, and that was really important to us. I mean, to be able to control the program in that type of way at scale. >> Yeah, by the way, the reason we went as fast it was a deliberate choice and you'll talk to plenty of folks who have a five year journey to the public cloud. And the reason we wanted to move as fast as we did and Jon talked about some of it, we wanted to get the capabilities to the business as quickly as we could. The pace of innovation was such that we had to offer native cloud capabilities we had to offer quickly. We also knew that by compressing the time it took to get to the cloud, we could focus the organization get it done as economically as possible but then lift all boats with the tide and move the organization forward in terms of the skills and the capabilities that we need to deliver modern outcomes. >> So, you know, we talk about impacts internally, obviously with your processes, but beyond that, not just scientists not just chemists, but to your, I mean, millions of customers, right? We're talking, you know, globally here. What kind of impacts can you see that directly relate to them, and benefits that they're receiving by this massive technical move you've made? >> Pfizer's mission is breakthroughs that change patient lives. I mean, the work that we do the work that everybody does within Pfizer is about delivering therapies that, you know provide health outcomes that make people live longer, live healthier lives. For us, modernizing our infrastructure means that we can enable the work of scientists to find novel therapies faster or find things that perhaps couldn't have been found any other way without some of the modern technologies that we're bringing to bear. Saving money within infrastructure and IT is treasure that we can pour back into the important areas of research or development or manufacturing. We're also able to, you know, offer an ecosystem and a capability in which we connect with patients differently through digital mechanisms. And modern cloud enables that, you know, using modern digital experiences and customer experience, and patient experience platforms means that we can use wearable devices and mobile technologies and connect to people in different ways and offer solutions that just didn't exist a couple years ago. >> And so, I mean, you're talking about IoT stuff too, right? >> 100%. >> It's way out on the edge and personal mobile, in a mobile environment. And so challenges in terms of you know, data governance and compliance and security, all these things, right? They come into play because it's personal health information. So how, as you've taken them, you know to this public cloud environment how much of a factor are those considerations? Because, you know, this is not just a product a service, it's a live human being. >> Yeah. I mean, you start with that, you think about it through the process and you think about it afterwards, right, I mean, that has to be a core factor in every stage of the program, and it was. >> So in, in terms of where you are now, then, okay, it's not over. >> It's never over. >> I mean, you know, as good as you are today and as fast as you are and as accurate and as efficient. >> Yeah. >> Got to get better, right? You got to stay competitive. >> Yeah. >> So where do you find that? Because, you know, with powers being what they are with speed and what it is how much more is there to squeeze out of this rock? >> There's a lot more to squeeze out of the rock. If you think about what we've done over the past year it's about creating sort of a new minimum viable product for infrastructure. So we've sort of raised the bar and created an environment upon which we can continue to innovate that innovation is going to continue sort of forever at this point. You know, the next focus for us is how to identify the business processes that deliver the greatest value ultimately to our patients. And use the modern platform that we've just built to improve those processes to deliver things faster, deliver new capabilities. Pfizer is making a huge investment in digital medicines therapies that are delivered through smart devices through wearables using, as I said technology that didn't exist before. That wouldn't be possible without the platform that we've built. So over the past year, we've come a long way but I think that we've effectively set the table for all of the things that are yet to come. >> So, Jon, how do you then, as you've learned a lot about life science or, and certainly Pfizer with what they're up to, how do you then apply, you know, what you know about their world to what you know about the tech world and make it actionable for growth to make it actionable for, for future expansion? >> Yeah, I mean, we start by doing it together, right? I think that's a really important part. Accenture brings a wealth of knowledge, both industry experience and expertise, technology experience and expertise. We work together with our clients like Pfizer with our partners like AWS to bring the best across that power of three to meet clients where they're at to understand where they want to go, and then create a bespoke approach that meets their business needs. And that's effectively what we're doing now, right? I mean, if you think about the phase that we've just went through, I mean, a couple of fast facts here no pun intended, right? 7,800 server instances across 11 operating system versions 7,500 databases across 20 database versions, right? 4,700 applications, 350,000 migration activities managed across an eight month period. >> In eight months. >> Yeah. But that's not the goal, right? The goal is now to take, to Wes' point that platform that's been developed and leverage that to the benefit of the business ultimately to the benefit of the patient. >> You know, why them, we have we've talked a lot about Pfizer, but why Accenture? What, what, what's, 'cause it's got to be a two way street, right? >> We've had a long partnership with Accenture. Accenture supports a huge component of our application environment at Pfizer and has for quite a long time. Look, we didn't make it easy on them. We put them up against a large number of world class SIs. But look, Accenture brought, you know, sort of what I think of as the trifecta here. They brought the technical capabilities and knowledge of the AWS environment. They brought the ability to really understand the business outcomes that we were trying to achieve and a program leadership capability that, you know I think is world class. And Jon talked about myNav, you know, we recognized that doing what we were trying to do in the time that we were doing it required new machinery, new analytics and data capabilities that just didn't exist. Automation didn't exist. Some people experience capabilities that would allow us to interface with application owners and users at a velocity and a pace and a scale that just hasn't been seen before at Pfizer. Accenture brought all three of those things together and I think they did a great job helping us get to where we need to be. >> When you hear Jon rattle through the stats like he just did, right? We talk about all, I mean, not that I'm going to ask you to pat yourself on the back but do you ever, >> He should. >> Does it blow your mind a little bit, honestly that you're talking about that magnitude of activity in that compressed period of time? That's extraordinary. >> It's 75% of our global IT footprint now in the public cloud, which is fantastic. I mean, look, I think the timing was right. I think Pfizer is in a little bit of a unique position coming off of COVID. We are incredibly motivated to keep the pace up, I mean across all lines of business. So, you know what we found is a really willing leadership team, executive leadership team, digital leadership team to endorse a change of this magnitude. >> Well, it's a great success story. It's beyond impressive. So congratulations to both you on that front and certainly you wish you continued success down the road as well. >> Thank you. >> Thank you gentlemen. >> Thank you. >> Good job. >> Pfizer, and boy, you talk about a job well done. Just spectacular. All right, you are watching our coverage here on theCUBE, we're at the AWS re:Invent 22 show. This is Executive Summit sponsored by Accenture and you're watching theCUBE the leader in high tech coverage.

Published Date : Dec 1 2022

SUMMARY :

and it's been great to be Jon, good to see you as well. Pfizer, we make drugs, right? has changed dramatically in the 18 years to where you were and And over the past couple years, and what you need it to how to best you know, And the past three years, on the team that we ask to to different people depending on Traumatic in some and the modern capabilities that we need. and pace that we've done anywhere else. And you know, many of with a lot of people and made sure that we could get the capabilities to the that directly relate to them, I mean, the work that we do of you know, data governance in every stage of the program, and it was. So in, in terms of where you are now, and as fast as you are and You got to stay competitive. that deliver the greatest value across that power of three to and leverage that to the of the AWS environment. of activity in that in the public cloud, which is fantastic. and certainly you wish Pfizer, and boy, you

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
PfizerORGANIZATION

0.99+

JonPERSON

0.99+

JohnPERSON

0.99+

AccentureORGANIZATION

0.99+

Jon HarrisonPERSON

0.99+

4,700 applicationsQUANTITY

0.99+

75%QUANTITY

0.99+

five yearQUANTITY

0.99+

AWSORGANIZATION

0.99+

last yearDATE

0.99+

Wes BarnesPERSON

0.99+

20 database versionsQUANTITY

0.99+

7,500 databasesQUANTITY

0.99+

10 yearsQUANTITY

0.99+

eight monthQUANTITY

0.99+

100%QUANTITY

0.99+

threeQUANTITY

0.99+

Saturday afternoonDATE

0.99+

Las VegasLOCATION

0.99+

18 yearsQUANTITY

0.99+

WesPERSON

0.99+

VenetianLOCATION

0.98+

bothQUANTITY

0.98+

eight monthsQUANTITY

0.98+

pandemicEVENT

0.97+

10 years agoDATE

0.97+

10%QUANTITY

0.97+

10,000 plus discrete activitiesQUANTITY

0.96+

two wayQUANTITY

0.95+

todayDATE

0.95+

oneQUANTITY

0.94+

7,800 server instancesQUANTITY

0.93+

millions of customersQUANTITY

0.93+

couple years agoDATE

0.9+

Executive SummitEVENT

0.9+

COVID-19 vaccineOTHER

0.89+

COVID programTITLE

0.88+

past three yearsDATE

0.87+

past yearDATE

0.87+

every single personQUANTITY

0.86+

11 operating system versionsQUANTITY

0.85+

AWSEVENT

0.8+

350,000 migration activitiesQUANTITY

0.79+

Digital Hosting SolutionsORGANIZATION

0.77+

re:Invent 22EVENT

0.75+

North AmericanOTHER

0.71+

yearsDATE

0.69+

Wes'PERSON

0.67+

Jon Bakke, MariaDB | AWS re:Invent 2022


 

(bright upbeat music) >> Welcome back everyone to theCUBE's live coverage here in Las Vegas for wall-to-wall coverage. It is re:Invent 2022, our 10th year with theCUBE. Dave and I started this journey 10 years ago here at re:Invent. There are two sets, here, a set upstairs. Great content, I'm here with Paul Gillin, my cohost. Paul's out reporting on the floor, doing some interviews. Paul, what do you think so far? It's pretty crazy activity going on here. >> Well, the activity hasn't declined at all. I mean here we are in day three of the show and it's just as busy out there as it was in day one. And there's just an energy here that you can feel, it's palpable. There is a lot of activity around developers, a lot around data. Which actually brings us a good segue into our next guest because one of the leaders in data management in the cloud is MariaDB. And John Bakke is the CRO at MariaDB, and here to talk to us about your cloud version and how open source is going for you. >> Yeah, thanks for having me. >> Paul: Thanks for joining us. >> To get the update on the product, what do you guys do on the relation to AWS? How's that going? Give us a quick update. >> In the relational database? >> No, no. The relationship with AWS >> Oh, with AWS? >> And SkySQL, what's the update? >> There's no relationship that we have that's more important than the AWS relationship. We're building our cloud, our premier cloud service called SkySQL on AWS. And they offer the best in class infrastructure for a SaaS company to build what they're building. And for us, it's a database service, right? And then beyond that, they help you from the business side, right? They try to get you lined up in the marketplace and make it possible for you to work best with customers. And then from a customer perspective, they're super helpful in not only finding prospective customers, but making that customer successful. 'Cause everybody's got a vested interest in the outcome. Right? >> Yeah, a little tongue twister there. Relational data-based relationship. We've got relational databases, we've got unstructured, data is at the center of the value proposition. Swami's keynote today and the Adam CEO's keynote, data and security dominated the keynotes >> John: Yes. >> and the conversations. So, this is real. The customers are really wanting to accelerate the developer experience, >> John: Yep. >> Developer pipe lining, more code faster, more horsepower under the hood. But this data conversation, it just never goes away. The world's keeping on coming around. >> John: It never goes away. I've been in this business for almost 30 years and we're still talking about the same key factors, right? Reliability, availability, performance, security. These things are pervasive in the data management because it's such a critical aspect to success. >> Yeah, in this case of SkySQL, you have both a transactional and an analytical engine in one. >> John: That's correct. >> Right? >> John: Yep. >> And that was a, what has the customer adoption been like of that hybrid, or I guess not a hybrid, but a dual function? >> Yeah. So the thing that makes that important is that instead of having siloed services, you have integrated data services. And a lot of times when you ask a question that's analytical it might depend on a transaction. And so, that makes the entire experience best for the developer, right? So, to take that further, we also, in SkySQL, offer a geospatial offering that integrates with all of that. And then we even take it further than that with distributed database with Xpand or ready to be Xpand. >> A lot of discussion. Geospatial announcement today on stage, just the diversity of data, and your experience in the industry. There's not the one database that rule them all anymore. There's a lot of databases out there. How are customers dealing with, I won't say database for all, 'Cause you need databases. And then you've got real time transactional, you got batch going on, you got streaming data, all kinds of data use cases now, all kind of having to be rolled together. What's your reaction? What's your take on the state of data and databases? >> Yeah, yeah, yeah. So when I started in this business, there were four databases, and now there's 400 databases. And the best databases really facilitate great application development. So having as many of those services in real time or in analytics as possible means that you are a database for everyone or for all users, right? And customers don't want to use multiple databases. Sometimes they feel like they're forced to do that, but if you're like MariaDB, then you offer all of those capabilities in an integrated way that makes the developer move faster. >> Amazon made a number of announcements this morning in the data management area, including geospatial support on RDS, I believe. How do you, I guess, coordinate yourself, your sales message with their sales message, given that you are partners, but they are competing with you in some ways? >> Yeah, there's always some cooperatition, I guess, that happens with AWS in the various product silos that they're offering their customers. For us, we're one of thousands of obviously partners that they have. And we're out there trying to do what our customers want, which is to have those services integrated, not glued together with a variety of different integration software. We want it integrated in the service so that it's one data provision, data capability for the application developer. It makes for a better experience for the developer in the end. >> On the customer side, what's the big activity? I mean, you got the on-premises database, you've got the cloud. When should a customer decide, or what's the signals to them that they should either move to the cloud, or change, be distributed? What are some of the forcing functions? What does the mark look like? >> Yeah, I've come a long way on this, but my opinion is that every customer should be in the cloud. And the reason simply is the economies that are involved, the pace of execution, the resilience and dependability of the cloud, Amazon being the leader in that space. So if you were to ask me, right now is the time to be in SkySQL because it's the premier data service in the cloud. So I would take my customer out of their on-prem and put them all in AWS, on SkySQL, if I could. Not everybody's ready for that, but my opinion is that the security is there, the reliability, the privacy, all of the things that maybe are legacy concerns, it's all been proven to be adequate and probably even better because of all of the economies of scale that you get out of being in the cloud just generally. >> Now, MariaDB, you started on-premise though. You still have a significant customer base on-premise. What, if anything are you doing to encourage them to migrate to the cloud? >> Well, so we have hundreds and hundreds of customers as MariaDB, and we weren't the first database company to put their database in the cloud, but watching it unfold helped us realize that we're going to put MariaDB in its best form factor in SkySQL. It's the only place you could get the enterprise version of MariaDB in a cloud service, right? So when we look at our customers on-prem, we're constantly telling them, obviously, that we have a cloud service. When they subscribe, we show them the efficiencies and the economies, and we do get customers that are moving. We had a customer go to Telefonica over in the UK that moved from an on-premise to manage their wifi services across Europe. And they're very happy. They were one of our very first SkySQL customers. And that has routinely proven itself to be a path towards not only a better operation for the customer, they're up more, they have fewer outages because they're not inflicting their own self wounds that they have in their own data center. They're running on world class infrastructure on world class databases. >> What are some of those self wounds? Is it personnel, kind of manual mistakes, just outages, reliability? What's the real cause, and then what's the benefit alternative in the cloud that is outside? >> Yeah. I mean, I think, when you repeat the same database implementation over and over on the infrastructure, it gets tested thousands and thousands of times. Whereas if I'm a database team and I install it once, I've tested it one time, and I can't account for all of the things that might happen in the future. So the benefit of the cloud is that you just get that repeat ability that happens and all of the sort of the kinks and bugs and issues are worked out of the system. And that's why it's just fundamentally better. We get 99.9999% uptime because all of those mistakes have been made, solved, and fixed. >> Fully managed, obviously. >> Yes. Right. >> Huge benefit. >> John: Right. >> And people are moving, it's just a great benefit. >> John: Yeah. >> So I'm a fan obviously. I think it's a great way to go. I got to ask about the security though, because big conversation here is security. What's the security posture? What's the security story to customers with SkySQL and MariaDB? >> Right, right, right. So we've taken the server, which was the initial product that MariaDB was founded upon, right? And we've come a long way over the several years that we've been in business. In SkySQL, we have SOC 2 compliance, for example. So we've gone through commercial certifications to make sure that customers can depend that we are following processes, we have technology in place in order to secure and protect their data. And in that environment, it is repeatable. So every time a customer uses our DBaaS infrastructure, databases a service infrastructure called SkySQL, they're benefiting from all of the testing that's been done. They go there and do that themselves, they would've to go through months and months of processes in order to reach the same level of protection. >> Now MariaDB is distributed by design. Is that right? >> Yes. So we have a distributed database, it's called Xpand, MariaDB Xpand. And it's an option inside of SkySQL. It's the same cost as MariaDB server, but Xpand is distributed. And the easiest way to understand what distributed database is is to understand what it is not first. What it is not is like every other cloud database. So most of the databases strangely in the cloud are not distributed databases. They have one single database node in a cluster that is where all of the changes and rights happen. And that creates a bottleneck in the database. And that's why there's difficulties in scale. AWS actually talked about this in the keynote which is the difficulty around multi writer in the cloud. And that's what Xpand does. And it spreads out the reads and the rights to make it scalable, more performant, and more resilient. One node goes down, still stays up, but you get the benefit of the consistency and the parallelization that happens in Xpand. >> So when would a customer choose Xpand versus SkySQL Vanilla? >> So we have, I would say a lot of times, but the profile of our customers are typically like financial services, trade stores. We have Samsung Cloud, 500,000 transactions per second in an expand cluster where they run sort of their Samsung cloud for their mobile device unit. We have many customers like that where it's a commercial facing website often or a service where the brand depends on uptime. Okay. So if you're in exchange or if you are a mobile device company or an IOT company, you need those databases to be working all the time and scale broadly and have high performance. >> So you have resiliency built in essentially? >> Yes, yeah. And that's the major benefit of it. It hasn't been solved by anybody other than us in the cloud to be quite honest with you. >> That's a differentiator for sure. >> It is a huge differentiator, and there are a lot of interested parties. We're going to see that be the next discussion probably next year when we come back is, what's the state of distributed database? Because it's really become really the tip of the spear with the database industry right now. >> And what's the benefits of that? Just quickly describe why that's important? >> Obviously the performance and the resilience are the two we just talked about, but also the efficiency. So if you have a multi-node cluster of a single master database, that gets replicated four times, five times over, five times the cost. And so we're taking cost out, adding performance in. And so, you're really seeing a revolution there because you're getting a lot more for a lot less. And whenever you do that, you win the game. Right? >> Awesome. Yeah, that's true. And it seems like, okay, that might be more costly but you're not replicating. >> That's right. >> That's the key. >> Replicating just enough to be resilient but not excessively to be overly redundant. Right. >> Yeah. I find that the conversation this year is starting to unpack some of these cloud native embedded capabilities inside AWS. So are you guys doing more around, on the customer side, around marketplace? Are you guys, how do people consume products? >> Yeah. It's really both. So sometimes they come to us from AWS. AWS might say, "Hey, you know what," "we don't really have an answer." And that's specifically true on the expand side. They don't really have that in their list of databases yet. Right. Hopefully, we'll get out in front of them. But they oftentimes come through our front door where they're a MariaDB customer already, right? There's over a hundred thousand production systems with MariaDB in the world, and hundreds of thousands of users of the database. So they know our brand, not quite as well as AWS, but they know our brand... >> You've got a customer base. >> We do. Right. I mean people love MariaDB. They just think it's the database that they use for application development all the time. And when they see us release an offering like Xpand just a few years ago, they're interested, they want to use that. They want to see how that works. And then when they take it into production and it works as advertised, of course, success happens. Right? >> Well great stuff, John. Great to have you on theCUBE. Paul, I guess time we do the Insta challenge here. New format on theCUBE, we usually say at the end, summarize what's most important story for you or show, what's the bumper sticker? We kind of put it around more of an Instagram reel. What's your sizzle reel? What's your thought leadership statement? 30 seconds >> John: Thought leadership. >> John? >> So the thought leadership is really in scaling the cloud to the next generation. We believe MariaDB's Xpand product will be the the technology that fronts the next wave of database solutions in the cloud. And AWS has become instrumental in helping us do that with their infrastructure and all the help that they give us, I think at the end of the day, when the story on Xpand is written, it's going to be a very fun ride over the next few years. >> John, thank you. CRO, chief revenue officer of MariaDB, great to have you on. >> Thank you. >> 34-year veteran or so in databases. (laughs) >> You're putting a lot of age on me. I'm 29. I'm 29 again. (all laugh) >> I just graduated high school and I've been doing this for 10 years. Great to have you on theCUBE. Thanks for coming on. >> Thanks guys. Yeah. >> Thanks for sharing. >> Appreciate it. >> I'm John Furrier with Paul Gillin here live on the floor, wall-to-wall coverage. We're already into like 70 videos already. Got a whole another day, finish out day three. Keep watching theCUBE, thanks for watching. We'll be right back. (calm music)

Published Date : Dec 1 2022

SUMMARY :

Paul's out reporting on the And John Bakke is the CRO at MariaDB, the relation to AWS? than the AWS relationship. data is at the center of and the conversations. it just never goes away. in the data management and an analytical engine in one. And so, that makes the entire experience just the diversity of data, And the best databases in the data management area, in the various product silos What are some of the forcing functions? and dependability of the cloud, What, if anything are you doing and the economies, and I can't account for all of the things And people are moving, What's the security posture? And in that environment, it is repeatable. Is that right? So most of the databases but the profile of our customers the major benefit of it. really the tip of the spear and the resilience And it seems like, but not excessively to I find that the conversation So sometimes they come to us from AWS. development all the time. the Insta challenge here. and all the help that they give us, MariaDB, great to have you on. in databases. I'm 29. Great to have you on theCUBE. Yeah. here live on the floor,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Paul GillinPERSON

0.99+

AWSORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

DavePERSON

0.99+

five timesQUANTITY

0.99+

Jon BakkePERSON

0.99+

PaulPERSON

0.99+

hundredsQUANTITY

0.99+

400 databasesQUANTITY

0.99+

JohnPERSON

0.99+

EuropeLOCATION

0.99+

UKLOCATION

0.99+

10 yearsQUANTITY

0.99+

thousandsQUANTITY

0.99+

Las VegasLOCATION

0.99+

John BakkePERSON

0.99+

twoQUANTITY

0.99+

TelefonicaORGANIZATION

0.99+

four timesQUANTITY

0.99+

next yearDATE

0.99+

MariaDBTITLE

0.99+

10 years agoDATE

0.99+

99.9999%QUANTITY

0.99+

two setsQUANTITY

0.99+

30 secondsQUANTITY

0.99+

SwamiPERSON

0.99+

34-yearQUANTITY

0.99+

SkySQLTITLE

0.99+

bothQUANTITY

0.98+

John FurrierPERSON

0.98+

10th yearQUANTITY

0.98+

oneQUANTITY

0.98+

XpandTITLE

0.98+

70 videosQUANTITY

0.98+

theCUBEORGANIZATION

0.98+

first databaseQUANTITY

0.98+

almost 30 yearsQUANTITY

0.97+

day oneQUANTITY

0.97+

todayDATE

0.97+

one timeQUANTITY

0.97+

SamsungORGANIZATION

0.97+

OneQUANTITY

0.97+

day threeQUANTITY

0.97+

over a hundred thousand production systemsQUANTITY

0.97+

this yearDATE

0.97+

29QUANTITY

0.97+

firstQUANTITY

0.96+

AdamPERSON

0.95+

Wes Barnes and Jon Harrison Final


 

(mellow music) >> Oh, welcome back to theCUBE. We continue our coverage here at AWS re:Invent 22. We're in the Venetian in Las Vegas, and this place is hopping. I'm tell you what. It is a nearly standing room only that exhibit floor is jam packed, and it's been great to be along for the ride here on Accenture's sponsorship at the Executive summit as well. We'll talk about Pfizer today, you know them quite well, one of the largest biopharmaceutical companies in the world but their tech footprint is impressive, to say the least. And to talk more about that is Wes Barnes, senior Director of Pfizer's Digital Hosting Solutions. Wes, good to see you, sir. >> Good to meet you, John. >> And Jon Harrison, the North American lead for Infrastructure and engineering at Accenture. Jon, good to see you as well. >> Good to see you as well. >> Thanks for joining us. >> Happy to be here. >> Alright, so let's jump in. Pfizer, we make drugs, right? >> Pharmaceuticals. >> Yes. >> Among the most preeminent, as I said biopharms in the world. But your tech capabilities and your tech focus as we were talking about earlier, has changed dramatically in the 18 years that you've been there. >> Yep. >> Now, talk about that evolution a little bit to where you were and what you have to be now. >> Yeah, yeah. It's interesting. When I started at Pfizer, IT was an enabling function. It was akin to HR or our facilities function. And over the past couple years, it's dramatically changed. Where Digital now is really at the center of everything we do across Pfizer. You know it really is a core strategic element of our business. >> Yeah. And those elements that you were talking about, just in terms of whether it's research, whether it's your patients, I don't want to go through the laundry lists the litany of things, but the touch points with data and what you need it to do for you in terms of you know, computations, what you, the list is long. It's pretty impressive. >> Yeah, yeah, for sure yeah. >> I mean, shed some light on that for us. >> We cannot release a medicine without the use of technology. And if you think about research now, a huge component of our research is computational chemistry. Manufacturing medicines now is a practice in using data and analytics and predictive machine learning and analytics capabilities to help us determine how to best you know, apply the capabilities to deliver the outcomes that we need. The way in which we connect with patients and payers now is wholly digital. So it's an entirely different way of operating than it was 10 years ago. >> And the past three years, pretty remarkable in many respects, to say the least, I would think, I mean, John, you've seen what Pfizer's been up to, talk about maybe just this, the recent past and all that has happened and what they've been able to do. >> Yeah, I mean, what is so exciting to me about working with a company like Pfizer and working in life sciences more broadly is the impact that they make on patients around the world world, right? I mean, think about those past three years and Pfizer stepped up and met the moment for all of us, right? And as we talk a little bit about the role that we played together with Pfizer with AWS in their journey to the cloud, it's so motivating for myself personally it's so motivating for every single person on the team that we ask to spend nights and weekends migrating things to the cloud, creating new capabilities, knowing that at the end of the day, the work that they're doing is making the world a healthier place. >> Yeah, we talk so much about modernization now, right? And it's, but it kind of means different things to different people depending on where you're coming into the game, right? If you've been smart and been planning all along then this is not a dramatic shift in some cases though, for others it is. Right? >> Yeah. >> Traumatic in some cases for some people. >> For sure. >> For Pfizer, I mean talk about how do you see modernization and what does it mean to your operations? >> Following our success of the COVID program of 2021, I mean it became evident to us that, you know we needed to maintain a new pace of innovation and in fact try to find ways to accelerate that pace of innovation. And as I said earlier everything we do at Pfizer is centered around digital. But despite that, and despite 10 years of consolidating infrastructure and moving towards modern technology, last year, only 10% of Pfizer's infrastructure was in the native public cloud. So we had a problem to solve. In fact, I remember, you know, we had to build up our clinical systems to support the volume of work that we were doing for COVID-19 vaccine. We were rolling things into our data center to build up the capacity to achieve what we needed to achieve. Moving to the public cloud became more imperative to try to achieve the scale and the modern capabilities that we need. >> And so where did you come into play here with this? Because obviously as a partner you're right alongside for the ride but you saw these inherent challenges that they had and how did Accenture answer the bell there? >> Well, so look, I mean we saw Pfizer react to the pandemic. We saw them seize the moment. We talked together about how IT needed to move quicker and quicker towards the cloud to unlock capabilities that would serve Pfizer's business well into the future. And together we laid out some pretty ambitious goals. I mean, really moving at a velocity in a pace that I think for both Accenture and AWS surpassed the velocity and pace that we've done anywhere else. >> Yeah, right, yeah. >> So we've set out on an ambitious plan together. You know, I was kind of reflecting about some of the successes, what went well what didn't in preparation for re:Invent. And you know, many of the folks that'll listen to this will remember the old days of moving data centers when you'd have a war room you'd have a conference bridge open the whole time. Someone would be running around the tile floor in the data center, do a task, call back up to the bridge and say, what do I do next, right? Then when I think about what we did together at Pfizer in moving towards the public cloud, I mean, we had weekends most weekends where we were running a wave with 10,000 plus discrete activities. >> Yeah. >> Wow. >> Right, so that old model doesn't scale. >> Right. >> And we really anchored, >> You have a very crowded data center with a lot of people running into each other. >> You'd have a whole lot of people running around. But we really anchored to an Accenture capability that we call myNav Migrate. I know you guys have talked about it here before so I won't go into that. But what we found is that we approached this problem of velocity not as a technical problem to solve for but as a loading and optimization problem of resources. Right, thought about it just a little bit different way and made sure that we could programmatically control command and control of the program in a way that people didn't have to wait around all Saturday afternoon to be notified that their next activity was ready, right? They could go out, they could live their day and they could get a notification from the platform that says, hey it's about your turn. Right, they could claim it they could do it, they could finish it, and that was really important to us. I mean, to be able to control the program in that type of way at scale. >> Yeah, by the way, the reason we went as fast it was a deliberate choice and you'll talk to plenty of folks who have a five year journey to the public cloud. And the reason we wanted to move as fast as we did and Jon talked about some of it, we wanted to get the capabilities to the business as quickly as we could. The pace of innovation was such that we had to offer native cloud capabilities we had to offer quickly. We also knew that by compressing the time it took to get to the cloud, we could focus the organization get it done as economically as possible but then lift all boats with the tide and move the organization forward in terms of the skills and the capabilities that we need to deliver modern outcomes. >> So, you know, we talk about impacts internally, obviously with your processes, but beyond that, not just scientists not just chemists, but to your, I mean, millions of customers, right? We're talking, you know, globally here. What kind of impacts can you see that directly relate to them, and benefits that they're receiving by this massive technical move you've made? >> Pfizer's mission is breakthroughs that change patient lives. I mean, the work that we do the work that everybody does within Pfizer is about delivering therapies that, you know provide health outcomes that make people live longer, live healthier lives. For us, modernizing our infrastructure means that we can enable the work of scientists to find novel therapies faster or find things that perhaps couldn't have been found any other way without some of the modern technologies that we're bringing to bear. Saving money within infrastructure and IT is treasure that we can pour back into the important areas of research or development or manufacturing. We're also able to, you know, offer an ecosystem and a capability in which we connect with patients differently through digital mechanisms. And modern cloud enables that, you know, using modern digital experiences and customer experience, and patient experience platforms means that we can use wearable devices and mobile technologies and connect to people in different ways and offer solutions that just didn't exist a couple years ago. >> And so, I mean, you're talking about IoT stuff too, right? >> 100%. >> It's way out on the edge and personal mobile, in a mobile environment. And so challenges in terms of you know, data governance and compliance and security, all these things, right? They come into play because it's personal health information. So how, as you've taken them, you know to this public cloud environment how much of a factor are those considerations? Because, you know, this is not just a product a service, it's a live human being. >> Yeah. I mean, you start with that, you think about it through the process and you think about it afterwards, right, I mean, that has to be a core factor in every stage of the program, and it was. >> So in, in terms of where you are now, then, okay, it's not over. >> It's never over. >> I mean, you know, as good as you are today and as fast as you are and as accurate and as efficient. >> Yeah. >> Got to get better, right? You got to stay competitive. >> Yeah. >> So where do you find that? Because, you know, with powers being what they are with speed and what it is how much more is there to squeeze out of this rock? >> There's a lot more to squeeze out of the rock. If you think about what we've done over the past year it's about creating sort of a new minimum viable product for infrastructure. So we've sort of raised the bar and created an environment upon which we can continue to innovate that innovation is going to continue sort of forever at this point. You know, the next focus for us is how to identify the business processes that deliver the greatest value ultimately to our patients. And use the modern platform that we've just built to improve those processes to deliver things faster, deliver new capabilities. Pfizer is making a huge investment in digital medicines therapies that are delivered through smart devices through wearables using, as I said technology that didn't exist before. That wouldn't be possible without the platform that we've built. So over the past year, we've come a long way but I think that we've effectively set the table for all of the things that are yet to come. >> So, Jon, how do you then, as you've learned a lot about life science or, and certainly Pfizer with what they're up to, how do you then apply, you know, what you know about their world to what you know about the tech world and make it actionable for growth to make it actionable for, for future expansion? >> Yeah, I mean, we start by doing it together, right? I think that's a really important part. Accenture brings a wealth of knowledge, both industry experience and expertise, technology experience and expertise. We work together with our clients like Pfizer with our partners like AWS to bring the best across that power of three to meet clients where they're at to understand where they want to go, and then create a bespoke approach that meets their business needs. And that's effectively what we're doing now, right? I mean, if you think about the phase that we've just went through, I mean, a couple of fast facts here no pun intended, right? 7,800 server instances across 11 operating system versions 7,500 databases across 20 database versions, right? 4,700 applications, 350,000 migration activities managed across an eight month period. >> In eight months. >> Yeah. But that's not the goal, right? The goal is now to take, to Wes' point that platform that's been developed and leverage that to the benefit of the business ultimately to the benefit of the patient. >> You know, why them, we have we've talked a lot about Pfizer, but why Accenture? What, what, what's, 'cause it's got to be a two way street, right? >> We've had a long partnership with Accenture. Accenture supports a huge component of our application environment at Pfizer and has for quite a long time. Look, we didn't make it easy on them. We put them up against a large number of world class SIs. But look, Accenture brought, you know, sort of what I think of as the trifecta here. They brought the technical capabilities and knowledge of the AWS environment. They brought the ability to really understand the business outcomes that we were trying to achieve and a program leadership capability that, you know I think is world class. And Jon talked about myNav, you know, we recognized that doing what we were trying to do in the time that we were doing it required new machinery, new analytics and data capabilities that just didn't exist. Automation didn't exist. Some people experience capabilities that would allow us to interface with application owners and users at a velocity and a pace and a scale that just hasn't been seen before at Pfizer. Accenture brought all three of those things together and I think they did a great job helping us get to where we need to be. >> When you hear Jon rattle through the stats like he just did, right? We talk about all, I mean, not that I'm going to ask you to pat yourself on the back but do you ever, >> He should. >> Does it blow your mind a little bit, honestly that you're talking about that magnitude of activity in that compressed period of time? That's extraordinary. >> It's 75% of our global IT footprint now in the public cloud, which is fantastic. I mean, look, I think the timing was right. I think Pfizer is in a little bit of a unique position coming off of COVID. We are incredibly motivated to keep the pace up, I mean across all lines of business. So, you know what we found is a really willing leadership team, executive leadership team, digital leadership team to endorse a change of this magnitude. >> Well, it's a great success story. It's beyond impressive. So congratulations to both you on that front and certainly you wish you continued success down the road as well. >> Thank you. >> Thank you gentlemen. >> Thank you. >> Good job. >> Pfizer, and boy, you talk about a job well done. Just spectacular. All right, you are watching our coverage here on theCUBE, we're at the AWS re:Invent 22 show. This is Executive Summit sponsored by Accenture and you're watching theCUBE the leader in high tech coverage.

Published Date : Nov 30 2022

SUMMARY :

and it's been great to be Jon, good to see you as well. Pfizer, we make drugs, right? has changed dramatically in the 18 years to where you were and And over the past couple years, and what you need it to how to best you know, And the past three years, on the team that we ask to to different people depending on Traumatic in some and the modern capabilities that we need. and pace that we've done anywhere else. And you know, many of with a lot of people and made sure that we could get the capabilities to the that directly relate to them, I mean, the work that we do of you know, data governance in every stage of the program, and it was. So in, in terms of where you are now, and as fast as you are and You got to stay competitive. that deliver the greatest value across that power of three to and leverage that to the of the AWS environment. of activity in that in the public cloud, which is fantastic. and certainly you wish Pfizer, and boy, you

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
PfizerORGANIZATION

0.99+

JonPERSON

0.99+

JohnPERSON

0.99+

AccentureORGANIZATION

0.99+

Jon HarrisonPERSON

0.99+

4,700 applicationsQUANTITY

0.99+

75%QUANTITY

0.99+

five yearQUANTITY

0.99+

AWSORGANIZATION

0.99+

last yearDATE

0.99+

Wes BarnesPERSON

0.99+

20 database versionsQUANTITY

0.99+

7,500 databasesQUANTITY

0.99+

10 yearsQUANTITY

0.99+

eight monthQUANTITY

0.99+

100%QUANTITY

0.99+

threeQUANTITY

0.99+

Saturday afternoonDATE

0.99+

Las VegasLOCATION

0.99+

18 yearsQUANTITY

0.99+

WesPERSON

0.99+

VenetianLOCATION

0.98+

bothQUANTITY

0.98+

eight monthsQUANTITY

0.98+

pandemicEVENT

0.97+

10 years agoDATE

0.97+

10%QUANTITY

0.97+

10,000 plus discrete activitiesQUANTITY

0.96+

two wayQUANTITY

0.95+

todayDATE

0.95+

oneQUANTITY

0.94+

7,800 server instancesQUANTITY

0.93+

millions of customersQUANTITY

0.93+

couple years agoDATE

0.9+

COVID-19 vaccineOTHER

0.89+

COVID programTITLE

0.88+

past three yearsDATE

0.87+

past yearDATE

0.87+

every single personQUANTITY

0.86+

11 operating system versionsQUANTITY

0.85+

350,000 migration activitiesQUANTITY

0.79+

Digital Hosting SolutionsORGANIZATION

0.77+

re:Invent 22EVENT

0.75+

North AmericanOTHER

0.71+

yearsDATE

0.69+

Wes'PERSON

0.67+

theCUBETITLE

0.66+

PfizerPERSON

0.63+

2021TITLE

0.61+

InventEVENT

0.61+

SiliconANGLE Report: Reporters Notebook with Adrian Cockcroft | AWS re:Invent 2022


 

(soft techno upbeat music) >> Hi there. Welcome back to Las Vegas. This is Dave Villante with Paul Gillon. Reinvent day one and a half. We started last night, Monday, theCUBE after dark. Now we're going wall to wall. Today. Today was of course the big keynote, Adam Selipsky, kind of the baton now handing, you know, last year when he did his keynote, he was very new. He was sort of still getting his feet wet and finding his guru swing. Settling in a little bit more this year, learning a lot more, getting deeper into the tech, but of course, sharing the love with other leaders like Peter DeSantis. Tomorrow's going to be Swamy in the keynote. Adrian Cockcroft is here. Former AWS, former network Netflix CTO, currently an analyst. You got your own firm now. You're out there. Great to see you again. Thanks for coming on theCUBE. >> Yeah, thanks. >> We heard you on at Super Cloud, you gave some really good insights there back in August. So now as an outsider, you come in obviously, you got to be impressed with the size and the ecosystem and the energy. Of course. What were your thoughts on, you know what you've seen so far, today's keynotes, last night Peter DeSantis, what stood out to you? >> Yeah, I think it's great to be back at Reinvent again. We're kind of pretty much back to where we were before the pandemic sort of shut it down. This is a little, it's almost as big as the, the largest one that we had before. And everyone's turned up. It just feels like we're back. So that's really good to see. And it's a slightly different style. I think there were was more sort of video production things happening. I think in this keynote, more storytelling. I'm not sure it really all stitched together very well. Right. Some of the stories like, how does that follow that? So there were a few things there and some of there were spelling mistakes on the slides, you know that ELT instead of ETL and they spelled ZFS wrong and something. So it just seemed like there was, I'm not quite sure just maybe a few things were sort of rushed at the last minute. >> Not really AWS like, was it? It's kind of remind the Patriots Paul, you know Bill Belichick's teams are fumbling all over the place. >> That's right. That's right. >> Part of it may be, I mean the sort of the market. They have a leader in marketing right now but they're going to have a CMO. So that's sort of maybe as lack of a single threaded leader for this thing. Everything's being shared around a bit more. So maybe, I mean, it's all fixable and it's mine. This is minor stuff. I'm just sort of looking at it and going there's a few things that looked like they were not quite as good as they could have been in the way it was put together. Right? >> But I mean, you're taking a, you know a year of not doing Reinvent. Yeah. Being isolated. You know, we've certainly seen it with theCUBE. It's like, okay, it's not like riding a bike. You know, things that, you know you got to kind of relearn the muscle memories. It's more like golf than is bicycle riding. >> Well I've done AWS keynotes myself. And they are pretty much scrambled. It looks nice, but there's a lot of scrambling leading up to when it actually goes. Right? And sometimes you can, you sometimes see a little kind of the edges of that, and sometimes it's much more polished. But you know, overall it's pretty good. I think Peter DeSantis keynote yesterday was a lot of really good meat there. There was some nice presentations, and some great announcements there. And today I was, I thought I was a little disappointed with some of the, I thought they could have been more. I think the way Andy Jesse did it, he crammed more announcements into his keynote, and Adam seems to be taking sort of a bit more of a measured approach. There were a few things he picked up on and then I'm expecting more to be spread throughout the rest of the day. >> This was more poetic. Right? He took the universe as the analogy for data, the ocean for security. Right? The Antarctic was sort of. >> Yeah. It looked pretty, >> yeah. >> But I'm not sure that was like, we're not here really to watch nature videos >> As analysts and journalists, You're like, come on. >> Yeah, >> Give it the meat >> That was kind the thing, yeah, >> It has always been the AWS has always been Reinvent has always been a shock at our approach. 100, 150 announcements. And they're really, that kind of pressure seems to be off them now. Their position at the top of the market seems to be unshakeable. There's no clear competition that's creeping up behind them. So how does that affect the messaging you think that AWS brings to market when it doesn't really have to prove that it's a leader anymore? It can go after maybe more of the niche markets or fix the stuff that's a little broken more fine tuning than grandiose statements. >> I think so AWS for a long time was so far out that they basically said, "We don't think about the competition, we are listen to the customers." And that was always the statement that works as long as you're always in the lead, right? Because you are introducing the new idea to the customer. Nobody else got there first. So that was the case. But in a few areas they aren't leading. Right? You could argue in machine learning, not necessarily leading in sustainability. They're not leading and they don't want to talk about some of these areas and-- >> Database. I mean arguably, >> They're pretty strong there, but the areas when you are behind, it's like they kind of know how to play offense. But when you're playing defense, it's a different set of game. You're playing a different game and it's hard to be good at both. I think and I'm not sure that they're really used to following somebody into a market and making a success of that. So there's something, it's a little harder. Do you see what I mean? >> I get opinion on this. So when I say database, David Foyer was two years ago, predicted AWS is going to have to converge somehow. They have no choice. And they sort of touched on that today, right? Eliminating ETL, that's one thing. But Aurora to Redshift. >> Yeah. >> You know, end to end. I'm not sure it's totally, they're fully end to end >> That's a really good, that is an excellent piece of work, because there's a lot of work that it eliminates. There's are clear pain points, but then you've got sort of the competing thing, is like the MongoDB and it's like, it's just a way with one database keeps it simple. >> Snowflake, >> Or you've got on Snowflake maybe you've got all these 20 different things you're trying to integrate at AWS, but it's kind of like you have a bag of Lego bricks. It's my favorite analogy, right? You want a toy for Christmas, you want a toy formula one racing car since that seems to be the theme, right? >> Okay. Do you want the fully built model that you can play with right now? Or do you want the Lego version that you have to spend three days building. Right? And AWS is the Lego technique thing. You have to spend some time building it, but once you've built it, you can evolve it, and you'll still be playing those are still good bricks years later. Whereas that prebuilt to probably broken gathering dust, right? So there's something about having an vulnerable architecture which is harder to get into, but more durable in the long term. And so AWS tends to play the long game in many ways. And that's one of the elements that they do that and that's good, but it makes it hard to consume for enterprise buyers that are used to getting it with a bow on top. And here's the solution. You know? >> And Paul, that was always Andy Chassy's answer to when we would ask him, you know, all these primitives you're going to make it simpler. You see the primitives give us the advantage to turn on a dime in the marketplace. And that's true. >> Yeah. So you're saying, you know, you take all these things together and you wrap it up, and you put a snowflake on top, and now you've got a simple thing or a Mongo or Mongo atlas or whatever. So you've got these layered platforms now which are making it simpler to consume, but now you're kind of, you know, you're all stuck in that ecosystem, you know, so it's like what layer of abstractions do you want to tie yourself to, right? >> The data bricks coming at it from more of an open source approach. But it's similar. >> We're seeing Amazon direct more into vertical markets. They spotlighted what Goldman Sachs is doing on their platform. They've got a variety of platforms that are supposedly targeted custom built for vertical markets. How do successful do you see that play being? Is this something that the customers you think are looking for, a fully integrated Amazon solution? >> I think so. There's usually if you look at, you know the MongoDB or data stacks, or the other sort of or elastic, you know, they've got the specific solution with the people that really are developing the core technology, there's open source equivalent version. The AWS is running, and it's usually maybe they've got a price advantage or it's, you know there's some data integration in there or it's somehow easier to integrate but it's not stopping those companies from growing. And what it's doing is it's endorsing that platform. So if you look at the collection of databases that have been around over the last few years, now you've got basically Elastic Mongo and Cassandra, you know the data stacks as being endorsed by the cloud vendors. These are winners. They're going to be around for a very long time. You can build yourself on that architecture. But what happened to Couch base and you know, a few of the other ones, you know, they don't really fit. Like how you going to bait? If you are now becoming an also ran, because you didn't get cloned by the cloud vendor. So the customers are going is that a safe place to be, right? >> But isn't it, don't they want to encourage those partners though in the name of building the marketplace ecosystem? >> Yeah. >> This is huge. >> But certainly the platform, yeah, the platform encourages people to do more. And there's always room around the edge. But the mainstream customers like that really like spending the good money, are looking for something that's got a long term life to it. Right? They're looking for a long commitment to that technology and that it's going to be invested in and grow. And the fact that the cloud providers are adopting and particularly AWS is adopting some of these technologies means that is a very long term commitment. You can base, you know, you can bet your future architecture on that for a decade probably. >> So they have to pick winners. >> Yeah. So it's sort of picking winners. And then if you're the open source company that's now got AWS turning up, you have to then leverage it and use that as a way to grow the market. And I think Mongo have done an excellent job of that. I mean, they're top level sponsors of Reinvent, and they're out there messaging that and doing a good job of showing people how to layer on top of AWS and make it a win-win both sides. >> So ever since we've been in the business, you hear the narrative hardware's going to die. It's just, you know, it's commodity and there's some truth to that. But hardware's actually driving good gross margins for the Cisco's of the world. Storage companies have always made good margins. Servers maybe not so much, 'cause Intel sucked all the margin out of it. But let's face it, AWS makes most of its money. We know on compute, it's got 25 plus percent operating margins depending on the seasonality there. What do you think happens long term to the infrastructure layer discussion? Okay, commodity cloud, you know, we talk about super cloud. Do you think that AWS, and the other cloud vendors that infrastructure, IS gets commoditized and they have to go up market or you see that continuing I mean history would say that still good margins in hardware. What are your thoughts on that? >> It's not commoditizing, it's becoming more specific. We've got all these accelerators and custom chips now, and this is something, this almost goes back. I mean, I was with some micro systems 20,30 years ago and we developed our own chips and HP developed their own chips and SGI mips, right? We were like, the architectures were all squabbling of who had the best processor chips and it took years to get chips that worked. Now if you make a chip and it doesn't work immediately, you screwed up somewhere right? It's become the technology of building these immensely complicated powerful chips that has become commoditized. So the cost of building a custom chip, is now getting to the point where Apple and Amazon, your Apple laptop has got full custom chips your phone, your iPhone, whatever and you're getting Google making custom chips and we've got Nvidia now getting into CPUs as well as GPUs. So we're seeing that the ability to build a custom chip, is becoming something that everyone is leveraging. And the cost of doing that is coming down to startups are doing it. So we're going to see many, many more, much more innovation I think, and this is like Intel and AMD are, you know they've got the compatibility legacy, but of the most powerful, most interesting new things I think are going to be custom. And we're seeing that with Graviton three particular in the three E that was announced last night with like 30, 40% whatever it was, more performance for HPC workloads. And that's, you know, the HPC market is going to have to deal with cloud. I mean they are starting to, and I was at Supercomputing a few weeks ago and they are tiptoeing around the edge of cloud, but those supercomputers are water cold. They are monsters. I mean you go around supercomputing, there are plumbing vendors on the booth. >> Of course. Yeah. >> Right? And they're highly concentrated systems, and that's really the only difference, is like, is it water cooler or echo? The rest of the technology stack is pretty much off the shelf stuff with a few tweets software. >> You point about, you know, the chips and what AWS is doing. The Annapurna acquisition. >> Yeah. >> They're on a dramatically different curve now. I think it comes down to, again, David Floyd's premise, really comes down to volume. The arm wafer volumes are 10 x those of X 86, volume always wins. And the economics of semis. >> That kind of got us there. But now there's also a risk five coming along if you, in terms of licensing is becoming one of the bottlenecks. Like if the cost of building a chip is really low, then it comes down to licensing costs and do you want to pay the arm license And the risk five is an open source chip set which some people are starting to use for things. So your dis controller may have a risk five in it, for example, nowadays, those kinds of things. So I think that's kind of the the dynamic that's playing out. There's a lot of innovation in hardware to come in the next few years. There's a thing called CXL compute express link which is going to be really interesting. I think that's probably two years out, before we start seeing it for real. But it lets you put glue together entire rack in a very flexible way. So just, and that's the entire industry coming together around a single standard, the whole industry except for Amazon, in fact just about. >> Well, but maybe I think eventually they'll get there. Don't use system on a chip CXL. >> I have no idea whether I have no knowledge about whether going to do anything CXL. >> Presuming I'm not trying to tap anything confidential. It just makes sense that they would do a system on chip. It makes sense that they would do something like CXL. Why not adopt the standard, if it's going to be as the cost. >> Yeah. And so that was one of the things out of zip computing. The other thing is the low latency networking with the elastic fabric adapter EFA and the extensions to that that were announced last night. They doubled the throughput. So you get twice the capacity on the nitro chip. And then the other thing was this, this is a bit technical, but this scalable datagram protocol that they've got which basically says, if I want to send a message, a packet from one machine to another machine, instead of sending it over one wire, I consider it over 16 wires in parallel. And I will just flood the network with all the packets and they can arrive in any order. This is why it isn't done normally. TCP is in order, the packets come in order they're supposed to, but this is fully flooding them around with its own fast retry and then they get reassembled at the other end. So they're not just using this now for HPC workloads. They've turned it on for TCP for just without any change to your application. If you are trying to move a large piece of data between two machines, and you're just pushing it down a network, a single connection, it takes it from five gigabits per second to 25 gigabits per second. A five x speed up, with a protocol tweak that's run by the Nitro, this is super interesting. >> Probably want to get all that AIML that stuff is going on. >> Well, the AIML stuff is leveraging it underneath, but this is for everybody. Like you're just copying data around, right? And you're limited, "Hey this is going to get there five times faster, pushing a big enough chunk of data around." So this is turning on gradually as the nitro five comes out, and you have to enable it at the instance level. But it's a super interesting announcement from last night. >> So the bottom line bumper sticker on commoditization is what? >> I don't think so. I mean what's the APIs? Your arm compatible, your Intel X 86 compatible or your maybe risk five one day compatible in the cloud. And those are the APIs, right? That's the commodity level. And the software is now, the software ecosystem is super portable across those as we're seeing with Apple moving from Intel to it's really not an issue, right? The software and the tooling is all there to do that. But underneath that, we're going to see an arms race between the top providers as they all try and develop faster chips for doing more specific things. We've got cranium for training, that instance has they announced it last year with 800 gigabits going out of a single instance, 800 gigabits or no, but this year they doubled it. Yeah. So 1.6 terabytes out of a single machine, right? That's insane, right? But what you're doing is you're putting together hundreds or thousands of those to solve the big machine learning training problems. These super, these enormous clusters that they're being formed for doing these massive problems. And there is a market now, for these incredibly large supercomputer clusters built for doing AI. That's all bandwidth limited. >> And you think about the timeframe from design to tape out. >> Yeah. >> Is just getting compressed It's relative. >> It is. >> Six is going the other way >> The tooling is all there. Yeah. >> Fantastic. Adrian, always a pleasure to have you on. Thanks so much. >> Yeah. >> Really appreciate it. >> Yeah, thank you. >> Thank you Paul. >> Cheers. All right. Keep it right there everybody. Don't forget, go to thecube.net, you'll see all these videos. Go to siliconangle.com, We've got features with Adam Selipsky, we got my breaking analysis, we have another feature with MongoDB's, Dev Ittycheria, Ali Ghodsi, as well Frank Sluman tomorrow. So check that out. Keep it right there. You're watching theCUBE, the leader in enterprise and emerging tech, right back. (soft techno upbeat music)

Published Date : Nov 30 2022

SUMMARY :

Great to see you again. and the ecosystem and the energy. Some of the stories like, It's kind of remind the That's right. I mean the sort of the market. the muscle memories. kind of the edges of that, the analogy for data, As analysts and journalists, So how does that affect the messaging always in the lead, right? I mean arguably, and it's hard to be good at both. But Aurora to Redshift. You know, end to end. of the competing thing, but it's kind of like you And AWS is the Lego technique thing. to when we would ask him, you know, and you put a snowflake on top, from more of an open source approach. the customers you think a few of the other ones, you know, and that it's going to and doing a good job of showing people and the other cloud vendors the HPC market is going to Yeah. and that's really the only difference, the chips and what AWS is doing. And the economics of semis. So just, and that's the entire industry Well, but maybe I think I have no idea whether if it's going to be as the cost. and the extensions to that AIML that stuff is going on. and you have to enable And the software is now, And you think about the timeframe Is just getting compressed Yeah. Adrian, always a pleasure to have you on. the leader in enterprise

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Adam SelipskyPERSON

0.99+

David FloydPERSON

0.99+

Peter DeSantisPERSON

0.99+

PaulPERSON

0.99+

Ali GhodsiPERSON

0.99+

Adrian CockcroftPERSON

0.99+

AWSORGANIZATION

0.99+

Frank SlumanPERSON

0.99+

Paul GillonPERSON

0.99+

AmazonORGANIZATION

0.99+

AppleORGANIZATION

0.99+

Andy ChassyPERSON

0.99+

Las VegasLOCATION

0.99+

AdamPERSON

0.99+

Dev IttycheriaPERSON

0.99+

Andy JessePERSON

0.99+

Dave VillantePERSON

0.99+

AugustDATE

0.99+

two machinesQUANTITY

0.99+

Bill BelichickPERSON

0.99+

10QUANTITY

0.99+

CiscoORGANIZATION

0.99+

todayDATE

0.99+

last yearDATE

0.99+

1.6 terabytesQUANTITY

0.99+

AMDORGANIZATION

0.99+

Goldman SachsORGANIZATION

0.99+

hundredsQUANTITY

0.99+

one machineQUANTITY

0.99+

three daysQUANTITY

0.99+

AdrianPERSON

0.99+

800 gigabitsQUANTITY

0.99+

TodayDATE

0.99+

iPhoneCOMMERCIAL_ITEM

0.99+

David FoyerPERSON

0.99+

two yearsQUANTITY

0.99+

GoogleORGANIZATION

0.99+

yesterdayDATE

0.99+

this yearDATE

0.99+

SnowflakeTITLE

0.99+

NvidiaORGANIZATION

0.99+

five timesQUANTITY

0.99+

oneQUANTITY

0.99+

NetflixORGANIZATION

0.99+

thecube.netOTHER

0.99+

IntelORGANIZATION

0.99+

fiveQUANTITY

0.99+

both sidesQUANTITY

0.99+

MongoORGANIZATION

0.99+

ChristmasEVENT

0.99+

last nightDATE

0.99+

HPORGANIZATION

0.98+

25 plus percentQUANTITY

0.98+

thousandsQUANTITY

0.98+

20,30 years agoDATE

0.98+

pandemicEVENT

0.98+

bothQUANTITY

0.98+

two years agoDATE

0.98+

twiceQUANTITY

0.98+

tomorrowDATE

0.98+

X 86COMMERCIAL_ITEM

0.98+

AntarcticLOCATION

0.98+

PatriotsORGANIZATION

0.98+

siliconangle.comOTHER

0.97+

Shinji Kim, Select Star | AWS re:Invent 2022


 

(upbeat music) >> It's theCUBE live in Las Vegas, covering AWS re:Invent 2022. This is the first full day of coverage. We will be here tomorrow and Thursday but we started last night. So hopefully you've caught some of those interviews. Lisa Martin here in Vegas with Paul Gillin. Paul, it's great to be back. We just saw a tweet from a very reliable source saying that there are upwards of 70,000 people here at rei:Invent '22 >> I think there's 70,000 people just in that aisle right there. >> I think so. It's been great so far we've gotten, what are some of the things that you have been excited about today? >> Data, I just see data everywhere, which very much relates to our next guest. Companies realizing the value of data and the strategic value of data, beginning to treat it as an asset rather than just exhaust. I see a lot of focus on app development here and building scalable applications now. Developers have to get over that, have to sort of reorient themselves toward building around the set of cloud native primitives which I think we'll see some amazing applications come out of that. >> Absolutely, we will. We're pleased to welcome back one of our alumni to the program. Shinji Kim joins us, the CEO and founder of Select Star. Welcome back Shinji. It's great to have you. >> Thanks Lisa, great to be back. >> So for the audience who may not know much about Select Star before we start digging into all of the good stuff give us a little overview about what the company does and what differentiates you. >> Sure, so Select Star is an automated data discovery platform. We act like it's Google for data scientists, data analysts and data engineers to help find and understand their data better. Lot of companies today, like what you mentioned, Paul, have 100s and 1000s of database tables now swimming through large volumes of data and variety of data today and it's getting harder and harder for people that wants to utilize data make decisions around data and analyze data to truly have the full context of where this data came from, who do you think that's inside the company or what other analysis might have been done? So Select Star's role in this case is we connect different data warehouses BI tools, wherever the data is actually being used inside the company, bringing out all the usage analytics and the pipeline and the models in one place so anyone can search through what's available and how the data has been created, used and being analyzed within the company. So that's why we call it it's kind of like your Google for data. >> What are some of the biggest challenges to doing that? I mean you've got data squirreled away in lots of corners of the organization, Excel spreadsheets, thumb drives, cloud storage accounts. How granular do you get and what's the difficulty of finding all this data? >> So today we focus primarily on lot of cloud data warehouses and data lakes. So this includes data warehouses like Redshift, Snowflake (indistinct), Databricks, S3 buckets, where a lot of the data from different sources are arriving. Because this is a one area where a lot of analysis are now being done. This is a place where you can join other data sets within the same infrastructural umbrella. And so that is one portion that we always integrate with. The other part that we also integrate a lot with are the BI tools. So whether that's (indistinct) where you are running analysis, building reports, and dashboards. We will pull out how those are, which analysis has been done and which business stakeholders are consuming that data through those tools. So you also mentioned about the differentiation. I would say one of the biggest differentiation that we have in the market today is that we are more in the cloud. So it's very cloud native, fully managed SaaS service and it's really focused on user experience of how easily anyone can really search and understand data through Select Star. In the past, data catalogs as a sector has been primarily focused on inventorizing all your enterprise data which are in many disciplinary forces. So it was more focused on technical aspect of the metadata. At the same time now this enterprise data catalog is important and is needed for even smaller companies because they are dealing with ton of data. Another part that we also see is more of democratization of data. Many different types of users are utilizing data whether they are fully technical or not. So we had basically emphasis around how to make our user interface as intuitive as possible for business users or non-technical users but also bring out as much context as possible from the metadata and the laws that we have access to, to bring out these insights for our customers. >> Got it. What was the impetus or the catalyst to launch the business just a couple of years ago? >> Yeah, so prior to this I had another data startup called Concord Systems. We focused on distributed stream processing framework. I sold the company to Akamai which is now called ... and the product is now called IoT Edge Connect. Through Akamai I started working with a lot of enterprises in automotive and consumer electronics and this is where I saw lot of the issues starting to happen when enterprises are starting to try to use the data. Collection of data, storage of data, processing of data with the help of lot of cloud providers, scaling that is not going to be a challenge as much anymore. At the same time now lot of enterprises, what I realized is a lot of enterprises were sitting on top of ton of data that they may not know how to utilize it or know even how to give the access to because they are not 100% sure what's really inside. And more and more companies, as they are building up their cloud data warehouse infrastructure they're starting to run into the same issue. So this is a part that I felt like was missing gap in the market that I wanted to fulfill and that's why I started the company. >> I'm fascinated with some of the mechanics of doing that. In March of 2020 when lockdowns were happening worldwide you're starting new a company, you have to get funding, you have to hire people, you don't have a team in place presumably. So you have to build that as free to core. How did you do all that? (Shinji laughs) >> Yeah, that was definitely a lot of work just starting from scratch. But I've been brewing this idea, I would say three four months prior. I had a few other ideas. Basically after Akamai I took some time off and then when I decided I wanted to start another company there were a number of ideas that I was toying around with. And so late 2019 I was talking to a lot of different potential customers and users to learn a little bit more about whether my hypothesis around data discovery was true or not. And that kind of led into starting to build prototypes and designs and showing them around to see if there is an interest. So it's only after all those validations and conversations in place that I truly decided that I was going to start another company and it just happened to be at the timing of end of February, early March. So that's kind of how it happened. At the same time, I'm very lucky that I was able to have had number of investors that I kept in touch with and I kept them posted on how this process was going and that's why I think during the pandemic it was definitely not an easy thing to raise our initial seed round but we were able to close it and then move on to really start building the product in 2020. >> Now you were also entering a market that's there's quite a few competitors already in that market. What has been your strategy for getting a foot in the door, getting some name recognition for your company other than being on the queue? >> Yes, this is certainly part of it. So I think there are a few things. One is when I was doing my market research and even today there are a lot of customers out there looking for an easier, faster, time to value solution. >> Yes. >> In the market. Today, existing players and legacy players have a whole suite of platform. However, the implementation time for those platforms take six months or longer and they don't necessarily are built for lot of users to use. They are built for database administrators or more technical people to use so that they end up finding their data governance project not necessarily succeeding or getting as much value out of it as they were hoping for. So this is an area that we really try to fill the gaps in because for us from day one you will be able to see all the usage analysis, how your data models look like, and the analysis right up front. And this is one part that a lot of our customers really like and also some of those customers have moved from the legacy players to Select Star's floor. >> Interesting, so you're actually taking business from some of the legacy guys and girls that may not be able to move as fast and quickly as you can. But I'd love to hear, every company these days has to be a data company, whether it's a grocery store or obviously a bank or a car dealership, there's no choice anymore. As consumers, we have this expectation that we're going to be able to get what we want, self-service. So these companies have to figure out where all the data is, what's the insides, what does it say, how can they act on that quickly? And that's a big challenge to enable organizations to be able to see what it is that they have, where's the value, where's the liability as well. Give me a favorite customer story example that you think really highlights the value of what Select Star is delivering. >> Sure, so one customer that we helped and have been working with closely is Pitney Bowes. It's one of the oldest companies, 100 year old company in logistics and manufacturing. They have ton of IoT data they collect from parcels and all the tracking and all the manufacturing that they run. They have recently, I would say a couple years ago moved to a cloud data warehouse. And this is where their challenge around managing data have really started because they have many different teams accessing the data warehouses but maybe different teams creating different things that might have been created before and it's not clear to the other teams and there is no single source of truth that they could manage. So for them, as they were starting to look into implementing data mesh architecture they adopted Select Star. And they have a, as being a very large and also mature company they have considered a lot of other legacy solutions in the market as well. But they decided to give it a try with select Star mainly because all of the automated version of data modeling and the documentation that we were able to provide upfront. And with all that, with the implementation of Select Star now they claim that they save more than 30 hours a month of every person that they have in the data management team. And we have a case study about that. So this is like one place where we see it save a lot of time for the data team as well as all the consumers that data teams serve. >> I have to ask you this as a successful woman in technology, a field that has not been very inviting to women over the years, what do you think this industry has to do better in terms of bringing along girls and young women, particularly in secondary school to encourage them to pursue careers in science and technology? >> Like what could they do better? >> What could this industry do? What is this industry, these 70,000 people here need to do better? Of which maybe 15% are female. >> Yeah, so actually I do see a lot more women and minority in data analytics field which is always great to see, also like bridging the gap between technology and the business point of view. If anything as a takeaway I feel like just making more opportunities for everyone to participate is always great. I feel like there has been, or you know just like being in the industry, a lot of people tends to congregate with people that they know or more closed groups but having more inclusive open groups that is inviting regardless of the level or gender I think is definitely something that needs to be encouraged more just overall in the industry. >> I agree. I think the inclusivity is so important but it also needs to be intentional. We've done a lot of chatting with women in tech lately and we've been talking about this very topic and that they all talk about the inclusivity, diversity, equity but it needs to be intentional by companies to be able to do that. >> Right, and I think in a way if you were to put it as like women in tech then I feel like that's also making it more explosive. I think it's better when it's focused on the industry problem or like the subject matter, but then intentionally inviting more women and minority to participate so that there's more exchange with more diverse attendees in the AWS. >> That's a great point and I hope to your 0.1 day that we're able to get there, but we don't have to call out women in tech but it is just so much more even playing field. And I hope like you that we're on our way to doing that but it's amazing that Paul brought up that you started the company during the pandemic. Also as a female founder getting funding is incredibly difficult. So kudos to you. >> Thank you. >> For all the successes that you've had. Tell us what's next for Select Star before we get to that last question. >> Yeah, we have a lot of exciting features that have been recently released and also coming up. First and foremost we have an auto documentation feature that we recently released. We have a fairly sophisticated data lineage function that parses through activity log and sequel queries to give you what the data pipeline models look like. This allows you to tell what is the dependency of different tables and dashboards so you can plan what your migration or any changes that might happen in the data warehouse so that nothing breaks whenever these changes happen. We went one step further to that to understand how the data replication actually happens and based on that we are now able to detect which are the duplicated data sets and how each different field might have changed their data values. And if the data actually stays the same then we can also propagate the same documentation as well as tagging. So this is particularly useful if you are doing like a PII tagging, you just mark one thing once and based on the data model we will also have the rest of the PII that it's associated with. So that's one part. The second part is more on the security and data governance front. So we are really seeing policy based access control where you can define who can see what data in the catalog based on their team tags and how you want to define the model. So this allows more enterprises to be able to have different teams to work together. And last one at least we have more integrations that we are releasing. We have an upgraded integration now with Redshift so that there's an easy cloud formation template to get it set up, but we now have not added Databricks, and power BI as well. So there are lots of stuff coming up. >> Man, you have accomplished a lot in two and a half years Shinji, my goodness! Last question for you, describing Select Star in a bumper sticker, what would that bumper sticker say? >> So this is on our website, but yes, automated data catalog in 15 minutes would be what I would call. >> 15 minutes. That's awesome. Thank you so much for joining us back on the program reintroducing our audience to Select Star. And again, congratulations on the successes that you've had. You have to come back because what you're creating is a flywheel and I can't wait to see where it goes. >> Awesome, thanks so much for having me here. >> Oh, our pleasure. Shinji Kim and Paul Gillin, I'm Lisa Martin. You're watching theCUBE, the leader in live enterprise and emerging tech coverage. (upbeat music)

Published Date : Nov 30 2022

SUMMARY :

This is the first full day of coverage. just in that aisle right there. of the things that you have and the strategic value of data, and founder of Select Star. So for the audience who may not know and how the data has been created, used of the organization, Excel in the market today is that or the catalyst to launch the business I sold the company to Akamai the mechanics of doing that. and it just happened to be for getting a foot in the door, time to value solution. and the analysis right up front. and girls that may not and the documentation that we here need to do better? and the business point of view. and that they all talk and minority to participate and I hope to your 0.1 day For all the successes that you've had. and based on that we are now able to So this is on our website, the successes that you've had. much for having me here. the leader in live enterprise

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
LisaPERSON

0.99+

Paul GillinPERSON

0.99+

Lisa MartinPERSON

0.99+

PaulPERSON

0.99+

March of 2020DATE

0.99+

VegasLOCATION

0.99+

2020DATE

0.99+

six monthsQUANTITY

0.99+

Concord SystemsORGANIZATION

0.99+

late 2019DATE

0.99+

100sQUANTITY

0.99+

Las VegasLOCATION

0.99+

15%QUANTITY

0.99+

FirstQUANTITY

0.99+

Select StarORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

70,000 peopleQUANTITY

0.99+

15 minutesQUANTITY

0.99+

Shinji KimPERSON

0.99+

second partQUANTITY

0.99+

AkamaiORGANIZATION

0.99+

end of FebruaryDATE

0.99+

ThursdayDATE

0.99+

100%QUANTITY

0.99+

last nightDATE

0.99+

ExcelTITLE

0.99+

AWSORGANIZATION

0.99+

early MarchDATE

0.99+

two and a half yearsQUANTITY

0.99+

tomorrowDATE

0.99+

one customerQUANTITY

0.99+

firstQUANTITY

0.98+

one partQUANTITY

0.98+

0.1 dayQUANTITY

0.98+

todayDATE

0.98+

oneQUANTITY

0.98+

one placeQUANTITY

0.98+

one areaQUANTITY

0.98+

Select StarORGANIZATION

0.98+

TodayDATE

0.97+

OneQUANTITY

0.96+

Pitney BowesORGANIZATION

0.96+

RedshiftORGANIZATION

0.96+

more than 30 hours a monthQUANTITY

0.96+

ShinjiPERSON

0.95+

pandemicEVENT

0.95+

one portionQUANTITY

0.94+

single sourceQUANTITY

0.93+

one stepQUANTITY

0.93+

DatabricksORGANIZATION

0.93+

rei:Invent '22EVENT

0.89+

couple of years agoDATE

0.88+

SnowflakeORGANIZATION

0.88+

one thingQUANTITY

0.86+

100 year oldQUANTITY

0.82+

couple years agoDATE

0.82+

onceQUANTITY

0.82+

three four months priorDATE

0.78+

ton of dataQUANTITY

0.78+

each different fieldQUANTITY

0.76+

Select StarTITLE

0.75+

1000s of database tablesQUANTITY

0.75+

re:Invent 2022EVENT

0.72+

select StarORGANIZATION

0.71+

ton of IoT dataQUANTITY

0.7+

SelectORGANIZATION

0.7+

day oneQUANTITY

0.68+

RedshiftTITLE

0.67+

StarORGANIZATION

0.64+

Invent 2022EVENT

0.64+

Shireesh Thota, SingleStore & Hemanth Manda, IBM | AWS re:Invent 2022


 

>>Good evening everyone and welcome back to Sparkly Sin City, Las Vegas, Nevada, where we are here with the cube covering AWS Reinvent for the 10th year in a row. John Furrier has been here for all 10. John, we are in our last session of day one. How does it compare? >>I just graduated high school 10 years ago. It's exciting to be, here's been a long time. We've gotten a lot older. My >>Got your brain is complex. You've been a lot in there. So fast. >>Graduated eight in high school. You know how it's No. All good. This is what's going on. This next segment, wrapping up day one, which is like the the kickoff. The Mondays great year. I mean Tuesdays coming tomorrow big days. The announcements are all around the kind of next gen and you're starting to see partnering and integration is a huge part of this next wave cuz API's at the cloud, next gen cloud's gonna be deep engineering integration and you're gonna start to see business relationships and business transformation scale a horizontally, not only across applications but companies. This has been going on for a while, covering it. This next segment is gonna be one of those things that we're gonna look at as something that's gonna happen more and more on >>Yeah, I think so. It's what we've been talking about all day. Without further ado, I would like to welcome our very exciting guest for this final segment, trust from single store. Thank you for being here. And we also have him on from IBM Data and ai. Y'all are partners. Been partners for about a year. I'm gonna go out on a limb only because their legacy and suspect that a few people, a few more people might know what IBM does versus what a single store does. So why don't you just give us a little bit of background so everybody knows what's going on. >>Yeah, so single store is a relational database. It's a foundational relational systems, but the thing that we do the best is what we call us realtime analytics. So we have these systems that are legacy, which which do operations or analytics. And if you wanted to bring them together, like most of the applications want to, it's really a big hassle. You have to build an ETL pipeline, you'd have to duplicate the data. It's really faulty systems all over the place and you won't get the insights really quickly. Single store is trying to solve that problem elegantly by having an architecture that brings both operational and analytics in one place. >>Brilliant. >>You guys had a big funding now expanding men. Sequel, single store databases, 46 billion again, databases. We've been saying this in the queue for 12 years have been great and recently not one database will rule the world. We know that. That's, everyone knows that databases, data code, cloud scale, this is the convergence now of all that coming together where data, this reinvent is the theme. Everyone will be talking about end to end data, new kinds of specialized services, faster performance, new kinds of application development. This is the big part of why you guys are working together. Explain the relationship, how you guys are partnering and engineering together. >>Yeah, absolutely. I think so ibm, right? I think we are mainly into hybrid cloud and ai and one of the things we are looking at is expanding our ecosystem, right? Because we have gaps and as opposed to building everything organically, we want to partner with the likes of single store, which have unique capabilities that complement what we have. Because at the end of the day, customers are looking for an end to end solution that's also business problems. And they are very good at real time data analytics and hit staff, right? Because we have transactional databases, analytical databases, data lakes, but head staff is a gap that we currently have. And by partnering with them we can essentially address the needs of our customers and also what we plan to do is try to integrate our products and solutions with that so that when we can deliver a solution to our customers, >>This is why I was saying earlier, I think this is a a tell sign of what's coming from a lot of use cases where people are partnering right now you got the clouds, a bunch of building blocks. If you put it together yourself, you can build a durable system, very stable if you want out of the box solution, you can get that pre-built, but you really can't optimize. It breaks, you gotta replace it. High level engineering systems together is a little bit different, not just buying something out of the box. You guys are working together. This is kind of an end to end dynamic that we're gonna hear a lot more about at reinvent from the CEO ofs. But you guys are doing it across companies, not just with aws. Can you guys share this new engineering business model use case? Do you agree with what I'm saying? Do you think that's No, exactly. Do you think John's crazy, crazy? I mean I all discourse, you got out of the box, engineer it yourself, but then now you're, when people do joint engineering project, right? They're different. Yeah, >>Yeah. No, I mean, you know, I think our partnership is a, is a testament to what you just said, right? When you think about how to achieve realtime insights, the data comes into the system and, and the customers and new applications want insights as soon as the data comes into the system. So what we have done is basically build an architecture that enables that we have our own storage and query engine indexing, et cetera. And so we've innovated in our indexing in our database engine, but we wanna go further than that. We wanna be able to exploit the innovation that's happening at ibm. A very good example is, for instance, we have a native connector with Cognos, their BI dashboards right? To reason data very natively. So we build a hyper efficient system that moves the data very efficiently. A very other good example is embedded ai. >>So IBM of course has built AI chip and they have basically advanced quite a bit into the embedded ai, custom ai. So what we have done is, is as a true marriage between the engineering teams here, we make sure that the data in single store can natively exploit that kind of goodness. So we have taken their libraries. So if you have have data in single store, like let's imagine if you have Twitter data, if you wanna do sentiment analysis, you don't have to move the data out model, drain the model outside, et cetera. We just have the pre-built embedded AI libraries already. So it's a, it's a pure engineering manage there that kind of opens up a lot more insights than just simple analytics and >>Cost by the way too. Moving data around >>Another big theme. Yeah. >>And latency and speed is everything about single store and you know, it couldn't have happened without this kind of a partnership. >>So you've been at IBM for almost two decades, don't look it, but at nearly 17 years in how has, and maybe it hasn't, so feel free to educate us. How has, how has IBM's approach to AI and ML evolved as well as looking to involve partnerships in the ecosystem as a, as a collaborative raise the water level together force? >>Yeah, absolutely. So I think when we initially started ai, right? I think we are, if you recollect Watson was the forefront of ai. We started the whole journey. I think our focus was more on end solutions, both horizontal and vertical. Watson Health, which is more vertically focused. We were also looking at Watson Assistant and Watson Discovery, which were more horizontally focused. I think it it, that whole strategy of the world period of time. Now we are trying to be more open. For example, this whole embedable AI that CICE was talking about. Yeah, it's essentially making the guts of our AI libraries, making them available for partners and ISVs to build their own applications and solutions. We've been using it historically within our own products the past few years, but now we are making it available. So that, how >>Big of a shift is that? Do, do you think we're seeing a more open and collaborative ecosystem in the space in general? >>Absolutely. Because I mean if you think about it, in my opinion, everybody is moving towards AI and that's the future. And you have two option. Either you build it on your own, which is gonna require significant amount of time, effort, investment, research, or you partner with the likes of ibm, which has been doing it for a while, right? And it has the ability to scale to the requirements of all the enterprises and partners. So you have that option and some companies are picking to do it on their own, but I believe that there's a huge amount of opportunity where people are looking to partner and source what's already available as opposed to investing from the scratch >>Classic buy versus build analysis for them to figure out, yeah, to get into the game >>And, and, and why reinvent the wheel when we're all trying to do things at, at not just scale but orders of magnitude faster and and more efficiently than we were before. It, it makes sense to share, but it's, it is, it does feel like a bit of a shift almost paradigm shift in, in the culture of competition versus how we're gonna creatively solve these problems. There's room for a lot of players here, I think. And yeah, it's, I don't >>Know, it's really, I wanted to ask if you don't mind me jumping in on that. So, okay, I get that people buy a bill I'm gonna use existing or build my own. The decision point on that is, to your point about the path of getting the path of AI is do I have the core competency skills, gap's a big issue. So, okay, the cube, if you had ai, we'd take it cuz we don't have any AI engineers around yet to build out on all the linguistic data we have. So we might use your ai but I might say this to then and we want to have a core competency. How do companies get that core competency going while using and partnering with, with ai? What you guys, what do you guys see as a way for them to get going? Because I think some people probably want to have core competency of >>Ai. Yeah, so I think, again, I think I, I wanna distinguish between a solution which requires core competency. You need expertise on the use case and you need expertise on your industry vertical and your customers versus the foundational components of ai, which are like, which are agnostic to the core competency, right? Because you take the foundational piece and then you further train it and define it for your specific use case. So we are not saying that we are experts in all the industry verticals. What we are good at is like foundational components, which is what we wanna provide. Got it. >>Yeah, that's the hard deep yes. Heavy lift. >>Yeah. And I can, I can give a color to that question from our perspective, right? When we think about what is our core competency, it's about databases, right? But there's a, some biotic relationship between data and ai, you know, they sort of like really move each other, right? You >>Need, they kind of can't have one without the other. You can, >>Right? And so the, the question is how do we make sure that we expand that, that that relationship where our customers can operationalize their AI applications closer to the data, not move the data somewhere else and do the modeling and then training somewhere else and dealing with multiple systems, et cetera. And this is where this kind of a cross engineering relationship helps. >>Awesome. Awesome. Great. And then I think companies are gonna want to have that baseline foundation and then start hiring in learning. It's like driving the car. You get the keys when you're ready to go. >>Yeah, >>Yeah. Think I'll give you a simple example, right? >>I want that turnkey lifestyle. We all do. Yeah, >>Yeah. Let me, let me just give you a quick analogy, right? For example, you can, you can basically make the engines and the car on your own or you can source the engine and you can make the car. So it's, it's basically an option that you can decide. The same thing with airplanes as well, right? Whether you wanna make the whole thing or whether you wanna source from someone who is already good at doing that piece, right? So that's, >>Or even create a new alloy for that matter. I mean you can take it all the way down in that analogy, >>Right? Is there a structural change and how companies are laying out their architecture in this modern era as we start to see this next let gen cloud emerge, teams, security teams becoming much more focused data teams. Its building into the DevOps into the developer pipeline, seeing that trend. What do you guys see in the modern data stack kind of evolution? Is there a data solutions architect coming? Do they exist yet? Is that what we're gonna see? Is it data as code automation? How do you guys see this landscape of the evolving persona? >>I mean if you look at the modern data stack as it is defined today, it is too detailed, it's too OSes and there are way too many layers, right? There are at least five different layers. You gotta have like a storage you replicate to do real time insights and then there's a query layer, visualization and then ai, right? So you have too many ETL pipelines in between, too many services, too many choke points, too many failures, >>Right? Etl, that's the dirty three letter word. >>Say no to ETL >>Adam Celeste, that's his quote, not mine. We hear that. >>Yeah. I mean there are different names to it. They don't call it etl, we call it replication, whatnot. But the point is hassle >>Data is getting more hassle. More >>Hassle. Yeah. The data is ultimately getting replicated in the modern data stack, right? And that's kind of one of our thesis at single store, which is that you'd have to converge not hyper specialize and conversation and convergence is possible in certain areas, right? When you think about operational analytics as two different aspects of the data pipeline, it is possible to bring them together. And we have done it, we have a lot of proof points to it, our customer stories speak to it and that is one area of convergence. We need to see more of it. The relationship with IBM is sort of another step of convergence wherein the, the final phases, the operation analytics is coming together and can we take analytics visualization with reports and dashboards and AI together. This is where Cognos and embedded AI comes into together, right? So we believe in single store, which is really conversions >>One single path. >>A shocking, a shocking tie >>Back there. So, so obviously, you know one of the things we love to joke about in the cube cuz we like to goof on the old enterprise is they solve complexity by adding more complexity. That's old. Old thinking. The new thinking is put it under the covers, abstract the way the complexities and make it easier. That's right. So how do you guys see that? Because this end to end story is not getting less complicated. It's actually, I believe increasing and complication complexity. However there's opportunities doing >>It >>More faster to put it under the covers or put it under the hood. What do you guys think about the how, how this new complexity gets managed or in this new data world we're gonna be coming in? >>Yeah, so I think you're absolutely right. It's the world is becoming more complex, technology is becoming more complex and I think there is a real need and it's not just from coming from us, it's also coming from the customers to simplify things. So our approach around AI is exactly that because we are essentially providing libraries, just like you have Python libraries, there are libraries now you have AI libraries that you can go infuse and embed deeply within applications and solutions. So it becomes integrated and simplistic for the customer point of view. From a user point of view, it's, it's very simple to consume, right? So that's what we are doing and I think single store is doing that with data, simplifying data and we are trying to do that with the rest of the portfolio, specifically ai. >>It's no wonder there's a lot of synergy between the two companies. John, do you think they're ready for the Instagram >>Challenge? Yes, they're ready. Uhoh >>Think they're ready. So we're doing a bit of a challenge. A little 32nd off the cuff. What's the most important takeaway? This could be your, think of it as your thought leadership sound bite from AWS >>2023 on Instagram reel. I'm scrolling. That's the Instagram, it's >>Your moment to stand out. Yeah, exactly. Stress. You look like you're ready to rock. Let's go for it. You've got that smile, I'm gonna let you go. Oh >>Goodness. You know, there is, there's this quote from astrophysics, space moves matter, a matter tells space how to curve. They have that kind of a relationship. I see the same between AI and data, right? They need to move together. And so AI is possible only with right data and, and data is meaningless without good insights through ai. They really have that kind of relationship and you would see a lot more of that happening in the future. The future of data and AI are combined and that's gonna happen. Accelerate a lot faster. >>Sures, well done. Wow. Thank you. I am very impressed. It's tough hacks to follow. You ready for it though? Let's go. Absolutely. >>Yeah. So just, just to add what is said, right, I think there's a quote from Rob Thomas, one of our leaders at ibm. There's no AI without ia. Essentially there's no AI without information architecture, which essentially data. But I wanna add one more thing. There's a lot of buzz around ai. I mean we are talking about simplicity here. AI in my opinion is three things and three things only. Either you use AI to predict future for forecasting, use AI to automate things. It could be simple, mundane task, it would be complex tasks depending on how exactly you want to use it. And third is to optimize. So predict, automate, optimize. Anything else is buzz. >>Okay. >>Brilliantly said. Honestly, I think you both probably hit the 32nd time mark that we gave you there. And the enthusiasm loved your hunger on that. You were born ready for that kind of pitch. I think they both nailed it for the, >>They nailed it. Nailed it. Well done. >>I I think that about sums it up for us. One last closing note and opportunity for you. You have a V 8.0 product coming out soon, December 13th if I'm not mistaken. You wanna give us a quick 15 second preview of that? >>Super excited about this. This is one of the, one of our major releases. So we are evolving the system on multiple dimensions on enterprise and governance and programmability. So there are certain features that some of our customers are aware of. We have made huge performance gains in our JSON access. We made it easy for people to consume, blossom on OnPrem and hybrid architectures. There are multiple other things that we're gonna put out on, on our site. So it's coming out on December 13th. It's, it's a major next phase of our >>System. And real quick, wasm is the web assembly moment. Correct. And the new >>About, we have pioneers in that we, we be wasm inside the engine. So you could run complex modules that are written in, could be C, could be rushed, could be Python. Instead of writing the the sequel and SQL as a store procedure, you could now run those modules inside. I >>Wanted to get that out there because at coupon we covered that >>Savannah Bay hot topic. Like, >>Like a blanket. We covered it like a blanket. >>Wow. >>On that glowing note, Dre, thank you so much for being here with us on the show. We hope to have both single store and IBM back on plenty more times in the future. Thank all of you for tuning in to our coverage here from Las Vegas in Nevada at AWS Reinvent 2022 with John Furrier. My name is Savannah Peterson. You're watching the Cube, the leader in high tech coverage. We'll see you tomorrow.

Published Date : Nov 29 2022

SUMMARY :

John, we are in our last session of day one. It's exciting to be, here's been a long time. So fast. The announcements are all around the kind of next gen So why don't you just give us a little bit of background so everybody knows what's going on. It's really faulty systems all over the place and you won't get the This is the big part of why you guys are working together. and ai and one of the things we are looking at is expanding our ecosystem, I mean I all discourse, you got out of the box, When you think about how to achieve realtime insights, the data comes into the system and, So if you have have data in single store, like let's imagine if you have Twitter data, if you wanna do sentiment analysis, Cost by the way too. Yeah. And latency and speed is everything about single store and you know, it couldn't have happened without this kind and maybe it hasn't, so feel free to educate us. I think we are, So you have that option and some in, in the culture of competition versus how we're gonna creatively solve these problems. So, okay, the cube, if you had ai, we'd take it cuz we don't have any AI engineers around yet You need expertise on the use case and you need expertise on your industry vertical and Yeah, that's the hard deep yes. you know, they sort of like really move each other, right? You can, And so the, the question is how do we make sure that we expand that, You get the keys when you're ready to I want that turnkey lifestyle. So it's, it's basically an option that you can decide. I mean you can take it all the way down in that analogy, What do you guys see in the modern data stack kind of evolution? I mean if you look at the modern data stack as it is defined today, it is too detailed, Etl, that's the dirty three letter word. We hear that. They don't call it etl, we call it replication, Data is getting more hassle. When you think about operational analytics So how do you guys see that? What do you guys think about the how, is exactly that because we are essentially providing libraries, just like you have Python libraries, John, do you think they're ready for the Instagram Yes, they're ready. A little 32nd off the cuff. That's the Instagram, You've got that smile, I'm gonna let you go. and you would see a lot more of that happening in the future. I am very impressed. I mean we are talking about simplicity Honestly, I think you both probably hit the 32nd time mark that we gave you there. They nailed it. I I think that about sums it up for us. So we are evolving And the new So you could run complex modules that are written in, could be C, We covered it like a blanket. On that glowing note, Dre, thank you so much for being here with us on the show.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JohnPERSON

0.99+

IBMORGANIZATION

0.99+

Savannah PetersonPERSON

0.99+

December 13thDATE

0.99+

Shireesh ThotaPERSON

0.99+

Las VegasLOCATION

0.99+

Adam CelestePERSON

0.99+

Rob ThomasPERSON

0.99+

46 billionQUANTITY

0.99+

12 yearsQUANTITY

0.99+

John FurrierPERSON

0.99+

three thingsQUANTITY

0.99+

15 secondQUANTITY

0.99+

TwitterORGANIZATION

0.99+

PythonTITLE

0.99+

10th yearQUANTITY

0.99+

two companiesQUANTITY

0.99+

thirdQUANTITY

0.99+

32nd timeQUANTITY

0.99+

bothQUANTITY

0.99+

tomorrowDATE

0.99+

32ndQUANTITY

0.99+

single storeQUANTITY

0.99+

TuesdaysDATE

0.99+

AWSORGANIZATION

0.99+

oneQUANTITY

0.98+

10 years agoDATE

0.98+

SingleStoreORGANIZATION

0.98+

Single storeQUANTITY

0.98+

Hemanth MandaPERSON

0.98+

DrePERSON

0.97+

eightQUANTITY

0.96+

two optionQUANTITY

0.96+

day oneQUANTITY

0.96+

one more thingQUANTITY

0.96+

one databaseQUANTITY

0.95+

two different aspectsQUANTITY

0.95+

MondaysDATE

0.95+

InstagramORGANIZATION

0.95+

IBM DataORGANIZATION

0.94+

10QUANTITY

0.94+

about a yearQUANTITY

0.94+

CICEORGANIZATION

0.93+

three letterQUANTITY

0.93+

todayDATE

0.93+

one placeQUANTITY

0.93+

WatsonTITLE

0.93+

One lastQUANTITY

0.92+

CognosORGANIZATION

0.91+

Watson AssistantTITLE

0.91+

nearly 17 yearsQUANTITY

0.9+

Watson HealthTITLE

0.89+

Las Vegas, NevadaLOCATION

0.89+

awsORGANIZATION

0.86+

one areaQUANTITY

0.86+

SQLTITLE

0.86+

One single pathQUANTITY

0.85+

two decadesQUANTITY

0.8+

five different layersQUANTITY

0.8+

Invent 2022EVENT

0.77+

JSONTITLE

0.77+

Justin Borgman, Starburst & Ashwin Patil, Deloitte | AWS re:Invent 2022


 

(electronic music) (graphics whoosh) (graphics tinkle) >> Welcome to Las Vegas! It's theCUBE live at AWS re:Invent '22. Lisa Martin here with Dave Vellante. Dave, it is not only great to be back, but this re:Invent seems to be bigger than last year for sure. >> Oh, definitely. I'd say it's double last year. I'd say it's comparable to 2019. Maybe even a little bigger, I've heard it's the largest re:Invent ever. And we're going to talk data, one of our favorite topics. >> We're going to talk data products. We have some great guests. One of them is an alumni who's back with us. Justin Borgman, the CEO of Starburst, and Ashwin Patil also joins us, Principal AI and Data Engineering at Deloitte. Guys, welcome to the program. >> Thank you. >> Together: Thank you. >> Justin, define data products. Give us the scoop, what's goin' on with Starburst. But define data products and the value in it for organizations of productizing data. >> Mm-hmm. So, data products are curated data sets that are able to span across multiple data sets. And I think that's what's makes it particularly unique, is you can span across multiple data sources to create federated data products that allow you to really bring together the business value that you're seeking. And I think ultimately, what's driving the interest in data products is a desire to ultimately facilitate self-service consumption within the enterprise. I think that's the holy grail that we've all been building towards. And data products represents a framework for sort of how you would do that. >> So, monetization is not necessarily a criterion? >> Not necessarily. (Dave's voice drowns) >> But it could be. >> It could be. It can be internal data products or external data products. And in either case, it's really intended to facilitate easier discovery and consumption of data. >> Ashwin, bringing you into the conversation, talk about some of the revenue drivers that data products can help organizations to unlock. >> Sure. Like Justin said, there are internal and external revenue drivers. So internally, a lot of clients are focused around, hey, how do I make the most out of my modernization platform? So, a lot of them are thinking about what AI, what analytics, what can they run to drive consumption? And when you think about consumption, consumption typically requires data from across the enterprise, right? And data from the enterprise is sometimes fragmented in pieces, in places. So, we've gone from being data in too many places to now, data products, helping bring all of that together, and really aid, drive business decisions faster with more data and more accuracy, right? Externally, a lot of that has got to do with how the ecosystems are evolving for data products that use not only company data, but also the ecosystem data that includes customers, that include suppliers and vendors. >> I mean, conceptually, data products, you could say have been around a long time. When I think of financial services, I think that's always been a data product in a sense. But suddenly, there's a lot more conversation about it. There's data mesh, there's data fabric, we could talk about that too, but why do you think now it's coming to the fore again? >> Yeah, I mean, I think it's because historically, there's always been this disconnect between the people that understand data infrastructure, and the people who know the right questions to ask of the data. Generally, these have been two very distinct groups. And so, the interest in data mesh as you mentioned, and data products as a foundational element of it, is really centered around how do we bring these groups together? How do we get the people who know the data the best to participate in the process of creating data to be consumed? Ultimately, again, trying to facilitate greater self-service consumption. And I think that's the real beauty behind it. And I think increasingly, in today's world, people are realizing the data will always be decentralized to some degree. That notion of bringing everything together into one single database has never really been successfully achieved, and is probably even further from the truth at this point in time, given you've got data on-prem and multiple clouds, and multiple different systems. And so, data products and data mesh represents, again, a framework for you to sort of think about data that lives everywhere. >> We did a session this summer with (chuckles) Justin and I, and some others on the data lies. And that was one of the good ol' lies, right? There's a single source of truth. >> Justin: Right. >> And all that is, we've probably never been further from the single source of truth. But actually, you're suggesting that there's maybe multiple truths that the same data can support. Is that a right way to think about it? >> Yeah, exactly. And I think ultimately, you want a single point of access that gives you, at your fingertips, everything that your organization knows about its business today. And that's really what data products aims to do, is sort of curate that for you, and provide high quality data sets that you can trust, that you can now self-service to answer your business question. >> One of the things that, oh, go ahead. >> No, no, I was just going to say, I mean, if you pivot it from the way the usage of data has changed, right? Traditionally, IT has been in the business of providing data to the business users. Today, with more self-service being driven, we want business users to be the drivers of consumption, right? So if you take that backwards one step, it's basically saying, what data do I need to support my business needs, such that IT doesn't always have to get involved in providing that data, or providing the reports on top of that data? So, the data products concept, I think supports that thinking of business-led technology-enabled, or IT-enabled really well. >> Business led. One of the things that Adam Zelinsky talked with John Furrier about just a week or so ago in their pre re:Invent interview, was talking about the role of the data analyst going away. That everybody in an organization, regardless of function, will be able to eventually be a data analyst, and need to evaluate and analyze data for their roles. Talk about data products as a facilitator of that democratization. >> Yeah. We are seeing more and more the concept of citizen data scientists. We are seeing more and more citizens AI. What we are seeing is a general trend, as we move towards self-service, there is going to be a need for business users to be able to access data when they want, how they want, and merge data across the enterprise in ways that they haven't done before, right? Technology today, through products like data products, right, provides you the access to do that. And that's why we are going to see this movement of people of seeing people become more and more self-service oriented, where you're going to democratize the use of AI and analytics into the business users. >> Do you think, when you talk to a data analyst, by the way, about that, he or she will be like, yeah, mm, maybe, good luck with that. So, do ya think maybe there's a sort of an interim step? Because we've had these highly, ZeMac lays this out very well. We've had these highly-centralized, highly-specialized teams. The premise being, oh, that's less expensive. Perhaps data analysts, like functions, get put into the line of business. Do you see that as a bridge or a stepping stone? Because it feels like it's quite a distance between what a data analyst does today, and this nirvana that we talk about. What are your thoughts on that? >> Yeah, I mean, I think there's possibly a new role around a data product manager. Much the way you have product managers in the products you actually build to sell, you might need data product managers to help facilitate and curate the high quality data products that others can consume. And I think that becomes an interesting and important, a skill set. Much the way that data scientist was created as a occupation, if you will, maybe 10 years ago, when previously, those were statisticians, or other names. >> Right. A big risk that many clients are seeing around data products is, how do you drive governance? And to that, to the point that Justin's making, we are going to see that role evolve where governance in the world, where data products are getting democratized is going to become increasingly important in terms of how are data products being generated, how is the propensity of data products towards a more governed environment being managed? And that's going to continue to play an important role as data products evolve. >> Okay, so how do you guys fit, because you take ZeMac's four principles, domain ownership, data as product. And that creates two problems. Governance. (chuckles) Right? How do you automate, and self-service, infrastructure and automated governance. >> Yep. >> Tell us what role Starburst plays in solving all of those, but the latter two in particular. >> Yeah. Well, we're working on all four of those dimensions to some degree, but I think ultimately, what we're focused today is the governance piece, providing fine-grained access controls, which is so important, if you're going to have a single point of access, you better have a way of controlling who has access to what. But secondly, data products allows you to really abstract away or decouple where the data is stored from the business meaning of the data. And I think that's what's so key here is, if we're going to ultimately democratize data as we've talked about, we need to change the conversation from a very storage-centric world, like, oh, that table lives in this system or that system, or that system. And make it much more about the data, and the value that it represents. And I think that's what data products aims to do. >> What about data fabric? I have to say, I'm confused by data fabric. I read this, I feel like Gartner just threw it in there to muck it up. And say, no, no, we get to make up the terms, but I've read data mesh versus data fabric, is data fabric just more sort of the physical infrastructure? And data mesh is more of an organizational construct, or how do you see it? >> Yeah, I'm happy to take that or. So, I mean, to me, it's a little bit of potato potato. I think there are some subtle differences. Data fabric is a little bit more about data movement. Whereas, I think data mesh is a little bit more about accessing the data where it lies. But they're both trying to solve the similar problem, which is that we have data in a wide variety of different data sets. And for us to actually analyze it, we need to have a single view. >> Because Gartner hype cycle says data mesh is DOA- >> Justin: I know. >> Which I think is complete BS, I think is real. You talk to customers that are doing it, they're doing it on AWS, they're trying to extend it across clouds, I mean, it's a real trend. I mean, anyway, that's how I see it. >> Yeah. I feel the word data fabric many a times gets misused. Because when you think about the digitization movement that happened, started almost a decade ago, many companies tried to digitize or create digital twins of their systems into the data world, right? So, everything has an underlying data fabric that replicates what's happening transactionally, or otherwise in the real world. What data mesh does is creates structure that works complimentary to the data fabric, that then lends itself to data products, right? So to me, data products becomes a medium, which drives the connection between data mesh and data fabric into the real world for usage and consumption. >> You should write for Gartner. (all laugh) That's the best explanation I've heard. That made sense! >> That really did. That was excellent. So, when we think about any company these days has to be a data company, whether it's your grocery store, a gas station, a car dealer, what can companies do to start productizing their data, so that they can actually unlock new revenue streams, new routes to market? What are some steps and recommendations that you have? Justin, we'll start with you. >> Sure. I would say the first thing is find data that is ultimately valuable to the consumers within your business, and create a product of it. And the way you do that at Starburst is allow you to essentially create a view of your data that can span multiple data sources. So again, we're decoupling where the data lives. That might be a table that lives in a traditional data warehouse, a table that lives in an operational system like Mongo, a table that lives in a data lake. And you can actually join those together, and represent it as a view, and now make it easily consumable. And so, the end user doesn't need to know, did that live in a data warehouse, an operational database, or a data lake? I'm just accessing that. And I think that's a great, easy way to start in your journey. Because I think if you absorb all the elements of data mesh at once, it can feel overwhelming. And I think that's a great way to start. >> Irrespective of physical location. >> Yes. >> Right? >> Precisely. Yep, multicloud, hybrid cloud, you name it. >> And when you think about the broader landscape, right? For the traditionally, companies that only looked at internal data as a way of driving business decisions. More and more, as things evolve into industry, clouds, or ecosystem data, and companies start going beyond their four walls in terms of the data that they manage or the data that they use to make decisions, I think data products are going to play more and more an important part in that construct where you don't govern all the data that our entities within that ecosystem will govern parts of their data, but that data lives together in the form of data products that are governed somewhat centrally. I mean, kind of like a blockchain system, but not really. >> Justin, for our folks here, as we kind of wrap the segment here, what's the bumper sticker for Starburst, and how you're helping organizations to really be able to build data products that add value to their organization? >> I would say analytics anywhere. Our core ethos is, we want to give you the ability to access data wherever it lives, and understand your business holistically. And our query engine allows you to do that from a query perspective, and data products allows you to bring that up a level and make it consumable. >> Make it consumable. Ashwin, last question for you, here we are, day one of re:Invent, loads of people behind us. Tomorrow all the great keynotes kick up. What are you hoping to take away from re:Invent '22? >> Well, I'm hoping to understand how all of these different entities that are represented here connect with each other, right? And to me, Starburst is an important player in terms of how do you drive connectivity. And to me, as we help plans from a Deloitte perspective, drive that business value, connectivity across all of the technology players is extremely important part. So, integration across those technology players is what I'm trying to get from re:Invent here. >> And so, you guys do, you're dot connectors. (Ashwin chuckles) >> Exactly, excellent. Guys, thank you so much for joining David and me on the program tonight. We appreciate your insights, your time, and probably the best explanation of data fabric versus data mesh. (Justin chuckles) And data products that we've maybe ever had on the show! We appreciate your time, thank you. >> Together: Thank you- >> Thanks, guys. >> All right. For our guests and Dave Vellante, I'm Lisa Martin, you're watching theCUBE, the leader in enterprise and emerging tech coverage. (electronic music)

Published Date : Nov 29 2022

SUMMARY :

Dave, it is not only great to be back, I've heard it's the Justin Borgman, the CEO of Starburst, and the value in it for that are able to span really intended to facilitate into the conversation, And data from the enterprise coming to the fore again? And so, the interest in data mesh and some others on the data lies. And all that is, we've And I think ultimately, you want data do I need to support One of the things that Adam Zelinsky and merge data across the enterprise into the line of business. in the products you And that's going to continue And that creates two problems. all of those, but the data products aims to do. And data mesh is more of an about accessing the data where it lies. You talk to customers that are doing it, and data fabric into the real world That's the best explanation I've heard. recommendations that you have? And the way you do that cloud, you name it. in terms of the data that they manage the ability to access Tomorrow all the great keynotes kick up. And to me, as we help plans And so, you guys do, And data products that we've the leader in enterprise

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Dave VellantePERSON

0.99+

JustinPERSON

0.99+

Lisa MartinPERSON

0.99+

Dave VellantePERSON

0.99+

Justin BorgmanPERSON

0.99+

DavePERSON

0.99+

Adam ZelinskyPERSON

0.99+

Las VegasLOCATION

0.99+

StarburstORGANIZATION

0.99+

GartnerORGANIZATION

0.99+

2019DATE

0.99+

John FurrierPERSON

0.99+

DeloitteORGANIZATION

0.99+

AWSORGANIZATION

0.99+

TomorrowDATE

0.99+

Ashwin PatilPERSON

0.99+

twoQUANTITY

0.99+

last yearDATE

0.99+

AshwinPERSON

0.99+

TodayDATE

0.98+

MongoORGANIZATION

0.98+

OneQUANTITY

0.98+

two very distinct groupsQUANTITY

0.98+

two problemsQUANTITY

0.98+

tonightDATE

0.97+

single sourceQUANTITY

0.97+

bothQUANTITY

0.97+

10 years agoDATE

0.96+

todayDATE

0.96+

single pointQUANTITY

0.96+

fourQUANTITY

0.95+

this summerDATE

0.94+

a decade agoDATE

0.93+

one stepQUANTITY

0.92+

one single databaseQUANTITY

0.91+

secondlyQUANTITY

0.89+

StarburstTITLE

0.87+

ZeMacORGANIZATION

0.86+

oneQUANTITY

0.84+

doubleQUANTITY

0.84+

single viewQUANTITY

0.83+

reEVENT

0.83+

first thingQUANTITY

0.81+

re:InventEVENT

0.8+

re:Invent '22EVENT

0.78+

a week or so agoDATE

0.72+

four principlesQUANTITY

0.68+

Invent 2022EVENT

0.67+

InventEVENT

0.57+

CEOPERSON

0.56+

onceQUANTITY

0.53+

The Truth About MySQL HeatWave


 

>>When Oracle acquired my SQL via the Sun acquisition, nobody really thought the company would put much effort into the platform preferring to focus all the wood behind its leading Oracle database, Arrow pun intended. But two years ago, Oracle surprised many folks by announcing my SQL Heatwave a new database as a service with a massively parallel hybrid Columbia in Mary Mary architecture that brings together transactional and analytic data in a single platform. Welcome to our latest database, power panel on the cube. My name is Dave Ante, and today we're gonna discuss Oracle's MySQL Heat Wave with a who's who of cloud database industry analysts. Holgar Mueller is with Constellation Research. Mark Stammer is the Dragon Slayer and Wikibon contributor. And Ron Westfall is with Fu Chim Research. Gentlemen, welcome back to the Cube. Always a pleasure to have you on. Thanks for having us. Great to be here. >>So we've had a number of of deep dive interviews on the Cube with Nip and Aggarwal. You guys know him? He's a senior vice president of MySQL, Heatwave Development at Oracle. I think you just saw him at Oracle Cloud World and he's come on to describe this is gonna, I'll call it a shock and awe feature additions to to heatwave. You know, the company's clearly putting r and d into the platform and I think at at cloud world we saw like the fifth major release since 2020 when they first announced MySQL heat wave. So just listing a few, they, they got, they taken, brought in analytics machine learning, they got autopilot for machine learning, which is automation onto the basic o l TP functionality of the database. And it's been interesting to watch Oracle's converge database strategy. We've contrasted that amongst ourselves. Love to get your thoughts on Amazon's get the right tool for the right job approach. >>Are they gonna have to change that? You know, Amazon's got the specialized databases, it's just, you know, the both companies are doing well. It just shows there are a lot of ways to, to skin a cat cuz you see some traction in the market in, in both approaches. So today we're gonna focus on the latest heat wave announcements and we're gonna talk about multi-cloud with a native MySQL heat wave implementation, which is available on aws MySQL heat wave for Azure via the Oracle Microsoft interconnect. This kind of cool hybrid action that they got going. Sometimes we call it super cloud. And then we're gonna dive into my SQL Heatwave Lake house, which allows users to process and query data across MyQ databases as heatwave databases, as well as object stores. So, and then we've got, heatwave has been announced on AWS and, and, and Azure, they're available now and Lake House I believe is in beta and I think it's coming out the second half of next year. So again, all of our guests are fresh off of Oracle Cloud world in Las Vegas. So they got the latest scoop. Guys, I'm done talking. Let's get into it. Mark, maybe you could start us off, what's your opinion of my SQL Heatwaves competitive position? When you think about what AWS is doing, you know, Google is, you know, we heard Google Cloud next recently, we heard about all their data innovations. You got, obviously Azure's got a big portfolio, snowflakes doing well in the market. What's your take? >>Well, first let's look at it from the point of view that AWS is the market leader in cloud and cloud services. They own somewhere between 30 to 50% depending on who you read of the market. And then you have Azure as number two and after that it falls off. There's gcp, Google Cloud platform, which is further way down the list and then Oracle and IBM and Alibaba. So when you look at AWS and you and Azure saying, hey, these are the market leaders in the cloud, then you start looking at it and saying, if I am going to provide a service that competes with the service they have, if I can make it available in their cloud, it means that I can be more competitive. And if I'm compelling and compelling means at least twice the performance or functionality or both at half the price, I should be able to gain market share. >>And that's what Oracle's done. They've taken a superior product in my SQL heat wave, which is faster, lower cost does more for a lot less at the end of the day and they make it available to the users of those clouds. You avoid this little thing called egress fees, you avoid the issue of having to migrate from one cloud to another and suddenly you have a very compelling offer. So I look at what Oracle's doing with MyQ and it feels like, I'm gonna use a word term, a flanking maneuver to their competition. They're offering a better service on their platforms. >>All right, so thank you for that. Holger, we've seen this sort of cadence, I sort of referenced it up front a little bit and they sat on MySQL for a decade, then all of a sudden we see this rush of announcements. Why did it take so long? And and more importantly is Oracle, are they developing the right features that cloud database customers are looking for in your view? >>Yeah, great question, but first of all, in your interview you said it's the edit analytics, right? Analytics is kind of like a marketing buzzword. Reports can be analytics, right? The interesting thing, which they did, the first thing they, they, they crossed the chasm between OTP and all up, right? In the same database, right? So major engineering feed very much what customers want and it's all about creating Bellevue for customers, which, which I think is the part why they go into the multi-cloud and why they add these capabilities. And they certainly with the AI capabilities, it's kind of like getting it into an autonomous field, self-driving field now with the lake cost capabilities and meeting customers where they are, like Mark has talked about the e risk costs in the cloud. So that that's a significant advantage, creating value for customers and that's what at the end of the day matters. >>And I believe strongly that long term it's gonna be ones who create better value for customers who will get more of their money From that perspective, why then take them so long? I think it's a great question. I think largely he mentioned the gentleman Nial, it's largely to who leads a product. I used to build products too, so maybe I'm a little fooling myself here, but that made the difference in my view, right? So since he's been charged, he's been building things faster than the rest of the competition, than my SQL space, which in hindsight we thought was a hot and smoking innovation phase. It kind of like was a little self complacent when it comes to the traditional borders of where, where people think, where things are separated between OTP and ola or as an example of adjacent support, right? Structured documents, whereas unstructured documents or databases and all of that has been collapsed and brought together for building a more powerful database for customers. >>So I mean it's certainly, you know, when, when Oracle talks about the competitors, you know, the competitors are in the, I always say they're, if the Oracle talks about you and knows you're doing well, so they talk a lot about aws, talk a little bit about Snowflake, you know, sort of Google, they have partnerships with Azure, but, but in, so I'm presuming that the response in MySQL heatwave was really in, in response to what they were seeing from those big competitors. But then you had Maria DB coming out, you know, the day that that Oracle acquired Sun and, and launching and going after the MySQL base. So it's, I'm, I'm interested and we'll talk about this later and what you guys think AWS and Google and Azure and Snowflake and how they're gonna respond. But, but before I do that, Ron, I want to ask you, you, you, you can get, you know, pretty technical and you've probably seen the benchmarks. >>I know you have Oracle makes a big deal out of it, publishes its benchmarks, makes some transparent on on GI GitHub. Larry Ellison talked about this in his keynote at Cloud World. What are the benchmarks show in general? I mean, when you, when you're new to the market, you gotta have a story like Mark was saying, you gotta be two x you know, the performance at half the cost or you better be or you're not gonna get any market share. So, and, and you know, oftentimes companies don't publish market benchmarks when they're leading. They do it when they, they need to gain share. So what do you make of the benchmarks? Have their, any results that were surprising to you? Have, you know, they been challenged by the competitors. Is it just a bunch of kind of desperate bench marketing to make some noise in the market or you know, are they real? What's your view? >>Well, from my perspective, I think they have the validity. And to your point, I believe that when it comes to competitor responses, that has not really happened. Nobody has like pulled down the information that's on GitHub and said, Oh, here are our price performance results. And they counter oracles. In fact, I think part of the reason why that hasn't happened is that there's the risk if Oracle's coming out and saying, Hey, we can deliver 17 times better query performance using our capabilities versus say, Snowflake when it comes to, you know, the Lakehouse platform and Snowflake turns around and says it's actually only 15 times better during performance, that's not exactly an effective maneuver. And so I think this is really to oracle's credit and I think it's refreshing because these differentiators are significant. We're not talking, you know, like 1.2% differences. We're talking 17 fold differences, we're talking six fold differences depending on, you know, where the spotlight is being shined and so forth. >>And so I think this is actually something that is actually too good to believe initially at first blush. If I'm a cloud database decision maker, I really have to prioritize this. I really would know, pay a lot more attention to this. And that's why I posed the question to Oracle and others like, okay, if these differentiators are so significant, why isn't the needle moving a bit more? And it's for, you know, some of the usual reasons. One is really deep discounting coming from, you know, the other players that's really kind of, you know, marketing 1 0 1, this is something you need to do when there's a real competitive threat to keep, you know, a customer in your own customer base. Plus there is the usual fear and uncertainty about moving from one platform to another. But I think, you know, the traction, the momentum is, is shifting an Oracle's favor. I think we saw that in the Q1 efforts, for example, where Oracle cloud grew 44% and that it generated, you know, 4.8 billion and revenue if I recall correctly. And so, so all these are demonstrating that's Oracle is making, I think many of the right moves, publishing these figures for anybody to look at from their own perspective is something that is, I think, good for the market and I think it's just gonna continue to pay dividends for Oracle down the horizon as you know, competition intens plots. So if I were in, >>Dave, can I, Dave, can I interject something and, and what Ron just said there? Yeah, please go ahead. A couple things here, one discounting, which is a common practice when you have a real threat, as Ron pointed out, isn't going to help much in this situation simply because you can't discount to the point where you improve your performance and the performance is a huge differentiator. You may be able to get your price down, but the problem that most of them have is they don't have an integrated product service. They don't have an integrated O L T P O L A P M L N data lake. Even if you cut out two of them, they don't have any of them integrated. They have multiple services that are required separate integration and that can't be overcome with discounting. And the, they, you have to pay for each one of these. And oh, by the way, as you grow, the discounts go away. So that's a, it's a minor important detail. >>So, so that's a TCO question mark, right? And I know you look at this a lot, if I had that kind of price performance advantage, I would be pounding tco, especially if I need two separate databases to do the job. That one can do, that's gonna be, the TCO numbers are gonna be off the chart or maybe down the chart, which you want. Have you looked at this and how does it compare with, you know, the big cloud guys, for example, >>I've looked at it in depth, in fact, I'm working on another TCO on this arena, but you can find it on Wiki bod in which I compared TCO for MySEQ Heat wave versus Aurora plus Redshift plus ML plus Blue. I've compared it against gcps services, Azure services, Snowflake with other services. And there's just no comparison. The, the TCO differences are huge. More importantly, thefor, the, the TCO per performance is huge. We're talking in some cases multiple orders of magnitude, but at least an order of magnitude difference. So discounting isn't gonna help you much at the end of the day, it's only going to lower your cost a little, but it doesn't improve the automation, it doesn't improve the performance, it doesn't improve the time to insight, it doesn't improve all those things that you want out of a database or multiple databases because you >>Can't discount yourself to a higher value proposition. >>So what about, I wonder ho if you could chime in on the developer angle. You, you followed that, that market. How do these innovations from heatwave, I think you used the term developer velocity. I've heard you used that before. Yeah, I mean, look, Oracle owns Java, okay, so it, it's, you know, most popular, you know, programming language in the world, blah, blah blah. But it does it have the, the minds and hearts of, of developers and does, where does heatwave fit into that equation? >>I think heatwave is gaining quickly mindshare on the developer side, right? It's not the traditional no sequel database which grew up, there's a traditional mistrust of oracles to developers to what was happening to open source when gets acquired. Like in the case of Oracle versus Java and where my sql, right? And, but we know it's not a good competitive strategy to, to bank on Oracle screwing up because it hasn't worked not on Java known my sequel, right? And for developers, it's, once you get to know a technology product and you can do more, it becomes kind of like a Swiss army knife and you can build more use case, you can build more powerful applications. That's super, super important because you don't have to get certified in multiple databases. You, you are fast at getting things done, you achieve fire, develop velocity, and the managers are happy because they don't have to license more things, send you to more trainings, have more risk of something not being delivered, right? >>So it's really the, we see the suite where this best of breed play happening here, which in general was happening before already with Oracle's flagship database. Whereas those Amazon as an example, right? And now the interesting thing is every step away Oracle was always a one database company that can be only one and they're now generally talking about heat web and that two database company with different market spaces, but same value proposition of integrating more things very, very quickly to have a universal database that I call, they call the converge database for all the needs of an enterprise to run certain application use cases. And that's what's attractive to developers. >>It's, it's ironic isn't it? I mean I, you know, the rumor was the TK Thomas Curian left Oracle cuz he wanted to put Oracle database on other clouds and other places. And maybe that was the rift. Maybe there was, I'm sure there was other things, but, but Oracle clearly is now trying to expand its Tam Ron with, with heatwave into aws, into Azure. How do you think Oracle's gonna do, you were at a cloud world, what was the sentiment from customers and the independent analyst? Is this just Oracle trying to screw with the competition, create a little diversion? Or is this, you know, serious business for Oracle? What do you think? >>No, I think it has lakes. I think it's definitely, again, attriting to Oracle's overall ability to differentiate not only my SQL heat wave, but its overall portfolio. And I think the fact that they do have the alliance with the Azure in place, that this is definitely demonstrating their commitment to meeting the multi-cloud needs of its customers as well as what we pointed to in terms of the fact that they're now offering, you know, MySQL capabilities within AWS natively and that it can now perform AWS's own offering. And I think this is all demonstrating that Oracle is, you know, not letting up, they're not resting on its laurels. That's clearly we are living in a multi-cloud world, so why not just make it more easy for customers to be able to use cloud databases according to their own specific, specific needs. And I think, you know, to holder's point, I think that definitely lines with being able to bring on more application developers to leverage these capabilities. >>I think one important announcement that's related to all this was the JSON relational duality capabilities where now it's a lot easier for application developers to use a language that they're very familiar with a JS O and not have to worry about going into relational databases to store their J S O N application coding. So this is, I think an example of the innovation that's enhancing the overall Oracle portfolio and certainly all the work with machine learning is definitely paying dividends as well. And as a result, I see Oracle continue to make these inroads that we pointed to. But I agree with Mark, you know, the short term discounting is just a stall tag. This is not denying the fact that Oracle is being able to not only deliver price performance differentiators that are dramatic, but also meeting a wide range of needs for customers out there that aren't just limited device performance consideration. >>Being able to support multi-cloud according to customer needs. Being able to reach out to the application developer community and address a very specific challenge that has plagued them for many years now. So bring it all together. Yeah, I see this as just enabling Oracles who ring true with customers. That the customers that were there were basically all of them, even though not all of them are going to be saying the same things, they're all basically saying positive feedback. And likewise, I think the analyst community is seeing this. It's always refreshing to be able to talk to customers directly and at Oracle cloud there was a litany of them and so this is just a difference maker as well as being able to talk to strategic partners. The nvidia, I think partnerships also testament to Oracle's ongoing ability to, you know, make the ecosystem more user friendly for the customers out there. >>Yeah, it's interesting when you get these all in one tools, you know, the Swiss Army knife, you expect that it's not able to be best of breed. That's the kind of surprising thing that I'm hearing about, about heatwave. I want to, I want to talk about Lake House because when I think of Lake House, I think data bricks, and to my knowledge data bricks hasn't been in the sites of Oracle yet. Maybe they're next, but, but Oracle claims that MySQL, heatwave, Lakehouse is a breakthrough in terms of capacity and performance. Mark, what are your thoughts on that? Can you double click on, on Lakehouse Oracle's claims for things like query performance and data loading? What does it mean for the market? Is Oracle really leading in, in the lake house competitive landscape? What are your thoughts? >>Well, but name in the game is what are the problems you're solving for the customer? More importantly, are those problems urgent or important? If they're urgent, customers wanna solve 'em. Now if they're important, they might get around to them. So you look at what they're doing with Lake House or previous to that machine learning or previous to that automation or previous to that O L A with O ltp and they're merging all this capability together. If you look at Snowflake or data bricks, they're tacking one problem. You look at MyQ heat wave, they're tacking multiple problems. So when you say, yeah, their queries are much better against the lake house in combination with other analytics in combination with O ltp and the fact that there are no ETLs. So you're getting all this done in real time. So it's, it's doing the query cross, cross everything in real time. >>You're solving multiple user and developer problems, you're increasing their ability to get insight faster, you're having shorter response times. So yeah, they really are solving urgent problems for customers. And by putting it where the customer lives, this is the brilliance of actually being multicloud. And I know I'm backing up here a second, but by making it work in AWS and Azure where people already live, where they already have applications, what they're saying is, we're bringing it to you. You don't have to come to us to get these, these benefits, this value overall, I think it's a brilliant strategy. I give Nip and Argo wallet a huge, huge kudos for what he's doing there. So yes, what they're doing with the lake house is going to put notice on data bricks and Snowflake and everyone else for that matter. Well >>Those are guys that whole ago you, you and I have talked about this. Those are, those are the guys that are doing sort of the best of breed. You know, they're really focused and they, you know, tend to do well at least out of the gate. Now you got Oracle's converged philosophy, obviously with Oracle database. We've seen that now it's kicking in gear with, with heatwave, you know, this whole thing of sweets versus best of breed. I mean the long term, you know, customers tend to migrate towards suite, but the new shiny toy tends to get the growth. How do you think this is gonna play out in cloud database? >>Well, it's the forever never ending story, right? And in software right suite, whereas best of breed and so far in the long run suites have always won, right? So, and sometimes they struggle again because the inherent problem of sweets is you build something larger, it has more complexity and that means your cycles to get everything working together to integrate the test that roll it out, certify whatever it is, takes you longer, right? And that's not the case. It's a fascinating part of what the effort around my SQL heat wave is that the team is out executing the previous best of breed data, bringing us something together. Now if they can maintain that pace, that's something to to, to be seen. But it, the strategy, like what Mark was saying, bring the software to the data is of course interesting and unique and totally an Oracle issue in the past, right? >>Yeah. But it had to be in your database on oci. And but at, that's an interesting part. The interesting thing on the Lake health side is, right, there's three key benefits of a lakehouse. The first one is better reporting analytics, bring more rich information together, like make the, the, the case for silicon angle, right? We want to see engagements for this video, we want to know what's happening. That's a mixed transactional video media use case, right? Typical Lakehouse use case. The next one is to build more rich applications, transactional applications which have video and these elements in there, which are the engaging one. And the third one, and that's where I'm a little critical and concerned, is it's really the base platform for artificial intelligence, right? To run deep learning to run things automatically because they have all the data in one place can create in one way. >>And that's where Oracle, I know that Ron talked about Invidia for a moment, but that's where Oracle doesn't have the strongest best story. Nonetheless, the two other main use cases of the lake house are very strong, very well only concern is four 50 terabyte sounds long. It's an arbitrary limitation. Yeah, sounds as big. So for the start, and it's the first word, they can make that bigger. You don't want your lake house to be limited and the terabyte sizes or any even petabyte size because you want to have the certainty. I can put everything in there that I think it might be relevant without knowing what questions to ask and query those questions. >>Yeah. And you know, in the early days of no schema on right, it just became a mess. But now technology has evolved to allow us to actually get more value out of that data. Data lake. Data swamp is, you know, not much more, more, more, more logical. But, and I want to get in, in a moment, I want to come back to how you think the competitors are gonna respond. Are they gonna have to sort of do a more of a converged approach? AWS in particular? But before I do, Ron, I want to ask you a question about autopilot because I heard Larry Ellison's keynote and he was talking about how, you know, most security issues are human errors with autonomy and autonomous database and things like autopilot. We take care of that. It's like autonomous vehicles, they're gonna be safer. And I went, well maybe, maybe someday. So Oracle really tries to emphasize this, that every time you see an announcement from Oracle, they talk about new, you know, autonomous capabilities. It, how legit is it? Do people care? What about, you know, what's new for heatwave Lakehouse? How much of a differentiator, Ron, do you really think autopilot is in this cloud database space? >>Yeah, I think it will definitely enhance the overall proposition. I don't think people are gonna buy, you know, lake house exclusively cause of autopilot capabilities, but when they look at the overall picture, I think it will be an added capability bonus to Oracle's benefit. And yeah, I think it's kind of one of these age old questions, how much do you automate and what is the bounce to strike? And I think we all understand with the automatic car, autonomous car analogy that there are limitations to being able to use that. However, I think it's a tool that basically every organization out there needs to at least have or at least evaluate because it goes to the point of it helps with ease of use, it helps make automation more balanced in terms of, you know, being able to test, all right, let's automate this process and see if it works well, then we can go on and switch on on autopilot for other processes. >>And then, you know, that allows, for example, the specialists to spend more time on business use cases versus, you know, manual maintenance of, of the cloud database and so forth. So I think that actually is a, a legitimate value proposition. I think it's just gonna be a case by case basis. Some organizations are gonna be more aggressive with putting automation throughout their processes throughout their organization. Others are gonna be more cautious. But it's gonna be, again, something that will help the overall Oracle proposition. And something that I think will be used with caution by many organizations, but other organizations are gonna like, hey, great, this is something that is really answering a real problem. And that is just easing the use of these databases, but also being able to better handle the automation capabilities and benefits that come with it without having, you know, a major screwup happened and the process of transitioning to more automated capabilities. >>Now, I didn't attend cloud world, it's just too many red eyes, you know, recently, so I passed. But one of the things I like to do at those events is talk to customers, you know, in the spirit of the truth, you know, they, you know, you'd have the hallway, you know, track and to talk to customers and they say, Hey, you know, here's the good, the bad and the ugly. So did you guys, did you talk to any customers my SQL Heatwave customers at, at cloud world? And and what did you learn? I don't know, Mark, did you, did you have any luck and, and having some, some private conversations? >>Yeah, I had quite a few private conversations. The one thing before I get to that, I want disagree with one point Ron made, I do believe there are customers out there buying the heat wave service, the MySEQ heat wave server service because of autopilot. Because autopilot is really revolutionary in many ways in the sense for the MySEQ developer in that it, it auto provisions, it auto parallel loads, IT auto data places it auto shape predictions. It can tell you what machine learning models are going to tell you, gonna give you your best results. And, and candidly, I've yet to meet a DBA who didn't wanna give up pedantic tasks that are pain in the kahoo, which they'd rather not do and if it's long as it was done right for them. So yes, I do think people are buying it because of autopilot and that's based on some of the conversations I had with customers at Oracle Cloud World. >>In fact, it was like, yeah, that's great, yeah, we get fantastic performance, but this really makes my life easier and I've yet to meet a DBA who didn't want to make their life easier. And it does. So yeah, I've talked to a few of them. They were excited. I asked them if they ran into any bugs, were there any difficulties in moving to it? And the answer was no. In both cases, it's interesting to note, my sequel is the most popular database on the planet. Well, some will argue that it's neck and neck with SQL Server, but if you add in Mariah DB and ProCon db, which are forks of MySQL, then yeah, by far and away it's the most popular. And as a result of that, everybody for the most part has typically a my sequel database somewhere in their organization. So this is a brilliant situation for anybody going after MyQ, but especially for heat wave. And the customers I talk to love it. I didn't find anybody complaining about it. And >>What about the migration? We talked about TCO earlier. Did your t does your TCO analysis include the migration cost or do you kind of conveniently leave that out or what? >>Well, when you look at migration costs, there are different kinds of migration costs. By the way, the worst job in the data center is the data migration manager. Forget it, no other job is as bad as that one. You get no attaboys for doing it. Right? And then when you screw up, oh boy. So in real terms, anything that can limit data migration is a good thing. And when you look at Data Lake, that limits data migration. So if you're already a MySEQ user, this is a pure MySQL as far as you're concerned. It's just a, a simple transition from one to the other. You may wanna make sure nothing broke and every you, all your tables are correct and your schema's, okay, but it's all the same. So it's a simple migration. So it's pretty much a non-event, right? When you migrate data from an O LTP to an O L A P, that's an ETL and that's gonna take time. >>But you don't have to do that with my SQL heat wave. So that's gone when you start talking about machine learning, again, you may have an etl, you may not, depending on the circumstances, but again, with my SQL heat wave, you don't, and you don't have duplicate storage, you don't have to copy it from one storage container to another to be able to be used in a different database, which by the way, ultimately adds much more cost than just the other service. So yeah, I looked at the migration and again, the users I talked to said it was a non-event. It was literally moving from one physical machine to another. If they had a new version of MySEQ running on something else and just wanted to migrate it over or just hook it up or just connect it to the data, it worked just fine. >>Okay, so every day it sounds like you guys feel, and we've certainly heard this, my colleague David Foyer, the semi-retired David Foyer was always very high on heatwave. So I think you knows got some real legitimacy here coming from a standing start, but I wanna talk about the competition, how they're likely to respond. I mean, if your AWS and you got heatwave is now in your cloud, so there's some good aspects of that. The database guys might not like that, but the infrastructure guys probably love it. Hey, more ways to sell, you know, EC two and graviton, but you're gonna, the database guys in AWS are gonna respond. They're gonna say, Hey, we got Redshift, we got aqua. What's your thoughts on, on not only how that's gonna resonate with customers, but I'm interested in what you guys think will a, I never say never about aws, you know, and are they gonna try to build, in your view a converged Oola and o LTP database? You know, Snowflake is taking an ecosystem approach. They've added in transactional capabilities to the portfolio so they're not standing still. What do you guys see in the competitive landscape in that regard going forward? Maybe Holger, you could start us off and anybody else who wants to can chime in, >>Happy to, you mentioned Snowflake last, we'll start there. I think Snowflake is imitating that strategy, right? That building out original data warehouse and the clouds tasking project to really proposition to have other data available there because AI is relevant for everybody. Ultimately people keep data in the cloud for ultimately running ai. So you see the same suite kind of like level strategy, it's gonna be a little harder because of the original positioning. How much would people know that you're doing other stuff? And I just, as a former developer manager of developers, I just don't see the speed at the moment happening at Snowflake to become really competitive to Oracle. On the flip side, putting my Oracle hat on for a moment back to you, Mark and Iran, right? What could Oracle still add? Because the, the big big things, right? The traditional chasms in the database world, they have built everything, right? >>So I, I really scratched my hat and gave Nipon a hard time at Cloud world say like, what could you be building? Destiny was very conservative. Let's get the Lakehouse thing done, it's gonna spring next year, right? And the AWS is really hard because AWS value proposition is these small innovation teams, right? That they build two pizza teams, which can be fit by two pizzas, not large teams, right? And you need suites to large teams to build these suites with lots of functionalities to make sure they work together. They're consistent, they have the same UX on the administration side, they can consume the same way, they have the same API registry, can't even stop going where the synergy comes to play over suite. So, so it's gonna be really, really hard for them to change that. But AWS super pragmatic. They're always by themselves that they'll listen to customers if they learn from customers suite as a proposition. I would not be surprised if AWS trying to bring things closer together, being morely together. >>Yeah. Well how about, can we talk about multicloud if, if, again, Oracle is very on on Oracle as you said before, but let's look forward, you know, half a year or a year. What do you think about Oracle's moves in, in multicloud in terms of what kind of penetration they're gonna have in the marketplace? You saw a lot of presentations at at cloud world, you know, we've looked pretty closely at the, the Microsoft Azure deal. I think that's really interesting. I've, I've called it a little bit of early days of a super cloud. What impact do you think this is gonna have on, on the marketplace? But, but both. And think about it within Oracle's customer base, I have no doubt they'll do great there. But what about beyond its existing install base? What do you guys think? >>Ryan, do you wanna jump on that? Go ahead. Go ahead Ryan. No, no, no, >>That's an excellent point. I think it aligns with what we've been talking about in terms of Lakehouse. I think Lake House will enable Oracle to pull more customers, more bicycle customers onto the Oracle platforms. And I think we're seeing all the signs pointing toward Oracle being able to make more inroads into the overall market. And that includes garnishing customers from the leaders in, in other words, because they are, you know, coming in as a innovator, a an alternative to, you know, the AWS proposition, the Google cloud proposition that they have less to lose and there's a result they can really drive the multi-cloud messaging to resonate with not only their existing customers, but also to be able to, to that question, Dave's posing actually garnish customers onto their platform. And, and that includes naturally my sequel but also OCI and so forth. So that's how I'm seeing this playing out. I think, you know, again, Oracle's reporting is indicating that, and I think what we saw, Oracle Cloud world is definitely validating the idea that Oracle can make more waves in the overall market in this regard. >>You know, I, I've floated this idea of Super cloud, it's kind of tongue in cheek, but, but there, I think there is some merit to it in terms of building on top of hyperscale infrastructure and abstracting some of the, that complexity. And one of the things that I'm most interested in is industry clouds and an Oracle acquisition of Cerner. I was struck by Larry Ellison's keynote, it was like, I don't know, an hour and a half and an hour and 15 minutes was focused on healthcare transformation. Well, >>So vertical, >>Right? And so, yeah, so you got Oracle's, you know, got some industry chops and you, and then you think about what they're building with, with not only oci, but then you got, you know, MyQ, you can now run in dedicated regions. You got ADB on on Exadata cloud to customer, you can put that OnPrem in in your data center and you look at what the other hyperscalers are, are doing. I I say other hyperscalers, I've always said Oracle's not really a hyperscaler, but they got a cloud so they're in the game. But you can't get, you know, big query OnPrem, you look at outposts, it's very limited in terms of, you know, the database support and again, that that will will evolve. But now you got Oracle's got, they announced Alloy, we can white label their cloud. So I'm interested in what you guys think about these moves, especially the industry cloud. We see, you know, Walmart is doing sort of their own cloud. You got Goldman Sachs doing a cloud. Do you, you guys, what do you think about that and what role does Oracle play? Any thoughts? >>Yeah, let me lemme jump on that for a moment. Now, especially with the MyQ, by making that available in multiple clouds, what they're doing is this follows the philosophy they've had the past with doing cloud, a customer taking the application and the data and putting it where the customer lives. If it's on premise, it's on premise. If it's in the cloud, it's in the cloud. By making the mice equal heat wave, essentially a plug compatible with any other mice equal as far as your, your database is concern and then giving you that integration with O L A P and ML and Data Lake and everything else, then what you've got is a compelling offering. You're making it easier for the customer to use. So I look the difference between MyQ and the Oracle database, MyQ is going to capture market more market share for them. >>You're not gonna find a lot of new users for the Oracle debate database. Yeah, there are always gonna be new users, don't get me wrong, but it's not gonna be a huge growth. Whereas my SQL heatwave is probably gonna be a major growth engine for Oracle going forward. Not just in their own cloud, but in AWS and in Azure and on premise over time that eventually it'll get there. It's not there now, but it will, they're doing the right thing on that basis. They're taking the services and when you talk about multicloud and making them available where the customer wants them, not forcing them to go where you want them, if that makes sense. And as far as where they're going in the future, I think they're gonna take a page outta what they've done with the Oracle database. They'll add things like JSON and XML and time series and spatial over time they'll make it a, a complete converged database like they did with the Oracle database. The difference being Oracle database will scale bigger and will have more transactions and be somewhat faster. And my SQL will be, for anyone who's not on the Oracle database, they're, they're not stupid, that's for sure. >>They've done Jason already. Right. But I give you that they could add graph and time series, right. Since eat with, Right, Right. Yeah, that's something absolutely right. That's, that's >>A sort of a logical move, right? >>Right. But that's, that's some kid ourselves, right? I mean has worked in Oracle's favor, right? 10 x 20 x, the amount of r and d, which is in the MyQ space, has been poured at trying to snatch workloads away from Oracle by starting with IBM 30 years ago, 20 years ago, Microsoft and, and, and, and didn't work, right? Database applications are extremely sticky when they run, you don't want to touch SIM and grow them, right? So that doesn't mean that heat phase is not an attractive offering, but it will be net new things, right? And what works in my SQL heat wave heat phases favor a little bit is it's not the massive enterprise applications which have like we the nails like, like you might be only running 30% or Oracle, but the connections and the interfaces into that is, is like 70, 80% of your enterprise. >>You take it out and it's like the spaghetti ball where you say, ah, no I really don't, don't want to do all that. Right? You don't, don't have that massive part with the equals heat phase sequel kind of like database which are more smaller tactical in comparison, but still I, I don't see them taking so much share. They will be growing because of a attractive value proposition quickly on the, the multi-cloud, right? I think it's not really multi-cloud. If you give people the chance to run your offering on different clouds, right? You can run it there. The multi-cloud advantages when the Uber offering comes out, which allows you to do things across those installations, right? I can migrate data, I can create data across something like Google has done with B query Omni, I can run predictive models or even make iron models in different place and distribute them, right? And Oracle is paving the road for that, but being available on these clouds. But the multi-cloud capability of database which knows I'm running on different clouds that is still yet to be built there. >>Yeah. And >>That the problem with >>That, that's the super cloud concept that I flowed and I I've always said kinda snowflake with a single global instance is sort of, you know, headed in that direction and maybe has a league. What's the issue with that mark? >>Yeah, the problem with the, with that version, the multi-cloud is clouds to charge egress fees. As long as they charge egress fees to move data between clouds, it's gonna make it very difficult to do a real multi-cloud implementation. Even Snowflake, which runs multi-cloud, has to pass out on the egress fees of their customer when data moves between clouds. And that's really expensive. I mean there, there is one customer I talked to who is beta testing for them, the MySQL heatwave and aws. The only reason they didn't want to do that until it was running on AWS is the egress fees were so great to move it to OCI that they couldn't afford it. Yeah. Egress fees are the big issue but, >>But Mark the, the point might be you might wanna root query and only get the results set back, right was much more tinier, which been the answer before for low latency between the class A problem, which we sometimes still have but mostly don't have. Right? And I think in general this with fees coming down based on the Oracle general E with fee move and it's very hard to justify those, right? But, but it's, it's not about moving data as a multi-cloud high value use case. It's about doing intelligent things with that data, right? Putting into other places, replicating it, what I'm saying the same thing what you said before, running remote queries on that, analyzing it, running AI on it, running AI models on that. That's the interesting thing. Cross administered in the same way. Taking things out, making sure compliance happens. Making sure when Ron says I don't want to be American anymore, I want to be in the European cloud that is gets migrated, right? So tho those are the interesting value use case which are really, really hard for enterprise to program hand by hand by developers and they would love to have out of the box and that's yet the innovation to come to, we have to come to see. But the first step to get there is that your software runs in multiple clouds and that's what Oracle's doing so well with my SQL >>Guys. Amazing. >>Go ahead. Yeah. >>Yeah. >>For example, >>Amazing amount of data knowledge and, and brain power in this market. Guys, I really want to thank you for coming on to the cube. Ron Holger. Mark, always a pleasure to have you on. Really appreciate your time. >>Well all the last names we're very happy for Romanic last and moderator. Thanks Dave for moderating us. All right, >>We'll see. We'll see you guys around. Safe travels to all and thank you for watching this power panel, The Truth About My SQL Heat Wave on the cube. Your leader in enterprise and emerging tech coverage.

Published Date : Nov 1 2022

SUMMARY :

Always a pleasure to have you on. I think you just saw him at Oracle Cloud World and he's come on to describe this is doing, you know, Google is, you know, we heard Google Cloud next recently, They own somewhere between 30 to 50% depending on who you read migrate from one cloud to another and suddenly you have a very compelling offer. All right, so thank you for that. And they certainly with the AI capabilities, And I believe strongly that long term it's gonna be ones who create better value for So I mean it's certainly, you know, when, when Oracle talks about the competitors, So what do you make of the benchmarks? say, Snowflake when it comes to, you know, the Lakehouse platform and threat to keep, you know, a customer in your own customer base. And oh, by the way, as you grow, And I know you look at this a lot, to insight, it doesn't improve all those things that you want out of a database or multiple databases So what about, I wonder ho if you could chime in on the developer angle. they don't have to license more things, send you to more trainings, have more risk of something not being delivered, all the needs of an enterprise to run certain application use cases. I mean I, you know, the rumor was the TK Thomas Curian left Oracle And I think, you know, to holder's point, I think that definitely lines But I agree with Mark, you know, the short term discounting is just a stall tag. testament to Oracle's ongoing ability to, you know, make the ecosystem Yeah, it's interesting when you get these all in one tools, you know, the Swiss Army knife, you expect that it's not able So when you say, yeah, their queries are much better against the lake house in You don't have to come to us to get these, these benefits, I mean the long term, you know, customers tend to migrate towards suite, but the new shiny bring the software to the data is of course interesting and unique and totally an Oracle issue in And the third one, lake house to be limited and the terabyte sizes or any even petabyte size because you want keynote and he was talking about how, you know, most security issues are human I don't think people are gonna buy, you know, lake house exclusively cause of And then, you know, that allows, for example, the specialists to And and what did you learn? The one thing before I get to that, I want disagree with And the customers I talk to love it. the migration cost or do you kind of conveniently leave that out or what? And when you look at Data Lake, that limits data migration. So that's gone when you start talking about So I think you knows got some real legitimacy here coming from a standing start, So you see the same And you need suites to large teams to build these suites with lots of functionalities You saw a lot of presentations at at cloud world, you know, we've looked pretty closely at Ryan, do you wanna jump on that? I think, you know, again, Oracle's reporting I think there is some merit to it in terms of building on top of hyperscale infrastructure and to customer, you can put that OnPrem in in your data center and you look at what the So I look the difference between MyQ and the Oracle database, MyQ is going to capture market They're taking the services and when you talk about multicloud and But I give you that they could add graph and time series, right. like, like you might be only running 30% or Oracle, but the connections and the interfaces into You take it out and it's like the spaghetti ball where you say, ah, no I really don't, global instance is sort of, you know, headed in that direction and maybe has a league. Yeah, the problem with the, with that version, the multi-cloud is clouds And I think in general this with fees coming down based on the Oracle general E with fee move Yeah. Guys, I really want to thank you for coming on to the cube. Well all the last names we're very happy for Romanic last and moderator. We'll see you guys around.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MarkPERSON

0.99+

Ron HolgerPERSON

0.99+

RonPERSON

0.99+

Mark StammerPERSON

0.99+

IBMORGANIZATION

0.99+

Ron WestfallPERSON

0.99+

RyanPERSON

0.99+

AWSORGANIZATION

0.99+

DavePERSON

0.99+

WalmartORGANIZATION

0.99+

Larry EllisonPERSON

0.99+

MicrosoftORGANIZATION

0.99+

AlibabaORGANIZATION

0.99+

OracleORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

Holgar MuellerPERSON

0.99+

AmazonORGANIZATION

0.99+

Constellation ResearchORGANIZATION

0.99+

Goldman SachsORGANIZATION

0.99+

17 timesQUANTITY

0.99+

twoQUANTITY

0.99+

David FoyerPERSON

0.99+

44%QUANTITY

0.99+

1.2%QUANTITY

0.99+

4.8 billionQUANTITY

0.99+

JasonPERSON

0.99+

UberORGANIZATION

0.99+

Fu Chim ResearchORGANIZATION

0.99+

Dave AntePERSON

0.99+

Oracle Announces MySQL HeatWave on AWS


 

>>Oracle continues to enhance my sequel Heatwave at a very rapid pace. The company is now in its fourth major release since the original announcement in December 2020. 1 of the main criticisms of my sequel, Heatwave, is that it only runs on O. C I. Oracle Cloud Infrastructure and as a lock in to Oracle's Cloud. Oracle recently announced that heat wave is now going to be available in AWS Cloud and it announced its intent to bring my sequel Heatwave to Azure. So my secret heatwave on AWS is a significant TAM expansion move for Oracle because of the momentum AWS Cloud continues to show. And evidently the Heatwave Engineering team has taken the development effort from O. C I. And is bringing that to A W S with a number of enhancements that we're gonna dig into today is senior vice president. My sequel Heatwave at Oracle is back with me on a cube conversation to discuss the latest heatwave news, and we're eager to hear any benchmarks relative to a W S or any others. Nippon has been leading the Heatwave engineering team for over 10 years and there's over 100 and 85 patents and database technology. Welcome back to the show and good to see you. >>Thank you. Very happy to be back. >>Now for those who might not have kept up with the news, uh, to kick things off, give us an overview of my sequel, Heatwave and its evolution. So far, >>so my sequel, Heat Wave, is a fully managed my secret database service offering from Oracle. Traditionally, my secret has been designed and optimised for transaction processing. So customers of my sequel then they had to run analytics or when they had to run machine learning, they would extract the data out of my sequel into some other database for doing. Unlike processing or machine learning processing my sequel, Heat provides all these capabilities built in to a single database service, which is my sequel. He'd fake So customers of my sequel don't need to move the data out with the same database. They can run transaction processing and predicts mixed workloads, machine learning, all with a very, very good performance in very good price performance. Furthermore, one of the design points of heat wave is is a scale out architecture, so the system continues to scale and performed very well, even when customers have very large late assignments. >>So we've seen some interesting moves by Oracle lately. The collaboration with Azure we've we've covered that pretty extensively. What was the impetus here for bringing my sequel Heatwave onto the AWS cloud? What were the drivers that you considered? >>So one of the observations is that a very large percentage of users of my sequel Heatwave, our AWS users who are migrating of Aurora or so already we see that a good percentage of my secret history of customers are migrating from GWS. However, there are some AWS customers who are still not able to migrate the O. C. I to my secret heat wave. And the reason is because of, um, exorbitant cost, which was charges. So in order to migrate the workload from AWS to go see, I digress. Charges are very high fees which becomes prohibitive for the customer or the second example we have seen is that the latency of practising a database which is outside of AWS is very high. So there's a class of customers who would like to get the benefits of my secret heatwave but were unable to do so and with this support of my secret trip inside of AWS, these customers can now get all the grease of the benefits of my secret he trip without having to pay the high fees or without having to suffer with the poorly agency, which is because of the ws architecture. >>Okay, so you're basically meeting the customer's where they are. So was this a straightforward lifted shift from from Oracle Cloud Infrastructure to AWS? >>No, it is not because one of the design girls we have with my sequel, Heatwave is that we want to provide our customers with the best price performance regardless of the cloud. So when we decided to offer my sequel, he headed west. Um, we have optimised my sequel Heatwave on it as well. So one of the things to point out is that this is a service with the data plane control plane and the console are natively running on AWS. And the benefits of doing so is that now we can optimise my sequel Heatwave for the E. W s architecture. In addition to that, we have also announced a bunch of new capabilities as a part of the service which will also be available to the my secret history of customers and our CI, But we just announced them and we're offering them as a part of my secret history of offering on AWS. >>So I just want to make sure I understand that it's not like you just wrapped your stack in a container and stuck it into a W s to be hosted. You're saying you're actually taking advantage of the capabilities of the AWS cloud natively? And I think you've made some other enhancements as well that you're alluding to. Can you maybe, uh, elucidate on those? Sure. >>So for status, um, we have taken the mind sequel Heatwave code and we have optimised for the It was infrastructure with its computer network. And as a result, customers get very good performance and price performance. Uh, with my secret he trade in AWS. That's one performance. Second thing is, we have designed new interactive counsel for the service, which means that customers can now provision there instances with the council. But in addition, they can also manage their schemas. They can. Then court is directly from the council. Autopilot is integrated. The council we have introduced performance monitoring, so a lot of capabilities which we have introduced as a part of the new counsel. The third thing is that we have added a bunch of new security features, uh, expose some of the security features which were part of the My Secret Enterprise edition as a part of the service, which gives customers now a choice of using these features to build more secure applications. And finally, we have extended my secret autopilot for a number of old gpus cases. In the past, my secret autopilot had a lot of capabilities for Benedict, and now we have augmented my secret autopilot to offer capabilities for elderly people. Includes as well. >>But there was something in your press release called Auto thread. Pooling says it provides higher and sustained throughput. High concerns concerns concurrency by determining Apple number of transactions, which should be executed. Uh, what is that all about? The auto thread pool? It seems pretty interesting. How does it affect performance? Can you help us understand that? >>Yes, and this is one of the capabilities of alluding to which we have added in my secret autopilot for transaction processing. So here is the basic idea. If you have a system where there's a large number of old EP transactions coming into it at a high degrees of concurrency in many of the existing systems of my sequel based systems, it can lead to a state where there are few transactions executing, but a bunch of them can get blocked with or a pilot tried pulling. What we basically do is we do workload aware admission control and what this does is it figures out, what's the right scheduling or all of these algorithms, so that either the transactions are executing or as soon as something frees up, they can start executing, so there's no transaction which is blocked. The advantage to the customer of this capability is twofold. A get significantly better throughput compared to service like Aurora at high levels of concurrency. So at high concurrency, for instance, uh, my secret because of this capability Uh oh, thread pulling offers up to 10 times higher compared to Aurora, that's one first benefit better throughput. The second advantage is that the true part of the system never drops, even at high levels of concurrency, whereas in the case of Aurora, the trooper goes up, but then, at high concurrency is, let's say, starting, uh, level of 500 or something. It depends upon the underlying shit they're using the troopers just dropping where it's with my secret heatwave. The truth will never drops. Now, the ramification for the customer is that if the truth is not gonna drop, the user can start off with a small shape, get the performance and be a show that even the workload increases. They will never get a performance, which is worse than what they're getting with lower levels of concurrency. So this let's leads to customers provisioning a shape which is just right for them. And if they need, they can, uh, go with the largest shape. But they don't like, you know, over pay. So those are the two benefits. Better performance and sustain, uh, regardless of the level of concurrency. >>So how do we quantify that? I know you've got some benchmarks. How can you share comparisons with other cloud databases especially interested in in Amazon's own databases are obviously very popular, and and are you publishing those again and get hub, as you have done in the past? Take us through the benchmarks. >>Sure, So benchmarks are important because that gives customers a sense of what performance to expect and what price performance to expect. So we have run a number of benchmarks. And yes, all these benchmarks are available on guitar for customers to take a look at. So we have performance results on all the three castle workloads, ol DB Analytics and Machine Learning. So let's start with the Rdp for Rdp and primarily because of the auto thread pulling feature. We show that for the IPCC for attended dataset at high levels of concurrency, heatwave offers up to 10 times better throughput and this performance is sustained, whereas in the case of Aurora, the performance really drops. So that's the first thing that, uh, tend to alibi. Sorry, 10 gigabytes. B B C c. I can come and see the performance are the throughput is 10 times better than Aurora for analytics. We have done a comparison of my secret heatwave in AWS and compared with Red Ship Snowflake Googled inquiry, we find that the price performance of my secret heatwave compared to read ship is seven times better. So my sequel, Heat Wave in AWS, provides seven times better price performance than red ship. That's a very, uh, interesting results to us. Which means that customers of Red Shift are really going to take the service seriously because they're gonna get seven times better price performance. And this is all running in a W s so compared. >>Okay, carry on. >>And then I was gonna say, compared to like, Snowflake, uh, in AWS offers 10 times better price performance. And compared to Google, ubiquity offers 12 times better price performance. And this is based on a four terabyte p PCH workload. Results are available on guitar, and then the third category is machine learning and for machine learning, uh, for training, the performance of my secret heatwave is 25 times faster compared to that shit. So all the three workloads we have benchmark's results, and all of these scripts are available on YouTube. >>Okay, so you're comparing, uh, my sequel Heatwave on AWS to Red Shift and snowflake on AWS. And you're comparing my sequel Heatwave on a W s too big query. Obviously running on on Google. Um, you know, one of the things Oracle is done in the past when you get the price performance and I've always tried to call fouls you're, like, double your price for running the oracle database. Uh, not Heatwave, but Oracle Database on a W s. And then you'll show how it's it's so much cheaper on on Oracle will be like Okay, come on. But they're not doing that here. You're basically taking my sequel Heatwave on a W s. I presume you're using the same pricing for whatever you see to whatever else you're using. Storage, um, reserved instances. That's apples to apples on A W s. And you have to obviously do some kind of mapping for for Google, for big query. Can you just verify that for me, >>we are being more than fair on two dimensions. The first thing is, when I'm talking about the price performance for analytics, right for, uh, with my secret heat rape, the cost I'm talking about from my secret heat rape is the cost of running transaction processing, analytics and machine learning. So it's a fully loaded cost for the case of my secret heatwave. There has been I'm talking about red ship when I'm talking about Snowflake. I'm just talking about the cost of these databases for running, and it's only it's not, including the source database, which may be more or some other database, right? So that's the first aspect that far, uh, trip. It's the cost for running all three kinds of workloads, whereas for the competition, it's only for running analytics. The second thing is that for these are those services whether it's like shit or snowflakes, That's right. We're talking about one year, fully paid up front cost, right? So that's what most of the customers would pay for. Many of the customers would pay that they will sign a one year contract and pay all the costs ahead of time because they get a discount. So we're using that price and the case of Snowflake. The costs were using is their standard edition of price, not the Enterprise edition price. So yes, uh, more than in this competitive. >>Yeah, I think that's an important point. I saw an analysis by Marx Tamer on Wiki Bond, where he was doing the TCO comparisons. And I mean, if you have to use two separate databases in two separate licences and you have to do et yelling and all the labour associated with that, that that's that's a big deal and you're not even including that aspect in in your comparison. So that's pretty impressive. To what do you attribute that? You know, given that unlike, oh, ci within the AWS cloud, you don't have as much control over the underlying hardware. >>So look hard, but is one aspect. Okay, so there are three things which give us this advantage. The first thing is, uh, we have designed hateful foreign scale out architecture. So we came up with new algorithms we have come up with, like, uh, one of the design points for heat wave is a massively partitioned architecture, which leads to a very high degree of parallelism. So that's a lot of hype. Each were built, So that's the first part. The second thing is that although we don't have control over the hardware, but the second design point for heat wave is that it is optimised for commodity cloud and the commodity infrastructure so we can have another guys, what to say? The computer we get, how much network bandwidth do we get? How much of, like objects to a brand that we get in here? W s. And we have tuned heat for that. That's the second point And the third thing is my secret autopilot, which provides machine learning based automation. So what it does is that has the users workload is running. It learns from it, it improves, uh, various premieres in the system. So the system keeps getting better as you learn more and more questions. And this is the third thing, uh, as a result of which we get a significant edge over the competition. >>Interesting. I mean, look, any I SV can go on any cloud and take advantage of it. And that's, uh I love it. We live in a new world. How about machine learning workloads? What? What did you see there in terms of performance and benchmarks? >>Right. So machine learning. We offer three capabilities training, which is fully automated, running in France and explanations. So one of the things which many of our customers told us coming from the enterprise is that explanations are very important to them because, uh, customers want to know that. Why did the the system, uh, choose a certain prediction? So we offer explanations for all models which have been derailed by. That's the first thing. Now, one of the interesting things about training is that training is usually the most expensive phase of machine learning. So we have spent a lot of time improving the performance of training. So we have a bunch of techniques which we have developed inside of Oracle to improve the training process. For instance, we have, uh, metal and proxy models, which really give us an advantage. We use adaptive sampling. We have, uh, invented in techniques for paralysing the hyper parameter search. So as a result of a lot of this work, our training is about 25 times faster than that ship them health and all the data is, uh, inside the database. All this processing is being done inside the database, so it's much faster. It is inside the database. And I want to point out that there is no additional charge for the history of customers because we're using the same cluster. You're not working in your service. So all of these machine learning capabilities are being offered at no additional charge inside the database and as a performance, which is significantly faster than that, >>are you taking advantage of or is there any, uh, need not need, but any advantage that you can get if two by exploiting things like gravity. John, we've talked about that a little bit in the past. Or trainee. Um, you just mentioned training so custom silicon that AWS is doing, you're taking advantage of that. Do you need to? Can you give us some insight >>there? So there are two things, right? We're always evaluating What are the choices we have from hybrid perspective? Obviously, for us to leverage is right and like all the things you mention about like we have considered them. But there are two things to consider. One is he is a memory system. So he favours a big is the dominant cost. The processor is a person of the cost, but memory is the dominant cost. So what we have evaluated and found is that the current shape which we are using is going to provide our customers with the best price performance. That's the first thing. The second thing is that there are opportunities at times when we can use a specialised processor for vaccinating the world for a bit. But then it becomes a matter of the cost of the customer. Advantage of our current architecture is on the same hardware. Customers are getting very good performance. Very good, energetic performance in a very good machine learning performance. If you will go with the specialised processor, it may. Actually, it's a machine learning, but then it's an additional cost with the customers we need to pay. So we are very sensitive to the customer's request, which is usually to provide very good performance at a very low cost. And we feel is that the current design we have as providing customers very good performance and very good price performance. >>So part of that is architectural. The memory intensive nature of of heat wave. The other is A W s pricing. If AWS pricing were to flip, it might make more sense for you to take advantage of something like like cranium. Okay, great. Thank you. And welcome back to the benchmarks benchmarks. Sometimes they're artificial right there. A car can go from 0 to 60 in two seconds. But I might not be able to experience that level of performance. Do you? Do you have any real world numbers from customers that have used my sequel Heatwave on A W s. And how they look at performance? >>Yes, absolutely so the my Secret service on the AWS. This has been in Vera for, like, since November, right? So we have a lot of customers who have tried the service. And what actually we have found is that many of these customers, um, planning to migrate from Aurora to my secret heat rape. And what they find is that the performance difference is actually much more pronounced than what I was talking about. Because with Aurora, the performance is actually much poorer compared to uh, like what I've talked about. So in some of these cases, the customers found improvement from 60 times, 240 times, right? So he travels 100 for 240 times faster. It was much less expensive. And the third thing, which is you know, a noteworthy is that customers don't need to change their applications. So if you ask the top three reasons why customers are migrating, it's because of this. No change to the application much faster, and it is cheaper. So in some cases, like Johnny Bites, what they found is that the performance of their applications for the complex storeys was about 60 to 90 times faster. Then we had 60 technologies. What they found is that the performance of heat we have compared to Aurora was 100 and 39 times faster. So, yes, we do have many such examples from real workloads from customers who have tried it. And all across what we find is if it offers better performance, lower cost and a single database such that it is compatible with all existing by sequel based applications and workloads. >>Really impressive. The analysts I talked to, they're all gaga over heatwave, and I can see why. Okay, last question. Maybe maybe two and one. Uh, what's next? In terms of new capabilities that customers are going to be able to leverage and any other clouds that you're thinking about? We talked about that upfront, but >>so in terms of the capabilities you have seen, like they have been, you know, non stop attending to the feedback from the customers in reacting to it. And also, we have been in a wedding like organically. So that's something which is gonna continue. So, yes, you can fully expect that people not dressed and continue to in a way and with respect to the other clouds. Yes, we are planning to support my sequel. He tripped on a show, and this is something that will be announced in the near future. Great. >>All right, Thank you. Really appreciate the the overview. Congratulations on the work. Really exciting news that you're moving my sequel Heatwave into other clouds. It's something that we've been expecting for some time. So it's great to see you guys, uh, making that move, and as always, great to have you on the Cube. >>Thank you for the opportunity. >>All right. And thank you for watching this special cube conversation. I'm Dave Volonte, and we'll see you next time.

Published Date : Sep 14 2022

SUMMARY :

The company is now in its fourth major release since the original announcement in December 2020. Very happy to be back. Now for those who might not have kept up with the news, uh, to kick things off, give us an overview of my So customers of my sequel then they had to run analytics or when they had to run machine So we've seen some interesting moves by Oracle lately. So one of the observations is that a very large percentage So was this a straightforward lifted shift from No, it is not because one of the design girls we have with my sequel, So I just want to make sure I understand that it's not like you just wrapped your stack in So for status, um, we have taken the mind sequel Heatwave code and we have optimised Can you help us understand that? So this let's leads to customers provisioning a shape which is So how do we quantify that? So that's the first thing that, So all the three workloads we That's apples to apples on A W s. And you have to obviously do some kind of So that's the first aspect And I mean, if you have to use two So the system keeps getting better as you learn more and What did you see there in terms of performance and benchmarks? So we have a bunch of techniques which we have developed inside of Oracle to improve the training need not need, but any advantage that you can get if two by exploiting We're always evaluating What are the choices we have So part of that is architectural. And the third thing, which is you know, a noteworthy is that In terms of new capabilities that customers are going to be able so in terms of the capabilities you have seen, like they have been, you know, non stop attending So it's great to see you guys, And thank you for watching this special cube conversation.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VolontePERSON

0.99+

December 2020DATE

0.99+

AmazonORGANIZATION

0.99+

JohnPERSON

0.99+

FranceLOCATION

0.99+

AWSORGANIZATION

0.99+

10 timesQUANTITY

0.99+

two thingsQUANTITY

0.99+

OracleORGANIZATION

0.99+

HeatwaveTITLE

0.99+

100QUANTITY

0.99+

60 timesQUANTITY

0.99+

one yearQUANTITY

0.99+

12 timesQUANTITY

0.99+

GWSORGANIZATION

0.99+

60 technologiesQUANTITY

0.99+

first partQUANTITY

0.99+

240 timesQUANTITY

0.99+

two separate licencesQUANTITY

0.99+

third categoryQUANTITY

0.99+

second advantageQUANTITY

0.99+

0QUANTITY

0.99+

seven timesQUANTITY

0.99+

two secondsQUANTITY

0.99+

twoQUANTITY

0.99+

AppleORGANIZATION

0.99+

seven timesQUANTITY

0.99+

GoogleORGANIZATION

0.99+

oneQUANTITY

0.99+

25 timesQUANTITY

0.99+

second pointQUANTITY

0.99+

NovemberDATE

0.99+

85 patentsQUANTITY

0.99+

second thingQUANTITY

0.99+

AuroraTITLE

0.99+

third thingQUANTITY

0.99+

EachQUANTITY

0.99+

second exampleQUANTITY

0.99+

10 gigabytesQUANTITY

0.99+

three thingsQUANTITY

0.99+

OneQUANTITY

0.99+

two benefitsQUANTITY

0.99+

one aspectQUANTITY

0.99+

first aspectQUANTITY

0.98+

two separate databasesQUANTITY

0.98+

over 10 yearsQUANTITY

0.98+

fourth major releaseQUANTITY

0.98+

39 timesQUANTITY

0.98+

first thingQUANTITY

0.98+

Heat WaveTITLE

0.98+

Digging into HeatWave ML Performance


 

(upbeat music) >> Hello everyone. This is Dave Vellante. We're diving into the deep end with AMD and Oracle on the topic of mySQL HeatWave performance. And we want to explore the important issues around machine learning. As applications become more data intensive and machine intelligence continues to evolve, workloads increasingly are seeing a major shift where data and AI are being infused into applications. And having a database that simplifies the convergence of transaction and analytics data without the need to context, switch and move data out of and into different data stores. And eliminating the need to perform extensive ETL operations is becoming an industry trend that customers are demanding. At the same time, workloads are becoming more automated and intelligent. And to explore these issues further, we're happy to have back in theCUBE Nipun Agarwal, who's the Senior Vice President of mySQL HeatWave and Kumaran Siva, who's the Corporate Vice President Strategic Business Development at AMD. Gents, hello again. Welcome back. >> Hello. Hi Dave. >> Thank you, Dave. >> Okay. Nipun, obviously machine learning has become a must have for analytics offerings. It's integrated into mySQL HeatWave. Why did you take this approach and not the specialized database approach as many competitors do right tool for the right job? >> Right? So, there are a lot of customers of mySQL who have the need to run machine learning on the data which is store in mySQL database. So in the past, customers would need to extract the data out of mySQL and they would take it to a specialized service for running machine learning. Now, the reason we decided to incorporate machine learning inside the database, there are multiple reasons. One, customers don't need to move the data. And if they don't need to move the data, it is more secure because it's protected by the same access controlled mechanisms as rest of the data There is no need for customers to manage multiple services. But in addition to that, when we run the machine learning inside the database customers are able to leverage the same service the same hardware, which has been provisioned for OTP analytics and use machine learning capabilities at no additional charge. So from a customer's perspective, they get the benefits that it is a single database. They don't need to manage multiple services. And it is offered at no additional charge. And then as another aspect, which is kind of hard to learn which is based on the IP, the work we have done it is also significantly faster than what customers would get by having a separate service. >> Just to follow up on that. How are you seeing customers use HeatWaves machine learning capabilities today? How is that evolving? >> Right. So one of the things which, you know customers very often want to do is to train their models based on the data. Now, one of the things is that data in a database or in a transaction database changes quite rapidly. So we have introduced support for auto machine learning as a part of HeatWave ML. And what it does is that it fully automates the process of training. And this is something which is very important to database users, very important to mySQL users that they don't really want to hire or data scientists or specialists for doing training. So that's the first part that training in HeatWave ML is fully automated. Doesn't require the user to provide any like specific parameters, just the source data and the task which they want to train. The second aspect is the training is really fast. So the training is really fast. The benefit is that customers can retrain quite often. They can make sure that the model is up to date with any changes which have been made to their transaction database. And as a result of the models being up to date, the accuracy of the prediction is high. Right? So that's the first aspect, which is training. The second aspect is inference, which customers run once they have the models trained. And the third thing, which is perhaps been the most sought after request from the mySQL customers is the ability to provide explanations. So, HeatWave ML provides explanations for any model which has been generated or trained by HeatWave ML. So these are the three capabilities- training, inference and explanations. And this whole process is completely automated, doesn't require a specialist or a data scientist. >> Yeah, that's nice. I mean, training obviously very popular today. I've said inference I think is going to explode in the coming decade. And then of course, AI explainable AI is a very important issue. Kumaran, what are the relevant capabilities of the AMD chips that are used in OCI to support HeatWave ML? Are they different from say the specs for HeatWave in general? >> So, actually they aren't. And this is one of the key features of this architecture or this implementation that is really exciting. Um, there with HeatWave ML, you're using the same CPU. And by the way, it's not a GPU, it's a CPU for both for all three of the functions that Nipun just talked about- inference, training and explanation all done on CPU. You know, bigger picture with the capabilities we bring here we're really providing a balance, you know between the CPU cores, memory and the networking. And what that allows you to do here is be able to feed the CPU cores appropriately. And within the cores, we have these AVX instruc... extensions in with the Zen 2 and Zen 3 cores. We had AVX 2, and then with the Zen 4 core coming out we're going to have AVX 512. But we were able to with that balance of being able to bring in the data and utilize the high memory bandwidth and then use the computation to its maximum we're able to provide, you know, build pride enough AI processing that we are able to get the job done. And then we're built to build a fit into that larger pipeline that that we build out here with the HeatWave. >> Got it. Nipun you know, you and I every time we have a conversation we've got to talk benchmarks. So you've done machine learning benchmarks with HeatWave. You might even be the first in the industry to publish you know, transparent, open ML benchmarks on GitHub. I mean, I, I wouldn't know for sure but I've not seen that as common. Can you describe the benchmarks and the data sets that you used here? >> Sure. So what we did was we took a bunch of open data sets for two categories of tasks- classification and regression. So we took about a dozen data sets for classification and about six for regression. So to give an example, the kind of data sets we used for classifications like the airlines data set, hex sensors bank, right? So these are open data sets. And what we did was for on these data sets we did a comparison of what would it take to train using HeatWave ML? And then the other service we compared with is that RedShift ML. So, there were two observations. One is that with HeatWave ML, the user does not need to provide any tuning parameters, right? The HeatWave ML using RML fully generates a train model, figures out what are the right algorithms? What are the right features? What are the right hyper parameters and sets, right? So no need for any manual intervention not so the case with Redshift ML. The second thing is the performance, right? So the performance of HeatWave ML aggregate on these 12 data sets for classification and the six data sets on regression. On an average, it is 25 times faster than Redshift ML. And note that Redshift ML in turn involves SageMaker, right? So on an average, HeatWave ML provides 25 times better performance for training. And the other point to note is that there is no need for any human intervention. That's fully automated. But in the case of Redshift ML, many of these data sets did not even complete in the set duration. If you look at price performance, one of the things again I want to highlight is because of the fact that AMD does pretty well in all kinds of workloads. We are able to use the same cluster users and use the same cluster for analytics, for OTP or for machine learning. So there is no additional cost for customers to run HeatWave ML if they have provision HeatWave. But assuming a user is provisioning a HeatWave cluster only to run HeatWave ML, right? That's the case, even in that case the price performance advantage of HeatWave ML over Redshift ML is 97 times, right? So 25 times faster at 1% of the cost compared to Redshift ML And all these scripts and all this information is available on GitHub for customers to try to modify and like, see, like what are the advantages they would get on their workloads? >> Every time I hear these numbers, I shake my head. I mean, they're just so overwhelming. Um, and so we'll see how the competition responds when, and if they respond. So, but thank you for sharing those results. Kumaran, can you elaborate on how the specs that you talked about earlier contribute to HeatWave ML's you know, benchmark results. I'm particularly interested in scalability, you know Typically things degrade as you push the system harder. What are you seeing? >> No, I think, I think it's good. Look, yeah. That's by those numbers, just blow me, blow my head too. That's crazy good performance. So look from, from an AMD perspective, we have really built an architecture. Like if you think about the chiplet architecture to begin with, it is fundamentally, you know, it's kind of scaling by design, right? And, and one of the things that we've done here is been able to work with, with the HeatWave team and heat well ML team, and then been able to, to within within the CPU package itself, be able to scale up to take very efficient use of all of the course. And then of course, work with them on how you go between nodes. So you can have these very large systems that can run ML very, very efficiently. So it's really, you know, building on the building blocks of the chiplet architecture and how scaling happens there. >> Yeah. So it's you're saying it's near linear scaling or essentially. >> So, let Nipun comment on that. >> Yeah. >> Is it... So, how about as cluster sizes grow, Nipun? >> Right. >> What happens there? >> So one of the design points for HeatWave is scale out architecture, right? So as you said, that as we add more data set or increase the size of the data, or we add the number of nodes to the cluster, we want the performance to scale. So we show that we have near linear scale factor, or nearly near scale scalability for SQL workloads in the case of HeatWave ML, as well. As users add more nodes to the cluster so the size of the cluster the performance of HeatWave ML improves. So I was giving you this example that HeatWave ML is 25 times faster compared to Redshift ML. Well, that was on a cluster size of two. If you increase the cluster size of HeatWave ML to a larger number. But I think the number is 16. The performance advantage over Redshift ML increases from 25 times faster to 45 times faster. So what that means is that on a cluster size of 16 nodes HeatWave ML is 45 times faster for training these again, dozen data sets. So this shows that HeatWave ML skills better than the computation. >> So you're saying adding nodes offsets any management complexity that you would think of as getting in the way. Is that right? >> Right. So one is the management complexity and which is why by features like last customers can scale up or scale down, you know, very easily. The second aspect is, okay What gives us this advantage, right, of scalability? Or how are we able to scale? Now, the techniques which we use for HeatWave ML scalability are a bit different from what we use for SQL processing. So in the case of HeatWave ML, they really like, you know, three, two trade offs which we have to be careful about. One is the accuracy. Because we want to provide better performance for machine learning without compromising on the accuracy. So accuracy would require like more synchronization if you have multiple threads. But if you have too much of synchronization that can slow down the degree of patterns that we get. Right? So we have to strike a fine balance. So what we do is that in HeatWave ML, there are different phases of training, like algorithm selection, feature selection, hyper probability training. Each of these phases is analyzed. And for instance, one of the ways techniques we use is that if you're trying to figure out what's the optimal hyper parameter to be used? We start up with the search space. And then each of the VMs gets a part of the search space. And then we synchronize only when needed, right? So these are some of the techniques which we have developed over the years. And there are actually paper's filed, research publications filed on this. And this is what we do to achieve good scalability. And what that results to the customer is that if they have some amount of training time and they want to make it better they can just provision a larger cluster and they will get better performance. >> Got it. Thank you. Kumaran, when I think of machine learning, machine intelligence, AI, I think GPU but you're not using GPU. So how are you able to get this type of performance or price performance without using GPU's? >> Yeah, definitely. So yeah, that's a good point. And you think about what is going on here and you consider the whole pipeline that Nipun has just described in terms of how you get you know, your training, your algorithms And using the mySQL pieces of it to get to the point where the AI can be effective. In that process what happens is you have to have a lot of memory to transactions. A lot of memory bandwidth comes into play. And then bringing all that data together, feeding the actual complex that does the AI calculations that in itself could be the bottleneck, right? And you can have multiple bottlenecks along the way. And I think what you see in the AMD architecture for epic for this use case is the balance. And the fact that you are able to do the pre-processing, the AI, and then the post-processing all kind of seamlessly together, that has a huge value. And that goes back to what Nipun was saying about using the same infrastructure, gets you the better TCO but it also gets you gets you better performance. And that's because of the fact that you're bringing the data to the computation. So the computation in this case is not strictly the bottleneck. It's really about how you pull together what you need and to do the AI computation. And that is, that's probably a more, you know, it's a common case. And so, you know, you're going to start I think the least start to see this especially for inference applications. But in this case we're doing both inference explanation and training. All using the the CPU in the same OCI infrastructure. >> Interesting. Now Nipun, is the secret sauce for HeatWave ML performance different than what we've discussed before you and I with with HeatWave generally? Is there some, you know, additive engine additive that you're putting in? >> Right? Yes. The secret sauce is indeed different, right? Just the way I was saying that for SQL processing. The reason we get very good performance and price performance is because we have come up with new algorithms which help the SQL process can scale out. Similarly for HeatWave ML, we have come up with new IP, new like algorithms. One example is that we use meta-learn proxy models, right? That's the technique we use for automating the training process, right? So think of this meta-learn proxy models to be like, you know using machine learning for machine learning training. And this is an IP which we developed. And again, we have published the results and the techniques. But having such kind of like techniques is what gives us a better performance. Similarly, another thing which we use is adaptive sampling that you can have a large data set. But we intelligently sample to figure out that how can we train on a small subset without compromising on the accuracy? So, yes, there are many techniques that you have developed specifically for machine learning which is what gives us the better performance, better price performance, and also better scalability. >> What about mySQL autopilot? Is there anything that differs from HeatWave ML that is relevant? >> Okay. Interesting you should ask. So mySQL Autopilot is think of it to be an application using machine learning. So mySQL Autopilot uses machine learning to automate various aspects of the database service. So for instance, if you want to figure out that what's the right partitioning scheme to partition the data in memory? We use machine learning techniques to figure out that what's the right, the best column based on the user's workload to partition the data in memory Or given a workload, if you want to figure out what is the right cluster size to provision? That's something we use mySQL autopilot for. And I want to highlight that we don't aware of any other database service which provides this level of machine learning based automation which customers get with mySQL Autopilot. >> Hmm. Interesting. Okay. Last question for both of you. What are you guys working on next? What can customers expect from this collaboration specifically in this space? Maybe Nipun, you can start and then Kamaran can bring us home. >> Sure. So there are two things we are working on. One is based on the feedback we have gotten from customers, we are going to keep making the machine learning capabilities richer in HeatWave ML. That's one dimension. And the second thing is which Kamaran was alluding to earlier, We are looking at the next generation of like processes coming from AMD. And we will be seeing as to how we can more benefit from these processes whether it's the size of the L3 cache, the memory bandwidth, the network bandwidth, and such or the newer effects. And make sure that we leverage the all the greatness which the new generation of processes will offer. >> It's like an engineering playground. Kumaran, let's give you the final word. >> No, that's great. Now look with the Zen 4 CPU cores, we're also bringing in AVX 512 instruction capability. Now our implementation is a little different. It was in, in Rome and Milan, too where we use a double pump implementation. What that means is, you know, we take two cycles to do these instructions. But the key thing there is we don't lower our speed of the CPU. So there's no noisy neighbor effects. And it's something that OCI and the HeatWave has taken full advantage of. And so like, as we go out in time and we see the Zen 4 core, we can... we see up to 96 CPUs that that's going to work really well. So we're collaborating closely with, with OCI and with the HeatWave team here to make sure that we can take advantage of that. And we're also going to upgrade the memory subsystem to get to 12 channels of DDR 5. So it should be, you know there should be a fairly significant boost in absolute performance. But more important or just as importantly in TCO value for the customers, the end customers who are going to adopt this great service. >> I love their relentless innovation guys. Thanks so much for your time. We're going to have to leave it there. Appreciate it. >> Thank you, David. >> Thank you, David. >> Okay. Thank you for watching this special presentation on theCUBE. Your leader in enterprise and emerging tech coverage.

Published Date : Sep 14 2022

SUMMARY :

And eliminating the need and not the specialized database approach So in the past, customers How are you seeing customers use So one of the things of the AMD chips that are used in OCI And by the way, it's not and the data sets that you used here? And the other point to note elaborate on how the specs And, and one of the things or essentially. So, how about as So one of the design complexity that you would So in the case of HeatWave ML, So how are you able to get And the fact that you are Nipun, is the secret sauce That's the technique we use for automating of the database service. What are you guys working on next? And the second thing is which Kamaran Kumaran, let's give you the final word. OCI and the HeatWave We're going to have to leave it there. and emerging tech coverage.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

RomeLOCATION

0.99+

DavePERSON

0.99+

DavidPERSON

0.99+

OCIORGANIZATION

0.99+

Nipun AgarwalPERSON

0.99+

MilanLOCATION

0.99+

45 timesQUANTITY

0.99+

25 timesQUANTITY

0.99+

12 channelsQUANTITY

0.99+

OracleORGANIZATION

0.99+

AMDORGANIZATION

0.99+

Zen 4COMMERCIAL_ITEM

0.99+

KumaranPERSON

0.99+

HeatWaveORGANIZATION

0.99+

Zen 3COMMERCIAL_ITEM

0.99+

second aspectQUANTITY

0.99+

Kumaran SivaPERSON

0.99+

12 data setsQUANTITY

0.99+

first aspectQUANTITY

0.99+

97 timesQUANTITY

0.99+

Zen 2COMMERCIAL_ITEM

0.99+

bothQUANTITY

0.99+

firstQUANTITY

0.99+

OneQUANTITY

0.99+

two thingsQUANTITY

0.99+

oneQUANTITY

0.99+

EachQUANTITY

0.99+

1%QUANTITY

0.99+

two cyclesQUANTITY

0.99+

three capabilitiesQUANTITY

0.99+

third thingQUANTITY

0.99+

eachQUANTITY

0.99+

AVX 2COMMERCIAL_ITEM

0.99+

AVX 512COMMERCIAL_ITEM

0.99+

second thingQUANTITY

0.99+

Redshift MLTITLE

0.99+

six data setsQUANTITY

0.98+

HeatWaveTITLE

0.98+

mySQL AutopilotTITLE

0.98+

twoQUANTITY

0.98+

NipunPERSON

0.98+

two categoriesQUANTITY

0.98+

mySQLTITLE

0.98+

two observationsQUANTITY

0.98+

first partQUANTITY

0.98+

mySQL autopilotTITLE

0.98+

threeQUANTITY

0.97+

SQLTITLE

0.97+

One exampleQUANTITY

0.97+

single databaseQUANTITY

0.95+

16QUANTITY

0.95+

todayDATE

0.95+

about sixQUANTITY

0.95+

HeatWavesORGANIZATION

0.94+

about a dozen data setsQUANTITY

0.94+

16 nodesQUANTITY

0.93+

mySQL HeatWaveTITLE

0.93+

AMD Oracle Partnership Elevates MySQLHeatwave


 

(upbeat music) >> For those of you who've been following the cloud database space, you know that MySQL HeatWave has been on a technology tear over the last 24 months with Oracle claiming record breaking benchmarks relative to other database platforms. So far, those benchmarks remain industry leading as competitors have chosen not to respond, perhaps because they don't feel the need to, or maybe they don't feel that doing so would serve their interest. Regardless, the HeatWave team at Oracle has been very aggressive about its performance claims, making lots of noise, challenging the competition to respond, publishing their scripts to GitHub. But so far, there are no takers, but customers seem to be picking up on these moves by Oracle and it's likely the performance numbers resonate with them. Now, the other area we want to explore, which we haven't thus far, is the engine behind HeatWave and that is AMD. AMD's epic processors have been the powerhouse on OCI, running MySQL HeatWave since day one. And today we're going to explore how these two technology companies are working together to deliver these performance gains and some compelling TCO metrics. In fact, a recent Wikibon analysis from senior analyst Marc Staimer made some TCO comparisons in OLAP workloads relative to AWS, Snowflake, GCP, and Azure databases, you can find that research on wikibon.com. And with that, let me introduce today's guest, Nipun Agarwal senior vice president of MySQL HeatWave and Kumaran Siva, who's the corporate vice president for strategic business development at AMD. Welcome to theCUBE gentlemen. >> Welcome. Thank you. >> Thank you, Dave. >> Hey Nipun, you and I have talked a lot about this. You've been on theCUBE a number of times talking about MySQL HeatWave. But for viewers who may not have seen those episodes maybe you could give us an overview of HeatWave and how it's different from competitive cloud database offerings. >> Sure. So MySQL HeatWave is a fully managed MySQL database service offering from Oracle. It's a single database, which can be used to run transactional processing, analytics and machine learning workloads. So, in the past, MySQL has been designed and optimized for transaction processing. So customers of MySQL when they had to run, analytics machine learning, would need to extract the data out of MySQL, into some other database or service, to run analytics or machine learning. MySQL HeatWave offers a single database for running all kinds of workloads so customers don't need to extract data into some of the database. In addition to having a single database, MySQL HeatWave is also very performant compared to one up databases and also it is very price competitive. So the advantages are; single database, very performant, and very good price performance. >> Yes. And you've published some pretty impressive price performance numbers against competitors. Maybe you could describe those benchmarks and highlight some of the results, please. >> Sure. So one thing to notice that the performance of any database is going to like vary, the performance advantage is going to vary based on, the size of the data and the specific workloads, so the mileage varies, that's the first thing to know. So what we have done is, we have published multiple benchmarks. So we have benchmarks on PPCH or PPCDS and we have benchmarks on different data sizes because based on the customer's workload, the mileage is going to vary, so we want to give customers a broad range of comparisons so that they can decide for themselves. So in a specific case, where we are running on a 30 terabyte PPCH workload, HeatWave is about 18 times better price performance compared to Redshift. 18 times better compared to Redshift, about 33 times better price performance, compared to Snowflake, and 42 times better price performance compared to Google BigQuery. So, this is on 30 Terabyte PPCH. Now, if the data size is different, or the workload is different, the characteristics may vary slightly but this is just to give a flavor of the kind of performance advantage MySQL HeatWave offers. >> And then my last question before we bring in Kumaran. We've talked about the secret sauce being the tight integration between hardware and software, but would you add anything to that? What is that secret sauce in HeatWave that enables you to achieve these performance results and what does it mean for customers? >> So there are three parts to this. One is HeatWave has been designed with a scale out architecture in mind. So we have invented and implemented new algorithms for skill out query processing for analytics. The second aspect is that HeatWave has been really optimized for cloud, commodity cloud, and that's where AMD comes in. So for instance, many of the partitioning schemes we have for processing HeatWave, we optimize them for the L3 cache of the AMD processor. The thing which is very important to our customers is not just the sheer performance but the price performance, and that's where we have had a very good partnership with AMD because not only does AMD help us provide very good performance, but the price performance, right? And that all these numbers which I was showing, big part of it is because we are running on AMD which provides very good price performance. So that's the second aspect. And the third aspect is, MySQL autopilot, which provides machine learning based automation. So it's really these three things, a combination of new algorithms, design for scale out query processing, optimized for commodity cloud hardware, specifically AMD processors, and third, MySQL auto pilot which gives us this performance advantage. >> Great, thank you. So that's a good segue for AMD and Kumaran. So Kumaran, what is AMD bringing to the table? What are the, like, for instance, relevance specs of the chips that are used in Oracle cloud infrastructure and what makes them unique? >> Yeah, thanks Dave. That's a good question. So, OCI is a great customer of ours. They use what we call the top of stack devices meaning that they have the highest core count and they also are very, very fast cores. So these are currently Zen 3 cores. I think the HeatWave product is right now deployed on Zen 2 but will shortly be also on the Zen 3 core as well. But we provide in the case of OCI 64 cores. So that's the largest devices that we build. What actually happens is, because these large number of CPUs in a single package and therefore increasing the density of the node, you end up with this fantastic TCO equation and the cost per performance, the cost per for deployed services like HeatWave actually ends up being extraordinarily competitive and that's a big part of the contribution that we're bringing in here. >> So Zen 3 is the AMD micro architecture which you introduced, I think in 2017, and it's the basis for EPIC, which is sort of the enterprise grade that you really attacked the enterprise with. Maybe you could elaborate a little bit, double click on how your chips contribute specifically to HeatWave's, price performance results. >> Yeah, absolutely. So in the case of HeatWave, so as Nipun alluded to, we have very large L3 caches, right? So in our very, very top end parts just like the Milan X devices, we can go all the way up to like 768 megabytes of L3 cache. And that gives you just enormous performance and performance gains. And that's part of what we're seeing with HeatWave today and that not that they're currently on the second generation ROM based product, 'cause it's a 7,002 based product line running with the 64 cores. But as time goes on, they'll be adopting the next generation Milan as well. And the other part of it too is, as our chip led architecture has evolved, we know, so from the first generation Naples way back in 2017, we went from having multiple memory domains and a sort of NUMA architecture at the time, today we've really optimized that architecture. We use a common I/O Die that has all of the memory channels attached to it. And what that means is that, these scale out applications like HeatWave, are able to really scale very efficiently as they go from a small domain of CPUs to, for example the entire chip, all 64 cores that scaling, is been a key focus for AMD and being able to design and build architectures that can take advantage of that and then have applications like HeatWave that scale so well on it, has been, a key aim of ours. >> And Gen 3 moving up the Italian countryside. Nipun, you've taken the somewhat unusual step of posting the benchmark parameters, making them public on GitHub. Now, HeatWave is relatively new. So people felt that when Oracle gained ownership of MySQL it would let it wilt on the vine in favor of Oracle database, so you lost some ground and now, you're getting very aggressive with HeatWave. What's the reason for publishing those benchmark parameters on GitHub? >> So, the main reason for us to publish price performance numbers for HeatWave is to communicate to our customers a sense of what are the benefits they're going to get when they use HeatWave. But we want to be very transparent because as I said the performance advantages for the customers may vary, based on the data size, based on the specific workloads. So one of the reasons for us to publish, all these scripts on GitHub is for transparency. So we want customers to take a look at the scripts, know what we have done, and be confident that we stand by the numbers which we are publishing, and they're very welcome, to try these numbers themselves. In fact, we have had customers who have downloaded the scripts from GitHub and run them on our service to kind of validate. The second aspect is in some cases, they may be some deviations from what we are publishing versus what the customer would like to run in the production deployments so it provides an easy way, for customers to take the scripts, modify them in some ways which may suit their real world scenario and run to see what the performance advantages are. So that's the main reason, first, is transparency, so the customers can see what we are doing, because of the comparison, and B, if they want to modify it to suit their needs, and then see what is the performance of HeatWave, they're very welcome to do so. >> So have customers done that? Have they taken the benchmarks? And I mean, if I were a competitor, honestly, I wouldn't get into that food fight because of the impressive performance, but unless I had to, I mean, have customers picked up on that, Nipun? >> Absolutely. In fact, we have had many customers who have benchmarked the performance of MySQL HeatWave, with other services. And the fact that the scripts are available, gives them a very good starting point, and then they've also tweaked those queries in some cases, to see what the Delta would be. And in some cases, customers got back to us saying, hey the performance advantage of HeatWave is actually slightly higher than what was published and what is the reason. And the reason was, when the customers were trying, they were trying on the latest version of the service, and our benchmark results were posted let's say, two months back. So the service had improved in those two to three months and customers actually saw better performance. So yes, absolutely. We have seen customers download the scripts, try them and also modify them to some extent and then do the comparison of HeatWave with other services. >> Interesting. Maybe a question for both of you how is the competition responding to this? They haven't said, "Hey, we're going to come up "with our own benchmarks." Which is very common, you oftentimes see that. Although, for instance, Snowflake hasn't responded to data bricks, so that's not their game, but if the customers are actually, putting a lot of faith in the benchmarks and actually using that for buying decisions, then it's inevitable. But how have you seen the competition respond to the MySQL HeatWave and AMD combo? >> So maybe I can take the first track from the database service standpoint. When customers have more choice, it is invariably advantages for the customer because then the competition is going to react, right? So the way we have seen the reaction is that we do believe, that the other database services are going to take a closer eye to the price performance, right? Because if you're offering such good price performance, the vendors are already looking at it. And, you know, instances where they have offered let's say discount to the customers, to kind of at least like close the gap to some extent. And the second thing would be in terms of the capability. So like one of the things which I should have mentioned even early on, is that not only does MySQL HeatWave on AMD, provide very good price performance, say on like a small cluster, but it's all the way up to a cluster size of 64 nodes, which has about 1000 cores. So the point is, that HeatWave performs very well, both on a small system, as well as a huge scale out. And this is again, one of those things which is a differentiation compared to other services so we expect that even other database services will have to improve their offerings to provide the same good scale factor, which customers are now starting to expectancy, with MySQL HeatWave. >> Kumaran, anything you'd add to that? I mean, you guys are an arms dealer, you love all your OEMs, but at the same time, you've got chip competitors, Silicon competitors. How do you see the competitive-- >> I'd say the broader answer and the big picture for AMD, we're very maniacally focused on our customers, right? And OCI and Oracle are huge and important customers for us, and this particular use cases is extremely interesting both in that it takes advantage, very well of our architecture and it pulls out some of the value that AMD bring. I think from a big picture standpoint, our aim is to execute, to build to bring out generations of CPUs, kind of, you know, do what we say and say, sorry, say what we do and do what we say. And from that point of view, we're hitting, the schedules that we say, and being able to bring out the latest technology and bring it in a TCO value proposition that generationally keeps OCI and HeatWave ahead. That's the crux of our partnership here. >> Yeah, the execution's been obvious for the last several years. Kumaran, staying with you, how would you characterize the collaboration between, the AMD engineers and the HeatWave engineering team? How do you guys work together? >> No, I'd say we're in a very, very deep collaboration. So, there's a few aspects where, we've actually been working together very closely on the code and being able to optimize for both the large L3 cache that AMD has, and so to be able to take advantage of that. And then also, to be able to take advantage of the scaling. So going between, you know, our architecture is chip like based, so we have these, the CPU cores on, we call 'em CCDs and the inter CCD communication, there's opportunities to optimize an application level and that's something we've been engaged with. In the broader engagement, we are going back now for multiple generations with OCI, and there's a lot of input that now, kind of resonates in the product line itself. And so we value this very close collaboration with HeatWave and OCI. >> Yeah, and the cadence, Nip, and you and I have talked about this quite a bit. The cadence has been quite rapid. It's like this constant cycle every couple of months I turn around, is something new on HeatWave. But for question again, for both of you, what new things do you think that organizations, customers, are going to be able to do with MySQL HeatWave if you could look out next 12 to 18 months, is there anything you can share at this time about future collaborations? >> Right, look, 12 to 18 months is a long time. There's going to be a lot of innovation, a lot of new capabilities coming out on in MySQL HeatWave. But even based on what we are currently offering, and the trend we are seeing is that customers are bringing, more classes of workloads. So we started off with OLTP for MySQL, then it went to analytics. Then we increased it to mixed workloads, and now we offer like machine learning as alike. So one is we are seeing, more and more classes of workloads come to MySQL HeatWave. And the second is a scale, that kind of data volumes people are using HeatWave for, to process these mixed workloads, analytics machine learning OLTP, that's increasing. Now, along the way we are making it simpler to use, we are making it more cost effective use. So for instance, last time, when we talked about, we had introduced this real time elasticity and that's something which is a very, very popular feature because customers want the ability to be able to scale out, or scale down very efficiently. That's something we provided. We provided support for compression. So all of these capabilities are making it more efficient for customers to run a larger part of their workloads on MySQL HeatWave, and we will continue to make it richer in the next 12 to 18 months. >> Thank you. Kumaran, anything you'd add to that, we'll give you the last word as we got to wrap it. >> No, absolutely. So, you know, next 12 to 18 months we will have our Zen 4 CPUs out. So this could potentially go into the next generation of the OCI infrastructure. This would be with the Genoa and then Bergamo CPUs taking us to 96 and 128 cores with 12 channels at DDR five. This capability, you know, when applied to an application like HeatWave, you can see that it'll open up another order of magnitude potentially of use cases, right? And we're excited to see what customers can do do with that. It certainly will make, kind of the, this service, and the cloud in general, that this cloud migration, I think even more attractive. So we're pretty excited to see how things evolve in this period of time. >> Yeah, the innovations are coming together. Guys, thanks so much, we got to leave it there really appreciate your time. >> Thank you. >> All right, and thank you for watching this special Cube conversation, this is Dave Vellante, and we'll see you next time. (soft calm music)

Published Date : Sep 14 2022

SUMMARY :

and it's likely the performance Thank you. and how it's different from So the advantages are; single and highlight some of the results, please. the first thing to know. We've talked about the secret sauce So for instance, many of the relevance specs of the chips that are used and that's a big part of the contribution and it's the basis for EPIC, So in the case of HeatWave, of posting the benchmark parameters, So one of the reasons for us to publish, So the service had improved how is the competition responding to this? So the way we have seen the but at the same time, and the big picture for AMD, for the last several years. and so to be able to Yeah, and the cadence, and the trend we are seeing is we'll give you the last and the cloud in general, Yeah, the innovations we'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Marc StaimerPERSON

0.99+

Dave VellantePERSON

0.99+

NipunPERSON

0.99+

OracleORGANIZATION

0.99+

2017DATE

0.99+

DavePERSON

0.99+

OCIORGANIZATION

0.99+

Zen 3COMMERCIAL_ITEM

0.99+

7,002QUANTITY

0.99+

KumaranPERSON

0.99+

second aspectQUANTITY

0.99+

Nipun AgarwalPERSON

0.99+

AMDORGANIZATION

0.99+

12QUANTITY

0.99+

64 coresQUANTITY

0.99+

768 megabytesQUANTITY

0.99+

twoQUANTITY

0.99+

MySQLTITLE

0.99+

third aspectQUANTITY

0.99+

12 channelsQUANTITY

0.99+

Kumaran SivaPERSON

0.99+

HeatWaveORGANIZATION

0.99+

96QUANTITY

0.99+

18 timesQUANTITY

0.99+

BergamoORGANIZATION

0.99+

three partsQUANTITY

0.99+

DeltaORGANIZATION

0.99+

three monthsQUANTITY

0.99+

MySQL HeatWaveTITLE

0.99+

42 timesQUANTITY

0.99+

bothQUANTITY

0.99+

18 monthsQUANTITY

0.99+

Zen 2COMMERCIAL_ITEM

0.99+

oneQUANTITY

0.99+

GitHubORGANIZATION

0.99+

OneQUANTITY

0.98+

second generationQUANTITY

0.98+

single databaseQUANTITY

0.98+

128 coresQUANTITY

0.98+

18 monthsQUANTITY

0.98+

three thingsQUANTITY

0.98+

SC22 Karan Batta, Kris Rice


 

>> Welcome back to Supercloud22, #Supercloud22. This is Dave Vellante. In 2019 Oracle and Microsoft announced a collaboration to bring interoperability between OCI, Oracle Cloud Infrastructure and Azure Clouds. It was Oracle's initial foray into so-called multi-cloud and we're joined by Karan Batta, who's the Vice President for Product Management at OCI. And Kris Rice is the Vice President of Software Development at Oracle Database. And we're going to talk about how this technology's evolving and whether it fits our view of what we call supercloud. Welcome gentlemen, thank you. >> Thanks for having us. >> So you recently just last month announced the new service. It extends on the initial partnership with Microsoft Oracle interconnect with Azure, and you refer to this as a secure private link between the two clouds, it cross 11 regions around the world, under two milliseconds data transmission sounds pretty cool. It enables customers to run Microsoft applications against data stored in Oracle databases without any loss in efficiency or presumably performance. So we use this term supercloud to describe a service or sets of services built on hyper scale infrastructure that leverages the core primitives and APIs of an individual cloud platform, but abstracts that underlying complexity to create a continuous experience across more than one cloud. Is that what you've done? >> Absolutely. I think it starts at the top layer in terms of just making things very simple for the customer, right. I think at the end of the day we want to enable true workloads running across two different clouds where you're potentially running maybe the app layer in one and the database layer or the back in another. And the integration I think starts with, you know, making it ease of use. Right. So you can start with things like, okay can you log into your second or your third cloud with the first cloud provider's credentials? Can you make calls against another cloud using another cloud's APIs? Can you peer the networks together? Can you make it seamless? I think those are all the components that are sort of, they're kind of the ingredients to making a multi-cloud or supercloud experience successful. >> Oh, thank you for that, Karan. So I guess there's a question for Chris is I'm trying to understand what you're really solving for? What specific customer problems are you focused on? What's the service optimized for presumably it's database but maybe you could double click on that. >> Sure. So, I mean, of course it's database. So it's a super fast network so that we can split the workload across two different clouds leveraging the best from both, but above the networking, what we had to do do is we had to think about what a true multi-cloud or what you're calling supercloud experience would be it's more than just making the network bites flow. So what we did is we took a look as Karan hinted at right, is where is my identity? Where is my observability? How do I connect these things across how it feels native to that other cloud? >> So what kind of engineering do you have to do to make that work? It's not just plugging stuff together. Maybe you could explain a little bit more detail, the the resources that you had to bring to bear and the technology behind the architecture. >> Sure. I think, it starts with actually, what our goal was, right? Our goal was to actually provide customers with a fully managed experience. What that means is we had to basically create a brand new service. So, we have obviously an Azure like portal and an experience that allows customers to do this but under the covers, we actually have a fully managed service that manages the networking layer, the physical infrastructure, and it actually calls APIs on both sides of the fence. It actually manages your Azure resources, creates them but it also interacts with OCI at the same time. And under the covers this service actually takes Azure primitives as inputs. And then it sort of like essentially translates them to OCI action. So, we actually truly integrated this as a service that's essentially built as a PaaS layer on top of these two clouds. >> So, the customer doesn't really care or know maybe they know cuz they might be coming through, an Azure experience, but you can run work on either Azure and or OCI. And it's a common experience across those clouds. Is that correct? >> That's correct. So like you said, the customer does know that they know there is a relationship with both clouds but thanks to all the things we built there's this thing we invented we created called a multi-cloud control plane. This control plane does operate against both clouds at the same time to make it as seamless as possible so that maybe they don't notice, you know, the power of the interconnect is extremely fast networking, as fast as what we could see inside a single cloud. If you think about how big a data center might be from edge to edge in that cloud, going across the interconnect makes it so that that workload is not important that it's spanning two clouds anymore. >> So you say extremely fast networking. I remember I used to, I wrote a piece a long time ago. Larry Ellison loves InfiniBand. I presume we've moved on from them, but maybe not. What is that interconnect? >> Yeah, so it's funny you mentioned interconnect you know, my previous history comes from Edge PC where we actually inside OCI today, we've moved from Infinite Band as is part of Exadata's core to what we call Rocky V two. So that's just another RDMA network. We actually use it very successfully, not just for Exadata but we use it for our standard computers that we provide to high performance computing customers. >> And the multi-cloud control plane runs. Where does that live? Does it live on OCI? Does it live on Azure? Yes? >> So it does it lives on our side. Our side of the house as part of our Oracle OCI control plane. And it is the veneer that makes these two clouds possible so that we can wire them together. So it knows how to take those Azure primitives and the OCI primitives and wire them at the appropriate levels together. >> Now I want to talk about this PaaS layer. Part of supercloud, we said to actually make it work you're going to have to have a super PaaS. I know we're taking this this term a little far but it's still it's instructive in that, what we surmised was you're probably not going to just use off the shelf, plain old vanilla PaaS, you're actually going to have a purpose built PaaS to solve for the specific problem. So as an example, if you're solving for ultra low latency, which I think you're doing, you're probably no offense to my friends at Red Hat but you're probably not going to develop this on OpenShift, but tell us about that PaaS layer or what we call the super PaaS layer. >> Go ahead, Chris. >> Well, so you're right. We weren't going to build it out on OpenShift. So we have Oracle OCI, you know, the standard is Terraform. So the back end of everything we do is based around Terraform. Today, what we've done is we built that control plane and it will be API drivable, it'll be drivable from the UI and it will let people operate and create primitives across both sides. So you can, you mentioned developers, developers love automation, right, because it makes our lives easy. We will be able to automate a multi-cloud workload from ground up config is code these days. So we can config an entire multi-cloud experience from one place. >> So, double click Chris on that developer experience. What is that like? They're using the same tool set irrespective of, which cloud we're running on is, and it's specific to this service or is it more generic, across other Oracle services? >> There's two parts to that. So one is the, we've only onboarded a portion. So the database portfolio and other services will be coming into this multi-cloud. For the majority of Oracle cloud, the automation, the config layer is based on Terraform. So using Terraform, anyone can configure everything from a mid-tier to an Exadata, all the way soup to nuts from smallest thing possible to the largest. What we've not done yet is integrated truly with the Azure API, from command line drivable. That is coming in the future. It is on the roadmap, it is coming. Then they could get into one tool but right now they would have half their automation for the multi-cloud config on the Azure tool set and half on the OCI tool set. >> But we're not crazy saying from a roadmap standpoint that will provide some benefit to developers and is a reasonable direction for the industry generally but Oracle and Microsoft specifically. >> Absolutely. I'm a developer at heart. And so one of the things we want to make sure is that developers' lives are as easy as possible. >> And is there a metadata management layer or intelligence that you've built in to optimize for performance or low latency or cost across the respective clouds? >> Yeah, definitely. I think, latency's going to be an important factor. The service that we've initially built isn't going to serve, the sort of the tens of microseconds but most applications that are sort of in, running on top of the enterprise applications that are running on top of the database are in the several millisecond range. And we've actually done a lot of work on the networking pairing side to make sure that when we launch these resources across the two clouds we actually picked the right trial site. We picked the right region we pick the right availability zone or domain. So we actually do the due diligence under the cover so the customer doesn't have to do the trial and error and try to find the right latency range. And this is actually one of the big reasons why we only launch the service on the interconnect regions. Even though we have close to, I think close to 40 regions at this point in OCI, this service is only built for the regions that we have an interconnect relationship with Microsoft. >> Okay, so you started with Microsoft in 2019. You're going deeper now in that relationship, is there any reason that you couldn't, I mean technically what would you have to do to go to other clouds? You talked about understanding the primitives and leveraging the primitives of Azure. Presumably if you wanted to do this with AWS or Google or Alibaba, you would have to do similar engineering work, is that correct? Or does what you've developed just kind of poured over to any cloud? >> Yeah, that's absolutely correct Dave. I think Chris talked a lot about the multi-cloud control plane, right? That's essentially the control plane that goes and does stuff on other clouds. We would have to essentially go and build that level of integration into the other clouds. And I think, as we get more popularity and as more products come online through these services I think we'll listen to what customers want. Whether it's, maybe it's the other way around too, Dave maybe it's the fact that they want to use Oracle cloud but they want to use other complimentary services within Oracle cloud. So I think it can go both ways. I think, the market and the customer base will dictate that. >> Yeah. So if I understand that correctly, somebody from another cloud Google cloud could say, Hey we actually want to run this service on OCI cuz we want to expand our market. And if TK gets together with his old friends and figures that out but then we're just, hypothesizing here. But, like you said, it can go both ways. And then, and I have another question related to that. So, multi clouds. Okay, great. Supercloud. How about the Edge? Do you ever see a day where that becomes part of the equation? Certainly the near Edge would, you know, a Home Depot or Lowe's store or a bank, but what about the far Edge, the tiny Edge. Can you talk about the Edge and where that fits in your vision? >> Yeah, absolutely. I think Edge is a interestingly, it's getting fuzzier and fuzzier day by day. I think, the term. Obviously every cloud has their own sort of philosophy in what Edge is, right. We have our own. It starts from, if you do want to do far Edge, we have devices like red devices, which is our ruggedized servers that talk back to our control plane in OCI. You could deploy those things unlike, into war zones and things like that underground. But then we also have things like clouded customer where customers can actually deploy components of our infrastructure like compute or Exadata into a facility where they only need that certain capability. And then a few years ago we launched, what's now called Dedicated Region. And that actually is a different take on Edge in some sense where you get the entire capability of our public commercial region, but within your facility. So imagine if a customer was to essentially point a finger on a commercial map and say, Hey, look, that region is just mine. Essentially that's the capability that we're providing to our customers, where if you have a white space if you have a facility, if you're exiting out of your data center space, you could essentially place an OCI region within your confines behind your firewall. And then you could interconnect that to a cloud provider if you wanted to, and get the same multi-cloud capability that you get in a commercial region. So we have all the spectrums of possibilities here. >> Guys, super interesting discussion. It's very clear to us that the next 10 years of cloud ain't going to be like the last 10. There's a whole new layer. Developing, data is a big key to that. We see industries getting involved. We obviously didn't get into the Oracle Cerner acquisitions. It's a little too early for that but we've actually predicted that companies like Cerner and you're seeing it with Goldman Sachs and Capital One they're actually building services on the cloud. So this is a really exciting new area and really appreciate you guys coming on the Supercloud22 event and sharing your insights. Thanks for your time. >> Thanks for having us. >> Okay. Keep it right there. #Supercloud22. We'll be right back with more great content right after this short break. (lighthearted marimba music)

Published Date : Aug 10 2022

SUMMARY :

And Kris Rice is the Vice President that leverages the core primitives And the integration I think What's the service optimized but above the networking, the resources that you on both sides of the fence. So, the customer at the same time to make So you say extremely fast networking. computers that we provide And the multi-cloud control plane runs. And it is the veneer that So as an example, if you're So the back end of everything we do and it's specific to this service and half on the OCI tool set. for the industry generally And so one of the things on the interconnect regions. and leveraging the primitives of Azure. of integration into the other clouds. of the equation? that talk back to our services on the cloud. with more great content

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Karan BattaPERSON

0.99+

ChrisPERSON

0.99+

OracleORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

OCIORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

AlibabaORGANIZATION

0.99+

Dave VellantePERSON

0.99+

Capital OneORGANIZATION

0.99+

Goldman SachsORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Kris RicePERSON

0.99+

KaranPERSON

0.99+

CernerORGANIZATION

0.99+

LoweORGANIZATION

0.99+

2019DATE

0.99+

secondQUANTITY

0.99+

DavePERSON

0.99+

two partsQUANTITY

0.99+

11 regionsQUANTITY

0.99+

Larry EllisonPERSON

0.99+

oneQUANTITY

0.99+

bothQUANTITY

0.99+

two cloudsQUANTITY

0.99+

Supercloud22EVENT

0.99+

both sidesQUANTITY

0.99+

Home DepotORGANIZATION

0.99+

Red HatORGANIZATION

0.99+

third cloudQUANTITY

0.99+

last monthDATE

0.98+

one placeQUANTITY

0.98+

both waysQUANTITY

0.98+

SupercloudORGANIZATION

0.98+

OpenShiftTITLE

0.98+

TodayDATE

0.98+

one toolQUANTITY

0.98+

ExadataORGANIZATION

0.97+

more than one cloudQUANTITY

0.97+

first cloudQUANTITY

0.96+

AzureTITLE

0.96+

Edge PCORGANIZATION

0.96+

EdgeORGANIZATION

0.96+

10QUANTITY

0.96+

two different cloudsQUANTITY

0.96+

Oracle DatabaseORGANIZATION

0.96+

TKPERSON

0.95+

both cloudsQUANTITY

0.95+

under two millisecondsQUANTITY

0.95+

40 regionsQUANTITY

0.94+

todayDATE

0.94+

single cloudQUANTITY

0.93+

Vice PresidentPERSON

0.91+

TerraformORGANIZATION

0.91+

PaaSTITLE

0.91+

InfiniBandORGANIZATION

0.91+

tens of microsecondsQUANTITY

0.9+

#Supercloud22EVENT

0.9+

Oracle Cloud InfrastructureORGANIZATION

0.9+

doubleQUANTITY

0.88+

OCICOMMERCIAL_ITEM

0.88+

AzureORGANIZATION

0.86+