Ed Walsh & Thomas Hazel | A New Database Architecture for Supercloud
(bright music) >> Hi, everybody, this is Dave Vellante, welcome back to Supercloud 2. Last August, at the first Supercloud event, we invited the broader community to help further define Supercloud, we assessed its viability, and identified the critical elements and deployment models of the concept. The objectives here at Supercloud too are, first of all, to continue to tighten and test the concept, the second is, we want to get real world input from practitioners on the problems that they're facing and the viability of Supercloud in terms of applying it to their business. So on the program, we got companies like Walmart, Sachs, Western Union, Ionis Pharmaceuticals, NASDAQ, and others. And the third thing that we want to do is we want to drill into the intersection of cloud and data to project what the future looks like in the context of Supercloud. So in this segment, we want to explore the concept of data architectures and what's going to be required for Supercloud. And I'm pleased to welcome one of our Supercloud sponsors, ChaosSearch, Ed Walsh is the CEO of the company, with Thomas Hazel, who's the Founder, CTO, and Chief Scientist. Guys, good to see you again, thanks for coming into our Marlborough studio. >> Always great. >> Great to be here. >> Okay, so there's a little debate, I'm going to put you right in the spot. (Ed chuckling) A little debate going on in the community started by Bob Muglia, a former CEO of Snowflake, and he was at Microsoft for a long time, and he looked at the Supercloud definition, said, "I think you need to tighten it up a little bit." So, here's what he came up with. He said, "A Supercloud is a platform that provides a programmatically consistent set of services hosted on heterogeneous cloud providers." So he's calling it a platform, not an architecture, which was kind of interesting. And so presumably the platform owner is going to be responsible for the architecture, but Dr. Nelu Mihai, who's a computer scientist behind the Cloud of Clouds Project, he chimed in and responded with the following. He said, "Cloud is a programming paradigm supporting the entire lifecycle of applications with data and logic natively distributed. Supercloud is an open architecture that integrates heterogeneous clouds in an agnostic manner." So, Ed, words matter. Is this an architecture or is it a platform? >> Put us on the spot. So, I'm sure you have concepts, I would say it's an architectural or design principle. Listen, I look at Supercloud as a mega trend, just like cloud, just like data analytics. And some companies are using the principle, design principles, to literally get dramatically ahead of everyone else. I mean, things you couldn't possibly do if you didn't use cloud principles, right? So I think it's a Supercloud effect, you're able to do things you're not able to. So I think it's more a design principle, but if you do it right, you get dramatic effect as far as customer value. >> So the conversation that we were having with Muglia, and Tristan Handy of dbt Labs, was, I'll set it up as the following, and, Thomas, would love to get your thoughts, if you have a CRM, think about applications today, it's all about forms and codifying business processes, you type a bunch of stuff into Salesforce, and all the salespeople do it, and this machine generates a forecast. What if you have this new type of data app that pulls data from the transaction system, the e-commerce, the supply chain, the partner ecosystem, et cetera, and then, without humans, actually comes up with a plan. That's their vision. And Muglia was saying, in order to do that, you need to rethink data architectures and database architectures specifically, you need to get down to the level of how the data is stored on the disc. What are your thoughts on that? Well, first of all, I'm going to cop out, I think it's actually both. I do think it's a design principle, I think it's not open technology, but open APIs, open access, and you can build a platform on that design principle architecture. Now, I'm a database person, I love solving the database problems. >> I'm waited for you to launch into this. >> Yeah, so I mean, you know, Snowflake is a database, right? It's a distributed database. And we wanted to crack those codes, because, multi-region, multi-cloud, customers wanted access to their data, and their data is in a variety of forms, all these services that you're talked about. And so what I saw as a core principle was cloud object storage, everyone streams their data to cloud object storage. From there we said, well, how about we rethink database architecture, rethink file format, so that we can take each one of these services and bring them together, whether distributively or centrally, such that customers can access and get answers, whether it's operational data, whether it's business data, AKA search, or SQL, complex distributed joins. But we had to rethink the architecture. I like to say we're not a first generation, or a second, we're a third generation distributed database on pure, pure cloud storage, no caching, no SSDs. Why? Because all that availability, the cost of time, is a struggle, and cloud object storage, we think, is the answer. >> So when you're saying no caching, so when I think about how companies are solving some, you know, pretty hairy problems, take MySQL Heatwave, everybody thought Oracle was going to just forget about MySQL, well, they come out with Heatwave. And the way they solve problems, and you see their benchmarks against Amazon, "Oh, we crush everybody," is they put it all in memory. So you said no caching? You're not getting performance through caching? How is that true, and how are you getting performance? >> Well, so five, six years ago, right? When you realize that cloud object storage is going to be everywhere, and it's going to be a core foundational, if you will, fabric, what would you do? Well, a lot of times the second generation say, "We'll take it out of cloud storage, put in SSDs or something, and put into cache." And that adds a lot of time, adds a lot of costs. But I said, what if, what if we could actually make the first read hot, the first read distributed joins and searching? And so what we went out to do was said, we can't cache, because that's adds time, that adds cost. We have to make cloud object storage high performance, like it feels like a caching SSD. That's where our patents are, that's where our technology is, and we've spent many years working towards this. So, to me, if you can crack that code, a lot of these issues we're talking about, multi-region, multicloud, different services, everybody wants to send their data to the data lake, but then they move it out, we said, "Keep it right there." >> You nailed it, the data gravity. So, Bob's right, the data's coming in, and you need to get the data from everywhere, but you need an environment that you can deal with all that different schema, all the different type of technology, but also at scale. Bob's right, you cannot use memory or SSDs to cache that, that doesn't scale, it doesn't scale cost effectively. But if you could, and what you did, is you made object storage, S3 first, but object storage, the only persistence by doing that. And then we get performance, we should talk about it, it's literally, you know, hundreds of terabytes of queries, and it's done in seconds, it's done without memory caching. We have concepts of caching, but the only caching, the only persistence, is actually when we're doing caching, we're just keeping another side-eye track of things on the S3 itself. So we're using, actually, the object storage to be a database, which is kind of where Bob was saying, we agree, but that's what you started at, people thought you were crazy. >> And maybe make it live. Don't think of it as archival or temporary space, make it live, real time streaming, operational data. What we do is make it smart, we see the data coming in, we uniquely index it such that you can get your use cases, that are search, observability, security, or backend operational. But we don't have to have this, I dunno, static, fixed, siloed type of architecture technologies that were traditionally built prior to Supercloud thinking. >> And you don't have to move everything, essentially, you can do it wherever the data lands, whatever cloud across the globe, you're able to bring it together, you get the cost effectiveness, because the only persistence is the cheapest storage persistent layer you can buy. But the key thing is you cracked the code. >> We had to crack the code, right? That was the key thing. >> That's where the plans are. >> And then once you do that, then everything else gets easier to scale, your architecture, across regions, across cloud. >> Now, it's a general purpose database, as Bob was saying, but we use that database to solve a particular issue, which is around operational data, right? So, we agree with Bob's. >> Interesting. So this brings me to this concept of data, Jimata Gan is one of our speakers, you know, we talk about data fabric, which is a NetApp, originally NetApp concept, Gartner's kind of co-opted it. But so, the basic concept is, data lives everywhere, whether it's an S3 bucket, or a SQL database, or a data lake, it's just a node on the data mesh. So in your view, how does this fit in with Supercloud? Ed, you've said that you've built, essentially, an enabler for that, for the data mesh, I think you're an enabler for the Supercloud-like principles. This is a big, chewy opportunity, and it requires, you know, a team approach. There's got to be an ecosystem, there's not going to be one Supercloud to rule them all, so where does the ecosystem fit into the discussion, and where do you fit into the ecosystem? >> Right, so we agree completely, there's not one Supercloud in effect, but we use Supercloud principles to build our platform, and then, you know, the ecosystem's going to be built on leveraging what everyone else's secret powers are, right? So our power, our superpower, based upon what we built is, we deal with, if you're having any scale, or cost effective scale issues, with data, machine generated data, like business observability or security data, we are your force multiplier, we will take that in singularly, just let it, simply put it in your object storage wherever it sits, and we give you uniformity access to that using OpenAPI access, SQL, or you know, Elasticsearch API. So, that's what we do, that's our superpower. So I'll play it into data mesh, that's a perfect, we are a node on a data mesh, but I'll play it in the soup about how, the ecosystem, we see it kind of playing, and we talked about it in just in the last couple days, how we see this kind of possibly. Short term, our superpowers, we deal with this data that's coming at these environments, people, customers, building out observability or security environments, or vendors that are selling their own Supercloud, I do observability, the Datadogs of the world, dot dot dot, the Splunks of the world, dot dot dot, and security. So what we do is we fit in naturally. What we do is a cost effective scale, just land it anywhere in the world, we deal with ingest, and it's a cost effective, an order of magnitude, or two or three order magnitudes more cost effective. Allows them, their customers are asking them to do the impossible, "Give me fast monitoring alerting. I want it snappy, but I want it to keep two years of data, (laughs) and I want it cost effective." It doesn't work. They're good at the fast monitoring alerting, we're good at the long-term retention. And yet there's some gray area between those two, but one to one is actually cheaper, so we would partner. So the first ecosystem plays, who wants to have the ability to, really, all the data's in those same environments, the security observability players, they can literally, just through API, drag our data into their point to grab. We can make it seamless for customers. Right now, we make it helpful to customers. Your Datadog, we make a button, easy go from Datadog to us for logs, save you money. Same thing with Grafana. But you can also look at ecosystem, those same vendors, it used to be a year ago it was, you know, its all about how can you grow, like it's growth at all costs, now it's about cogs. So literally we can go an environment, you supply what your customer wants, but we can help with cogs. And one-on one in a partnership is better than you trying to build on your own. >> Thomas, you were saying you make the first read fast, so you think about Snowflake. Everybody wants to talk about Snowflake and Databricks. So, Snowflake, great, but you got to get the data in there. All right, so that's, can you help with that problem? >> I mean we want simple in, right? And if you have to have structure in, you're not simple. So the idea that you have a simple in, data lake, schema read type philosophy, but schema right type performance. And so what I wanted to do, what we have done, is have that simple lake, and stream that data real time, and those access points of Search or SQL, to go after whatever business case you need, security observability, warehouse integration. But the key thing is, how do I make that click, click, click answer, and do it quickly? And so what we want to do is, that first read has to be fast. Why? 'Cause then you're going to do all this siloing, layers, complexity. If your first read's not fast, you're at a disadvantage, particularly in cost. And nobody says I want less data, but everyone has to, whether they say we're going to shorten the window, we're going to use AI to choose, but in a security moment, when you don't have that answer, you're in trouble. And that's why we are this service, this Supercloud service, if you will, providing access, well-known search, well-known SQL type access, that if you just have one access point, you're at a disadvantage. >> We actually talked about Snowflake and BigQuery, and a different platform, Data Bricks. That's kind of where we see the phase two of ecosystem. One is easy, the low-hanging fruit is observability and security firms. But the next one is, what we do, our super power is dealing with this messy data that schema is changing like night and day. Pipelines are tough, and it's changing all the time, but you want these things fast, and it's big data around the world. That's the next point, just use us alongside, or inside, one of their platforms, and now we get the best of both worlds. Our superpower is keeping this messy data as a streaming, okay, not a batch thing, allow you to do that. So, that's the second one. And then to be honest, the third one, which plays you to Supercloud, it also plays perfectly in the data mesh, is if you really go to the ultimate thing, what we have done is made object storage, S3, GCS, and blob storage, we made it a database. Put, get, complex query with big joins. You know, so back to your original thing, and Muglia teed it up perfectly, we've done that. Now imagine if that's an ecosystem, who would want that? If it's, again, it's uniform available across all the regions, across all the clouds, and it's right next to where you are building a service, or a client's trying, that's where the ecosystem, I think people are going to use Superclouds for their superpowers. We're really good at this, allows that short term. I think the Snowflakes and the Data Bricks are the medium term, you know? And then I think eventually gets to, hey, listen if you can make object storage fast, you can just go after it with simple SQL queries, or elastic. Who would want that? I think that's where people are going to leverage it. It's not going to be one Supercloud, and we leverage the super clouds. >> Our viewpoint is smart object storage can be programmable, and so we agree with Bob, but we're not saying do it here, do it here. This core, fundamental layer across regions, across clouds, that everyone has? Simple in. Right now, it's hard to get data in for access for analysis. So we said, simply, we'll automate the entire process, give you API access across regions, across clouds. And again, how do you do a distributed join that's fast? How do you do a distributed join that doesn't cost you an arm or a leg? And how do you do it at scale? And that's where we've been focused. >> So prior, the cloud object store was a niche. >> Yeah. >> S3 obviously changed that. How standard is, essentially, object store across the different cloud platforms? Is that a problem for you? Is that an easy thing to solve? >> Well, let's talk about it. I mean we've fundamentally, yeah we've extracted it, but fundamentally, cloud object storage, put, get, and list. That's why it's so scalable, 'cause it doesn't have all these other components. That complexity is where we have moved up, and provide direct analytical API access. So because of its simplicity, and costs, and security, and reliability, it can scale naturally. I mean, really, distributed object storage is easy, it's put-get anywhere, now what we've done is we put a layer of intelligence, you know, call it smart object storage, where access is simple. So whether it's multi-region, do a query across, or multicloud, do a query across, or hunting, searching. >> We've had clients doing Amazon and Google, we have some Azure, but we see Amazon and Google more, and it's a consistent service across all of them. Just literally put your data in the bucket of choice, or folder of choice, click a couple buttons, literally click that to say "that's hot," and after that, it's hot, you can see it. But we're not moving data, the data gravity issue, that's the other. That it's already natively flowing to these pools of object storage across different regions and clouds. We don't move it, we index it right there, we're spinning up stateless compute, back to the Supercloud concept. But now that allows us to do all these other things, right? >> And it's no longer just cheap and deep object storage. Right? >> Yeah, we make it the same, like you have an analytic platform regardless of where you're at, you don't have to worry about that. Yeah, we deal with that, we deal with a stateless compute coming up -- >> And make it programmable. Be able to say, "I want this bucket to provide these answers." Right, that's really the hope, the vision. And the complexity to build the entire stack, and then connect them together, we said, the fabric is cloud storage, we just provide the intelligence on top. >> Let's bring it back to the customers, and one of the things we're exploring in Supercloud too is, you know, is Supercloud a solution looking for a problem? Is a multicloud really a problem? I mean, you hear, you know, a lot of the vendor marketing says, "Oh, it's a disaster, because it's all different across the clouds." And I talked to a lot of customers even as part of Supercloud too, they're like, "Well, I solved that problem by just going mono cloud." Well, but then you're not able to take advantage of a lot of the capabilities and the primitives that, you know, like Google's data, or you like Microsoft's simplicity, their RPA, whatever it is. So what are customers telling you, what are their near term problems that they're trying to solve today, and how are they thinking about the future? >> Listen, it's a real problem. I think it started, I think this is a a mega trend, just like cloud. Just, cloud data, and I always add, analytics, are the mega trends. If you're looking at those, if you're not considering using the Supercloud principles, in other words, leveraging what I have, abstracting it out, and getting the most out of that, and then build value on top, I think you're not going to be able to keep up, In fact, no way you're going to keep up with this data volume. It's a geometric challenge, and you're trying to do linear things. So clients aren't necessarily asking, hey, for Supercloud, but they're really saying, I need to have a better mechanism to simplify this and get value across it, and how do you abstract that out to do that? And that's where they're obviously, our conversations are more amazed what we're able to do, and what they're able to do with our platform, because if you think of what we've done, the S3, or GCS, or object storage, is they can't imagine the ingest, they can't imagine how easy, time to glass, one minute, no matter where it lands in the world, querying this in seconds for hundreds of terabytes squared. People are amazed, but that's kind of, so they're not asking for that, but they are amazed. And then when you start talking on it, if you're an enterprise person, you're building a big cloud data platform, or doing data or analytics, if you're not trying to leverage the public clouds, and somehow leverage all of them, and then build on top, then I think you're missing it. So they might not be asking for it, but they're doing it. >> And they're looking for a lens, you mentioned all these different services, how do I bring those together quickly? You know, our viewpoint, our service, is I have all these streams of data, create a lens where they want to go after it via search, go after via SQL, bring them together instantly, no e-tailing out, no define this table, put into this database. We said, let's have a service that creates a lens across all these streams, and then make those connections. I want to take my CRM with my Google AdWords, and maybe my Salesforce, how do I do analysis? Maybe I want to hunt first, maybe I want to join, maybe I want to add another stream to it. And so our viewpoint is, it's so natural to get into these lake platforms and then provide lenses to get that access. >> And they don't want it separate, they don't want something different here, and different there. They want it basically -- >> So this is our industry, right? If something new comes out, remember virtualization came out, "Oh my God, this is so great, it's going to solve all these problems." And all of a sudden it just got to be this big, more complex thing. Same thing with cloud, you know? It started out with S3, and then EC2, and now hundreds and hundreds of different services. So, it's a complex matter for a lot of people, and this creates problems for customers, especially when you got divisions that are using different clouds, and you're saying that the solution, or a solution for the part of the problem, is to really allow the data to stay in place on S3, use that standard, super simple, but then give it what, Ed, you've called superpower a couple of times, to make it fast, make it inexpensive, and allow you to do that across clouds. >> Yeah, yeah. >> I'll give you guys the last word on that. >> No, listen, I think, we think Supercloud allows you to do a lot more. And for us, data, everyone says more data, more problems, more budget issue, everyone knows more data is better, and we show you how to do it cost effectively at scale. And we couldn't have done it without the design principles of we're leveraging the Supercloud to get capabilities, and because we use super, just the object storage, we're able to get these capabilities of ingest, scale, cost effectiveness, and then we built on top of this. In the end, a database is a data platform that allows you to go after everything distributed, and to get one platform for analytics, no matter where it lands, that's where we think the Supercloud concepts are perfect, that's where our clients are seeing it, and we're kind of excited about it. >> Yeah a third generation database, Supercloud database, however we want to phrase it, and make it simple, but provide the value, and make it instant. >> Guys, thanks so much for coming into the studio today, I really thank you for your support of theCUBE, and theCUBE community, it allows us to provide events like this and free content. I really appreciate it. >> Oh, thank you. >> Thank you. >> All right, this is Dave Vellante for John Furrier in theCUBE community, thanks for being with us today. You're watching Supercloud 2, keep it right there for more thought provoking discussions around the future of cloud and data. (bright music)
SUMMARY :
And the third thing that we want to do I'm going to put you right but if you do it right, So the conversation that we were having I like to say we're not a and you see their So, to me, if you can crack that code, and you need to get the you can get your use cases, But the key thing is you cracked the code. We had to crack the code, right? And then once you do that, So, we agree with Bob's. and where do you fit into the ecosystem? and we give you uniformity access to that so you think about Snowflake. So the idea that you have are the medium term, you know? and so we agree with Bob, So prior, the cloud that an easy thing to solve? you know, call it smart object storage, and after that, it's hot, you can see it. And it's no longer just you don't have to worry about And the complexity to and one of the things we're and how do you abstract it's so natural to get and different there. and allow you to do that across clouds. I'll give you guys and we show you how to do it but provide the value, I really thank you for around the future of cloud and data.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Walmart | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
NASDAQ | ORGANIZATION | 0.99+ |
Bob Muglia | PERSON | 0.99+ |
Thomas | PERSON | 0.99+ |
Thomas Hazel | PERSON | 0.99+ |
Ionis Pharmaceuticals | ORGANIZATION | 0.99+ |
Western Union | ORGANIZATION | 0.99+ |
Ed Walsh | PERSON | 0.99+ |
Bob | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Nelu Mihai | PERSON | 0.99+ |
Sachs | ORGANIZATION | 0.99+ |
Tristan Handy | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
two years | QUANTITY | 0.99+ |
Supercloud 2 | TITLE | 0.99+ |
first | QUANTITY | 0.99+ |
Last August | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
dbt Labs | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Ed | PERSON | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
Jimata Gan | PERSON | 0.99+ |
third one | QUANTITY | 0.99+ |
one minute | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
first generation | QUANTITY | 0.99+ |
third generation | QUANTITY | 0.99+ |
Grafana | ORGANIZATION | 0.99+ |
second generation | QUANTITY | 0.99+ |
second one | QUANTITY | 0.99+ |
hundreds of terabytes | QUANTITY | 0.98+ |
SQL | TITLE | 0.98+ |
five | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
Databricks | ORGANIZATION | 0.98+ |
a year ago | DATE | 0.98+ |
ChaosSearch | ORGANIZATION | 0.98+ |
Muglia | PERSON | 0.98+ |
MySQL | TITLE | 0.98+ |
both worlds | QUANTITY | 0.98+ |
third thing | QUANTITY | 0.97+ |
Marlborough | LOCATION | 0.97+ |
theCUBE | ORGANIZATION | 0.97+ |
today | DATE | 0.97+ |
Supercloud | ORGANIZATION | 0.97+ |
Elasticsearch | TITLE | 0.96+ |
NetApp | TITLE | 0.96+ |
Datadog | ORGANIZATION | 0.96+ |
One | QUANTITY | 0.96+ |
EC2 | TITLE | 0.96+ |
each one | QUANTITY | 0.96+ |
S3 | TITLE | 0.96+ |
one platform | QUANTITY | 0.95+ |
Supercloud 2 | EVENT | 0.95+ |
first read | QUANTITY | 0.95+ |
six years ago | DATE | 0.95+ |
Nir Zuk, Palo Alto Networks | An Architecture for Securing the Supercloud
(bright upbeat music) >> Welcome back, everybody, to the Supercloud 2. My name is Dave Vellante. And I'm pleased to welcome Nir Zuk. He's the founder and CTO of Palo Alto Networks. Nir, good to see you again. Welcome. >> Same here. Good to see you. >> So let's start with the right security architecture in the context of today's fragmented market. You've got a lot of different tools, you've got different locations, on-prem, you've got hardware and software. Tell us about the right security architecture from your standpoint. What's that look like? >> You know, the funny thing is using the word security in architecture rarely works together. (Dave chuckles) If you ask a typical information security person to step up to a whiteboard and draw their security architecture, they will look at you as if you fell from the moon. I mean, haven't you been here in the last 25 years? There's no security architecture. The architecture today is just buying a bunch of products and dropping them into the infrastructure at some relatively random way without really any guiding architecture. And that's a huge challenge in cybersecurity. It's always been, we've always tried to find ways to put an architecture into writing blueprints, whatever you want to call it, and it's always been difficult. Luckily, two things. First, there's something called zero trust, which we can talk a little bit about more, if you want, and zero trust among other things is really a way to create a security architecture, and second, because in the cloud, in the supercloud, we're starting from scratch, we can do things differently. We don't have to follow the way we've always done cybersecurity, again, buying random products, okay, maybe not random, maybe there is some thinking going into it by buying products, one of the other, dropping them in, and doing it over 20 years and ending up with a mess in the cloud, we have an opportunity to do it differently and really have an architecture. >> You know, I love talking to founders and particularly technical founders from StartupNation. I think I saw an article, I think it was Erie Levine, one of the founders or co-founders of Waze, and he had a t-shirt on, it said, "Fall in love with the problem, not the solution." Is that how you approached architecture? You talk about zero trust, it's a relatively new term, but was that in your head when you thought about forming the company? >> Yeah, so when I started Palo Alto Networks, exactly, by the way, 17 years ago, we got funded January, 2006, January 18th, 2006. The idea behind Palo Alto Networks was to create a security platform and over time take more and more cybersecurity functions and deliver them on top of that platform, by the way, as a service, SaaS. Everybody thought we were crazy trying to combine many functions into one platform, best of breed and defense in death and putting all your eggs in the same basket and a bunch of other slogans were flying around, and also everybody thought we were crazy asking customers to send information to the cloud in order to secure themselves. Of course, step forward 17 years, everything is now different. We changed the market. Almost all of cybersecurity today is delivered as SaaS and platforms are ruling more and more the world. And so again, the idea behind the platform was to over time take more and more cybersecurity functions and deliver them together, one brain, one decision being made for each and every packet or system call or file or whatever it is that you're making the decision about and it works really, really well. As a side effect, when you combine that with zero trust and you end up with, let's not call it an architecture yet. You end up with with something where any user, any location, both geographically as well as any location in terms of branch office, headquarters, home, coffee shop, hotel, whatever, so any user, any geographical location, any location, any connectivity method, whether it is SD1 or IPsec or Client VPN or Client SVPN or proxy or browser isolation or whatever and any application deployed anywhere, public cloud, private cloud, traditional data center, SaaS, you secure the same way. That's really zero trust, right? You secure everything, no matter who the user is, no matter where they are, no matter where they go, you secure them exactly the same way. You don't make any assumptions about the user or the application or the location or whatever, just because you trust nothing. And as a side effect, when you do that, you end up with a security architecture, the security architecture I just described. The same thing is true for securing applications. If you try to really think and not just act instinctively the way we usually do in cybersecurity and you say, I'm going to secure my traditional data center applications or private cloud applications and public cloud applications and my SaaS applications the same way, I'm not going to trust something just because it's deployed in the private data center. I'm not going to trust two components of an application or two applications talking to each other just because they're deployed in the same place versus if one component is deployed in one public cloud and the other component is deployed in another public cloud or private cloud or whatever. I'm going to secure all of them the same way without making any trust assumptions. You end up with an architecture for securing your applications, which is applicable for the supercloud. >> It was very interesting. There's a debate I want to pick up on what you said because you said don't call it an architecture yet. So Bob Muglia, I dunno if you know Bob, but he sort of started the debate, said, "Supercloud, think of it as a platform, not an architecture." And there are others that are saying, "No, no, if we do that, then we're going to have a bunch of more stove pipes. So there needs to be standard, almost a purist view. There needs to be a supercloud architecture." So how do you think about it? And it's a bit academic, I know, but do you think of this idea of a supercloud, this layer of value on top of the hyperscalers, do you think of that as a platform approach that each of the individual vendors are responsible for the architecture? Or is there some kind of overriding architecture of standards that needs to emerge to enable the supercloud? >> So we can talk academically or we can talk practically. >> Yeah, let's talk practically. That's who you are. (Dave laughs) >> Practically, this world is ruled by financial interests and none of the public cloud providers, especially the bigger they are has any interest of making it easy for anyone to go multi-cloud, okay? Also, on top of that, if we want to be even more practical, each of those large cloud providers, cloud scale providers have engineers and all these engineers think they're the best in the world, which they are and they all like to do things differently. So you can't expect things in AWS and in Azure and GCP and in the other clouds like Oracle and Ali and so on to be the same. They're not going to be the same. And some things can be abstracted. Maybe cloud storage or bucket storage can be abstracted with the layer that makes them look the same no matter where you're running. And some things cannot be abstracted and unfortunately will not be abstracted because the economical interest and the way engineers work won't let it happen. We as a third party provider, cybersecurity provider, and I'm sure other providers in other areas as well are trying or we're doing our best. We're not trying, we are doing our best, and it's pretty close to being the way you describe the top of your supercloud. We're building something that abstracts the underlying cloud such that securing each of these clouds, and by the way, I would add private cloud to it as well, looks exactly the same. So we use, almost always, whenever possible, the same terminology, no matter which cloud we're securing and the same policy and the same alerts and the same information and so on. And that's also very important because when you look at the people that actually end up using the product, security engineers and more importantly, SOC, security operations center analysts, they're not going to study the details of each and every cloud. It's just going to be too much. So we need to abstract it for them. >> Yeah, we agree by the way that the supercloud definition is inclusive of on-prem, you know, what you call private cloud. And I want to pick up on something else you said. I think you're right that abstracting and making consistent across clouds something like object storage, get put, you know, whether it's an S3 bucket or an Azure Blob, relatively speaking trivial. When you now bring that supercloud concept to something more complex like security, first of all, as a technically feasible and inferring the answer there is yes, and if so, what do you see as the main technical challenges of doing so? >> So it is feasible to the extent that the different cloud provide the same functionality. Then you step into a territory where different cloud providers have different paths services and different cloud providers do things a little bit differently and they have different sets of permissions and different logging that sometimes provides all the information and sometimes it doesn't. So you end up with some differences. And then the question is, do you abstract the lowest common dominator and that's all you support? Or do you find a way to be smarter than that? And yeah, whatever can be abstracted is abstracted and whatever cannot be abstracted, you find an easy way to represent that to your users, security engineers, security analysts, and so on, which is what I believe we do. >> And you do that by what? Inventing or developing technology that presents that experience to users? Could you be more specific there? >> Yeah, so different cloud providers call their storage in different names and you use different ways to configure them and the logs come out the same. So we normalize it. I mean, the keyword is probably normalization. Normalize it. And we try to, you know, then you have to pick a winner here and to use someone's terminology or you need to invent new terminology. So we try to use the terminology of the largest cloud provider so that we have a better chance of doing that but we can't always do that because they don't support everything that other cloud providers provide, but the important thing is, with or thanks to that normalization, our customers both on the engineering side and on the user side, operations side end up having to learn one terminology in order to set policies and understand attacks and investigate incidents. >> I wonder if I could pick your brain on what you see as the ideal deployment model to achieve this supercloud experience. For example, do you think instantiating your stack in multiple regions and multiple clouds is the right way to do it? Or is building a single global instance on top of the clouds a more preferable way? Are maybe other models we should consider? What do you see as the trade off of these different deployment models and which one is ideal in your view? >> Yeah, so first, when you deploy cloud security, you have to decide whether you're going to use agents or not. By agents, I mean something working, something running inside the workload. Inside a virtual machine on the container host attached to function, serverless function and so on and I, of course, recommend using agents because that enables prevention, it enables functionality you cannot get without agents but you have to choose that. Now, of course, if you choose agent, you need to deploy AWS agents in AWS and GCP agents in GCP and Azure agents in Azure and so on. Of course, you don't do it manually. You do it through the CICD pipeline. And then the second thing that you need to do is you need to connect with the consoles. Of course, that can be done over the internet no matter where your security instances is running. You can run it on premise, you can run it in one of the other different clouds. Of course, we don't run it on premise. We prefer not to run it on premise because if you're secured in cloud, you might as well run in the cloud. And then the question is, for example, do you run a separate instance for AWS for GCP or for Azure, or you want to run one instance for all of them in one of these clouds? And there are advantages and disadvantages. I think that from a security perspective, it's always better to run in one place because then when you collect the information, you get information from all the clouds and you can start looking for cross-cloud issues, incidents, attacks, and so on. The downside of that is that you need to send all the information to one of the clouds and you probably know that sending data out of the cloud costs a lot of money versus keeping it in the cloud. So theoretically, you can build an architecture where you keep the data for AWS in AWS, Azure in Azure, GCP in GCP, and then you try to run distributed queries. When you do that, you find out you'd end up paying more for the compute to do that than you would've paid for sending all the data to a central location. So we prefer the approach of running in one place, bringing all the data there, and running all the security, the machine learning or whatever, the rules or whatever it is that you're running in one place versus trying to create a distributed deployment in order to try to save some money on the data, the network data transfers. >> Yeah, thank you for that. That makes a lot of sense. And so basically, should we think about the next layer building security data lake, if you will, and then running machine learning on top of that if I can use that term of a data lake or a lake house? Is that sort of where you're headed? >> Yeah, look, the world is headed in that direction, not just the cybersecurity world. The world is headed from being rule-based to being data-based. So cybersecurity is not different and what we used to do with rules in the past, we're now doing with machine learning. So in the past, you would define rules saying, if you see this, this, and this, it's an attack. Now you just throw the data at the machine, I mean, I'm simplifying it, but you throw data at a machine. You'll tell the machine, find the attack in the data. It's not that simple. You need to build the right machine learning models. It needs to be done by people that are both cybersecurity experts and machine learning experts. We do it mostly with ex-military offensive people that take their offensive knowledge and translate it into machine learning models. But look, the world is moving in that direction and cybersecurity is moving in that direction as well. You need to collect a lot of data. Like I said, I prefer to see all the data in one place so that the machine learning can be much more efficient, pay for transferring the data, save money on the compute. >> I think the drop the mic quote it ignite that you had was within five years, your security operation is going to be AI-powered. And so you could probably apply that to virtually any job over the next five years. >> I don't know if any job. Certainly writing essays for school is automated already as we've seen with ChatGPT and potentially other things. By the way, we need to talk at some point about ChatGPT security. I don't want to think what happens when someone spends a lot of money on creating a lot of fake content and teaches ChatGPT the wrong answer to a question. We start seeing ChatGPT as the oracle of everything. We need to figure out what to do with the security of that. But yeah, things have to be automated in cybersecurity. They have to be automated. They're just too much data to deal with and it's just not even close to being good enough to wait for an incident to happen and then going investigate the incident based on the data that we have. It's better to look at all the data all the time, millions of events per second, and find those incidents before they happen. There's no way to do that without machine learning. >> I'd love to have you back and talk about ChatGPT. I know they're trying to put in some guardrails but there are a lot of unintended consequences, aren't there? >> Look, if they're not going to have a person filtering the data, then with enough money, you can create thousands or tens of thousands of pieces of articles or whatever that look real and teach the machine something that is totally wrong. >> We were talking about the hyper skills before and I agree with you. It's very unlikely they're going to get together, band together, and create these standards. But it's not a static market. It's a moving train, if you will. So assuming you're building this cross cloud experience which you are, what do you want from the hyperscalers? What do you want them to bring to the table? What is a technology supplier like Palo Alto Networks bring? In other words, where do you see ongoing as your unique value add and that moat that you're building and how will that evolve over time vis-a-vis the hyperscaler evolution? >> Yeah, look, we need APIs. The more data we have, the more access we have to more data, the less restricted the access is and the cheaper the access is to the data because someone has to pay today for some reason for accessing that data, the more secure their customers are going to be. So we need help and are helping by the way a lot, all of them in finding easy ways for customers to deploy things in the cloud, access data, and again, a lot of data, very diversified data and do it in a cost-effective way. >> And when we talk about the edge, I presume you look at the edge as just another data center or maybe it's the reverse. Maybe the data center is just another edge location, but you're seeing specific edge security solutions come out. I'm guessing that you would say, that's not what we want. Edge should be part of that architecture that we talked about earlier. Do you agree? >> Correct, it should be part of the architecture. I would also say that the edge provides an opportunity specifically for network security, whereas traditional network security would be deployed on premise. I'm talking about internet security but half network security market, and not just network security but also the other network intelligent functions like routing and QS. We're seeing a trend of pushing those to the edge of the cloud. So what you deploy on premise is technology for bringing packets to the edge of the cloud and then you run your security at the edge, whatever that edge is, whether it's a private edge or public edge, you run it in the edge. It's called SASE, Secure Access Services Edge, pronounced SASE. >> Nir, I got to thank you so much. You're such a clear thinker. I really appreciate you participating in Supercloud 2. >> Thank you. >> All right, keep it right there for more content covering the future of cloud and data. This is Dave Vellante for John Furrier. I'll be right back. (bright upbeat music)
SUMMARY :
Nir, good to see you again. Good to see you. in the context of today's and second, because in the cloud, Is that how you approached architecture? and my SaaS applications the same way, that each of the individual So we can talk academically That's who you are. and none of the public cloud providers, and if so, what do you see and that's all you support? and on the user side, operations side is the right way to do it? and then you try to run about the next layer So in the past, you would that you had was within five years, and teaches ChatGPT the I'd love to have you that look real and teach the machine and that moat that you're building and the cheaper the access is to the data I'm guessing that you would and then you run your Nir, I got to thank you so much. the future of cloud and data.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Bob Muglia | PERSON | 0.99+ |
January, 2006 | DATE | 0.99+ |
Erie Levine | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Palo Alto Networks | ORGANIZATION | 0.99+ |
Bob | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
Nir Zuk | PERSON | 0.99+ |
two applications | QUANTITY | 0.99+ |
Nir | PERSON | 0.99+ |
one component | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
StartupNation | ORGANIZATION | 0.99+ |
Waze | ORGANIZATION | 0.99+ |
First | QUANTITY | 0.99+ |
two components | QUANTITY | 0.99+ |
second thing | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
January 18th, 2006 | DATE | 0.99+ |
one platform | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
17 years ago | DATE | 0.98+ |
over 20 years | QUANTITY | 0.98+ |
Azure | TITLE | 0.98+ |
17 years | QUANTITY | 0.98+ |
ChatGPT | TITLE | 0.98+ |
each | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
two things | QUANTITY | 0.97+ |
one place | QUANTITY | 0.97+ |
one instance | QUANTITY | 0.96+ |
one brain | QUANTITY | 0.96+ |
today | DATE | 0.95+ |
zero trust | QUANTITY | 0.94+ |
single | QUANTITY | 0.94+ |
second | QUANTITY | 0.94+ |
GCP | TITLE | 0.92+ |
five years | QUANTITY | 0.91+ |
tens of thousands | QUANTITY | 0.91+ |
one decision | QUANTITY | 0.88+ |
last 25 years | DATE | 0.86+ |
SASE | TITLE | 0.85+ |
Supercloud | ORGANIZATION | 0.85+ |
ChatGPT | ORGANIZATION | 0.84+ |
one terminology | QUANTITY | 0.79+ |
zero | QUANTITY | 0.77+ |
millions of events per second | QUANTITY | 0.75+ |
S3 | COMMERCIAL_ITEM | 0.75+ |
SOC | ORGANIZATION | 0.72+ |
Azure Blob | TITLE | 0.72+ |
Ali | ORGANIZATION | 0.72+ |
Supercloud 2 | ORGANIZATION | 0.68+ |
Exploring a Supercloud Architecture | Supercloud2
(upbeat music) >> Welcome back everyone to Supercloud 2, live here in Palo Alto, our studio, where we're doing a live stage performance and virtually syndicating out around the world. I'm John Furrier with Dave Vellante, my co-host with the The Cube here. We've got Kit Colbert, the CTO of VM. We're doing a keynote on Cloud Chaos, the evolution of SuperCloud Architecture Kit. Great to see you, thanks for coming on. >> Yeah, thanks for having me back. It's great to be here for Supercloud 2. >> And so we're going to dig into it. We're going to do a Q&A. We're going to let you present. You got some slides. I really want to get this out there, it's really compelling story. Do the presentation and then we'll come back and discuss. Take it away. >> Yeah, well thank you. So, we had a great time at the original Supercloud event, since then, been talking to a lot of customers, and started to better formulate some of the thinking that we talked about last time So, let's jump into it. Just a few quick slides to sort of set the tone here. So, if we go to the the next slide, what that shows is the journey that we see customers on today, going from what we call Cloud First into this phase that many customers are stuck in, called Cloud Chaos, and where they want to get to, and this is the term customers actually use, we didn't make this up, we heard it from customers. This notion of Cloud Smart, right? How do they use cloud more effectively, more intelligently? Now, if you walk through this journey, customers start with Cloud First. They usually select a single cloud that they're going to standardize on, and when they do that, they have to build out a whole bunch of functionality around that cloud. Things you can see there on the screen, disaster recovery, security, how do they monitor it or govern it? Like, these are things that are non-negotiable, you've got to figure it out, and typically what they do is, they leverage solutions that are specific for that cloud, and that's fine when you have just one cloud. But if we build out here, what we see is that most customers are using more than just one, they're actually using multiple, not necessarily 10 or however many on the screen, but this is just as an example. And so what happens is, they have to essentially duplicate or replicate that stack they've built for each different cloud, and they do so in a kind of a siloed manner. This results in the Cloud Chaos term that that we talked about before. And this is where most businesses out there are, they're using two, maybe three public clouds. They've got some stuff on-prem and they've also got some stuff out at the edge. This is apps, data, et cetera. So, this is the situation, this is sort of that Cloud Chaos. So, the question is, how do we move from this phase to Cloud Smart? And this is where the architecture comes in. This is why architecture, I think, is so important. It's really about moving away from these single cloud services that just solve a problem for one cloud, to something we call a Cross-Cloud service. Something that can support a set of functionality across all clouds, and that means not just public clouds, but also private clouds, edge, et cetera, and when you evolve that across the board, what you get is this sort of Supercloud. This notion that we're talking about here, where you combine these cross-cloud services in many different categories. You can see some examples there on the screen. This is not meant to be a complete set of things, but just examples of what can be done. So, this is sort of the transition and transformation that we're talking about here, and I think the architecture piece comes in both for the individual cloud services as well as that Supercloud concept of how all those services come together. >> Great presentation., thanks for sharing. If you could pop back to that slide, on the Cloud Chaos one. I just want to get your thoughts on something there. This is like the layout of the stack. So, this slide here that I'm showing on the screen, that you presented, okay, take us through that complexity. This is the one where I wanted though, that looks like a spaghetti code mix. >> Yes. >> So, do you turn this into a Supercloud stack, right? Is that? >> well, I think it's, it's an evolving state that like, let's take one of these examples, like security. So, instead of implementing security individually in different ways, using different technologies, different tooling for each cloud, what you would do is say, "Hey, I want a single security solution that works across all clouds", right? A concrete example of this would be secure software supply chain. This is probably one of the top ones that I hear when I talk to customers. How do I know that the software I'm building is truly what I expect it to be, and not something that some hacker has gotten into, and polluted with malicious code? And what they do is that, typically today, their teams have gone off and created individual secure software supply chain solutions for each cloud. So, now they could say, "Hey, I can take a single implementation and just have different endpoints." It could go to Google, or AWS, or on-prem, or wherever have you, right? So, that's the sort of architectural evolution that we're talking about. >> You know, one of the things we hear, Dave, you've been on theCUBE all the time, and we, when we talk privately with customers who are asking us like, what's, what's going on? They have the same complaint, "I don't want to build a team, a dev team, for that stack." So, if you go back to that slide again, you'll see that, that illustrates the tech stack for the clouds and the clouds at the bottom. So, the number one complaint we hear, and I want to get your reaction to that, "I don't want to have a team to have to work on that. So, I'm going to pick one and then have a hedge secondary one, as a backup." Here, that's one, that's four, five, eight, ten, ten environments. >> Yeah, I got a lot. >> That's going to be the reality, so, what's the technical answer to that? >> Yeah, well first of all, let me just say, this picture is again not totally representative of reality oftentimes, because while that picture shows a solution for every cloud, oftentimes that's not the case. Oftentimes it's a line of business going off, starting to use a new cloud. They might solve one or two things, but usually not security, usually not some of these other things, right? So, I think from a technical standpoint, where you want to get to is, yes, that sort of common service, with a common operational team behind it, that is trained on that, that can work across clouds. And that's really I think the important evolution here, is that you don't need to replicate these operational teams, one for each cloud. You can actually have them more focused across all those clouds. >> Yeah, in fact, we were commenting on the opening today. Dave and I were talking about the benefits of the cloud. It's heterogeneous, which is a good thing, but it's complex. There's skill gaps and skill required, but at the end of the day, self-service of the cloud, and the elastic nature of it makes it the benefit. So, if you try to create too many common services, you lose the value of the cloud. So, what's the trade off, in your mind right now as customers start to look at okay, identity, maybe I'll have one single sign on, that's an obvious one. Other ones? What are the areas people are looking at from a combination, common set of services? Where do they start? What's the choices? What are some of the trade offs? 'Cause you can't do it everything. >> No, it's a great question. So, that's actually a really good point and as I answer your question, before I answer your question, the important point about that, as you saw here, you know, across cloud services or these set of Cross-Cloud services, the things that comprise the Supercloud, at least in my view, the point is not necessarily to completely abstract the underlying cloud. The point is to give a business optionality and choice, in terms of what it wants to abstract, and I think that gets to your question, is how much do you actually want to abstract from the underlying cloud? Now, what I find, is that typically speaking, cloud choice is driven at least from a developer or app team perspective, by the best of breed services. What higher level application type services do you need? A database or AI, you know, ML systems, for your application, and that's going to drive your choice of the cloud. So oftentimes, businesses I talk to, want to allow those services to shine through, but for other things that are not necessarily highly differentiated and yet are absolutely critical to creating a successful application, those are things that you want to standardize. Again, like things like security, the supply chain piece, cost management, like these things you need to, and you know, things like cogs become really, really important when you start operating at scale. So, those are the things in it that I see people wanting to focus on. >> So, there's a majority model. >> Yes. >> All right, and we heard of earlier from Walmart, who's fairly, you know, advanced, but at the same time their supercloud is pretty immature. So, what are you seeing in terms of supercloud momentum, crosscloud momentum? What's the starting point for customers? >> Yeah, so it's interesting, right, on that that three-tiered journey that I talked about, this Cloud Smart notion is, that is adoption of what you might call a supercloud or architecture, and most folks aren't there yet. Even the really advanced ones, even the really large ones, and I think it's because of the fact that, we as an industry are still figuring this out. We as an industry did not realize this sort of Cloud Chaos state could happen, right? We didn't, I think most folks thought they could standardize on one cloud and that'd be it, but as time has shown, that's simply not the case. As much as one might try to do that, that's not where you end up. So, I think there's two, there's two things here. Number one, for folks that are early in to the cloud, and are in this Cloud Chaos phase, we see the path out through standardization of these cross-cloud services through adoption of this sort of supercloud architecture, but the other thing I think is particularly exciting, 'cause I talked to a number of of businesses who are not yet in the Cloud Chaos phase. They're earlier on in the cloud journey, and I think the opportunity there is that they don't have to go through Cloud Chaos. They can actually skip that whole phase if they adopt this supercloud architecture from the beginning, and I think being thoughtful around that is really the key here. >> It's interesting, 'cause we're going to hear from Ionis Pharmaceuticals later, and they, yes there are multiple clouds, but the multiple clouds are largely separate, and so it's a business unit using that. So, they're not in Cloud Chaos, but they're not tapping the advantages that you could get for best of breed across those business units. So, to your point, they have an opportunity to actually build that architecture or take advantage of those cross-cloud services, prior to reaching cloud chaos. >> Well, I, actually, you know, I'd love to hear from them if, 'cause you say they're not in Cloud Chaos, but are they, I mean oftentimes I find that each BU, each line of business may feel like they're fine, in of themselves. >> Yes, exactly right, yes. >> But when you look at it from an overall company perspective, they're like, okay, things are pretty chaotic here. We don't have standardization, I don't, you know, like, again, security compliance, these things, especially in many regulated industries, become huge problems when you're trying to run applications across multiple clouds, but you don't have any of those company-wide standardizations. >> Well, this is a point. So, they have a big deal with AstraZeneca, who's got this huge ecosystem, they want to start sharing data across those ecosystem, and that's when they will, you know, that Cloud Chaos will, you know, come, come to fore, you would think. I want to get your take on something that Bob Muglia said earlier, which is, he kind of said, "Hey Dave, you guys got to tighten up your definition a little bit." So, he said a supercloud is a platform that provides programmatically consistent services hosted on heterogeneous cloud providers. So, you know, thank you, that was nice and simple. However others in the community, we're going to hear from Dr. Nelu Mihai later, says, no, no, wait a minute, it's got to be an architecture, not a platform. Where do you land on this architecture v. platform thing? >> I look at it as, I dunno if it's, you call it maturity or just kind of a time horizon thing, but for me when I hear the word platform, I typically think of a single vendor. A single vendor provides this platform. That's kind of the beauty of a platform, is that there is a simplicity usually consistency to it. >> They did the architecture. (laughing) >> Yeah, exactly but I mean, well, there's obviously architecture behind it, has to be, but you as a customer don't necessarily need to deal with that. Now, I think one of the opportunities with Supercloud is that it's not going to be, or there is no single vendor that can solve all these problems. It's got to be the industry coming together as a community, inter-operating, working together, and so, that's why, for me, I think about it as an architecture, that there's got to be these sort of, well-defined categories of functionality. There's got to be well-defined interfaces between those categories of functionality to enable modularity, to enable businesses to be able to pick and choose the right sorts of services, and then weave those together into an overall supercloud. >> Okay, so you're not pitching, necessarily the platform, you're saying, hey, we have an architecture that's open. I go back to something that Vittorio said on August 9th, with the first Supercloud, because as well, remember we talked about abstracting, but at the same time giving developers access to those primitives. So he said, and this, I think your answer sort of confirms this. "I want to have my cake eat it too and not gain weight." >> (laughing) Right. Well and I think that's where the platform aspect can eventually come, after we've gotten aligned architecture, you're going to start to naturally see some vendors step up to take on some of the remaining complexity there. So, I do see platforms eventually emerging here, but I think where we have to start as an industry is around aligning, okay, what does this definition mean? What does that architecture look like? How do we enable interoperability? And then we can take the next step. >> Because it depends too, 'cause I would say Snowflake has a platform, and they've just defined the architecture, but we're not talking about infrastructure here, obviously, we're talking about something else. >> Well, I think that the Snowflake talks about, what he talks about, security and data, you're going to start to see the early movement around areas that are very spanning oriented, and I think that's the beginning of the trend and I think there's going to be a lot more, I think on the infrastructure side. And to your point about the platform architecture, that's actually a really good thought exercise because it actually makes you think about what you're designing in the first place, and that's why I want to get your reaction. >> Quote from- >> Well I just have to interrupt since, later on, you're going to hear from near Nir Zuk of Palo Alto Network. He says architecture and security historically, they don't go hand in hand, 'cause it's a big mess. >> It depends if you're whacking the mole or you actually proactively building something. Well Kit, I want to get your reaction from a quote from someone in our community who said about Supercloud, you know, "The Supercloud's great, there are issues around computer science rigors, and customer requirements." So, there's some issues around the science itself as well as not just listen to the customer, 'cause if that's the case, we'd have a better database, a better Oracle, right, so, but there's other, this tech involved, new tech. We need an open architecture with universal data modeling interconnecting among them, connectivity is a part of security, and then, once we get through that gate, figuring out the technical, the data, and the customer requirements, they say "Supercloud should be a loosely coupled platform with open architecture, plug and play, specialized services, ready for optimization, automation that can stand the test of time." What's your reaction to that sentiment? You like it, is that, does that sound good? >> Yeah, no, broadly aligns with my thinking, I think, and what I see from talking with customers as well. I mean, I like the, again, the, you know, listening to customer needs, prioritizing those things, focusing on some of the connective tissue networking, and data and some of these aspects talking about the open architecture, the interoperability, those are all things I think are absolutely critical. And then, yeah, like I think at the end. >> On the computer science side, do you see some science and engineering things that need to be engineered differently? We heard databases are radically going to change and that are inadequate for the new architecture. What are some of the things like that, from a science standpoint? >> Yeah, yeah, yeah. Some of the more academic research type things. >> More tech, or more better tech or is it? >> Yeah, look, absolutely. I mean I think that there's a bunch around, certainly around the data piece, around, you know, there's issues of data gravity, data mobility. How do you want to do that in a way that's performant? There's definitely issues around security as well. Like how do you enable like trust in these environments, there's got to be some sort of hardware rooted trusts, and you know, a whole bunch of various types of aspects there. >> So, a lot of work still be done. >> Yes, I think so. And that's why I look at this as, this is not a one year thing, or you know, it's going to be multi-years, and I think again, it's about all of us in the industry working together to come to an aligned picture of what that looks like. >> So, as the world's moved from private cloud to public cloud and now Cross-cloud services, supercloud, metacloud, whatever you want to call it, how have you sort of changed the way engineering's organized, developers sort of approached the problem? Has it changed and how? >> Yeah, absolutely. So, you know, it's funny, we at VMware, going through the same challenges as our customers and you know, any business, right? We use multiple clouds, we got a big, of course, on-prem footprint. You know, what we're doing is similar to what I see in many other customers, which, you see the evolution of a platform team, and so the platform team is really in charge of trying to develop a lot of these underlying services to allow our lines of business, our product teams, to be able to move as quickly as possible, to focus on the building, while we help with a lot of the operational overheads, right? We maintain security, compliance, all these other things. We also deal with, yeah, just making the developer's life as simple as possible. So, they do need to know some stuff about, you know, each public cloud they're using, those public cloud services, but at the same, time we can abstract a lot of the details they don't need to be in. So, I think this sort of delineation or separation, I should say, between the underlying platform team and the product teams is a very, very common pattern. >> You know, I noticed the four layers you talked about were observability, infrastructure, security and developers, on that slide, the last slide you had at the top, that was kind of the abstraction key areas that you guys at VMware are working? >> Those were just some groupings that we've come up with, but we like to debate them. >> I noticed data's in every one of them. >> Yeah, yep, data is key. >> It's not like, so, back to the data questions that security is called out as a pillar. Observability is just kind of watching everything, but it's all pretty much data driven. Of the four layers that you see, I take that as areas that you can. >> Standardize. >> Consistently rely on to have standard services. >> Yes. >> Which one do you start with? What's the, is there order of operations? >> Well, that's, I mean. >> 'Cause I think infrastructure's number one, but you had observability, you need to know what's going on. >> Yeah, well it really, it's highly dependent. Again, it depends on the business that we talk to and what, I mean, it really goes back to, what are your business priorities, right? And we have some customers who may want to get out of a data center, they want to evacuate the data center, and so what they want is then, consistent infrastructure, so they can just move those applications up to the cloud. They don't want to have to refactor them and we'll do it later, but there's an immediate and sort of urgent problem that they have. Other customers I talk to, you know, security becomes top of mind, or maybe compliance, because they're in a regulated industry. So, those are the sort of services they want to prioritize. So, I would say there is no single right answer, no one size fits all. The point about this architecture is really around the optionality of it, as it allows you as a business to decide what's most important and where you want to prioritize. >> How about the deployment models kit? Do, does a customer have that flexibility from a deployment model standpoint or do I have to, you know, approach it a specific way? Can you address that? >> Yeah, I mean deployment models, you're talking about how they how they consume? >> So, for instance, yeah, running a control plane in the cloud. >> Got it, got it. >> And communicating elsewhere or having a single global instance or instantiating that instance, and? >> So, that's a good point actually, and you know, the white paper that we released back in August, around this sort of concept, the Cross-cloud service. This is some of the stuff we need to figure out as an industry. So, you know when we talk about a Cross-cloud service, we can mean actually any of the things you just talked about. It could be a single instance that runs, let's say in one public cloud, but it supports all of 'em. Or it could be one that's multi-instance and that runs in each of the clouds, and that customers can take dependencies on whichever one, depending on what their use cases are or the, even going further than that, there's a type of Cross-cloud service that could actually be instantiated even in an air gapped or offline environment, and we have many, many businesses, especially heavily regulated ones that have that requirement, so I think, you know. >> Global don't forget global, regions, locales. >> Yeah, there's all sorts of performance latency issues that can be concerned about. So, most services today are the former, there are single sort of instance or set of instances within a single cloud that support multiple clouds, but I think what we're doing and where we're going with, you know, things like what we see with Kubernetes and service meshes and all these things, will better enable folks to hit these different types of cross-cloud service architectures. So, today, you as a customer probably wouldn't have too much choice, but where we're going, you'll see a lot more choice in the future. >> If you had to summarize for folks watching the importance of Supercloud movement, multi-cloud, cross-cloud services, as an industry in flexible, 'cause I'm always riffing on the whole old school network protocol stacks that got disrupted by TCP/IP, that's a little bit dated, we got people on the chat that are like, you know, 20 years old that weren't even born then. So, but this is a, one of those inflection points that's once in a generation inflection point, I'm sure you agree. What scoped the order of magnitude of the change and the opportunity around the marketplace, the business models, the technology, and ultimately benefits the society. >> Yeah. Wow. Getting bigger. >> You have 10 seconds, go. >> I know. Yeah. (laughing) No, look, so I think it is what we're seeing is really the next phase of what you might call cloud, right? This notion of delivering services, the way they've been packaged together, traditionally by the hyperscalers is now being challenged. and what we're seeing is really opening that up to new levels of innovation, and I think that will be huge for businesses because it'll help meet them where they are. Instead of needing to contort the businesses to, you know, make it work with the technology, the technology will support the business and where it's going. Give people more optionality, more flexibility in order to get there, and I think in the end, for us as individuals, it will just make for better experiences, right? You can get better performance, better interactivity, given that devices are so much of what we do, and so much of what we interact with all the time. This sort of flexibility and optionality will fundamentally better for us as individuals in our experiences. >> And we're seeing that with ChatGPT, everyone's talking about, just early days. There'll be more and more of things like that, that are next gen, like obviously like, wow, that's a fall out of your chair moment. >> It'll be the next wave of innovation that's unleashed. >> All right, Kit Colbert, thanks for coming on and sharing and exploring the Supercloud architecture, Cloud Chaos, the Cloud Smart, there's a transition progression happening and it's happening fast. This is the supercloud wave. If you're not on this wave, you'll be driftwood. That's a Pat Gelsinger quote on theCUBE. This is theCUBE Be right back with more Supercloud coverage, here in Palo Alto after this break. (upbeat music) (upbeat music continues)
SUMMARY :
We've got Kit Colbert, the CTO of VM. It's great to be here for Supercloud 2. We're going to let you present. and when you evolve that across the board, This is like the layout of the stack. How do I know that the So, the number one complaint we hear, is that you don't need to replicate and the elastic nature of and I think that gets to your question, So, what are you seeing in terms but the other thing I think that you could get for best of breed Well, I, actually, you know, I don't, you know, like, and that's when they will, you know, That's kind of the beauty of a platform, They did the architecture. is that it's not going to be, but at the same time Well and I think that's and they've just defined the architecture, beginning of the trend Well I just have to and the customer requirements, focusing on some of the that need to be engineered differently? Some of the more academic and you know, a whole bunch or you know, it's going to be multi-years, of the details they don't need to be in. that we've come up with, Of the four layers that you see, to have standard services. but you had observability, you is really around the optionality of it, running a control plane in the cloud. and that runs in each of the clouds, Global don't forget and where we're going with, you know, and the opportunity of what you might call cloud, right? that are next gen, like obviously like, It'll be the next wave of and exploring the Supercloud architecture,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Bob Muglia | PERSON | 0.99+ |
Kit Colbert | PERSON | 0.99+ |
August 9th | DATE | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Pat Gelsinger | PERSON | 0.99+ |
10 seconds | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Ionis Pharmaceuticals | ORGANIZATION | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
AstraZeneca | ORGANIZATION | 0.99+ |
Nelu Mihai | PERSON | 0.99+ |
August | DATE | 0.99+ |
two things | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Supercloud | ORGANIZATION | 0.99+ |
Vittorio | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
one year | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
each | QUANTITY | 0.99+ |
Kit | PERSON | 0.99+ |
three | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
today | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
each cloud | QUANTITY | 0.98+ |
one cloud | QUANTITY | 0.97+ |
each cloud | QUANTITY | 0.97+ |
ten | QUANTITY | 0.97+ |
VMware | ORGANIZATION | 0.96+ |
five | QUANTITY | 0.96+ |
single cloud | QUANTITY | 0.96+ |
single | QUANTITY | 0.96+ |
each line | QUANTITY | 0.96+ |
supercloud wave | EVENT | 0.96+ |
single instance | QUANTITY | 0.95+ |
Palo Alto Network | ORGANIZATION | 0.95+ |
four | QUANTITY | 0.94+ |
eight | QUANTITY | 0.94+ |
single vendor | QUANTITY | 0.94+ |
Cloud Chaos | TITLE | 0.94+ |
Nir Zuk | PERSON | 0.94+ |
three-tiered | QUANTITY | 0.93+ |
Cloud First | TITLE | 0.91+ |
four layers | QUANTITY | 0.91+ |
Cloud Smart | TITLE | 0.91+ |
Supercloud | TITLE | 0.89+ |
single implementation | QUANTITY | 0.88+ |
Supercloud 2 | EVENT | 0.87+ |
first place | QUANTITY | 0.84+ |
single right answer | QUANTITY | 0.84+ |
once | QUANTITY | 0.83+ |
single sort | QUANTITY | 0.82+ |
Is Supercloud an Architecture or a Platform | Supercloud2
(electronic music) >> Hi everybody, welcome back to Supercloud 2. I'm Dave Vellante with my co-host John Furrier. We're here at our tricked out Palo Alto studio. We're going live wall to wall all day. We're inserting a number of pre-recorded interviews, folks like Walmart. We just heard from Nir Zuk of Palo Alto Networks, and I'm really pleased to welcome in David Flynn. David Flynn, you may know as one of the people behind Fusion-io, completely changed the way in which people think about storing data, accessing data. David Flynn now the founder and CEO of a company called Hammerspace. David, good to see you, thanks for coming on. >> David: Good to see you too. >> And Dr. Nelu Mihai is the CEO and founder of Cloud of Clouds. He's actually built a Supercloud. We're going to get into that. Nelu, thanks for coming on. >> Thank you, Happy New Year. >> Yeah, Happy New Year. So I'm going to start right off with a little debate that's going on in the community if you guys would bring out this slide. So Bob Muglia early today, he gave a definition of Supercloud. He felt like we had to tighten ours up a little bit. He said a Supercloud is a platform, underscoring platform, that provides programmatically consistent services hosted on heterogeneous cloud providers. Now, Nelu, we have this shared doc, and you've been in there. You responded, you said, well, hold on. Supercloud really needs to be an architecture, or else we're going to have this stove pipe of stove pipes, really. And then you went on with more detail, what's the information model? What's the execution model? How are users going to interact with Supercloud? So I start with you, why architecture? The inference is that a platform, the platform provider's responsible for the architecture? Why does that not work in your view? >> No, the, it's a very interesting question. So whenever I think about platform, what's the connotation, you think about monolithic system? Yeah, I mean, I don't know whether it's true or or not, but there is this connotation of of monolithic. On the other hand, if you look at what's a problem right now with HyperClouds, from the customer perspective, they're very complex. There is a heterogeneous world where actually every single one of this HyperClouds has their own architecture. You need rocket scientists to build a cloud applications. Always there is this contradiction between cost and performance. They fight each other. And I'm quoting here a former friend of mine from Bell Labs who work at AWS who used to say "Cloud is cheap as long as you don't use it too much." (group chuckles) So clearly we need something that kind of plays from the principle point of view the role of an operating system, that seats on top of this heterogeneous HyperCloud, and there's nothing wrong by having these proprietary HyperClouds, think about processors, think about operating system and so on, so forth. But in order to build a system that is simple enough, I think we need to go deeper and understand. >> So the argument, the counterargument to that, David, is you'll never get there. You need a proprietary system to get to market sooner, to solve today's problem. Now I don't know where you stand on this platform versus architecture. I haven't asked you, but. >> I think there are aspects of both for sure. I mean it needs to be an architecture in the sense that it's broad based and open and so forth. But you know, platform, you could say as long as people can instantiate it themselves, on their own infrastructure, as long as it's something that can be deployed as, you know, software defined, you don't want the concept of platform being the monolith, you know, combined hardware and software. So it really depends on what you're focused on when you're saying platform, you know, I'd say as long as they software defined thing, to where it can literally run anywhere. I mean, because I really think what we're talking about here is the original concept of cloud computing. The ability to run anything anywhere, without having to care about the physical infrastructure. And what we have today is not that, the cloud today is a big mainframe in the sky, that just happens to be large enough that once you select which region, generally you have enough resources. But, you know, nowadays you don't even necessarily have enough resources in one region. and then you're kind of stuck. So we haven't really gotten to that utility model of computing. And you're also asked to rewrite your application, you know, to abandon the conveniences of high performance file access. You got to rewrite it to use object storage stuff. We have to get away from that. >> Okay, I want to just drill on that, 'cause I think I like that point about, there's not enough availability, but on the developer cloud, the original AWS premise was targeting developers, 'cause at that time, you have to provision a Sun box get a Cisco DSU/CSU, now you get on the cloud. But I think you're giving up the scale question, 'cause I think right now, scale is huge, enterprise grade versus cloud for developers. >> That's Right. >> Because I mean look at, Amazon, Azure, they got compute, they got storage, they got queuing, and some stuff. If you're doing a startup, you throw your app up there, localhost to cloud, no big deal. It's the scale thing that gets me- >> And you can tell by the fact that, in regions that are under high demand, right, like in London or LA, at least with the clients we work with in the median entertainment space, it costs twice as much for the exact same cloud instances that do the exact same amount of work, as somewhere out in rural Canada. So why is it you have such a cost differential, it has to do with that supply and demand, and the fact that the clouds aren't really the ability to run anything anywhere. Even within the same cloud vendor, you're stuck in a specific region. >> And that was never the original promise, right? I mean it was, we turned it into that. But the original promise was get rid of the heavy lifting of IT. >> Not have to run your own, yeah, exactly. >> And then it became, wow, okay I can run anywhere. And then you know, it's like web 2.0. You know people say why Supercloud, you and I talked about this, why do you need a name for Supercloud? It's like web 2.0. >> It's what Cloud was supposed to be. >> It's what cloud was supposed to be, (group laughing and talking) exactly, right. >> Cloud was supposed to be run anything anywhere, or at least that's what we took it as. But you're right, originally it was just, oh don't have to run your own infrastructure, and you can choose somebody else's infrastructure. >> And you did that >> But you're still bound to that. >> Dave: And People said I want more, right? >> But how do we go from here? >> That's, that's actually, that's a very good point, because indeed when the first HyperClouds were designed, were designed really focus on customers. I think Supercloud is an opportunity to design in the right way. Also having in mind the computer science rigor. And we should take advantage of that, because in fact actually, if cloud would've been designed properly from the beginning, probably wouldn't have needed Supercloud. >> David: You wouldn't have to have been asked to rewrite your application. >> That's correct. (group laughs) >> To use REST interfaces to your storage. >> Revisist history is always a good one. But look, cloud is great. I mean your point is cloud is a good thing. Don't hold it back. >> It is a very good thing. >> Let it continue. >> Let it go as as it is. >> Yeah, let that thing continue to grow. Don't impose restrictions on the cloud. Just refactor what you need to for scale or enterprise grade or availability. >> And you would agree with that, is that true or is it problem you're solving? >> Well yeah, I mean it, what the cloud is doing is absolutely necessary. What the public cloud vendors are doing is absolutely necessary. But what's been missing is how to provide a consistent interface, especially to persistent data. And have it be available across different regions, and across different clouds. 'cause data is a highly localized thing in current architecture. It only exists as rendered by the storage system that you put it in. Whether that's a legacy thing like a NetApp or an Isilon or even a cloud data service. It's localized to a specific region of the cloud in which you put that. We have to delocalize data, and provide a consistent interface to it across all sites. That's high performance, local access, but to global data. >> And so Walmart earlier today described their, what we call Supercloud, they call it the Walmart cloud native platform. And they use this triplet model. They have AWS and Azure, no, oh sorry, no AWS. They have Azure and GCP and then on-prem, where all the VMs live. When you, you know, probe, it turns out that it's only stateless in the cloud. (John laughs) So, the state stuff- >> Well let's just admit it, there is no such thing as stateless, because even the application binaries and libraries are state. >> Well I'm happy that I'm hearing that. >> Yeah, okay. >> Because actually I have a lot of debate (indistinct). If you think about no software running on a (indistinct) machine is stateless. >> David: Exactly. >> This is something that was- >> David: And that's data that needs to be distributed and provided consistently >> (indistinct) >> Across all the clouds, >> And actually, it's a nonsense, but- >> Dave: So it's an illusion, okay. (group talks over each other) >> (indistinct) you guys talk about stateless. >> Well, see, people make the confusion between state and persistent state, okay. Persistent state it's a different thing. State is a different thing. So, but anyway, I want to go back to your point, because there's a lot of debate here. People are talking about data, some people are talking about logic, some people are talking about networking. In my opinion is this triplet, which is data logic and connectivity, that has equal importance. And actually depending on the application, can have the center of gravity moving towards data, moving towards what I call execution units or workloads. And connectivity is actually the most important part of it. >> David: (indistinct). >> Some people are saying move the logic towards the data, some other people, and you are saying actually, that no, you have to build a distributed data mesh. What I'm saying is actually, you have to consider all these three variables, all these vector in order to decide, based on application, what's the most important. Because sometimes- >> John: So the application chooses >> That's correct. >> Well it it's what operating systems were in the past, was principally the thing that runs and manages the jobs, the job scheduler, and the thing that provides your persistent data (indistinct). >> Okay. So we finally got operating system into the equation, thank you. (group laughs) >> Nelu: I actually have a PhD in operating system. >> Cause what we're talking about is an operating system. So forget platform or architecture, it's an operating environment. Let's use it as a general term. >> All right. I think that's about it for me. >> All right, let's take (indistinct). Nelu, I want ask you quick, 'cause I want to give a, 'cause I believe it's an operating system. I think it's going to be a reset, refactored. You wrote to me, "The model of Supercloud has to be open theoretical, has to satisfy the rigors of computer science, and customer requirements." So unique to today, if the OS is going to be refactored, it's not going to be, may or may not be Red Hat or somebody else. This new OS, obviously requirements are for customers too but is what's the computer science that is needed? Where are we, what's the missing? Where's the science in this shift? It's not your standard OS it's not like an- (group talks over each other) >> I would beg to differ. >> (indistinct) truly an operation environment. But the, if you think about, and make analogies, what you need when you design a distributed system, well you need an information model, yeah. You need to figure out how the data is located and distributed. You need a model for the execution units, and you need a way to describe the interactions between all these objects. And it is my opinion that we need to go deeper and formalize these operations in order to make a step forward. And when we design Supercloud, and design something that is better than the current HyperClouds. And actually that is when we design something better, you make a system more efficient and it's going to be better from the cost point of view, from the performance point of view. But we need to add some math into all this customer focus centering and I really admire AWS and their executive team focusing on the customer. But now it's time to go back and see, if we apply some computer science, if you try to formalize to build a theoretical model of cloud, can we build a system that is better than existing ones? >> So David, how do you- >> this is what I'm saying. >> That's a good question >> How do You see the operating system of a, or operating environment of a decentralized cloud? >> Well I think it's layered. I mean we have operating systems that can run systems quite efficiently. Linux has sort of one in the data center, but we're talking about a layer on top of that. And I think we're seeing the emergence of that. For example, on the job scheduling side of things, Kubernetes makes a really good example. You know, you break the workload into the most granular units of compute, the containerized microservice, and then you use a declarative model to state what is needed and give the system the degrees of freedom that it can choose how to instantiate it. Because the thing about these distributed systems, is that the complexity explodes, right? Running a piece of hardware, running a single server is not a problem, even with all the many cores and everything like that. It's when you start adding in the networking, and making it so that you have many of them. And then when it's going across whole different data centers, you know, so, at that level the way you solve this is not manually (group laughs) and not procedurally. You have to change the language so it's intent based, it's a declarative model, and what you're stating is what is intended, and you're leaving it to more advanced techniques, like machine learning to decide how to instantiate that service across the cluster, which is what Kubernetes does, or how to instantiate the data across the diverse storage infrastructure. And that's what we do. >> So that's a very good point because actually what has been neglected with HyperClouds is really optimization and automation. But in order to be able to do both of these things, you need, I'm going back and I'm stubborn, you need to have a mathematical model, a theoretical model because what does automation mean? It means that we have to put machines to do the work instead of us, and machines work with what? Formula, with algorithms, they don't work with services. So I think Supercloud is an opportunity to underscore the importance of optimization and automation- >> Totally agree. >> In HyperCloud, and actually by doing that, we can also have an interesting connotation. We are also contributing to save our planet, because if you think right now. we're consuming a lot of energy on this HyperClouds and also all this AI applications, and I think we can do better and build the same kind of application using less energy. >> So yeah, great point, love that call out, the- you know, Dave and I always joke about the old, 'cause we're old, we talk about, you know, (Nelu Laughs) old history, OS/2 versus DOS, okay, OS's, OS/2 is silly better, first threaded OS, DOS never went away. So how does legacy play into this conversation? Because I buy the theoretical, I love the conversation. Okay, I think it's an OS, totally see it that way myself. What's the blocker? Is there a legacy that drags it back? Is the anchor dragging from legacy? Is there a DOS OS/2 moment? Is there an opportunity to flip the script? This is- >> I think that's a perfect example of why we need to support the existing interfaces, Operating Systems, real operating systems like Linux, understands how to present data, it's called a file system, block devices, things that that plumb in there. And by, you know, going to a REST interface and S3 and telling people they have to rewrite their applications, you can't even consume your application binaries that way, the OS doesn't know how to pull that sort of thing. So we, to get to cloud, to get to the ability to host massive numbers of tenants within a centralized infrastructure, you know, we abandoned these lower level interfaces to the OS and we have to go back to that. It's the reason why DOS ultimately won, is it had the momentum of the install base. We're seeing the same thing here. Whatever it is, it has to be a real file system and not a come down file system >> Nelu, what's your reaction, 'cause you're in the theoretical bandwagon. Let's get your reaction. >> No, I think it's a good, I'll give, you made a good analogy between OS/2 and DOS, but I'll go even farther saying, if you think about the evolution operating system didn't stop the evolution of underlying microprocessors, hardware, and so on and so forth. On the contrary, it was a catalyst for that. So because everybody could develop their own hardware, without worrying that the applications on top of operating system are going to modify. The same thing is going to happen with Supercloud. You're going to have the AWSs, you're going to have the Azure and the the GCP continue to evolve in their own way proprietary. But if we create on top of it the right interface >> The open, this is why open is important. >> That's correct, because actually you're going to see sometime ago, everybody was saying, remember venture capitals were saying, "AWS killed the world, nobody's going to come." Now you see what Oracle is doing, and then you're going to see other players. >> It's funny, Amazon's trying to be more like Microsoft. Microsoft's trying to be more like Amazon and Google- Oracle's just trying to say they have cloud. >> That's, that's correct, (group laughs) so, my point is, you're going to see a multiplication of this HyperClouds and cloud technology. So, the system has to be open in order to accommodate what it is and what is going to come. Okay, so it's open. >> So the the legacy- so legacy is an opportunity, not a blocker in your mind. And you see- >> That's correct, I think we should allow them to continue to to to be their own actually. But maybe you're going to find a way to connect with it. >> Amazon's the processor, and they're on the 80 80 80 right? >> That's correct. >> You're saying you love people trying to get put to work. >> That's a good analogy. >> But, performance levels you say good luck, right? >> Well yeah, we have to be able to take traditional applications, high performance applications, those that consume file system and persistent data. Those things have to be able to run anywhere. You need to be able to put, put them onto, you know, more elastic infrastructure. So, we have to actually get cloud to where it lives up to its billing. >> And that's what you're solving for, with Hammerspace, >> That's what we're solving for, making it possible- >> Give me the bumper sticker. >> Solving for how do you have massive quantities of unstructured file data? At the end of the day, all data ultimately is unstructured data. Have that persistent data available, across any data center, within any cloud, within any region on-prem, at the edge. And have not just the same APIs, but have the exact same data sets, and not sucked over a straw remote, but at extreme high performance, local access. So how do you have local access to globally shared distributed data? And that's what we're doing. We are orchestrating data globally across all different forms of storage infrastructure, so you have a consistent access at the highest performance levels, at the lowest level innate built into the OS, how to consume it as (indistinct) >> So are you going into the- all the clouds and natively building in there, or are you off cloud? >> So This is software that can run on cloud instances and provide high performance file within the cloud. It can take file data that's on-prem. Again, it's software, it can run in virtual or on physical servers. And it abstracts the data from the existing storage infrastructure, and makes the data visible and consumable and orchestratable across any of it. >> And what's the elevator pitch for Cloud of Cloud, give that too. >> Well, Cloud of Clouds creates a theoretical model of cloud, and it describes every single object in the cloud. Where is data, execution units, and connectivity, with one single class of very simple object. And I can, I can give you (indistinct) >> And the problem that solves is what? >> The problem that solves is, it creates this mathematical model that is necessary in order to do other interesting things, such as optimization, using sata engines, using automation, applying ML for instance. Or deep learning to automate all this clouds, if you think about in the industrial field, we know how to manage and automate huge plants. Why wouldn't it do the same thing in cloud? It's the same thing you- >> That's what you mean by theoretical model. >> Nelu: That's correct. >> Lay out the architecture, almost the bones of skeleton or something, or, and then- >> That's correct, and then on top of it you can actually build a platform, You can create your services, >> when you say math, you mean you put numbers to it, you kind of index it. >> You quantify this thing and you apply mathematical- It's really about, I can disclose this thing. It's really about describing the cloud as a knowledge graph for every single object in the graph for node, an edge is a vector. And then once you have this model, then you can apply the field theory, and linear algebra to do operation with these vectors. And it's, this creates a very interesting opportunity to let the math do this thing for us. >> Okay, so what happens with hyperscale, or it's like AWS in your model. >> So in, in my model actually, >> Are they happy with this, or they >> I'm very happy with that. >> Will they be happy with you? >> We create an interface to every single HyperCloud. We actually, we don't need to interface with the thousands of APIs, but you know, if we have the 80 20 rule, and we map these APIs into this graph, and then every single operation that is done in this graph is done from the beginning, in an optimized manner and also automation ready. >> That's going to be great. David, I want us to go back to you before we close real quick. You've had a lot of experience, multiple ventures on the front end. You talked to a lot of customers who've been innovating. Where are the classic (indistinct)? Cause you, you used to sell and invent product around the old school enterprises with storage, you know that that trajectory storage is still critical to store the data. Where's the classic enterprise grade mindset right now? Those customers that were buying, that are buying storage, they're in the cloud, they're lifting and shifting. They not yet put the throttle on DevOps. When they look at this Supercloud thing, Are they like a deer in the headlights, or are they like getting it? What's the, what's the classic enterprise look like? >> You're seeing people at different stages of adoption. Some folks are trying to get to the cloud, some folks are trying to repatriate from the cloud, because they've realized it's better to own than to rent when you use a lot of it. And so people are at very different stages of the journey. But the one thing that's constant is that there's always change. And the change here has to do with being able to change the location where you're doing your computing. So being able to support traditional workloads in the cloud, being able to run things at the edge, and being able to rationalize where the data ought to exist, and with a declarative model, intent-based, business objective-based, be able to swipe a mouse and have the data get redistributed and positioned across different vendors, across different clouds, that, we're seeing that as really top of mind right now, because everybody's at some point on this journey, trying to go somewhere, and it involves taking their data with them. (John laughs) >> Guys, great conversation. Thanks so much for coming on, for John, Dave. Stay tuned, we got a great analyst power panel coming right up. More from Palo Alto, Supercloud 2. Be right back. (bouncy music)
SUMMARY :
and I'm really pleased to And Dr. Nelu Mihai is the CEO So I'm going to start right off On the other hand, if you look at what's So the argument, the of platform being the monolith, you know, but on the developer cloud, It's the scale thing that gets me- the ability to run anything anywhere. of the heavy lifting of IT. Not have to run your And then you know, it's like web 2.0. It's what Cloud It's what cloud was supposed to be, and you can choose somebody bound to that. Also having in mind the to rewrite your application. That's correct. I mean your point is Yeah, let that thing continue to grow. of the cloud in which you put that. So, the state stuff- because even the application binaries If you think about no software running on Dave: So it's an illusion, okay. (indistinct) you guys talk And actually depending on the application, that no, you have to build the job scheduler, and the thing the equation, thank you. a PhD in operating system. about is an operating system. I think I think it's going to and it's going to be better at that level the way you But in order to be able to and build the same kind of Because I buy the theoretical, the OS doesn't know how to Nelu, what's your reaction, of it the right interface The open, this is "AWS killed the world, to be more like Microsoft. So, the system has to be open So the the legacy- to continue to to to put to work. You need to be able to put, And have not just the same APIs, and makes the data visible and consumable for Cloud of Cloud, give that too. And I can, I can give you (indistinct) It's the same thing you- That's what you mean when you say math, and linear algebra to do Okay, so what happens with hyperscale, the thousands of APIs, but you know, the old school enterprises with storage, and being able to rationalize Stay tuned, we got a
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Nelu | PERSON | 0.99+ |
David Flynn | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
AWS | ORGANIZATION | 0.99+ |
London | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
LA | LOCATION | 0.99+ |
Bob Muglia | PERSON | 0.99+ |
OS/2 | TITLE | 0.99+ |
Nir Zuk | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Hammerspace | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Bell Labs | ORGANIZATION | 0.99+ |
Nelu Mihai | PERSON | 0.99+ |
DOS | TITLE | 0.99+ |
AWSs | ORGANIZATION | 0.99+ |
Palo Alto Networks | ORGANIZATION | 0.99+ |
twice | QUANTITY | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Canada | LOCATION | 0.99+ |
both | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Supercloud | ORGANIZATION | 0.99+ |
Nelu Laughs | PERSON | 0.98+ |
thousands | QUANTITY | 0.98+ |
first | QUANTITY | 0.97+ |
Linux | TITLE | 0.97+ |
HyperCloud | TITLE | 0.97+ |
Cloud of Cloud | TITLE | 0.97+ |
one | QUANTITY | 0.96+ |
Cloud of Clouds | ORGANIZATION | 0.95+ |
GCP | TITLE | 0.95+ |
Azure | TITLE | 0.94+ |
three variables | QUANTITY | 0.94+ |
one single class | QUANTITY | 0.94+ |
single server | QUANTITY | 0.94+ |
triplet | QUANTITY | 0.94+ |
one region | QUANTITY | 0.92+ |
NetApp | TITLE | 0.92+ |
DOS OS/2 | TITLE | 0.92+ |
Azure | ORGANIZATION | 0.92+ |
earlier today | DATE | 0.92+ |
Cloud of Clouds | TITLE | 0.91+ |
Lie 1, The Most Effective Data Architecture Is Centralized | Starburst
(bright upbeat music) >> In 2011, early Facebook employee and Cloudera co-founder Jeff Hammerbacher famously said, "The best minds of my generation are thinking about how to get people to click on ads, and that sucks!" Let's face it. More than a decade later, organizations continue to be frustrated with how difficult it is to get value from data and build a truly agile and data-driven enterprise. What does that even mean, you ask? Well, it means that everyone in the organization has the data they need when they need it in a context that's relevant to advance the mission of an organization. Now, that could mean cutting costs, could mean increasing profits, driving productivity, saving lives, accelerating drug discovery, making better diagnoses, solving supply chain problems, predicting weather disasters, simplifying processes, and thousands of other examples where data can completely transform people's lives beyond manipulating internet users to behave a certain way. We've heard the prognostications about the possibilities of data before and in fairness we've made progress, but the hard truth is the original promises of master data management, enterprise data warehouses, data marts, data hubs, and yes even data lakes were broken and left us wanting for more. Welcome to The Data Doesn't Lie... Or Does It? A series of conversations produced by theCUBE and made possible by Starburst Data. I'm your host, Dave Vellante, and joining me today are three industry experts. Justin Borgman is the co-founder and CEO of Starburst, Richard Jarvis is the CTO at EMIS Health, and Teresa Tung is cloud first technologist at Accenture. Today, we're going to have a candid discussion that will expose the unfulfilled, and yes, broken promises of a data past. We'll expose data lies: big lies, little lies, white lies, and hidden truths. And we'll challenge, age old data conventions and bust some data myths. We're debating questions like is the demise of a single source of truth inevitable? Will the data warehouse ever have feature parity with the data lake or vice versa? Is the so-called modern data stack simply centralization in the cloud, AKA the old guards model in new cloud close? How can organizations rethink their data architectures and regimes to realize the true promises of data? Can and will an open ecosystem deliver on these promises in our lifetimes? We're spanning much of the Western world today. Richard is in the UK, Teresa is on the West Coast, and Justin is in Massachusetts with me. I'm in theCUBE studios, about 30 miles outside of Boston. Folks, welcome to the program. Thanks for coming on. >> Thanks for having us. >> Okay, let's get right into it. You're very welcome. Now, here's the first lie. The most effective data architecture is one that is centralized with a team of data specialists serving various lines of business. What do you think Justin? >> Yeah, definitely a lie. My first startup was a company called Hadapt, which was an early SQL engine for IDU that was acquired by Teradata. And when I got to Teradata, of course, Teradata is the pioneer of that central enterprise data warehouse model. One of the things that I found fascinating was that not one of their customers had actually lived up to that vision of centralizing all of their data into one place. They all had data silos. They all had data in different systems. They had data on prem, data in the cloud. Those companies were acquiring other companies and inheriting their data architecture. So despite being the industry leader for 40 years, not one of their customers truly had everything in one place. So I think definitely history has proven that to be a lie. >> So Richard, from a practitioner's point of view, what are your thoughts? I mean, there's a lot of pressure to cut cost, keep things centralized, serve the business as best as possible from that standpoint. What does your experience show? >> Yeah, I mean, I think I would echo Justin's experience really that we as a business have grown up through acquisition, through storing data in different places sometimes to do information governance in different ways to store data in a platform that's close to data experts people who really understand healthcare data from pharmacies or from doctors. And so, although if you were starting from a greenfield site and you were building something brand new, you might be able to centralize all the data and all of the tooling and teams in one place. The reality is that businesses just don't grow up like that. And it's just really impossible to get that academic perfection of storing everything in one place. >> Teresa, I feel like Sarbanes-Oxley have kind of saved the data warehouse, right? (laughs) You actually did have to have a single version of the truth for certain financial data, but really for some of those other use cases I mentioned, I do feel like the industry has kind of let us down. What's your take on this? Where does it make sense to have that sort of centralized approach versus where does it make sense to maybe decentralize? >> I think you got to have centralized governance, right? So from the central team, for things like Sarbanes-Oxley, for things like security, for certain very core data sets having a centralized set of roles, responsibilities to really QA, right? To serve as a design authority for your entire data estate, just like you might with security, but how it's implemented has to be distributed. Otherwise, you're not going to be able to scale, right? So being able to have different parts of the business really make the right data investments for their needs. And then ultimately, you're going to collaborate with your partners. So partners that are not within the company, right? External partners. We're going to see a lot more data sharing and model creation. And so you're definitely going to be decentralized. >> So Justin, you guys last, jeez, I think it was about a year ago, had a session on data mesh. It was a great program. You invited Zhamak Dehghani. Of course, she's the creator of the data mesh. One of our fundamental premises is that you've got this hyper specialized team that you've got to go through if you want anything. But at the same time, these individuals actually become a bottleneck, even though they're some of the most talented people in the organization. So I guess, a question for you Richard. How do you deal with that? Do you organize so that there are a few sort of rock stars that build cubes and the like or have you had any success in sort of decentralizing with your constituencies that data model? >> Yeah. So we absolutely have got rockstar data scientists and data guardians, if you like. People who understand what it means to use this data, particularly the data that we use at EMIS is very private, it's healthcare information. And some of the rules and regulations around using the data are very complex and strict. So we have to have people who understand the usage of the data, then people who understand how to build models, how to process the data effectively. And you can think of them like consultants to the wider business because a pharmacist might not understand how to structure a SQL query, but they do understand how they want to process medication information to improve patient lives. And so that becomes a consulting type experience from a set of rock stars to help a more decentralized business who needs to understand the data and to generate some valuable output. >> Justin, what do you say to a customer or prospect that says, "Look, Justin. I got a centralized team and that's the most cost effective way to serve the business. Otherwise, I got duplication." What do you say to that? >> Well, I would argue it's probably not the most cost effective, and the reason being really twofold. I think, first of all, when you are deploying a enterprise data warehouse model, the data warehouse itself is very expensive, generally speaking. And so you're putting all of your most valuable data in the hands of one vendor who now has tremendous leverage over you for many, many years to come. I think that's the story at Oracle or Teradata or other proprietary database systems. But the other aspect I think is that the reality is those central data warehouse teams, as much as they are experts in the technology, they don't necessarily understand the data itself. And this is one of the core tenets of data mesh that Zhamak writes about is this idea of the domain owners actually know the data the best. And so by not only acknowledging that data is generally decentralized, and to your earlier point about Sarbanes-Oxley, maybe saving the data warehouse, I would argue maybe GDPR and data sovereignty will destroy it because data has to be decentralized for those laws to be compliant. But I think the reality is the data mesh model basically says data's decentralized and we're going to turn that into an asset rather than a liability. And we're going to turn that into an asset by empowering the people that know the data the best to participate in the process of curating and creating data products for consumption. So I think when you think about it that way, you're going to get higher quality data and faster time to insight, which is ultimately going to drive more revenue for your business and reduce costs. So I think that that's the way I see the two models comparing and contrasting. >> So do you think the demise of the data warehouse is inevitable? Teresa, you work with a lot of clients. They're not just going to rip and replace their existing infrastructure. Maybe they're going to build on top of it, but what does that mean? Does that mean the EDW just becomes less and less valuable over time or it's maybe just isolated to specific use cases? What's your take on that? >> Listen, I still would love all my data within a data warehouse. I would love it mastered, would love it owned by a central team, right? I think that's still what I would love to have. That's just not the reality, right? The investment to actually migrate and keep that up to date, I would say it's a losing battle. Like we've been trying to do it for a long time. Nobody has the budgets and then data changes, right? There's going to be a new technology that's going to emerge that we're going to want to tap into. There's going to be not enough investment to bring all the legacy, but still very useful systems into that centralized view. So you keep the data warehouse. I think it's a very, very valuable, very high performance tool for what it's there for, but you could have this new mesh layer that still takes advantage of the things I mentioned: the data products in the systems that are meaningful today, and the data products that actually might span a number of systems. Maybe either those that either source systems with the domains that know it best, or the consumer-based systems or products that need to be packaged in a way that'd be really meaningful for that end user, right? Each of those are useful for a different part of the business and making sure that the mesh actually allows you to use all of them. >> So, Richard, let me ask you. Take Zhamak's principles back to those. You got the domain ownership and data as product. Okay, great. Sounds good. But it creates what I would argue are two challenges: self-serve infrastructure, let's park that for a second, and then in your industry, one of the most regulated, most sensitive, computational governance. How do you automate and ensure federated governance in that mesh model that Teresa was just talking about? >> Well, it absolutely depends on some of the tooling and processes that you put in place around those tools to centralize the security and the governance of the data. And I think although a data warehouse makes that very simple 'cause it's a single tool, it's not impossible with some of the data mesh technologies that are available. And so what we've done at EMIS is we have a single security layer that sits on top of our data mesh, which means that no matter which user is accessing which data source, we go through a well audited, well understood security layer. That means that we know exactly who's got access to which data field, which data tables. And then everything that they do is audited in a very kind of standard way regardless of the underlying data storage technology. So for me, although storing the data in one place might not be possible, understanding where your source of truth is and securing that in a common way is still a valuable approach, and you can do it without having to bring all that data into a single bucket so that it's all in one place. And so having done that and investing quite heavily in making that possible has paid dividends in terms of giving wider access to the platform, and ensuring that only data that's available under GDPR and other regulations is being used by the data users. >> Yeah. So Justin, we always talk about data democratization, and up until recently, they really haven't been line of sight as to how to get there, but do you have anything to add to this because you're essentially doing analytic queries with data that's all dispersed all over. How are you seeing your customers handle this challenge? >> Yeah, I mean, I think data products is a really interesting aspect of the answer to that. It allows you to, again, leverage the data domain owners, the people who know the data the best, to create data as a product ultimately to be consumed. And we try to represent that in our product as effectively, almost eCommerce like experience where you go and discover and look for the data products that have been created in your organization, and then you can start to consume them as you'd like. And so really trying to build on that notion of data democratization and self-service, and making it very easy to discover and start to use with whatever BI tool you may like or even just running SQL queries yourself. >> Okay guys, grab a sip of water. After the short break, we'll be back to debate whether proprietary or open platforms are the best path to the future of data excellence. Keep it right there. (bright upbeat music)
SUMMARY :
has the data they need when they need it Now, here's the first lie. has proven that to be a lie. of pressure to cut cost, and all of the tooling have kind of saved the data So from the central team, for that build cubes and the like and to generate some valuable output. and that's the most cost effective way is that the reality is those of the data warehouse is inevitable? and making sure that the mesh one of the most regulated, most sensitive, and processes that you put as to how to get there, aspect of the answer to that. or open platforms are the best path
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Richard | PERSON | 0.99+ |
Justin Borgman | PERSON | 0.99+ |
Justin | PERSON | 0.99+ |
Richard Jarvis | PERSON | 0.99+ |
Teresa Tung | PERSON | 0.99+ |
Jeff Hammerbacher | PERSON | 0.99+ |
Teresa | PERSON | 0.99+ |
Teradata | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Massachusetts | LOCATION | 0.99+ |
Zhamak Dehghani | PERSON | 0.99+ |
UK | LOCATION | 0.99+ |
2011 | DATE | 0.99+ |
two challenges | QUANTITY | 0.99+ |
Hadapt | ORGANIZATION | 0.99+ |
40 years | QUANTITY | 0.99+ |
Starburst | ORGANIZATION | 0.99+ |
two models | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Sarbanes-Oxley | ORGANIZATION | 0.99+ |
Each | QUANTITY | 0.99+ |
first lie | QUANTITY | 0.99+ |
Accenture | ORGANIZATION | 0.99+ |
GDPR | TITLE | 0.99+ |
Today | DATE | 0.98+ |
today | DATE | 0.98+ |
SQL | TITLE | 0.98+ |
Starburst Data | ORGANIZATION | 0.98+ |
EMIS Health | ORGANIZATION | 0.98+ |
Cloudera | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
first startup | QUANTITY | 0.98+ |
one place | QUANTITY | 0.98+ |
about 30 miles | QUANTITY | 0.98+ |
One | QUANTITY | 0.97+ |
More than a decade later | DATE | 0.97+ |
EMIS | ORGANIZATION | 0.97+ |
single bucket | QUANTITY | 0.97+ |
first technologist | QUANTITY | 0.96+ |
three industry experts | QUANTITY | 0.96+ |
single tool | QUANTITY | 0.96+ |
single version | QUANTITY | 0.94+ |
Zhamak | PERSON | 0.92+ |
theCUBE | ORGANIZATION | 0.91+ |
single source | QUANTITY | 0.9+ |
West Coast | LOCATION | 0.87+ |
one vendor | QUANTITY | 0.84+ |
single security layer | QUANTITY | 0.81+ |
about a year ago | DATE | 0.75+ |
IDU | ORGANIZATION | 0.68+ |
Is | TITLE | 0.65+ |
a second | QUANTITY | 0.64+ |
EDW | ORGANIZATION | 0.57+ |
examples | QUANTITY | 0.55+ |
echo | COMMERCIAL_ITEM | 0.54+ |
twofold | QUANTITY | 0.5+ |
Lie | TITLE | 0.35+ |
Breaking Analysis: How JPMC is Implementing a Data Mesh Architecture on the AWS Cloud
>> From theCUBE studios in Palo Alto and Boston, bringing you data-driven insights from theCUBE and ETR. This is braking analysis with Dave Vellante. >> A new era of data is upon us, and we're in a state of transition. You know, even our language reflects that. We rarely use the phrase big data anymore, rather we talk about digital transformation or digital business, or data-driven companies. Many have come to the realization that data is a not the new oil, because unlike oil, the same data can be used over and over for different purposes. We still use terms like data as an asset. However, that same narrative, when it's put forth by the vendor and practitioner communities, includes further discussions about democratizing and sharing data. Let me ask you this, when was the last time you wanted to share your financial assets with your coworkers or your partners or your customers? Hello everyone, and welcome to this week's Wikibon Cube Insights powered by ETR. In this breaking analysis, we want to share our assessment of the state of the data business. We'll do so by looking at the data mesh concept and how a leading financial institution, JP Morgan Chase is practically applying these relatively new ideas to transform its data architecture. Let's start by looking at what is the data mesh. As we've previously reported many times, data mesh is a concept and set of principles that was introduced in 2018 by Zhamak Deghani who's director of technology at ThoughtWorks, it's a global consultancy and software development company. And she created this movement because her clients, who were some of the leading firms in the world had invested heavily in predominantly monolithic data architectures that had failed to deliver desired outcomes in ROI. So her work went deep into trying to understand that problem. And her main conclusion that came out of this effort was the world of data is distributed and shoving all the data into a single monolithic architecture is an approach that fundamentally limits agility and scale. Now a profound concept of data mesh is the idea that data architectures should be organized around business lines with domain context. That the highly technical and hyper specialized roles of a centralized cross functional team are a key blocker to achieving our data aspirations. This is the first of four high level principles of data mesh. So first again, that the business domain should own the data end-to-end, rather than have it go through a centralized big data technical team. Second, a self-service platform is fundamental to a successful architectural approach where data is discoverable and shareable across an organization and an ecosystem. Third, product thinking is central to the idea of data mesh. In other words, data products will power the next era of data success. And fourth data products must be built with governance and compliance that is automated and federated. Now there's lot more to this concept and there are tons of resources on the web to learn more, including an entire community that is formed around data mesh. But this should give you a basic idea. Now, the other point is that, in observing Zhamak Deghani's work, she is deliberately avoided discussions around specific tooling, which I think has frustrated some folks because we all like to have references that tie to products and tools and companies. So this has been a two-edged sword in that, on the one hand it's good, because data mesh is designed to be tool agnostic and technology agnostic. On the other hand, it's led some folks to take liberties with the term data mesh and claim mission accomplished when their solution, you know, maybe more marketing than reality. So let's look at JP Morgan Chase in their data mesh journey. Is why I got really excited when I saw this past week, a team from JPMC held a meet up to discuss what they called, data lake strategy via data mesh architecture. I saw that title, I thought, well, that's a weird title. And I wondered, are they just taking their legacy data lakes and claiming they're now transformed into a data mesh? But in listening to the presentation, which was over an hour long, the answer is a definitive no, not at all in my opinion. A gentleman named Scott Hollerman organized the session that comprised these three speakers here, James Reid, who's a divisional CIO at JPMC, Arup Nanda who is a technologist and architect and Serita Bakst who is an information architect, again, all from JPMC. This was the most detailed and practical discussion that I've seen to date about implementing a data mesh. And this is JP Morgan's their approach, and we know they're extremely savvy and technically sound. And they've invested, it has to be billions in the past decade on data architecture across their massive company. And rather than dwell on the downsides of their big data past, I was really pleased to see how they're evolving their approach and embracing new thinking around data mesh. So today, we're going to share some of the slides that they use and comment on how it dovetails into the concept of data mesh that Zhamak Deghani has been promoting, and at least as we understand it. And dig a bit into some of the tooling that is being used by JP Morgan, particularly around it's AWS cloud. So the first point is it's all about business value, JPMC, they're in the money business, and in that world, business value is everything. So Jr Reid, the CIO showed this slide and talked about their overall goals, which centered on a cloud first strategy to modernize the JPMC platform. I think it's simple and sensible, but there's three factors on which he focused, cut costs always short, you got to do that. Number two was about unlocking new opportunities, or accelerating time to value. But I was really happy to see number three, data reuse. That's a fundamental value ingredient in the slide that he's presenting here. And his commentary was all about aligning with the domains and maximizing data reuse, i.e. data is not like oil and making sure there's appropriate governance around that. Now don't get caught up in the term data lake, I think it's just how JP Morgan communicates internally. It's invested in the data lake concept, so they use water analogies. They use things like data puddles, for example, which are single project data marts or data ponds, which comprise multiple data puddles. And these can feed in to data lakes. And as we'll see, JPMC doesn't strive to have a single version of the truth from a data standpoint that resides in a monolithic data lake, rather it enables the business lines to create and own their own data lakes that comprise fit for purpose data products. And they do have a single truth of metadata. Okay, we'll get to that. But generally speaking, each of the domains will own end-to-end their own data and be responsible for those data products, we'll talk about that more. Now the genesis of this was sort of a cloud first platform, JPMC is leaning into public cloud, which is ironic since the early days, in the early days of cloud, all the financial institutions were like never. Anyway, JPMC is going hard after it, they're adopting agile methods and microservices architectures, and it sees cloud as a fundamental enabler, but it recognizes that on-prem data must be part of the data mesh equation. Here's a slide that starts to get into some of that generic tooling, and then we'll go deeper. And I want to make a couple of points here that tie back to Zhamak Deghani's original concept. The first is that unlike many data architectures, this puts data as products right in the fat middle of the chart. The data products live in the business domains and are at the heart of the architecture. The databases, the Hadoop clusters, the files and APIs on the left-hand side, they serve the data product builders. The specialized roles on the right hand side, the DBA's, the data engineers, the data scientists, the data analysts, we could have put in quality engineers, et cetera, they serve the data products. Because the data products are owned by the business, they inherently have the context that is the middle of this diagram. And you can see at the bottom of the slide, the key principles include domain thinking, an end-to-end ownership of the data products. They build it, they own it, they run it, they manage it. At the same time, the goal is to democratize data with a self-service as a platform. One of the biggest points of contention of data mesh is governance. And as Serita Bakst said on the Meetup, metadata is your friend, and she kind of made a joke, she said, "This sounds kind of geeky, but it's important to have a metadata catalog to understand where data resides and the data lineage in overall change management. So to me, this really past the data mesh stink test pretty well. Let's look at data as products. CIO Reid said the most difficult thing for JPMC was getting their heads around data product, and they spent a lot of time getting this concept to work. Here's the slide they use to describe their data products as it related to their specific industry. They set a common language and taxonomy is very important, and you can imagine how difficult that was. He said, for example, it took a lot of discussion and debate to define what a transaction was. But you can see at a high level, these three product groups around wholesale, credit risk, party, and trade and position data as products, and each of these can have sub products, like, party, we'll have to know your customer, KYC for example. So a key for JPMC was to start at a high level and iterate to get more granular over time. So lots of decisions had to be made around who owns the products and the sub-products. The product owners interestingly had to defend why that product should even exist, what boundaries should be in place and what data sets do and don't belong in the various products. And this was a collaborative discussion, I'm sure there was contention around that between the lines of business. And which sub products should be part of these circles? They didn't say this, but tying it back to data mesh, each of these products, whether in a data lake or a data hub or a data pond or data warehouse, data puddle, each of these is a node in the global data mesh that is discoverable and governed. And supporting this notion, Serita said that, "This should not be infrastructure-bound, logically, any of these data products, whether on-prem or in the cloud can connect via the data mesh." So again, I felt like this really stayed true to the data mesh concept. Well, let's look at some of the key technical considerations that JPM discussed in quite some detail. This chart here shows a diagram of how JP Morgan thinks about the problem, and some of the challenges they had to consider were how to write to various data stores, can you and how can you move data from one data store to another? How can data be transformed? Where's the data located? Can the data be trusted? How can it be easily accessed? Who has the right to access that data? These are all problems that technology can help solve. And to address these issues, Arup Nanda explained that the heart of this slide is the data in ingestor instead of ETL. All data producers and contributors, they send their data to the ingestor and the ingestor then registers the data so it's in the data catalog. It does a data quality check and it tracks the lineage. Then, data is sent to the router, which persists the data in the data store based on the best destination as informed by the registration. This is designed to be a flexible system. In other words, the data store for a data product is not fixed, it's determined at the point of inventory, and that allows changes to be easily made in one place. The router simply reads that optimal location and sends it to the appropriate data store. Nowadays you see the schema infer there is used when there is no clear schema on right. In this case, the data product is not allowed to be consumed until the schema is inferred, and then the data goes into a raw area, and the inferer determines the schema and then updates the inventory system so that the data can be routed to the proper location and properly tracked. So that's some of the detail of how the sausage factory works in this particular use case, it was very interesting and informative. Now let's take a look at the specific implementation on AWS and dig into some of the tooling. As described in some detail by Arup Nanda, this diagram shows the reference architecture used by this group within JP Morgan, and it shows all the various AWS services and components that support their data mesh approach. So start with the authorization block right there underneath Kinesis. The lake formation is the single point of entitlement and has a number of buckets including, you can see there the raw area that we just talked about, a trusted bucket, a refined bucket, et cetera. Depending on the data characteristics at the data catalog registration block where you see the glue catalog, that determines in which bucket the router puts the data. And you can see the many AWS services in use here, identity, the EMR, the elastic MapReduce cluster from the legacy Hadoop work done over the years, the Redshift Spectrum and Athena, JPMC uses Athena for single threaded workloads and Redshift Spectrum for nested types so they can be queried independent of each other. Now remember very importantly, in this use case, there is not a single lake formation, rather than multiple lines of business will be authorized to create their own lakes, and that creates a challenge. So how can that be done in a flexible and automated manner? And that's where the data mesh comes into play. So JPMC came up with this federated lake formation accounts idea, and each line of business can create as many data producer or consumer accounts as they desire and roll them up into their master line of business lake formation account. And they cross-connect these data products in a federated model. And these all roll up into a master glue catalog so that any authorized user can find out where a specific data element is located. So this is like a super set catalog that comprises multiple sources and syncs up across the data mesh. So again to me, this was a very well thought out and practical application of database. Yes, it includes some notion of centralized management, but much of that responsibility has been passed down to the lines of business. It does roll up to a master catalog, but that's a metadata management effort that seems compulsory to ensure federated and automated governance. As well at JPMC, the office of the chief data officer is responsible for ensuring governance and compliance throughout the federation. All right, so let's take a look at some of the suspects in this world of data mesh and bring in the ETR data. Now, of course, ETR doesn't have a data mesh category, there's no such thing as that data mesh vendor, you build a data mesh, you don't buy it. So, what we did is we use the ETR dataset to select and filter on some of the culprits that we thought might contribute to the data mesh to see how they're performing. This chart depicts a popular view that we often like to share. It's a two dimensional graphic with net score or spending momentum on the vertical axis and market share or pervasiveness in the data set on the horizontal axis. And we filtered the data on sectors such as analytics, data warehouse, and the adjacencies to things that might fit into data mesh. And we think that these pretty well reflect participation that data mesh is certainly not all compassing. And it's a subset obviously, of all the vendors who could play in the space. Let's make a few observations. Now as is often the case, Azure and AWS, they're almost literally off the charts with very high spending velocity and large presence in the market. Oracle you can see also stands out because much of the world's data lives inside of Oracle databases. It doesn't have the spending momentum or growth, but the company remains prominent. And you can see Google Cloud doesn't have nearly the presence in the dataset, but it's momentum is highly elevated. Remember that red dotted line there, that 40% line, anything over that indicates elevated spending momentum. Let's go to Snowflake. Snowflake is consistently shown to be the gold standard in net score in the ETR dataset. It continues to maintain highly elevated spending velocity in the data. And in many ways, Snowflake with its data marketplace and its data cloud vision and data sharing approach, fit nicely into the data mesh concept. Now, a caution, Snowflake has used the term data mesh in it's marketing, but in our view, it lacks clarity, and we feel like they're still trying to figure out how to communicate what that really is. But is really, we think a lot of potential there to that vision. Databricks is also interesting because the firm has momentum and we expect further elevated levels in the vertical axis in upcoming surveys, especially as it readies for its IPO. The firm has a strong product and managed service, and is really one to watch. Now we included a number of other database companies for obvious reasons like Redis and Mongo, MariaDB, Couchbase and Terradata. SAP as well is in there, but that's not all database, but SAP is prominent so we included them. As is IBM more of a database, traditional database player also with the big presence. Cloudera includes Hortonworks and HPE Ezmeral comprises the MapR business that HPE acquired. So these guys got the big data movement started, between Cloudera, Hortonworks which is born out of Yahoo, which was the early big data, sorry early Hadoop innovator, kind of MapR when it's kind of owned course, and now that's all kind of come together in various forms. And of course, we've got Talend and Informatica are there, they are two data integration companies that are worth noting. We also included some of the AI and ML specialists and data science players in the mix like DataRobot who just did a monster $250 million round. Dataiku, H2O.ai and ThoughtSpot, which is all about democratizing data and injecting AI, and I think fits well into the data mesh concept. And you know we put VMware Cloud in there for reference because it really is the predominant on-prem infrastructure platform. All right, let's wrap with some final thoughts here, first, thanks a lot to the JP Morgan team for sharing this data. I really want to encourage practitioners and technologists, go to watch the YouTube of that meetup, we'll include it in the link of this session. And thank you to Zhamak Deghani and the entire data mesh community for the outstanding work that you're doing, challenging the established conventions of monolithic data architectures. The JPM presentation, it gives you real credibility, it takes Data Mesh well beyond concept, it demonstrates how it can be and is being done. And you know, this is not a perfect world, you're going to start somewhere and there's going to be some failures, the key is to recognize that shoving everything into a monolithic data architecture won't support massive scale and agility that you're after. It's maybe fine for smaller use cases in smaller firms, but if you're building a global platform in a data business, it's time to rethink data architecture. Now much of this is enabled by the cloud, but cloud first doesn't mean cloud only, doesn't mean you'll leave your on-prem data behind, on the contrary, you have to include non-public cloud data in your Data Mesh vision just as JPMC has done. You've got to get some quick wins, that's crucial so you can gain credibility within the organization and grow. And one of the key takeaways from the JP Morgan team is, there is a place for dogma, like organizing around data products and domains and getting that right. On the other hand, you have to remain flexible because technologies is going to come, technology is going to go, so you got to be flexible in that regard. And look, if you're going to embrace the metaphor of water like puddles and ponds and lakes, we suggest maybe a little tongue in cheek, but still we believe in this, that you expand your scope to include data ocean, something John Furry and I have talked about and laughed about extensively in theCUBE. Data oceans, it's huge. It's the new data lake, go transcend data lake, think oceans. And think about this, just as we're evolving our language, we should be evolving our metrics. Much the last the decade of big data was around just getting the stuff to work, getting it up and running, standing up infrastructure and managing massive, how much data you got? Massive amounts of data. And there were many KPIs built around, again, standing up that infrastructure, ingesting data, a lot of technical KPIs. This decade is not just about enabling better insights, it's a more than that. Data mesh points us to a new era of data value, and that requires the new metrics around monetizing data products, like how long does it take to go from data product conception to monetization? And how does that compare to what it is today? And what is the time to quality if the business owns the data, and the business has the context? the quality that comes out of them, out of the shoot should be at a basic level, pretty good, and at a higher mark than out of a big data team with no business context. Automation, AI, and very importantly, organizational restructuring of our data teams will heavily contribute to success in the coming years. So we encourage you, learn, lean in and create your data future. Okay, that's it for now, remember these episodes, they're all available as podcasts wherever you listen, all you got to do is search, breaking analysis podcast, and please subscribe. Check out ETR's website at etr.plus for all the data and all the survey information. We publish a full report every week on wikibon.com and siliconangle.com. And you can get in touch with us, email me david.vellante@siliconangle.com, you can DM me @dvellante, or you can comment on my LinkedIn posts. This is Dave Vellante for theCUBE insights powered by ETR. Have a great week everybody, stay safe, be well, and we'll see you next time. (upbeat music)
SUMMARY :
This is braking analysis and the adjacencies to things
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
JPMC | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
Zhamak Deghani | PERSON | 0.99+ |
James Reid | PERSON | 0.99+ |
JP Morgan | ORGANIZATION | 0.99+ |
JP Morgan | ORGANIZATION | 0.99+ |
Cloudera | ORGANIZATION | 0.99+ |
Serita Bakst | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Scott Hollerman | PERSON | 0.99+ |
Hortonworks | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
40% | QUANTITY | 0.99+ |
JP Morgan Chase | ORGANIZATION | 0.99+ |
Serita | PERSON | 0.99+ |
Yahoo | ORGANIZATION | 0.99+ |
Arup Nanda | PERSON | 0.99+ |
each | QUANTITY | 0.99+ |
ThoughtWorks | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
david.vellante@siliconangle.com | OTHER | 0.99+ |
each line | QUANTITY | 0.99+ |
Terradata | ORGANIZATION | 0.99+ |
Redis | ORGANIZATION | 0.99+ |
$250 million | QUANTITY | 0.99+ |
first point | QUANTITY | 0.99+ |
three factors | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
MapR | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Informatica | ORGANIZATION | 0.99+ |
Talend | ORGANIZATION | 0.99+ |
John Furry | PERSON | 0.99+ |
Zhamak Deghani | PERSON | 0.99+ |
first platform | QUANTITY | 0.98+ |
YouTube | ORGANIZATION | 0.98+ |
fourth | QUANTITY | 0.98+ |
single | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
Third | QUANTITY | 0.97+ |
Couchbase | ORGANIZATION | 0.97+ |
three speakers | QUANTITY | 0.97+ |
two data | QUANTITY | 0.97+ |
first strategy | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
one place | QUANTITY | 0.96+ |
Jr Reid | PERSON | 0.96+ |
single lake | QUANTITY | 0.95+ |
SAP | ORGANIZATION | 0.95+ |
wikibon.com | OTHER | 0.95+ |
siliconangle.com | OTHER | 0.94+ |
Azure | ORGANIZATION | 0.93+ |
Breaking Analysis - How AWS is Revolutionizing Systems Architecture
from the cube studios in palo alto in boston bringing you data-driven insights from the cube and etr this is breaking analysis with dave vellante aws is pointing the way to a revolution in system architecture much in the same way that aws defined the cloud operating model last decade we believe it is once again leading in future systems design the secret sauce underpinning these innovations is specialized designs that break the stranglehold of inefficient and bloated centralized processing and allows aws to accommodate a diversity of workloads that span cloud data center as well as the near and far edge hello and welcome to this week's wikibon cube insights powered by etr in this breaking analysis we'll dig into the moves that aws has been making which we believe define the future of computing we'll also project what this means for customers partners and aws many competitors now let's take a look at aws's architectural journey the is revolution it started by giving easy access as we all know to virtual machines that could be deployed and decommissioned on demand amazon at the time used a highly customized version of zen that allowed multiple vms to run on one physical machine the hypervisor functions were controlled by x86 now according to werner vogels as much as 30 of the processing was wasted meaning it was supporting hypervisor functions and managing other parts of the system including the storage and networking these overheads led to aws developing custom asics that help to accelerate workloads now in 2013 aws began shipping custom chips and partnered with amd to announce ec2 c3 instances but as the as the aws cloud started to scale they really weren't satisfied with the performance gains that they were getting and they were hitting architectural barriers that prompted aws to start a partnership with anaperta labs this was back in 2014 and they launched then ec2 c4 instances in 2015. the asic in c4 optimized offload functions for storage and networking but still relied on intel xeon as the control point aws aws shelled out a reported 350 million dollars to acquire annapurna in 2015 which is a meager sum to acquire the secret sauce of its future system design this acquisition led to a modern version of project nitro in 2017 nitro nitro offload cards were first introduced in 2013 at this time aws introduced c5 instances and replaced zen with kvm and more tightly coupled the hypervisor with the asic vogels shared last year that this milestone offloaded the remaining components including the control plane the rest of the i o and enabled nearly a hundred percent of the processing to support customer workloads it also enabled a bare metal version of the compute that spawned the partnership the famous partnership with vmware to launch vmware cloud on aws then in 2018 aws took the next step and introduced graviton its custom designed arm-based chip this broke the dependency on x86 and launched a new era of architecture which now supports a wide variety of configurations to support data intensive workloads now these moves preceded other aws innovations including new chips optimized for machine learning and training and inferencing and all kinds of ai the bottom line is aws has architected an approach that offloaded the work currently done by the central processing unit in most general purpose workloads like in the data center it has set the stage in our view for the future allowing shared memory memory disaggregation and independent resources that can be configured to support workloads from the cloud all the way to the edge and nitro is the key to this architecture and to summarize aws nitro think of it as a set of custom hardware and software that runs on an arm-based platform from annapurna aws has moved the hypervisor the network the storage virtualization to dedicated hardware that frees up the cpu to run more efficiently this in our opinion is where the entire industry is headed so let's take a look at that this chart pulls data from the etr data set and lays out key players competing for the future of cloud data center and the edge now we've superimposed nvidia up top and intel they don't show up directly in the etr survey but they clearly are platform players in the mix we covered nvidia extensively in previous breaking analysis and won't go too deep there today but the data shows net scores on the vertical axis that's a measure of spending velocity and then it shows market share in the horizontal axis which is a measure of pervasiveness within the etr data set we're not going to dwell on the relative positions here rather let's comment on the players and start with aws we've laid out aws how they got here and we believe they are setting the direction for the future of the industry and aws is really pushing migration to its arm-based platforms pat morehead at the 6-5 summit spoke to dave brown who heads ec2 at aws and he talked extensively about migrating from x86 to aws's arm-based graviton 2. and he announced a new developer challenge to accelerate that migration to arm instances graviton instances and the end game for customers is a 40 better price performance so a customer running 100 server instances can do the same work with 60 servers now there's some work involved but for the by the customers to actually get there but the payoff if they can get 40 improvement in price performance is quite large imagine this aws currently offers 400 different ec2 instances last year as we reported sorry last year as we reported earlier this year nearly 50 percent of the new ec2 instances so nearly 50 percent of the new ec2 instances shipped in 2020 were arm based and aws is working hard to accelerate this pace it's very clear now let's talk about intel i'll just say it intel is finally responding in earnest and basically it's taking a page out of arm's playbook we're going to dig into that a bit today in 2015 intel paid 16.7 billion dollars for altera a maker of fpgas now also at the 6.5 summit nevin shenoy of intel presented details of what intel is calling an ipu it's infrastructure processing unit this is a departure from intel norms where everything is controlled by a central processing unit ipu's are essentially smart knicks as our dpus so don't get caught up in all the acronym soup as we've reported it's all about offloading work and disaggregating memory and evolving socs system-on-chip and sops system on package but just let this sink in a bit a bit for a moment intel's moves this past week it seems to us anyway are designed to create a platform that is nitro like and the basis of that platform is a 16.7 billion dollar acquisition just compare that to aws's 350 million dollar tuck-in of annapurna that is incredible now chenoy said in his presentation rough quote we've already deployed ipu's using fpgas in a in very high volume at microsoft azure and we've recently announced partnerships with baidu jd cloud and vmware so let's look at vmware vmware is the other you know really big platform player in this race in 2020 vmware announced project monterrey you might recall that it's based on the aforementioned fpgas from intel so vmware is in the mix and it chose to work with intel most likely for a variety of reasons one of the obvious ones is all the software that's running on on on vmware it's been built for x86 and there's a huge install base there the other is pat was heading vmware at the time and and you know when project monterey was conceived so i'll let you connect the dots if you like regardless vmware has a nitro like offering in our view its optionality however is limited by intel but at least it's in the game and appears to be ahead of the competition in this space aws notwithstanding because aws is clearly in the lead now what about microsoft and google suffice it to say that we strongly believe that despite the comments that intel made about shipping fpgas and volume to microsoft that both microsoft and google as well as alibaba will follow aws's lead and develop an arm-based platform like nitro we think they have to in order to keep pace with aws now what about the rest of the data center pack well dell has vmware so despite the split we don't expect any real changes there dell is going to leverage whatever vmware does and do it better than anyone else cisco is interesting in that it just revamped its ucs but we don't see any evidence that it has a nitro like plans in its roadmap same with hpe now both of these companies have history and capabilities around silicon cisco designs its own chips today for carrier class use cases and and hpe as we've reported probably has some remnants of the machine hanging around but both companies are very likely in our view to follow vmware's lead and go with an intel based design what about ibm well we really don't know we think the best thing ibm could do would be to move the ibm cloud of course to an arm-based nitro-like platform we think even the mainframe should move to arm as well i mean it's just too expensive to build a specialized mainframe cpu these days now oracle they're interesting if we were running oracle we would build an arm-based nitro-like database cloud where oracle the database runs cheaper faster and consumes less energy than any other platform that would would dare to run oracle and we'd go one step further and we would optimize for competitive databases in the oracle cloud so we would make oci run the table on all databases and be essentially the database cloud but you know back to sort of fpgas we're not overly excited about about the market amd is acquiring xi links for 35 billion dollars so i guess that's something to get excited about i guess but at least amd is using its inflated stock price to do the deal but we honestly we think that the arm ecosystem will will obliterate the fpga market by making it simpler and faster to move to soc with far better performance flexibility integration and mobility so again we're not too sanguine about intel's acquisition of altera and the moves that amd is making in in the long term now let's take a deeper look at intel's vision of the data center of the future here's a chart that intel showed depicting its vision of the future of the data center what you see is the ipu's which are intelligent nixed and they're embedded in the four blocks shown and they're communicating across a fabric now you have general purpose compute in the upper left and machine intelligent on the bottom left machine intelligence apps and up in the top right you see storage services and then the bottom right variation of alternative processors and this is intel's view of how to share resources and go from a world where everything is controlled by a central processing unit to a more independent set of resources that can work in parallel now gelsinger has talked about all the cool tech that this will allow intel to incorporate including pci and gen 5 and cxl memory interfaces and or cxl memory which are interfaces that enable memory sharing and disaggregation and 5g and 6g connectivity and so forth so that's intel's view of the future of the data center let's look at arm's vision of the future and compare them now there are definite similarities as you can see especially on the right hand side of this chart you've got the blocks of different process processor types these of course are programmable and you notice the high bandwidth memory the hbm3 plus the ddrs on the two sides kind of bookending the blocks that's shared across the entire system and it's connected by pcie gen 5 cxl or ccix multi-die socket so you know you may be looking to say okay two sets of block diagrams big deal well while there are similarities around disaggregation and i guess implied shared memory in the intel diagram and of course the use of advanced standards there are also some notable differences in particular arm is really already at the soc level whereas intel is talking about fpgas neoverse arms architecture is shipping in test mode and we'll have end market product by year end 2022 intel is talking about maybe 2024 we think that's aspirational or 2025 at best arm's road map is much more clear now intel said it will release more details in october so we'll pay attention then maybe we'll recalibrate at that point but it's clear to us that arm is way further along now the other major difference is volume intel is coming at this from a high data center perspective and you know presumably plans to push down market or out to the edge arm is coming at this from the edge low cost low power superior price performance arm is winning at the edge and based on the data that we shared earlier from aws it's clearly gaining ground in the enterprise history strongly suggests that the volume approach will win not only at the low end but eventually at the high end so we want to wrap by looking at what this means for customers and the partner ecosystem the first point we'd like to make is follow the consumer apps this capability the capabilities that we see in consumer apps like image processing and natural language processing and facial recognition and voice translation these inference capabilities that are going on today in mobile will find their way into the enterprise ecosystem ninety percent of the cost associated with machine learning in the cloud is around inference in the future most ai in the enterprise and most certainly at the edge will be inference it's not today because it's too expensive this is why aws is building custom chips for inferencing to drive costs down so it can increase adoption now the second point is we think that customers should start experimenting and see what you can do with arm-based platforms moore's law is accelerating at least the outcome of moore's law the doubling of performance every of the 18 to 24 months it's it's actually much higher than that now when you add up all the different components in these alternative processors just take a look at apple's a5 a15 chip and arm is in the lead in terms of performance price performance cost and energy consumption by moving some workloads onto graviton for example you'll see what types of cost savings you can drive for which applications and possibly generate new applications that you can deliver to your business put a couple engineers in the task and see what they can do in two or three weeks you might be surprised or you might say hey it's too early for us but you'll find out and you may strike gold we would suggest that you talk to your hybrid cloud provider as well and find out if they have a nitro we shared that vmware they've got a clear path as does dell because they're you know vmware cousins what about your other strategic suppliers what's their roadmap what's the time frame to move from where they are today to something that resembles nitro do they even think about that how do they think about that do they think it's important to get there so if if so or if not how are they thinking about reducing your costs and supporting your new workloads at scale now for isvs these consumer capabilities that we discussed earlier all these mobile and and automated systems and cars and and things like that biometrics another example they're going to find their way into your software and your competitors are porting to arm they're embedding these consumer-like capabilities into their apps are you we would strongly recommend that you take a look at that talk to your cloud suppliers and see what they can do to help you innovate run faster and cut costs okay that's it for now thanks to my collaborator david floyer who's been on this topic since early last decade thanks to the community for your comments and insights and hey thanks to patrick morehead and daniel newman for some timely interviews from your event nice job fellas remember i published each week on wikibon.com and siliconangle.com these episodes are all available as podcasts just search for breaking analysis podcasts you can always connect with me on twitter at d vallante or email me at david.velante at siliconangle.com i appreciate the comments on linkedin and clubhouse so follow us if you see us in a room jump in and let's riff on these topics and don't forget to check out etr.plus for all the survey data this is dave vellante for the cube insights powered by etr be well and we'll see you next time
SUMMARY :
and nitro is the key to this
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
2013 | DATE | 0.99+ |
2015 | DATE | 0.99+ |
dave brown | PERSON | 0.99+ |
2014 | DATE | 0.99+ |
2020 | DATE | 0.99+ |
2017 | DATE | 0.99+ |
david floyer | PERSON | 0.99+ |
60 servers | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
last year | DATE | 0.99+ |
18 | QUANTITY | 0.99+ |
microsoft | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
daniel newman | PERSON | 0.99+ |
35 billion dollars | QUANTITY | 0.99+ |
alibaba | ORGANIZATION | 0.99+ |
16.7 billion dollars | QUANTITY | 0.99+ |
16.7 billion dollar | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
second point | QUANTITY | 0.99+ |
ninety percent | QUANTITY | 0.99+ |
siliconangle.com | OTHER | 0.99+ |
october | DATE | 0.99+ |
350 million dollars | QUANTITY | 0.99+ |
dave vellante | PERSON | 0.99+ |
2024 | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
nvidia | ORGANIZATION | 0.99+ |
amd | ORGANIZATION | 0.99+ |
boston | LOCATION | 0.99+ |
first point | QUANTITY | 0.99+ |
both companies | QUANTITY | 0.99+ |
three weeks | QUANTITY | 0.99+ |
24 months | QUANTITY | 0.99+ |
apple | ORGANIZATION | 0.98+ |
30 | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
graviton | TITLE | 0.98+ |
each week | QUANTITY | 0.98+ |
nearly 50 percent | QUANTITY | 0.98+ |
aws | ORGANIZATION | 0.98+ |
earlier this year | DATE | 0.98+ |
100 server instances | QUANTITY | 0.98+ |
amazon | ORGANIZATION | 0.98+ |
two sides | QUANTITY | 0.98+ |
intel | ORGANIZATION | 0.98+ |
400 different | QUANTITY | 0.97+ |
early last decade | DATE | 0.97+ |
ORGANIZATION | 0.97+ | |
ORGANIZATION | 0.97+ | |
40 improvement | QUANTITY | 0.97+ |
x86 | TITLE | 0.96+ |
last decade | DATE | 0.96+ |
cisco | ORGANIZATION | 0.95+ |
oracle | ORGANIZATION | 0.95+ |
chenoy | PERSON | 0.95+ |
40 better | QUANTITY | 0.95+ |
vmware | ORGANIZATION | 0.95+ |
350 million dollar | QUANTITY | 0.94+ |
nitro | ORGANIZATION | 0.92+ |
UNLIST TILL 4/2 - The Next-Generation Data Underlying Architecture
>> Paige: Hello, everybody, and thank you for joining us today for the virtual Vertica BDC 2020. Today's breakout session is entitled, Vertica next generation architecture. I'm Paige Roberts, open social relationship Manager at Vertica, I'll be your host for this session. And joining me is Vertica Chief Architect, Chuck Bear, before we begin, I encourage you to submit questions or comments during the virtual session. You don't have to wait, just type your question or comment, in the question box that's below the slides and click submit. So as you think about it, go ahead and type it in, there'll be a Q&A session at the end of the presentation, where we'll answer as many questions, as we're able to during the time. Any questions that we don't get a chance to address, we'll do our best to answer offline. Or alternatively, you can visit the Vertica forums to post your questions there, after the session. Our engineering team is planning to join the forum and keep the conversation going, so you can, it's just sort of like the developers lounge would be in delight conference. It gives you a chance to talk to our engineering team. Also, as a reminder, you can maximize your screen by clicking the double arrow button in the lower right corner of the slide. And before you ask, yes, this virtual session is being recorded, and it will be available to view on demand this week, we'll send you a notification, as soon as it's ready. Okay, now, let's get started, over to you, Chuck. >> Chuck: Thanks for the introduction, Paige, Vertica vision is to help customers, get value from structured data. This vision is simple, it doesn't matter what vertical the customer is in. They're all analytics companies, it doesn't matter what the customers environment is, as data is generated everywhere. We also can't do this alone, we know that you need other tools and people to build a complete solution. You know our database is key to delivering on the vision because we need a database that scales. When you start a new database company, you aren't going to win against 30 year old products on features. But from day one, we had something else, an architecture built for analytics performance. This architecture was inspired by the C-store project, combining the best design ideas from academics and industry veterans like Dr. Mike Stonebreaker. Our storage is optimized for performance, we use many computers in parallel. After over 10 years of refinements against various customer workloads, much of the design held up and serendipitously, the fact that we don't store in place updates set Vertica up for success in the cloud as well. These days, there are other tools that embody some of these design ideas. But we have other strengths that are more important than the storage format, where the only good analytics database that runs both on premise and in the cloud, giving customers the option to migrate their workloads, in most convenient and economical environment, or a full data management solution, not just the query tool. Unlike some other choices, ours comes with integration with a sequel ecosystem and full professional support. We organize our product roadmap into four key pillars, plus the cross cutting concerns of open integration and performance and scale. We have big plans to strengthen Vertica, while staying true to our core. This presentation is primarily about the separation pillar, and performance and scale, I'll cover our plans for Eon, our data management architecture, Mart analytic clusters, or fifth generation query executer, and our data storage layer. Let's start with how Vertica manages data, one of the central design points for Vertica was shared nothing, a design that didn't utilize a dedicated hardware shared disk technology. This quote here is how Mike put it politely, but around the Vertica office, shared disk with an LMTB over Mike's dead body. And we did get some early field experience with shared disk, customers, well, in fact will learn on anything if you let them. There were misconfigurations that required certified experts, obscure bugs extent. Another thing about the shared nothing designed for commodity hardware though, and this was in the papers, is that all the data management features like fault tolerance, backup and elasticity have to be done in software. And no matter how much you do, procuring, configuring and maintaining the machines with disks is harder. The software configuration process to add more service may be simple, but capacity planning, racking and stacking is not. The original allure of shared storage returned, this time though, the complexity and economics are different. It's cheaper, even provision storage with a few clicks and only pay for what you need. It expands, contracts and brings the maintenance of the storage close to a team is good at it. But there's a key difference, it's an object store, an object stores don't support the API's and access patterns used by most database software. So another Vertica visionary Ben, set out to exploit Vertica storage organization, which turns out to be a natural fit for modern cloud shared storage. Because Vertica data files are written once and not updated, they match the object storage model perfectly. And so today we have Eon, Eon uses shared storage to hold Vertica data with local disk depot's that act as caches, ensuring that we can get the performance that our customers have come to expect. Essentially Eon in enterprise behave similarly, but we have the benefit of flexible storage. Today Eon has the features our customers expect, it's been developed in tune for years, we have successful customers such as Redpharma, and if you'd like to know more about Eon has helped them succeed in Amazon cloud, I highly suggest reading their case study, which you can find on vertica.com. Eon provides high availability and flexible scaling, sometimes on premise customers with local disks get a little jealous of how recovery and sub-clusters work in Eon. Though we operate on premise, particularly on pure storage, but enterprise also had strengths, the most obvious being that you don't need and short shared storage to run it. So naturally, our vision is to converge the two modes, back into a single Vertica. A Vertica that runs any combination of local disks and shared storage, with full flexibility and portability. This is easy to say, but over the next releases, here's what we'll do. First, we realize that the query executer, optimizer and client drivers and so on, are already the same. Just the transaction handling and data management is different. But there's already more going on, we have peer-to-peer depot operations and other internode transfers. And enterprise also has a network, we could just get files from remote nodes over that network, essentially mimicking the behavior and benefits of shared storage with the layer of software. The only difference at the end of it, will be which storage hold the master copy. In enterprise, the nodes can't drop the files because they're the master copy. Whereas in Eon they can be evicted because it's just the cache, the masters, then shared storage. And in keeping with versus current support for multiple storage locations, we can intermix these approaches at the table level. Getting there as a journey, and we've already taken the first steps. One of the interesting design ideas of the C-store paper is the idea that redundant copies, don't have to have the same physical organization. Different copies can be optimized for different queries, sorted in different ways. Of course, Mike also said to keep the recovery system simple, because it's hard to debug, whenever the recovery system is being used, it's always in a high pressure situation. This turns out to be a contradiction, and the latter idea was better. No down performing stuff, if you don't keep the storage the same. Recovery hardware if you have, to reorganize data in the process. Even query optimization is more complicated. So over the past couple releases, we got rid of non identical buddies. But the storage files can still diverge at the fifth level, because tuple mover operations are synchronized. The same record can end up in different files than different nodes. The next step in our journey, is to make sure both copies are identical. This will help with backup and restore as well, because the second copy doesn't need backed up, or if it is backed up, it appears identical to the deduplication that is going to look present in both backup systems. Simultaneously, we're improving the Vertica networking service to support this new access pattern. In conjunction with identical storage files, we will converge to a recovery system that instantaneous nodes can process queries immediately, by retrieving data they need over the network from the redundant copies as they do in Eon day with even higher performance. The final step then is to unify the catalog and transaction model. Related concepts such as segment and shard, local catalog and shard catalog will be coalesced, as they're really represented the same concepts all along, just in different modes. In the catalog, we'll make slight changes to the definition of a projection, which represents the physical storage organization. The new definition simplifies segmentation and introduces valuable granularities of sharding to support evolution over time, and offers a straightforward migration path for both Eon and enterprise. There's a lot more to our Eon story than just the architectural roadmap. If you missed yesterday's Vertica, in Eon mode presentation about supported cloud, on premise storage option, replays are available. Be sure to catch the upcoming presentation on sizing and configuring vertica and in beyond doors. As we've seen with Eon, Vertica can separate data storage from the compute nodes, allowing machines to quickly fill in for each other, to rebuild fault tolerance. But separating compute and storage is used for much, much more. We now offer powerful, flexible ways for Vertica to add servers and increase access to the data. Vertica nine, this feature is called sub-clusters. It allows computing capacity to be added quickly and incrementally, and isolates workloads from each other. If your exploratory analytics team needs direct access to the source data, they need a lot of machines and not the same number all the time, and you don't 100% trust the kind of queries and user defined functions, they might be using sub-clusters as the solution. While there's much more expensive information available in our other presentation. I'd like to point out the highlights of our latest sub-cluster best practices. We suggest having a primary sub-cluster, this is the one that runs all the time, if you're loading data around the clock. It should be sized for the ETL workloads and also determines the natural shard count. Additional read oriented secondary sub-clusters can be added for real time dashboards, reports and analytics. That way, subclusters can be added or deep provisioned, without disruption to other users. The sub-cluster features of Vertica 9.3 are working well for customers. Yesterday, the Trade Desk presented their use case for Vertica over 300,000 in 5 sub clusters running in the cloud. If you missed a presentation, check out the replay. But we have plans beyond sub-clusters, we're extending sub-clusters to real clusters. For the Vertica savvy, this means the clusters bump, share the same spread ring network. This will provide further isolation, allowing clusters to control their own independent data sets. While replicating all are part of the data from other clusters using a publish subscribe mechanism. Synchronizing data between clusters is a feature customers want to understand the real business for themselves. This vision effects are designed for ancillary aspects, how we should assign resource pools, security policies and balance client connection. We will be simplifying our data segmentation strategy, so that when data that originate in the different clusters meet, they'll still get fully optimized joins, even if those clusters weren't positioned with the same number of nodes per shard. Having a broad vision for data management is a key component to political success. But we also take pride in our execution strategy, when you start a new database from scratch as we did 15 years ago, you won't compete on features. Our key competitive points where speed and scale of analytics, we set a target of 100 x better query performance in traditional databases with path loads. Our storage architecture provides a solid foundation on which to build toward these goals. Every query starts with data retrieval, keeping data sorted, organized by column and compressed by using adaptive caching, to keep the data retrieval time in IO to the bare minimum theoretically required. We also keep the data close to where it will be processed, and you clusters the machines to increase throughput. We have partition pruning a robust optimizer evaluate active use segmentation as part of the physical database designed to keep records close to the other relevant records. So the solid foundation, but we also need optimal execution strategies and tactics. One execution strategy which we built for a long time, but it's still a source of pride, it's how we process expressions. Databases and other systems with general purpose expression evaluators, write a compound expression into a tree. Here I'm using A plus one times B as an example, during execution, if your CPU traverses the tree and compute sub-parts from the whole. Tree traversal often takes more compute cycles than the actual work to be done. Especially in evaluation is a very common operation, so something worth optimizing. One instinct that engineers have is to use what we call, just-in-time or JIT compilation, which means generating code form the CPU into the specific activity expression, and add them. This replaces the tree of boxes that are custom made box for the query. This approach has complexity bugs, but it can be made to work. It has other drawbacks though, it adds a lot to query setup time, especially for short queries. And it pretty much eliminate the ability of mere models, mere mortals to develop user defined functions. If you go back to the problem we're trying to solve, the source of the overhead is the tree traversal. If you increase the batch of records processed in each traversal step, this overhead is amortized until it becomes negligible. It's a perfect match for a columnar storage engine. This also sets the CPU up for efficiency. The CPUs look particularly good, at following the same small sequence of instructions in a tight loop. In some cases, the CPU may even be able to vectorize, and apply the same processing to multiple records to the same instruction. This approach is easy to implement and debug, user defined functions are possible, then generally aligned with the other complexities of implementing and improving a large system. More importantly, the performance, both in terms of query setup and record throughput is dramatically improved. You'll hear me say that we look at research and industry for inspiration. In this case, our findings in line with academic binding. If you'd like to read papers, I recommend everything you always wanted to know about compiled and vectorized queries, don't afraid to ask, so we did have this idea before we read that paper. However, not every decision we made in the Vertica executer that the test of time as well as the expression evaluator. For example, sorting and grouping aren't susceptible to vectorization because sort decisions interrupt the flow. We have used JIT compiling on that for years, and Vertica 401, and it provides modest setups, but we know we can do even better. But who we've embarked on a new design for execution engine, which I call EE five, because it's our best. It's really designed especially for the cloud, now I know what you're thinking, you're thinking, I just put up a slide with an old engine, a new engine, and a sleek play headed up into the clouds. But this isn't just marketing hype, here's what I mean, when I say we've learned lessons over the years, and then we're redesigning the executer for the cloud. And of course, you'll see that the new design works well on premises as well. These changes are just more important for the cloud. Starting with the network layer in the cloud, we can't count on all nodes being connected to the same switch. Multicast doesn't work like it does in a custom data center, so as I mentioned earlier, we're redesigning the network transfer layer for the cloud. Storage in the cloud is different, and I'm not referring here to the storage of persistent data, but to the storage of temporary data used only once during the course of query execution. Our new pattern is designed to take into account the strengths and weaknesses of cloud object storage, where we can't easily do a path. Moving on to memory, many of our access patterns are reasonably effective on bare metal machines, that aren't the best choice on cloud hyperbug that have overheads, page faults or big gap. Here again, we found we can improve performance, a bit on dedicated hardware, and even more in the cloud. Finally, and this is true in all environments, core counts have gone up. And not all of our algorithms take full advantage, there's a lot of ground to cover here. But I think sorting in the perfect example to illustrate these points, I mentioned that we use JIT in sorting. We're getting rid of JIT in favor of a data format that can be treated efficiently, independent of what the data types are. We've drawn on the best, most modern technology from academia and industry. We've got our own analysis and testing, you know what we chose, we chose parallel merge sort, anyone wants to take a guess when merge sort was invented. It was invented in 1948, or at least documented that way, like computing context. If you've heard me talk before, you know that I'm fascinated by how all the things I worked with as an engineer, were invented before I was born. And in Vertica , we don't use the newest technologies, we use the best ones. And what is noble about Vertica is the way we've combined the best ideas together into a cohesive package. So all kidding about the 1940s aside, or he redesigned is actually state of the art. How do we know the sort routine is state of the art? It turns out, there's a pretty credible benchmark or at the appropriately named historic sortbenchmark.org. Anyone with resources looking for fame for their product or academic paper can try to set the record. Record is last set in 2016 with Tencent Sort, 100 terabytes in 99 seconds. Setting the records it's hard, you have to come up with hundreds of machines on a dedicated high speed switching fabric. There's a lot to a distributed sort, there all have core sorting algorithms. The authors of the paper conveniently broke out of the time spent in their sort, 67 out of 99 seconds want to know local sorting. If we break this out, divided by two CPUs and each of 512 nodes, we find that each CPU so there's almost a gig and a half per second. This is for what's called an indy sort, like an Indy race car, is in general purpose. It only handles fixed hundred five records with 10 byte key. There is a record length can vary, then it's called daytona sort, a 10 set daytona sort, is a little slower. One point is 10 gigabytes per second per CPU, now for Verrtica, We have a wide variety ability in record sizes, and more interesting data types, but still no harm in setting us like phone numbers, comfortable to the world record. On my 2017 era AMD desktop CPU, the Vertica EE5 sort to store about two and a half gigabytes per second. Obviously, this test isn't apply to apples because they use their own open power chip. But the number of DRM channels is the same, so it's pretty close the number that says we've hit on the right approach. And it performs this way on premise, in the cloud, and we can adapt it to cloud temp space. So what's our roadmap for integrating EE5 into the product and compare replacing the query executed the database to replacing the crankshaft and other parts of the engine of a car while it's been driven. We've actually done it before, between Vertica three and a half and five, and then we never really stopped changing it, now we'll do it again. The first part in replacing with algorithm called storage merge, which combines sorted data from disk. The first time has was two that are in vertical in incoming 10.0 patch that will be EE5 or resegmented storage merge, and then convert sorting and grouping into do out. There the performance results so far, in cases where the Vertica execute is doing well today, simple environments with simple data patterns, such as this simple capitalistic query, there's a lot of speed up, when we ship the segmentation code, which didn't quite make the freeze as much like to bump longer term, what we do is grouping into the storage of large operations, we'll get to where we think we ought to be, given a theoretical minimum work the CPUs need to do. Now if we look at a case where the current execution isn't doing as well, we see there's a much stronger benefit to the code shipping in Vertica 10. In fact, it turns a chart bar sideways to try to help you see the difference better. This case also benefit from the improvements in 10 product point releases and beyond. They will not happening to the vertical query executer, That was just the taste. But now I'd like to switch to the roadmap first for our adapters layer. I'll start with a story about, how our storage access layer evolved. If you go back to the academic ideas, if you start paper that persuaded investors to fund Vertica, read optimized store was the part that had substantiation in the form of performance data. Much of the paper was speculative, but we tried to follow it anyway. That paper talked about the WS with RS, The rights are in the read store, and how they work together for transaction processing and how there was a supernova. In all honesty, Vertica engineers couldn't figure out from the paper what to do next, incase you want to try, and we asked them they would like, We never got enough clarification to build it that way. But here's what we built, instead. We built the ROS, read optimized store, introduction on steep major revision. It's sorted, ordered columnar and compressed that follows a table partitioning that worked even better than the we are as described in the paper. We also built the last byte optimized store, we built four versions of this over the years actually. But this was the best one, it's not a set of interrelated V tree. It's just an append only, insertion order remember your way here, am sorry, no compression, no base, no partitioning. There is, however, a tuple over which does what we call move out. Move the data from WOS to ROS, sorting and compressing. Let's take a moment to compare how they behave, when you load data directly to the ROS, there's a data parsing operation. Then we finished the sorting, and then compressing right out the columnar data files to stay storage. The next query through executes against the ROS and it runs as it should because the ROS is read optimized. Let's repeat the exercise for WOS, the load operation response before the sorting and compressing, and before the data is written to persistent storage. Now it's possible for a query to come along, and the query could be responsible for sorting the lost data in addition to its other processes. Effect on query isn't predictable until the TM comes along and writes the data to the ROS. Over the years, we've done a lot of comparisons between ROS and WOS. ROS has always been better for sustained load throughput, it achieves much higher records per second without pushing back against the client and hasn't Vertica for when we developed the first usable merge out algorithm. ROS has always been better for predictable query performance, the ROS has never had the same management complexity and limitations as WOS. You don't have to pick a memory size and figure out which transactions get to use the pool. A non persistent nature of ROS always cause headaches when there are unexpected cluster shutdowns. We also looked at field usage data, we found that few customers were using a lot, especially among those that studied the issue carefully. So how we set out on a mission to improve the ROS to the point where it was always better than both the WOS and the profit of the past. And now it's true, ROS is better than the WOS and the loss of a couple of years ago. We implemented storage bundling, better catalog object storage and better tuple mover merge outs. And now, after extensive Q&A and customer testing, we've now succeeded, and in Vertica 10, we've removed the whys. Let's talk for a moment about simplicity, one of the best things Mike Stonebreaker said is no knobs. Anyone want to guess how many knobs we got rid of, and we took the WOS out of the product. 22 were five knobs to control whether it didn't went to ROS as well. Six controlling the ROS itself, Six more to set policies for the typical remove out and so on. In my honest opinion is still wasn't enough control over to achieve excess in a multi tenant environment, the big reason to get rid of the WOS for simplicity. Make the lives of DBAs and users better, we have a long way to go, but we're doing it. On my desk, I keep a jar with the knob in it for each knob in Vertica. When developers add a knob to the product, they have to add a knob to the jar. When they remove a knob, they get to choose one to take out, We have a lot of work to do, but I'm thrilled to report that in 15 years 10 is the first release with a number of knobs ticked downward. Get back to the WOS, I've said the most important thing get rid of it for last. We're getting rid of it so we can deliver our vision of the future to our customer. Remember how he said an Eon and sub-clusters we got all these benefits from shared storage? Guess what can't live in shared storage, the WOS. Remember how it's been a big part of the future was keeping the copies that identical to the primary copy? Independent actions of the WOS took a little at the root of the divergence between copies of the data. You have to admit it when you're wrong. That was in the original design and held up to the a selling point of time, without onto the idea of a separate ROS and WOS for too long. In Vertica, 10, we can finally bid, good reagents. I've covered a lot of ground, so let's put all the pieces together. I've talked a lot about our vision and how we're achieving it. But we also still pay attention to tactical detail. We've been fine tuning our memory management model to enhance performance. That involves revisiting tens of thousands of satellite of code, much like painting the inside of a large building with small paintbrushes. We're getting results as shown in the chart in Vertica nine, concurrent monitoring queries use memory from the global catalog tool, and Vertica 10, they don't. This is only one example of an important detail we're improving. We've also reworked the monitoring tables without network messages into two parts. The increased data we're collecting and analyzing and our quality assurance processes, we're improving on everything. As the story goes, I still have my grandfather's axe, of course, my father had to replace the handle, and I had to replace the head. Along the same lines, we still have Mike Stonebreaker Vertica. We didn't replace the query optimizer twice the debate database designer and storage layer four times each. The query executed is and it's a free design, like charted out how our code has changed over the years. I found that we don't have much from a long time ago, I did some digging, and you know what we have left in 2007. We have the original curly braces, and a little bit of percent code for handling dates and times. To deliver on our mission to help customers get value from their structured data, with high performance of scale, and in diverse deployment environments. We have the sound architecture roadmap, reviews the best execution strategy and solid tactics. On the architectural front, we're converging in an enterprise, we're extending smart analytic clusters. In query processing, we're redesigning the execution engine for the cloud, as I've told you. There's a lot more than just the fast engine. that you want to learn about our new data support for complex data types, improvements to the query optimizer statistics, or extension to live aggregate projections and flatten tables. You should check out some of the other engineering talk that the big data conference. We continue to stay on top of the details from low level CPU and memory too, to the monitoring management, developing tighter feedback cycles between development, Q&A and customers. And don't forget to check out the rest of the pillars of our roadmap. We have new easier ways to get started with Vertica in the cloud. Engineers have been hard at work on machine learning and security. It's easier than ever to use Vertica with third Party product, as a variety of tools integrations continues to increase. Finally, the most important thing we can do, is to help people get value from structured data to help people learn more about Vertica. So hopefully I left plenty of time for Q&A at the end of this presentation. I hope to hear your questions soon.
SUMMARY :
and keep the conversation going, and apply the same processing to multiple records
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Mike | PERSON | 0.99+ |
Mike Stonebreaker | PERSON | 0.99+ |
2007 | DATE | 0.99+ |
Chuck Bear | PERSON | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
2016 | DATE | 0.99+ |
Paige Roberts | PERSON | 0.99+ |
Chuck | PERSON | 0.99+ |
second copy | QUANTITY | 0.99+ |
99 seconds | QUANTITY | 0.99+ |
67 | QUANTITY | 0.99+ |
100% | QUANTITY | 0.99+ |
1948 | DATE | 0.99+ |
Ben | PERSON | 0.99+ |
two modes | QUANTITY | 0.99+ |
Redpharma | ORGANIZATION | 0.99+ |
first time | QUANTITY | 0.99+ |
first steps | QUANTITY | 0.99+ |
Paige | PERSON | 0.99+ |
two parts | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
five knobs | QUANTITY | 0.99+ |
100 terabytes | QUANTITY | 0.99+ |
both copies | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
each knob | QUANTITY | 0.99+ |
WS | ORGANIZATION | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Eon | ORGANIZATION | 0.99+ |
1940s | DATE | 0.99+ |
today | DATE | 0.99+ |
One point | QUANTITY | 0.99+ |
first part | QUANTITY | 0.99+ |
fifth level | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
yesterday | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
Six | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
512 nodes | QUANTITY | 0.98+ |
ROS | TITLE | 0.98+ |
over 10 years | QUANTITY | 0.98+ |
Yesterday | DATE | 0.98+ |
15 years ago | DATE | 0.98+ |
twice | QUANTITY | 0.98+ |
sortbenchmark.org | OTHER | 0.98+ |
first release | QUANTITY | 0.98+ |
two CPUs | QUANTITY | 0.97+ |
Vertica 10 | TITLE | 0.97+ |
100 x | QUANTITY | 0.97+ |
WOS | TITLE | 0.97+ |
vertica.com | OTHER | 0.97+ |
10 byte | QUANTITY | 0.97+ |
this week | DATE | 0.97+ |
one | QUANTITY | 0.97+ |
5 sub clusters | QUANTITY | 0.97+ |
two | QUANTITY | 0.97+ |
one example | QUANTITY | 0.97+ |
over 300,000 | QUANTITY | 0.96+ |
Dr. | PERSON | 0.96+ |
One | QUANTITY | 0.96+ |
tens of thousands of satellite | QUANTITY | 0.96+ |
EE5 | COMMERCIAL_ITEM | 0.96+ |
fifth generation | QUANTITY | 0.96+ |
UNLIST TILL 4/2 - A Technical Overview of Vertica Architecture
>> Paige: Hello, everybody and thank you for joining us today on the Virtual Vertica BDC 2020. Today's breakout session is entitled A Technical Overview of the Vertica Architecture. I'm Paige Roberts, Open Source Relations Manager at Vertica and I'll be your host for this webinar. Now joining me is Ryan Role-kuh? Did I say that right? (laughs) He's a Vertica Senior Software Engineer. >> Ryan: So it's Roelke. (laughs) >> Paige: Roelke, okay, I got it, all right. Ryan Roelke. And before we begin, I want to be sure and encourage you guys to submit your questions or your comments during the virtual session while Ryan is talking as you think of them as you go along. You don't have to wait to the end, just type in your question or your comment in the question box below the slides and click submit. There'll be a Q and A at the end of the presentation and we'll answer as many questions as we're able to during that time. Any questions that we don't address, we'll do our best to get back to you offline. Now, alternatively, you can visit the Vertica forums to post your question there after the session as well. Our engineering team is planning to join the forums to keep the conversation going, so you can have a chat afterwards with the engineer, just like any other conference. Now also, you can maximize your screen by clicking the double arrow button in the lower right corner of the slides and before you ask, yes, this virtual session is being recorded and it will be available to view on demand this week. We'll send you a notification as soon as it's ready. Now, let's get started. Over to you, Ryan. >> Ryan: Thanks, Paige. Good afternoon, everybody. My name is Ryan and I'm a Senior Software Engineer on Vertica's Development Team. I primarily work on improving Vertica's query execution engine, so usually in the space of making things faster. Today, I'm here to talk about something that's more general than that, so we're going to go through a technical overview of the Vertica architecture. So the intent of this talk, essentially, is to just explain some of the basic aspects of how Vertica works and what makes it such a great database software and to explain what makes a query execute so fast in Vertica, we'll provide some background to explain why other databases don't keep up. And we'll use that as a starting point to discuss an academic database that paved the way for Vertica. And then we'll explain how Vertica design builds upon that academic database to be the great software that it is today. I want to start by sharing somebody's approximation of an internet minute at some point in 2019. All of the data on this slide is generated by thousands or even millions of users and that's a huge amount of activity. Most of the applications depicted here are backed by one or more databases. Most of this activity will eventually result in changes to those databases. For the most part, we can categorize the way these databases are used into one of two paradigms. First up, we have online transaction processing or OLTP. OLTP workloads usually operate on single entries in a database, so an update to a retail inventory or a change in a bank account balance are both great examples of OLTP operations. Updates to these data sets must be visible immediately and there could be many transactions occurring concurrently from many different users. OLTP queries are usually key value queries. The key uniquely identifies the single entry in a database for reading or writing. Early databases and applications were probably designed for OLTP workloads. This example on the slide is typical of an OLTP workload. We have a table, accounts, such as for a bank, which tracks information for each of the bank's clients. An update query, like the one depicted here, might be run whenever a user deposits $10 into their bank account. Our second category is online analytical processing or OLAP which is more about using your data for decision making. If you have a hardware device which periodically records how it's doing, you could analyze trends of all your devices over time to observe what data patterns are likely to lead to failure or if you're Google, you might log user search activity to identify which links helped your users find the answer. Analytical processing has always been around but with the advent of the internet, it happened at scales that were unimaginable, even just 20 years ago. This SQL example is something you might see in an OLAP workload. We have a table, searches, logging user activity. We will eventually see one row in this table for each query submitted by users. If we want to find out what time of day our users are most active, then we could write a query like this one on the slide which counts the number of unique users running searches for each hour of the day. So now let's rewind to 2005. We don't have a picture of an internet minute in 2005, we don't have the data for that. We also don't have the data for a lot of other things. The term Big Data is not quite yet on anyone's radar and The Cloud is also not quite there or it's just starting to be. So if you have a database serving your application, it's probably optimized for OLTP workloads. OLAP workloads just aren't mainstream yet and database engineers probably don't have them in mind. So let's innovate. It's still 2005 and we want to try something new with our database. Let's take a look at what happens when we do run an analytic workload in 2005. Let's use as a motivating example a table of stock prices over time. In our table, the symbol column identifies the stock that was traded, the price column identifies the new price and the timestamp column indicates when the price changed. We have several other columns which, we should know that they're there, but we're not going to use them in any example queries. This table is designed for analytic queries. We're probably not going to make any updates or look at individual rows since we're logging historical data and want to analyze changes in stock price over time. Our database system is built to serve OLTP use cases, so it's probably going to store the table on disk in a single file like this one. Notice that each row contains all of the columns of our data in row major order. There's probably an index somewhere in the memory of the system which will help us to point lookups. Maybe our system expects that we will use the stock symbol and the trade time as lookup keys. So an index will provide quick lookups for those columns to the position of the whole row in the file. If we did have an update to a single row, then this representation would work great. We would seek to the row that we're interested in, finding it would probably be very fast using the in-memory index. And then we would update the file in place with our new value. On the other hand, if we ran an analytic query like we want to, the data access pattern is very different. The index is not helpful because we're looking up a whole range of rows, not just a single row. As a result, the only way to find the rows that we actually need for this query is to scan the entire file. We're going to end up scanning a lot of data that we don't need and that won't just be the rows that we don't need, there's many other columns in this table. Many information about who made the transaction, and we'll also be scanning through those columns for every single row in this table. That could be a very serious problem once we consider the scale of this file. Stocks change a lot, we probably have thousands or millions or maybe even billions of rows that are going to be stored in this file and we're going to scan all of these extra columns for every single row. If we tried out our stocks use case behind the desk for the Fortune 500 company, then we're probably going to be pretty disappointed. Our queries will eventually finish, but it might take so long that we don't even care about the answer anymore by the time that they do. Our database is not built for the task we want to use it for. Around the same time, a team of researchers in the North East have become aware of this problem and they decided to dedicate their time and research to it. These researchers weren't just anybody. The fruits of their labor, which we now like to call the C-Store Paper, was published by eventual Turing Award winner, Mike Stonebraker, along with several other researchers from elite universities. This paper presents the design of a read-optimized relational DBMS that contrasts sharply with most current systems, which are write-optimized. That sounds exactly like what we want for our stocks use case. Reasoning about what makes our queries executions so slow brought our researchers to the Memory Hierarchy, which essentially is a visualization of the relative speeds of different parts of a computer. At the top of the hierarchy, we have the fastest data units, which are, of course, also the most expensive to produce. As we move down the hierarchy, components get slower but also much cheaper and thus you can have more of them. Our OLTP databases data is stored in a file on the hard disk. We scanned the entirety of this file, even though we didn't need most of the data and now it turns out, that is just about the slowest thing that our query could possibly be doing by over two orders of magnitude. It should be clear, based on that, that the best thing we can do to optimize our query's execution is to avoid reading unnecessary data from the disk and that's what the C-Store researchers decided to look at. The key innovation of the C-Store paper does exactly that. Instead of storing data in a row major order, in a large file on disk, they transposed the data and stored each column in its own file. Now, if we run the same select query, we read only the relevant columns. The unnamed columns don't factor into the table scan at all since we don't even open the files. Zooming out to an internet scale sized data set, we can appreciate the savings here a lot more. But we still have to read a lot of data that we don't need to answer this particular query. Remember, we had two predicates, one on the symbol column and one on the timestamp column. Our query is only interested in AAPL stock, but we're still reading rows for all of the other stocks. So what can we do to optimize our disk read even more? Let's first partition our data set into different files based on the timestamp date. This means that we will keep separate files for each date. When we query the stocks table, the database knows all of the files we have to open. If we have a simple predicate on the timestamp column, as our sample query does, then the database can use it to figure out which files we don't have to look at at all. So now all of our disk reads that we have to do to answer our query will produce rows that pass the timestamp predicate. This eliminates a lot of wasteful disk reads. But not all of them. We do have another predicate on the symbol column where symbol equals AAPL. We'd like to avoid disk reads of rows that don't satisfy that predicate either. And we can avoid those disk reads by clustering all the rows that match the symbol predicate together. If all of the AAPL rows are adjacent, then as soon as we see something different, we can stop reading the file. We won't see any more rows that can pass the predicate. Then we can use the positions of the rows we did find to identify which pieces of the other columns we need to read. One technique that we can use to cluster the rows is sorting. So we'll use the symbol column as a sort key for all of the columns. And that way we can reconstruct a whole row by seeking to the same row position in each file. It turns out, having sorted all of the rows, we can do a bit more. We don't have any more wasted disk reads but we can still be more efficient with how we're using the disk. We've clustered all of the rows with the same symbol together so we don't really need to bother repeating the symbol so many times in the same file. Let's just write the value once and say how many rows we have. This one length encoding technique can compress large numbers of rows into a small amount of space. In this example, we do de-duplicate just a few rows but you can imagine de-duplicating many thousands of rows instead. This encoding is great for reducing the amounts of disk we need to read at query time, but it also has the additional benefit of reducing the total size of our stored data. Now our query requires substantially fewer disk reads than it did when we started. Let's recap what the C-Store paper did to achieve that. First, we transposed our data to store each column in its own file. Now, queries only have to read the columns used in the query. Second, we partitioned the data into multiple file sets so that all rows in a file have the same value for the partition column. Now, a predicate on the partition column can skip non-matching file sets entirely. Third, we selected a column of our data to use as a sort key. Now rows with the same value for that column are clustered together, which allows our query to stop reading data once it finds non-matching rows. Finally, sorting the data this way enables high compression ratios, using one length encoding which minimizes the size of the data stored on the disk. The C-Store system combined each of these innovative ideas to produce an academically significant result. And if you used it behind the desk of a Fortune 500 company in 2005, you probably would've been pretty pleased. But it's not 2005 anymore and the requirements of a modern database system are much stricter. So let's take a look at how C-Store fairs in 2020. First of all, we have designed the storage layer of our database to optimize a single query in a single application. Our design optimizes the heck out of that query and probably some similar ones but if we want to do anything else with our data, we might be in a bit of trouble. What if we just decide we want to ask a different question? For example, in our stock example, what if we want to plot all the trade made by a single user over a large window of time? How do our optimizations for the previous query measure up here? Well, our data's partitioned on the trade date, that could still be useful, depending on our new query. If we want to look at a trader's activity over a long period of time, we would have to open a lot of files. But if we're still interested in just a day's worth of data, then this optimization is still an optimization. Within each file, our data is ordered on the stock symbol. That's probably not too useful anymore, the rows for a single trader aren't going to be clustered together so we will have to scan all of the rows in order to figure out which ones match. You could imagine a worse design but as it becomes crucial to optimize this new type of query, then we might have to go as far as reconfiguring the whole database. The next problem of one of scale. One server is probably not good enough to serve a database in 2020. C-Store, as described, runs on a single server and stores lots of files. What if the data overwhelms this small system? We could imagine exhausting the file system's inodes limit with lots of small files due to our partitioning scheme. Or we could imagine something simpler, just filling up the disk with huge volumes of data. But there's an even simpler problem than that. What if something goes wrong and C-Store crashes? Then our data is no longer available to us until the single server is brought back up. A third concern, another one of scalability, is that one deployment does not really suit all possible things and use cases we could imagine. We haven't really said anything about being flexible. A contemporary database system has to integrate with many other applications, which might themselves have pretty restricted deployment options. Or the demands imposed by our workloads have changed and the setup you had before doesn't suit what you need now. C-Store doesn't do anything to address these concerns. What the C-Store paper did do was lead very quickly to the founding of Vertica. Vertica's architecture and design are essentially all about bringing the C-Store designs into an enterprise software system. The C-Store paper was just an academic exercise so it didn't really need to address any of the hard problems that we just talked about. But Vertica, the first commercial database built upon the ideas of the C-Store paper would definitely have to. This brings us back to the present to look at how an analytic query runs in 2020 on the Vertica Analytic Database. Vertica takes the key idea from the paper, can we significantly improve query performance by changing the way our data is stored and give its users the tools to customize their storage layer in order to heavily optimize really important or commonly wrong queries. On top of that, Vertica is a distributed system which allows it to scale up to internet-sized data sets, as well as have better reliability and uptime. We'll now take a brief look at what Vertica does to address the three inadequacies of the C-Store system that we mentioned. To avoid locking into a single database design, Vertica provides tools for the database user to customize the way their data is stored. To address the shortcomings of a single node system, Vertica coordinates processing among multiple nodes. To acknowledge the large variety of desirable deployments, Vertica does not require any specialized hardware and has many features which smoothly integrate it with a Cloud computing environment. First, we'll look at the database design problem. We're a SQL database, so our users are writing SQL and describing their data in SQL way, the Create Table statement. Create Table is a logical description of what your data looks like but it doesn't specify the way that it has to be stored, For a single Create Table, we could imagine a lot of different storage layouts. Vertica adds some extensions to SQL so that users can go even further than Create Table and describe the way that they want the data to be stored. Using terminology from the C-Store paper, we provide the Create Projection statement. Create Projection specifies how table data should be laid out, including column encoding and sort order. A table can have multiple projections, each of which could be ordered on different columns. When you query a table, Vertica will answer the query using the projection which it determines to be the best match. Referring back to our stock example, here's a sample Create Table and Create Projection statement. Let's focus on our heavily optimized example query, which had predicates on the stock symbol and date. We specify that the table data is to be partitioned by date. The Create Projection Statement here is excellent for this query. We specify using the order by clause that the data should be ordered according to our predicates. We'll use the timestamp as a secondary sort key. Each projection stores a copy of the table data. If you don't expect to need a particular column in a projection, then you can leave it out. Our average price query didn't care about who did the trading, so maybe our projection design for this query can leave the trader column out entirely. If the question we want to ask ever does change, maybe we already have a suitable projection, but if we don't, then we can create another one. This example shows another projection which would be much better at identifying trends of traders, rather than identifying trends for a particular stock. Next, let's take a look at our second problem, that one, or excuse me, so how should you decide what design is best for your queries? Well, you could spend a lot of time figuring it out on your own, or you could use Vertica's Database Designer tool which will help you by automatically analyzing your queries and spitting out a design which it thinks is going to work really well. If you want to learn more about the Database Designer Tool, then you should attend the session Vertica Database Designer- Today and Tomorrow which will tell you a lot about what the Database Designer does and some recent improvements that we have made. Okay, now we'll move to our next problem. (laughs) The challenge that one server does not fit all. In 2020, we have several orders of magnitude more data than we had in 2005. And you need a lot more hardware to crunch it. It's not tractable to keep multiple petabytes of data in a system with a single server. So Vertica doesn't try. Vertica is a distributed system so will deploy multiple severs which work together to maintain such a high data volume. In a traditional Vertica deployment, each node keeps some of the data in its own locally-attached storage. Data is replicated so that there is a redundant copy somewhere else in the system. If any one node goes down, then the data that it served is still available on a different node. We'll also have it so that in the system, there's no special node with extra duties. All nodes are created equal. This ensures that there is no single point of failure. Rather than replicate all of your data, Vertica divvies it up amongst all of the nodes in your system. We call this segmentation. The way data is segmented is another parameter of storage customization and it can definitely have an impact upon query performance. A common way to segment data is by using a hash expression, which essentially randomizes the node that a row of data belongs to. But with a guarantee that the same data will always end up in the same place. Describing the way data is segmented is another part of the Create Projection Statement, as seen in this example. Here we segment on the hash of the symbol column so all rows with the same symbol will end up on the same node. For each row that we load into the system, we'll apply our segmentation expression. The result determines which segment the row belongs to and then we'll send the row to each node which holds the copy of that segment. In this example, our projection is marked KSAFE 1, so we will keep one redundant copy of each segment. When we load a row, we might find that its segment had copied on Node One and Node Three, so we'll send a copy of the row to each of those nodes. If Node One is temporarily disconnected from the network, then Node Three can serve the other copy of the segment so that the whole system remains available. The last challenge we brought up from the C-Store design was that one deployment does not fit all. Vertica's cluster design neatly addressed many of our concerns here. Our use of segmentation to distribute data means that a Vertica system can scale to any size of deployment. And since we lack any special hardware or nodes with special purposes, Vertica servers can run anywhere, on premise or in the Cloud. But let's suppose you need to scale out your cluster to rise to the demands of a higher workload. Suppose you want to add another node. This changes the division of the segmentation space. We'll have to re-segment every row in the database to find its new home and then we'll have to move around any data that belongs to a different segment. This is a very expensive operation, not something you want to be doing all that often. Traditional Vertica doesn't solve that problem especially well, but Vertica Eon Mode definitely does. Vertica's Eon Mode is a large set of features which are designed with a Cloud computing environment in mind. One feature of this design is elastic throughput scaling, which is the idea that you can smoothly change your cluster size without having to pay the expenses of shuffling your entire database. Vertica Eon Mode had an entire session dedicated to it this morning. I won't say any more about it here, but maybe you already attended that session or if you haven't, then I definitely encourage you to listen to the recording. If you'd like to learn more about the Vertica architecture, then you'll find on this slide links to several of the academic conference publications. These four papers here, as well as Vertica Seven Years Later paper which describes some of the Vertica designs seven years after the founding and also a paper about the innovations of Eon Mode and of course, the Vertica documentation is an excellent resource for learning more about what's going on in a Vertica system. I hope you enjoyed learning about the Vertica architecture. I would be very happy to take all of your questions now. Thank you for attending this session.
SUMMARY :
A Technical Overview of the Vertica Architecture. Ryan: So it's Roelke. in the question box below the slides and click submit. that the best thing we can do
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ryan | PERSON | 0.99+ |
Mike Stonebraker | PERSON | 0.99+ |
Ryan Roelke | PERSON | 0.99+ |
2005 | DATE | 0.99+ |
2020 | DATE | 0.99+ |
thousands | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
$10 | QUANTITY | 0.99+ |
Paige Roberts | PERSON | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
Paige | PERSON | 0.99+ |
Node Three | TITLE | 0.99+ |
Today | DATE | 0.99+ |
First | QUANTITY | 0.99+ |
each file | QUANTITY | 0.99+ |
Roelke | PERSON | 0.99+ |
each row | QUANTITY | 0.99+ |
Node One | TITLE | 0.99+ |
millions | QUANTITY | 0.99+ |
each hour | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
second category | QUANTITY | 0.99+ |
each column | QUANTITY | 0.99+ |
One technique | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
two predicates | QUANTITY | 0.99+ |
each node | QUANTITY | 0.99+ |
One server | QUANTITY | 0.99+ |
SQL | TITLE | 0.99+ |
C-Store | TITLE | 0.99+ |
second problem | QUANTITY | 0.99+ |
Ryan Role | PERSON | 0.99+ |
Third | QUANTITY | 0.99+ |
North East | LOCATION | 0.99+ |
each segment | QUANTITY | 0.99+ |
today | DATE | 0.98+ |
single entry | QUANTITY | 0.98+ |
each date | QUANTITY | 0.98+ |
ORGANIZATION | 0.98+ | |
one row | QUANTITY | 0.98+ |
one server | QUANTITY | 0.98+ |
single server | QUANTITY | 0.98+ |
single entries | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
20 years ago | DATE | 0.98+ |
two paradigms | QUANTITY | 0.97+ |
a day | QUANTITY | 0.97+ |
this week | DATE | 0.97+ |
billions of rows | QUANTITY | 0.97+ |
Vertica | TITLE | 0.97+ |
4/2 | DATE | 0.97+ |
single application | QUANTITY | 0.97+ |
each query | QUANTITY | 0.97+ |
Each projection | QUANTITY | 0.97+ |
Understanding Container Architecture - Wikibon Whiteboard
hello my name is Brian Grace Lee analysts with wiki Bond and on today's wiki bond whiteboards we're gonna begin to understand container architectures containers are really the big technology talked about these days especially for infrastructure teams there's a component of it that's both application and infrastructure but in this whiteboard we're really going to understand the basics of how it applies to the infrastructure and we're going to try and put it in the context of things that most infrastructure teams understand today which is virtualization so let's go ahead and begin what we've done and again this is for context we've tried to take a standard environment that people are used to seeing for virtualization and in this case we're going to use VMware as the example because obviously broadest market share and a lot of people understand what they do so let's talk about the basics of what happens here people understand what happens at the host level I've got servers within each server I've got a hypervisor so in VMware's case ESX or ESXi within that hypervisor I'm going to go ahead and create virtual machines so every single virtual machine has a copy of the full operating system and then within that virtual machine I've got a an operating an application itself for multiple applications so everybody understands that pretty well now how I manage those hype with those hypervisors and virtual machines is through a centralized control plane and that's called V Center and V Center may be a single instance it may be a clustered instance but think of it as the thing that's going to manage the scheduling of the resources and the management of those resources and it's really only focused on virtual machines okay now above that we're going to have if we're deploying applications I can either deploy them by hand or I may begin to deploy them through application templates so I may deploy the same type of application over and over again a web server a sequel database something else do that consistently I'm going to use some sort of typically a templating function and a lot of that can come in the management flame framework from something like V realize VMware via realized and then on top of that I'm going to have my applications whatever those might be sequel databases s AP Oracle Microsoft applications whatever those things might be so the key things I want you to understand is at the host level its hypervisor virtual machine full operating system and application and at the control plane it's this sort of structured format of V Center cluster V Center is going to make sure that virtual machines get deployed on to those hosts and it's going to keep track of where they are and make sure that they stay alive using things like VMware H a VM or V motion and VMware fault tolerance okay so now that we have that basic context in place let's take a look at how the container ecosystem is beginning to evolve and in this example we're gonna use docker because similar to VMware right now docker is the most frequently used container technology there are other ones in the marketplace but we're going to use docker just as an example the rest of what we talked about will be applicable whether it's docker core OS rocket or a number of the other container technologies that are out there so let's begin down at the host level just like we did over here in the simplest form I'm gonna have a host I'm gonna have a server we're not going to have a hypervisor we're just going to have the operating system today in most container environments that operating system is going to be Linux now there's a lot going on in the marketplace where this will eventually be Linux and Windows Microsoft is is working quite a bit on this but for right now let's just say that operating system is Linux ok I'm going to have my container runtime which in this case is his docker and you can think about that as sort of being like a hypervisor but it's almost a lightweight hypervisor and then that container runtime is going to create my containers themselves and each one of those containers now what's unique about this that's different from this environment is each one of those containers only uses they all share the same operating system so again all of your containers within a single host have to run the same operating system either all Linux or eventually would be all Windows they're going to use the bits that they need from that operating system so the net-net of it is it's a lighter-weight footprint I should be able to boot them quicker and the reason people get very very fixated on I can boot a container fast is because in this container environment the types of applications that I'm building tend to be more what they call ephemeral pieces of them are going to go away they're going to come back I'm gonna want to spin them up quickly if I have a scalable application spin them up or spin them down and so what you're looking for is a operating environment that will come up very very quickly so just to put that in context to spin up a virtual machine it may take three four minutes because of the operating system coming up to spin up a container usually is on the order of a second or a couple of seconds so big you know order of difference between there now the second piece that's really important and this is where a lot of people kind of get confused about what's going on in the container ecosystem is what happens at that control plane and the first thing to understand is when we talked about you know virtualized applications we tend to talk about very stateful sometimes they're called platform two sometimes they're called legacy applications but they're more or less stateful so the expectation is once you deploy them other than maybe Vee motioning them around for availability you're not scaling them up and down you don't expect them to fail frequently and so the scalability needed at the control plane is fairly well-defined maybe it's a thousand hosts or 10,000 hosts when we start dealing with containers the types of applications we deploy tend to be more what they call 12 factor applications sometimes you hear them call modular applications cloud native applications the idea being they're much more modular they tend to be more state less so the idea of maintaining state tends to get pushed somewhere else but they're designed for scale they're designed for mobile applications for real-time data applications and so the control plane unlike here which tends to be somewhat stateful and more confined in terms of scale has to be designed to be a distributed control plane it has to be designed to scale much much larger and so as part of that what we see is we're seeing technologies come out that sort of break up the things that were functions inside of a vCenter control plane into sort of distinct technologies that number one tend to scale more because they're written in distributed manner and number two they've got a certain amount of sort of mix-and-match that you can have with them depending on what your applications gonna do so let's talk through the basic things that are in here the first layer that you'll often hear about is clustering how do I cluster together sets of container hosts an example of this is docker swarm technology another example of this is something like Etsy D from core OS it's a technology to sort of figure out where my clusters of hosts are going to be the next layer is what's called service discovery if I'm deploying hundreds and hundreds or thousands of devices I want to you know containers I want to be able to figure out what services are available queuing services database services you know notification services the things that are out there I need to do that dynamically and automatically the next piece is going to be scheduling those containers just like vCenter is going to put it on the right host to make sure that it's load balanced properly there's a scheduling function to make sure that containers get deployed to the right container and then the next piece is what they call application scheduling so in these environments I don't tend to schedule my applications in these environments they could be a mix of batch applications Hadoop applications long running applications short running applications I need a more advanced intelligent scheduler to make sure that I'm getting the containers and the applications deployed on the right place and as efficient as possible and then on top of that I have my actual applications so the takeaway from this is at the host level some difference between how heavy a virtualized environment is going to be versus a container environment and that you want that to match how your application requirements are and if the control plane a more structured model for doing the functions that you need to manage the environment in a container model a more distributed model so with that I'm gonna go ahead and wrap that up we're going to get into some more depth in other videos we hope you enjoy these once again this has been a wiki bound whiteboard video you can find more information about all of our research and all the information about these technologies at wiki bond com and again if you want to follow me again my name is brian grace lee i'm at be grace lee on twitter or you can follow at wiki bon on twitter as well thank you and have a great day
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
10,000 hosts | QUANTITY | 0.99+ |
ESXi | TITLE | 0.99+ |
Brian Grace Lee | PERSON | 0.99+ |
second piece | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ESX | TITLE | 0.99+ |
Linux | TITLE | 0.99+ |
brian grace lee | PERSON | 0.99+ |
first layer | QUANTITY | 0.99+ |
Windows | TITLE | 0.98+ |
Oracle | ORGANIZATION | 0.98+ |
first thing | QUANTITY | 0.97+ |
VMware | TITLE | 0.97+ |
ORGANIZATION | 0.97+ | |
today | DATE | 0.96+ |
each server | QUANTITY | 0.95+ |
hundreds and | QUANTITY | 0.95+ |
12 factor | QUANTITY | 0.95+ |
a thousand hosts | QUANTITY | 0.95+ |
single instance | QUANTITY | 0.93+ |
wiki bond com | ORGANIZATION | 0.93+ |
three four minutes | QUANTITY | 0.93+ |
a couple of seconds | QUANTITY | 0.92+ |
Understanding Container Architecture | TITLE | 0.91+ |
each one | QUANTITY | 0.91+ |
single host | QUANTITY | 0.91+ |
thousands of devices | QUANTITY | 0.91+ |
a lot of people | QUANTITY | 0.89+ |
a lot of people | QUANTITY | 0.86+ |
a second | QUANTITY | 0.83+ |
wiki bond | TITLE | 0.8+ |
vCenter | TITLE | 0.8+ |
two | QUANTITY | 0.78+ |
both application | QUANTITY | 0.76+ |
wiki Bond | TITLE | 0.76+ |
hundreds | QUANTITY | 0.75+ |
docker | TITLE | 0.73+ |
Etsy D | TITLE | 0.72+ |
number one | QUANTITY | 0.7+ |
V | TITLE | 0.67+ |
grace lee | PERSON | 0.67+ |
every single virtual | QUANTITY | 0.65+ |
each one | QUANTITY | 0.64+ |
wiki bon | ORGANIZATION | 0.61+ |
V Center | TITLE | 0.57+ |
platform two | QUANTITY | 0.52+ |
Wikibon | ORGANIZATION | 0.51+ |
Center | COMMERCIAL_ITEM | 0.48+ |
Abdullah Abuzaid, Dell Technologies & Gil Hellmann, Wind River | MWC Barcelona 2023
(intro music) >> Narrator: "theCUBE's" live coverage is made possible by funding from Dell Technologies, creating technologies that drive human progress. (gentle music) >> Hey everyone, welcome back to "theCUBE," the leader in live and emerging tech coverage. As you well know, we are live at MWC23 in Barcelona, Spain. Lisa Martin with Dave Nicholson. Day three of our coverage, as you know, 'cause you've been watching the first two days. A lot of conversations about ecosystem, a lot about disruption in the telco industry. We're going to be talking about Open RAN. You've heard some of those great conversations, the complexities, the opportunities. Two guests join Dave and me. Abdullah Abuzaid, Technical Product Manager at Dell, and Gil Hellmann, VP Telecom Solutions Engineering and Architecture at Wind River. Welcome to the program guys. >> Thank you. >> Nice to be here. >> Let's talk a little bit about Dell and Wind River. We'll each ask you both the same question, and talk to us about how you're working together to really address the complexities that organizations are having when they're considering moving from a closed environment to an open environment. >> Definitely. Thank you for hosting us. By end of the day, the relationship between Dell and Wind River is not a new. We've been collaborating in the open ecosystem for long a time enough. And that's one of the, our partnership is a result of this collaboration where we've been trying to make more efficient operation in the ecosystem. The open environment ecosystem, it has the plus and a concern. The plus of simplicity, choice of multiple vendors, and then the concern of complexity managing these vendors. Especially if we look at examples for the Open RAN ecosystem, dealing with multiple vendors, trying to align them. It bring a lot of operational complexity and TCO challenges for our customers, from this outcome where we build our partnership with Wind River in order to help our customer to simplify, or run deployment, operation, and lifecycle management and sustain it. >> And who are the customers, by the way? >> Mainly the CSP customers who are targeting Open RAN and Virtual RAN deployments. That digital transformation moving towards unified cloud environment, or a seamless cloud experience from Core to RAN, these are the customers we are working with them. >> You'll give us your perspective, your thoughts on the partnership, and the capabilities that you're enabling, the CSPs with that. >> Sure. It's actually started last year here in Barcelona, when we set together, and started to look at the, you know, the industry, the adoption of Open RAN, and the challenges. And Open RAN brings a lot of possibilities and benefit, but it does bring a lot of challenges of reintegrating what you desegregate. In the past, you purchase everything from one vendor, they provide the whole solution. Now you open it, you have different layers. So if you're looking at Open RAN, you have, I like to look at it as three major layers, the management, application, and the infrastructure. And we're starting to look what are the challenges. And the challenges of integration, of complexity, knowledge that operator has with cloud infrastructure. And this is where we basically, Dell and Winder River set together and say, "How can we ease this? "How we can make it simpler?" And we decided to partner and bring a joint infrastructure solution to market, that's not only integrated at a lab at the factory level, but it basically comes with complete lifecycle management from the day zero deployment, through the day two operation, everything done through location, through Dell supported, working out of the box. So basically taking this whole infrastructure layer integration pain out, de-risking everything, and then continuing from there to work with the ecosystem vendor to reintegrate, validate the application, on top of this infrastructure. >> So what is the, what is the Wind River secret sauce in this, in this mix, for folks who aren't familiar with what Wind River does? >> Yes, absolutely. So Wind River, for many, many don't know, we're in business since 1981. So over 40 years. We specialize high performance, high reliability infrastructure. We touch every aspect of your day and your life. From the airplane that you fly, the cars, the medical equipment. And if we go into the telco, most of the telco equipment that it's not virtualized, not throughout the fight today, using our operating system. So from all the leading equipment manufacturers and even the smaller one. And as the world started to go into desegregation in cloud, Wind River started to look at this and say, "Okay, everything is evolving. Instead of a device that included the application, the hardware, everything fused together, it's now being decomposed. So instead of providing the operating environment to develop and deploy the application to the device manufacturer, now we're providing it basically to build the cloud. So to oversimplify, I call it a cloud OS, okay. It's a lot more than OS, it's an operating environment. But we took basically our experience, the same experience that, you know, we used in all those years with the telco equipment manufacturer, and brought it into the cloud. So we're basically providing solution to build an on-premises scalable cloud from the core all the way to the far edge, that doesn't compromise reliability, doesn't compromise performance, and address all the telco needs. >> So I, Abdullah, maybe you can a answer this. >> Yeah. >> What is the, what does the go-to-market motion look like, considering that you have two separate companies that can address customers directly, separately. What does that, what does that look like if you're approaching a possible customer who is, who's knocking on the door? >> How does that work? >> Exactly. And this effort is a Dell turnkey sales service offering, or solution offering to our customers. Where Dell, in collaboration with Wind River, we proactively validate, integrate, and productize the solution as engineered system, knock door on our customer who are trying to transform to Open RAN or open ecosystem. We can help you to go through that seamless experience, by pre-validating with whatever workload you want to introduce, enable zero touch provisioning, and during the day one deployment, and ensure we have sustainable lifecycle management throughout the lifecycle of the product in, in operate, in operational network, as well as having a unified single call of support from Dell side. >> Okay. So I was just going to ask you about support. So I'm a CSP, I have the solution, I go to Dell for support. >> Exactly. >> Okay. So start with Dell, and level one, level two. And if there are complex issues related to the cloud core itself, then Wind River will be on our back supporting us. >> Talk a little bit about a cust, a CSP example that is, is using the technology, and some of the outcomes that they're able to achieve. I'd love to get both of your perspectives on that. >> Vodafone is a great example. We're here in Barcelona. Vodafone is the first ora network in Europe, and it's using our joint solution. >> What are some of the, the outcomes that it's helping them to achieve? >> Faster time to market. As you see, they already started to deploy the ORAN in commercial network, and very successful in the trials that they did last year. We're also not stopping there. We're evolving, working with them together to improve like stuff around energy efficiency. So continue to optimize. So the outcome, it's just simplifying it, and you know, ready to go. Using experience that we have, Wind River is powering the first basically virtualized RAN 5G network in the world. This is with Verizon. We're at the very large scale. We started this deployment in late '20 and '19, the first site. And then through 2020 to 2022, we basically rolled in large scale. We have a lot of experience learning from it, which what we brought into the table when we partnered with Dell. A lot of experience from how you deploy at scale. Many sites from a central location, updates, upgrade. So the whole day two operation, and this is coming to bearing the solution that basically Vodafone is deploying now, and which allowed them... If I, if I look at my engagement with Verizon, started years before we started. And it took quite some time until we got stuff running. And if you look at the Vodafone time schedule, was significantly compressed compared to the Verizon first deployment. And I can tell you that there are other service providers that were announced here by KDI, for example. It's another one moving even faster. So it's accelerating the whole movement to Ora. >> We've heard a lot of acceleration talk this week. I'd love to get your perspective, Abdullah, talking about, you know, you, you just mentioned two huge names in Telco, Vodafone and Verizon. >> Yep. >> Talk a little bit about Dell's commitment to helping telecommunications companies really advance, accelerate innovation so that all of us on the other end have this thing that just works wherever we are 24 by 7. >> Not exactly. And this, we go back to the challenges in Open ecosystem. Managing multiple vendors at the same time, is a challenge for our customers. And that's why we are trying to simplify their life cycle by have, by being a trusted partner, working with our customer through all the journey. We started with Dish in their 5G deployment. Also with Vodafone. We're finding the right partners working with them proactively before getting into, in front of the customer to, we've done our homework, we are ready to simplify the process for you to go for it. If you look at the RAN in particular, we are talking with the 5g. We have ran the simplification, but they still have on the other side, limited resources and skillset can support it. So, bringing a pro, ahead of time engineer system, with a zero touch of provisioning enablement, and sustainable life cycle management, it lead to the faster time to market deployment, TCO savings, improved margins for our customers, and faster business revenue for their end users. >> Solid outcomes. >> And, and what you just just described, justifies the pain associated with disaggregating and reintegrating, which is the way that Gill referenced it, which I think is great because you're not, you're not, you're not re-aggregating, (laughs) you're reintegrating, and you're creating something that's better. >> Exactly. >> Moving forward. Otherwise, why would you do it? >> Exactly. And if you look at it, the player in the ecosystem, you have the vendors, you have the service integrators, you have the automation enablers, but kind of they are talking in silos. Everyone, this is my raci, this is what I'm responsible for. I, I'm not able, I don't want to get into something else while we are going the extra mile by working proactively in that ecosystem to... Let's bring brains together, find out what's one plus one can bring three for our customers, so we make it end-to-end seamless experience, not only on the technical part, but also on the business aspect side of it. >> So, so the partnership, it's about reducing the pen. I will say eliminating it. So this is the, the core of it. And you mentioned getting better coverage for your phone. I do want to point out that the phones are great, but if you look at the premises of a 5G network, it's to enable a lot more things that will touch your life that are beyond the consumer and the phone. Stuff like connected vehicles. So for example, something as simple as collision avoidance, the ability for the car that goes in front of you to be able to see what's happening and broadcast this information to the car behind that have no ability to see it. And basically affect our life in a way that makes our driving safer. And for this, you need a ultra low, reliable low latency communication. You need a 5G network. >> I'm glad you brought that up, because you know, we think about, "Well we just have to be connected all the time." But those are some of the emerging technologies that are going to be potentially lifesaving, and, and really life transforming that you guys are helping to enable. So, really great stuff there, but so much promise coming down the road. What's next for Dell and Wind River? And, and when you're in conversations with prospective CSP's, what is the superpower that you deliver together? I'd love to get both of your perspectives. >> So, if you look at it, number one, customers look at it, last savings and their day-to-day operation. In 5G nature, we are talking the introduction of ORAN. This is still picking up. But there is a mutualization and densification of ORAN. And this is where we're talking on monetizing my deployment. Then the third phase, we're talking sustainability and advanced service introduction. Where I want to move not only ORAN, I want to bring the edge at the same side, I want to define the advanced use cases of edge, where it enables me with this pre-work being done to deliver more services and better SLA services. By end of the day, 5G as a girl mentioned earlier, is not about a good better phone coverage, or a better speed robot, but what customized SLA's I can deliver. So it enables me to deliver different business streams to my end users. >> Yeah. >> So yeah. I will say there are two pens. One, it's the technology side. So for an example, energy efficiency. It's a very big pin point. And sustainability. So we work a lot around this, and basically to advance this. So if you look at the integrated solution today, it's very highly optimized for resource consumption. But to be able to more dynamically be able to change your power profile without compromising the SLA. So this is one side. The other side, it's about all those applications that will come to the 5G network to make our life better. It's about integrating, validating, certifying those applications. So, it's not just easy to deploy an ORAN network, but it's easy to deploy those applications. >> I'd be curious to get your perspective on the question of ROI in this, in this space. Specifically with the sort of the macro headwinds (clears throat) the economies of the world are facing right now, if you accept that. What does the ROI timeline look like when you're talking about moving towards ORAN, adopting VRAN, an amazing, you know, a plethora of new services that can be delivered, but will these operators have the appetite to take that, make that investment and take on that risk based upon the ROI time horizon? Any thoughts on that? >> Yeah. So if you look at the early days or ORAN introduction in particular, most of the entrepreneurs of ORAN and Virtual RAN ran into the challenges of not only the complexity of open ecosystem, but the integration, is like the redos of the work. And that's where we are trying to address it via pre-engineered system or building an engineer system proactively before getting it to the customers. Per our result or outcomes we get, we are talking about 30 to 50% savings on the optics. We are talking 110 ROI for our customers, simply because we are reducing the redos, the time spent to discover and explore. Because we've done that rework ahead of time, we found the optimization issues. Just for example, any customer can buy the same components from any multiple vendors, but how I can bring them together and give, deliver for me the best performance that I can fully utilize, that's, that's where it brings the value for our customer, and accelerate the deployment and the operation of the network. >> Do you have anything to add before we close in the next 30 seconds? >> Yeah. Yeah. (laughs) >> Absolutely. I would say, we start to see the data coming from two years of operation at scale. And the data supports performance. It's the same or better than traditional system. And the cost of operation, it's as good or better than traditional. Unfortunately, I can't provide more specific data. But the point is, when something is unknown in the beginning, of course you're more afraid, you take more conservative approach. Now the data starts to flow. And from here, the intention needs to go even better. So more efficiency, so cost less than traditional system, both to operate as well as to build up. But it's definitely the data that we have today says, the, ORAN system is at part, at the minimum. >> So, definite ROI there. Guys, thank you so much for joining Dave and me talking about how you're helping organizations not just address the complexities of moving from close to open, but to your point, eliminating them. We appreciate your time and, and your insights. >> Thank you. >> All right. For our guests and for Dave Nicholson, I'm Lisa Martin. You're watching "theCUBE," the leader in live and emerging tech coverage. Live from MWC23. We'll be back after a short break. (outro music)
SUMMARY :
that drive human progress. in the telco industry. and talk to us about how By end of the day, Mainly the CSP and the capabilities that you're enabling, In the past, you purchase From the airplane that you fly, the cars, you can a answer this. considering that you have and during the day one deployment, So I'm a CSP, I have the solution, issues related to the and some of the outcomes Vodafone is the first and this is coming to bearing the solution I'd love to get your Dell's commitment to helping front of the customer to, justifies the pain associated with Otherwise, why would you do it? but also on the business that are beyond the but so much promise coming down the road. By end of the day, 5G as and basically to advance this. of the macro headwinds the time spent to discover and explore. (laughs) Now the data starts to flow. not just address the the leader in live and
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Vodafone | ORGANIZATION | 0.99+ |
Telco | ORGANIZATION | 0.99+ |
Verizon | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Abdullah Abuzaid | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Wind River | ORGANIZATION | 0.99+ |
Barcelona | LOCATION | 0.99+ |
Abdullah | PERSON | 0.99+ |
Gil Hellmann | PERSON | 0.99+ |
last year | DATE | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
two years | QUANTITY | 0.99+ |
telco | ORGANIZATION | 0.99+ |
110 ROI | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
Two guests | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
1981 | DATE | 0.99+ |
2022 | DATE | 0.99+ |
Winder River | ORGANIZATION | 0.99+ |
first site | QUANTITY | 0.99+ |
two pens | QUANTITY | 0.99+ |
over 40 years | QUANTITY | 0.99+ |
KDI | ORGANIZATION | 0.99+ |
Gill | PERSON | 0.99+ |
both | QUANTITY | 0.99+ |
Open RAN | TITLE | 0.99+ |
Barcelona, Spain | LOCATION | 0.99+ |
today | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
late '20 | DATE | 0.99+ |
24 | QUANTITY | 0.98+ |
third phase | QUANTITY | 0.98+ |
7 | QUANTITY | 0.98+ |
first two days | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
one vendor | QUANTITY | 0.98+ |
50% | QUANTITY | 0.98+ |
two separate companies | QUANTITY | 0.98+ |
Open RAN | TITLE | 0.98+ |
one | QUANTITY | 0.97+ |
one side | QUANTITY | 0.97+ |
first deployment | QUANTITY | 0.97+ |
three major layers | QUANTITY | 0.97+ |
tel | ORGANIZATION | 0.96+ |
Day three | QUANTITY | 0.96+ |
Dish | ORGANIZATION | 0.95+ |
'19 | DATE | 0.95+ |
ORAN | TITLE | 0.95+ |
Jack Greenfield, Walmart | A Dive into Walmart's Retail Supercloud
>> Welcome back to SuperCloud2. This is Dave Vellante, and we're here with Jack Greenfield. He's the Vice President of Enterprise Architecture and the Chief Architect for the global technology platform at Walmart. Jack, I want to thank you for coming on the program. Really appreciate your time. >> Glad to be here, Dave. Thanks for inviting me and appreciate the opportunity to chat with you. >> Yeah, it's our pleasure. Now we call what you've built a SuperCloud. That's our term, not yours, but how would you describe the Walmart Cloud Native Platform? >> So WCNP, as the acronym goes, is essentially an implementation of Kubernetes for the Walmart ecosystem. And what that means is that we've taken Kubernetes off the shelf as open source, and we have integrated it with a number of foundational services that provide other aspects of our computational environment. So Kubernetes off the shelf doesn't do everything. It does a lot. In particular the orchestration of containers, but it delegates through API a lot of key functions. So for example, secret management, traffic management, there's a need for telemetry and observability at a scale beyond what you get from raw Kubernetes. That is to say, harvesting the metrics that are coming out of Kubernetes and processing them, storing them in time series databases, dashboarding them, and so on. There's also an angle to Kubernetes that gets a lot of attention in the daily DevOps routine, that's not really part of the open source deliverable itself, and that is the DevOps sort of CICD pipeline-oriented lifecycle. And that is something else that we've added and integrated nicely. And then one more piece of this picture is that within a Kubernetes cluster, there's a function that is critical to allowing services to discover each other and integrate with each other securely and with proper configuration provided by the concept of a service mesh. So Istio, Linkerd, these are examples of service mesh technologies. And we have gone ahead and integrated actually those two. There's more than those two, but we've integrated those two with Kubernetes. So the net effect is that when a developer within Walmart is going to build an application, they don't have to think about all those other capabilities where they come from or how they're provided. Those are already present, and the way the CICD pipelines are set up, it's already sort of in the picture, and there are configuration points that they can take advantage of in the primary YAML and a couple of other pieces of config that we supply where they can tune it. But at the end of the day, it offloads an awful lot of work for them, having to stand up and operate those services, fail them over properly, and make them robust. All of that's provided for. >> Yeah, you know, developers often complain they spend too much time wrangling and doing things that aren't productive. So I wonder if you could talk about the high level business goals of the initiative in terms of the hardcore benefits. Was the real impetus to tap into best of breed cloud services? Were you trying to cut costs? Maybe gain negotiating leverage with the cloud guys? Resiliency, you know, I know was a major theme. Maybe you could give us a sense of kind of the anatomy of the decision making process that went in. >> Sure, and in the course of answering your question, I think I'm going to introduce the concept of our triplet architecture which we haven't yet touched on in the interview here. First off, just to sort of wrap up the motivation for WCNP itself which is kind of orthogonal to the triplet architecture. It can exist with or without it. Currently does exist with it, which is key, and I'll get to that in a moment. The key drivers, business drivers for WCNP were developer productivity by offloading the kinds of concerns that we've just discussed. Number two, improving resiliency, that is to say reducing opportunity for human error. One of the challenges you tend to run into in a large enterprise is what we call snowflakes, lots of gratuitously different workloads, projects, configurations to the extent that by developing and using WCNP and continuing to evolve it as we have, we end up with cookie cutter like consistency across our workloads which is super valuable when it comes to building tools or building services to automate operations that would otherwise be manual. When everything is pretty much done the same way, that becomes much simpler. Another key motivation for WCNP was the ability to abstract from the underlying cloud provider. And this is going to lead to a discussion of our triplet architecture. At the end of the day, when one works directly with an underlying cloud provider, one ends up taking a lot of dependencies on that particular cloud provider. Those dependencies can be valuable. For example, there are best of breed services like say Cloud Spanner offered by Google or say Cosmos DB offered by Microsoft that one wants to use and one is willing to take the dependency on the cloud provider to get that functionality because it's unique and valuable. On the other hand, one doesn't want to take dependencies on a cloud provider that don't add a lot of value. And with Kubernetes, we have the opportunity, and this is a large part of how Kubernetes was designed and why it is the way it is, we have the opportunity to sort of abstract from the underlying cloud provider for stateless workloads on compute. And so what this lets us do is build container-based applications that can run without change on different cloud provider infrastructure. So the same applications can run on WCNP over Azure, WCNP over GCP, or WCNP over the Walmart private cloud. And we have a private cloud. Our private cloud is OpenStack based and it gives us some significant cost advantages as well as control advantages. So to your point, in terms of business motivation, there's a key cost driver here, which is that we can use our own private cloud when it's advantageous and then use the public cloud provider capabilities when we need to. A key place with this comes into play is with elasticity. So while the private cloud is much more cost effective for us to run and use, it isn't as elastic as what the cloud providers offer, right? We don't have essentially unlimited scale. We have large scale, but the public cloud providers are elastic in the extreme which is a very powerful capability. So what we're able to do is burst, and we use this term bursting workloads into the public cloud from the private cloud to take advantage of the elasticity they offer and then fall back into the private cloud when the traffic load diminishes to the point where we don't need that elastic capability, elastic capacity at low cost. And this is a very important paradigm that I think is going to be very commonplace ultimately as the industry evolves. Private cloud is easier to operate and less expensive, and yet the public cloud provider capabilities are difficult to match. >> And the triplet, the tri is your on-prem private cloud and the two public clouds that you mentioned, is that right? >> That is correct. And we actually have an architecture in which we operate all three of those cloud platforms in close proximity with one another in three different major regions in the US. So we have east, west, and central. And in each of those regions, we have all three cloud providers. And the way it's configured, those data centers are within 10 milliseconds of each other, meaning that it's of negligible cost to interact between them. And this allows us to be fairly agnostic to where a particular workload is running. >> Does a human make that decision, Jack or is there some intelligence in the system that determines that? >> That's a really great question, Dave. And it's a great question because we're at the cusp of that transition. So currently humans make that decision. Humans choose to deploy workloads into a particular region and a particular provider within that region. That said, we're actively developing patterns and practices that will allow us to automate the placement of the workloads for a variety of criteria. For example, if in a particular region, a particular provider is heavily overloaded and is unable to provide the level of service that's expected through our SLAs, we could choose to fail workloads over from that cloud provider to a different one within the same region. But that's manual today. We do that, but people do it. Okay, we'd like to get to where that happens automatically. In the same way, we'd like to be able to automate the failovers, both for high availability and sort of the heavier disaster recovery model between, within a region between providers and even within a provider between the availability zones that are there, but also between regions for the sort of heavier disaster recovery or maintenance driven realignment of workload placement. Today, that's all manual. So we have people moving workloads from region A to region B or data center A to data center B. It's clean because of the abstraction. The workloads don't have to know or care, but there are latency considerations that come into play, and the humans have to be cognizant of those. And automating that can help ensure that we get the best performance and the best reliability. >> But you're developing the dataset to actually, I would imagine, be able to make those decisions in an automated fashion over time anyway. Is that a fair assumption? >> It is, and that's what we're actively developing right now. So if you were to look at us today, we have these nice abstractions and APIs in place, but people run that machine, if you will, moving toward a world where that machine is fully automated. >> What exactly are you abstracting? Is it sort of the deployment model or, you know, are you able to abstract, I'm just making this up like Azure functions and GCP functions so that you can sort of run them, you know, with a consistent experience. What exactly are you abstracting and how difficult was it to achieve that objective technically? >> that's a good question. What we're abstracting is the Kubernetes node construct. That is to say a cluster of Kubernetes nodes which are typically VMs, although they can run bare metal in certain contexts, is something that typically to stand up requires knowledge of the underlying cloud provider. So for example, with GCP, you would use GKE to set up a Kubernetes cluster, and in Azure, you'd use AKS. We are actually abstracting that aspect of things so that the developers standing up applications don't have to know what the underlying cluster management provider is. They don't have to know if it's GCP, AKS or our own Walmart private cloud. Now, in terms of functions like Azure functions that you've mentioned there, we haven't done that yet. That's another piece that we have sort of on our radar screen that, we'd like to get to is serverless approach, and the Knative work from Google and the Azure functions, those are things that we see good opportunity to use for a whole variety of use cases. But right now we're not doing much with that. We're strictly container based right now, and we do have some VMs that are running in sort of more of a traditional model. So our stateful workloads are primarily VM based, but for serverless, that's an opportunity for us to take some of these stateless workloads and turn them into cloud functions. >> Well, and that's another cost lever that you can pull down the road that's going to drop right to the bottom line. Do you see a day or maybe you're doing it today, but I'd be surprised, but where you build applications that actually span multiple clouds or is there, in your view, always going to be a direct one-to-one mapping between where an application runs and the specific cloud platform? >> That's a really great question. Well, yes and no. So today, application development teams choose a cloud provider to deploy to and a location to deploy to, and they have to get involved in moving an application like we talked about today. That said, the bursting capability that I mentioned previously is something that is a step in the direction of automatic migration. That is to say we're migrating workload to different locations automatically. Currently, the prototypes we've been developing and that we think are going to eventually make their way into production are leveraging Istio to assess the load incoming on a particular cluster and start shedding that load into a different location. Right now, the configuration of that is still manual, but there's another opportunity for automation there. And I think a key piece of this is that down the road, well, that's a, sort of a small step in the direction of an application being multi provider. We expect to see really an abstraction of the fact that there is a triplet even. So the workloads are moving around according to whatever the control plane decides is necessary based on a whole variety of inputs. And at that point, you will have true multi-cloud applications, applications that are distributed across the different providers and in a way that application developers don't have to think about. >> So Walmart's been a leader, Jack, in using data for competitive advantages for decades. It's kind of been a poster child for that. You've got a mountain of IP in the form of data, tools, applications best practices that until the cloud came out was all On Prem. But I'm really interested in this idea of building a Walmart ecosystem, which obviously you have. Do you see a day or maybe you're even doing it today where you take what we call the Walmart SuperCloud, WCNP in your words, and point or turn that toward an external world or your ecosystem, you know, supporting those partners or customers that could drive new revenue streams, you know directly from the platform? >> Great questions, Dave. So there's really two things to say here. The first is that with respect to data, our data workloads are primarily VM basis. I've mentioned before some VMware, some straight open stack. But the key here is that WCNP and Kubernetes are very powerful for stateless workloads, but for stateful workloads tend to be still climbing a bit of a growth curve in the industry. So our data workloads are not primarily based on WCNP. They're VM based. Now that said, there is opportunity to make some progress there, and we are looking at ways to move things into containers that are currently running in VMs which are stateful. The other question you asked is related to how we expose data to third parties and also functionality. Right now we do have in-house, for our own use, a very robust data architecture, and we have followed the sort of domain-oriented data architecture guidance from Martin Fowler. And we have data lakes in which we collect data from all the transactional systems and which we can then use and do use to build models which are then used in our applications. But right now we're not exposing the data directly to customers as a product. That's an interesting direction that's been talked about and may happen at some point, but right now that's internal. What we are exposing to customers is applications. So we're offering our global integrated fulfillment capabilities, our order picking and curbside pickup capabilities, and our cloud powered checkout capabilities to third parties. And this means we're standing up our own internal applications as externally facing SaaS applications which can serve our partners' customers. >> Yeah, of course, Martin Fowler really first introduced to the world Zhamak Dehghani's data mesh concept and this whole idea of data products and domain oriented thinking. Zhamak Dehghani, by the way, is a speaker at our event as well. Last question I had is edge, and how you think about the edge? You know, the stores are an edge. Are you putting resources there that sort of mirror this this triplet model? Or is it better to consolidate things in the cloud? I know there are trade-offs in terms of latency. How are you thinking about that? >> All really good questions. It's a challenging area as you can imagine because edges are subject to disconnection, right? Or reduced connection. So we do place the same architecture at the edge. So WCNP runs at the edge, and an application that's designed to run at WCNP can run at the edge. That said, there are a number of very specific considerations that come up when running at the edge, such as the possibility of disconnection or degraded connectivity. And so one of the challenges we have faced and have grappled with and done a good job of I think is dealing with the fact that applications go offline and come back online and have to reconnect and resynchronize, the sort of online offline capability is something that can be quite challenging. And we have a couple of application architectures that sort of form the two core sets of patterns that we use. One is an offline/online synchronization architecture where we discover that we've come back online, and we understand the differences between the online dataset and the offline dataset and how they have to be reconciled. The other is a message-based architecture. And here in our health and wellness domain, we've developed applications that are queue based. So they're essentially business processes that consist of multiple steps where each step has its own queue. And what that allows us to do is devote whatever bandwidth we do have to those pieces of the process that are most latency sensitive and allow the queue lengths to increase in parts of the process that are not latency sensitive, knowing that they will eventually catch up when the bandwidth is restored. And to put that in a little bit of context, we have fiber lengths to all of our locations, and we have I'll just use a round number, 10-ish thousand locations. It's larger than that, but that's the ballpark, and we have fiber to all of them, but when the fiber is disconnected, When the disconnection happens, we're able to fall back to 5G and to Starlink. Starlink is preferred. It's a higher bandwidth. 5G if that fails. But in each of those cases, the bandwidth drops significantly. And so the applications have to be intelligent about throttling back the traffic that isn't essential, so that it can push the essential traffic in those lower bandwidth scenarios. >> So much technology to support this amazing business which started in the early 1960s. Jack, unfortunately, we're out of time. I would love to have you back or some members of your team and drill into how you're using open source, but really thank you so much for explaining the approach that you've taken and participating in SuperCloud2. >> You're very welcome, Dave, and we're happy to come back and talk about other aspects of what we do. For example, we could talk more about the data lakes and the data mesh that we have in place. We could talk more about the directions we might go with serverless. So please look us up again. Happy to chat. >> I'm going to take you up on that, Jack. All right. This is Dave Vellante for John Furrier and the Cube community. Keep it right there for more action from SuperCloud2. (upbeat music)
SUMMARY :
and the Chief Architect for and appreciate the the Walmart Cloud Native Platform? and that is the DevOps Was the real impetus to tap into Sure, and in the course And the way it's configured, and the humans have to the dataset to actually, but people run that machine, if you will, Is it sort of the deployment so that the developers and the specific cloud platform? and that we think are going in the form of data, tools, applications a bit of a growth curve in the industry. and how you think about the edge? and allow the queue lengths to increase for explaining the and the data mesh that we have in place. and the Cube community.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Jack Greenfield | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Jack | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Martin Fowler | PERSON | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
US | LOCATION | 0.99+ |
Zhamak Dehghani | PERSON | 0.99+ |
Today | DATE | 0.99+ |
each | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
today | DATE | 0.99+ |
two things | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
each step | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
early 1960s | DATE | 0.99+ |
Starlink | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.98+ |
a day | QUANTITY | 0.97+ |
GCP | TITLE | 0.97+ |
Azure | TITLE | 0.96+ |
WCNP | TITLE | 0.96+ |
10 milliseconds | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.94+ |
Cloud Spanner | TITLE | 0.94+ |
Linkerd | ORGANIZATION | 0.93+ |
triplet | QUANTITY | 0.92+ |
three cloud providers | QUANTITY | 0.91+ |
Cube | ORGANIZATION | 0.9+ |
SuperCloud2 | ORGANIZATION | 0.89+ |
two core sets | QUANTITY | 0.88+ |
John Furrier | PERSON | 0.88+ |
one more piece | QUANTITY | 0.86+ |
two public clouds | QUANTITY | 0.86+ |
thousand locations | QUANTITY | 0.83+ |
Vice President | PERSON | 0.8+ |
10-ish | QUANTITY | 0.79+ |
WCNP | ORGANIZATION | 0.75+ |
decades | QUANTITY | 0.75+ |
three different major regions | QUANTITY | 0.74+ |
Jack Greenfield, Walmart | A Dive into Walmart's Retail Supercloud
>> Welcome back to SuperCloud2. This is Dave Vellante, and we're here with Jack Greenfield. He's the Vice President of Enterprise Architecture and the Chief Architect for the global technology platform at Walmart. Jack, I want to thank you for coming on the program. Really appreciate your time. >> Glad to be here, Dave. Thanks for inviting me and appreciate the opportunity to chat with you. >> Yeah, it's our pleasure. Now we call what you've built a SuperCloud. That's our term, not yours, but how would you describe the Walmart Cloud Native Platform? >> So WCNP, as the acronym goes, is essentially an implementation of Kubernetes for the Walmart ecosystem. And what that means is that we've taken Kubernetes off the shelf as open source, and we have integrated it with a number of foundational services that provide other aspects of our computational environment. So Kubernetes off the shelf doesn't do everything. It does a lot. In particular the orchestration of containers, but it delegates through API a lot of key functions. So for example, secret management, traffic management, there's a need for telemetry and observability at a scale beyond what you get from raw Kubernetes. That is to say, harvesting the metrics that are coming out of Kubernetes and processing them, storing them in time series databases, dashboarding them, and so on. There's also an angle to Kubernetes that gets a lot of attention in the daily DevOps routine, that's not really part of the open source deliverable itself, and that is the DevOps sort of CICD pipeline-oriented lifecycle. And that is something else that we've added and integrated nicely. And then one more piece of this picture is that within a Kubernetes cluster, there's a function that is critical to allowing services to discover each other and integrate with each other securely and with proper configuration provided by the concept of a service mesh. So Istio, Linkerd, these are examples of service mesh technologies. And we have gone ahead and integrated actually those two. There's more than those two, but we've integrated those two with Kubernetes. So the net effect is that when a developer within Walmart is going to build an application, they don't have to think about all those other capabilities where they come from or how they're provided. Those are already present, and the way the CICD pipelines are set up, it's already sort of in the picture, and there are configuration points that they can take advantage of in the primary YAML and a couple of other pieces of config that we supply where they can tune it. But at the end of the day, it offloads an awful lot of work for them, having to stand up and operate those services, fail them over properly, and make them robust. All of that's provided for. >> Yeah, you know, developers often complain they spend too much time wrangling and doing things that aren't productive. So I wonder if you could talk about the high level business goals of the initiative in terms of the hardcore benefits. Was the real impetus to tap into best of breed cloud services? Were you trying to cut costs? Maybe gain negotiating leverage with the cloud guys? Resiliency, you know, I know was a major theme. Maybe you could give us a sense of kind of the anatomy of the decision making process that went in. >> Sure, and in the course of answering your question, I think I'm going to introduce the concept of our triplet architecture which we haven't yet touched on in the interview here. First off, just to sort of wrap up the motivation for WCNP itself which is kind of orthogonal to the triplet architecture. It can exist with or without it. Currently does exist with it, which is key, and I'll get to that in a moment. The key drivers, business drivers for WCNP were developer productivity by offloading the kinds of concerns that we've just discussed. Number two, improving resiliency, that is to say reducing opportunity for human error. One of the challenges you tend to run into in a large enterprise is what we call snowflakes, lots of gratuitously different workloads, projects, configurations to the extent that by developing and using WCNP and continuing to evolve it as we have, we end up with cookie cutter like consistency across our workloads which is super valuable when it comes to building tools or building services to automate operations that would otherwise be manual. When everything is pretty much done the same way, that becomes much simpler. Another key motivation for WCNP was the ability to abstract from the underlying cloud provider. And this is going to lead to a discussion of our triplet architecture. At the end of the day, when one works directly with an underlying cloud provider, one ends up taking a lot of dependencies on that particular cloud provider. Those dependencies can be valuable. For example, there are best of breed services like say Cloud Spanner offered by Google or say Cosmos DB offered by Microsoft that one wants to use and one is willing to take the dependency on the cloud provider to get that functionality because it's unique and valuable. On the other hand, one doesn't want to take dependencies on a cloud provider that don't add a lot of value. And with Kubernetes, we have the opportunity, and this is a large part of how Kubernetes was designed and why it is the way it is, we have the opportunity to sort of abstract from the underlying cloud provider for stateless workloads on compute. And so what this lets us do is build container-based applications that can run without change on different cloud provider infrastructure. So the same applications can run on WCNP over Azure, WCNP over GCP, or WCNP over the Walmart private cloud. And we have a private cloud. Our private cloud is OpenStack based and it gives us some significant cost advantages as well as control advantages. So to your point, in terms of business motivation, there's a key cost driver here, which is that we can use our own private cloud when it's advantageous and then use the public cloud provider capabilities when we need to. A key place with this comes into play is with elasticity. So while the private cloud is much more cost effective for us to run and use, it isn't as elastic as what the cloud providers offer, right? We don't have essentially unlimited scale. We have large scale, but the public cloud providers are elastic in the extreme which is a very powerful capability. So what we're able to do is burst, and we use this term bursting workloads into the public cloud from the private cloud to take advantage of the elasticity they offer and then fall back into the private cloud when the traffic load diminishes to the point where we don't need that elastic capability, elastic capacity at low cost. And this is a very important paradigm that I think is going to be very commonplace ultimately as the industry evolves. Private cloud is easier to operate and less expensive, and yet the public cloud provider capabilities are difficult to match. >> And the triplet, the tri is your on-prem private cloud and the two public clouds that you mentioned, is that right? >> That is correct. And we actually have an architecture in which we operate all three of those cloud platforms in close proximity with one another in three different major regions in the US. So we have east, west, and central. And in each of those regions, we have all three cloud providers. And the way it's configured, those data centers are within 10 milliseconds of each other, meaning that it's of negligible cost to interact between them. And this allows us to be fairly agnostic to where a particular workload is running. >> Does a human make that decision, Jack or is there some intelligence in the system that determines that? >> That's a really great question, Dave. And it's a great question because we're at the cusp of that transition. So currently humans make that decision. Humans choose to deploy workloads into a particular region and a particular provider within that region. That said, we're actively developing patterns and practices that will allow us to automate the placement of the workloads for a variety of criteria. For example, if in a particular region, a particular provider is heavily overloaded and is unable to provide the level of service that's expected through our SLAs, we could choose to fail workloads over from that cloud provider to a different one within the same region. But that's manual today. We do that, but people do it. Okay, we'd like to get to where that happens automatically. In the same way, we'd like to be able to automate the failovers, both for high availability and sort of the heavier disaster recovery model between, within a region between providers and even within a provider between the availability zones that are there, but also between regions for the sort of heavier disaster recovery or maintenance driven realignment of workload placement. Today, that's all manual. So we have people moving workloads from region A to region B or data center A to data center B. It's clean because of the abstraction. The workloads don't have to know or care, but there are latency considerations that come into play, and the humans have to be cognizant of those. And automating that can help ensure that we get the best performance and the best reliability. >> But you're developing the dataset to actually, I would imagine, be able to make those decisions in an automated fashion over time anyway. Is that a fair assumption? >> It is, and that's what we're actively developing right now. So if you were to look at us today, we have these nice abstractions and APIs in place, but people run that machine, if you will, moving toward a world where that machine is fully automated. >> What exactly are you abstracting? Is it sort of the deployment model or, you know, are you able to abstract, I'm just making this up like Azure functions and GCP functions so that you can sort of run them, you know, with a consistent experience. What exactly are you abstracting and how difficult was it to achieve that objective technically? >> that's a good question. What we're abstracting is the Kubernetes node construct. That is to say a cluster of Kubernetes nodes which are typically VMs, although they can run bare metal in certain contexts, is something that typically to stand up requires knowledge of the underlying cloud provider. So for example, with GCP, you would use GKE to set up a Kubernetes cluster, and in Azure, you'd use AKS. We are actually abstracting that aspect of things so that the developers standing up applications don't have to know what the underlying cluster management provider is. They don't have to know if it's GCP, AKS or our own Walmart private cloud. Now, in terms of functions like Azure functions that you've mentioned there, we haven't done that yet. That's another piece that we have sort of on our radar screen that, we'd like to get to is serverless approach, and the Knative work from Google and the Azure functions, those are things that we see good opportunity to use for a whole variety of use cases. But right now we're not doing much with that. We're strictly container based right now, and we do have some VMs that are running in sort of more of a traditional model. So our stateful workloads are primarily VM based, but for serverless, that's an opportunity for us to take some of these stateless workloads and turn them into cloud functions. >> Well, and that's another cost lever that you can pull down the road that's going to drop right to the bottom line. Do you see a day or maybe you're doing it today, but I'd be surprised, but where you build applications that actually span multiple clouds or is there, in your view, always going to be a direct one-to-one mapping between where an application runs and the specific cloud platform? >> That's a really great question. Well, yes and no. So today, application development teams choose a cloud provider to deploy to and a location to deploy to, and they have to get involved in moving an application like we talked about today. That said, the bursting capability that I mentioned previously is something that is a step in the direction of automatic migration. That is to say we're migrating workload to different locations automatically. Currently, the prototypes we've been developing and that we think are going to eventually make their way into production are leveraging Istio to assess the load incoming on a particular cluster and start shedding that load into a different location. Right now, the configuration of that is still manual, but there's another opportunity for automation there. And I think a key piece of this is that down the road, well, that's a, sort of a small step in the direction of an application being multi provider. We expect to see really an abstraction of the fact that there is a triplet even. So the workloads are moving around according to whatever the control plane decides is necessary based on a whole variety of inputs. And at that point, you will have true multi-cloud applications, applications that are distributed across the different providers and in a way that application developers don't have to think about. >> So Walmart's been a leader, Jack, in using data for competitive advantages for decades. It's kind of been a poster child for that. You've got a mountain of IP in the form of data, tools, applications best practices that until the cloud came out was all On Prem. But I'm really interested in this idea of building a Walmart ecosystem, which obviously you have. Do you see a day or maybe you're even doing it today where you take what we call the Walmart SuperCloud, WCNP in your words, and point or turn that toward an external world or your ecosystem, you know, supporting those partners or customers that could drive new revenue streams, you know directly from the platform? >> Great question, Steve. So there's really two things to say here. The first is that with respect to data, our data workloads are primarily VM basis. I've mentioned before some VMware, some straight open stack. But the key here is that WCNP and Kubernetes are very powerful for stateless workloads, but for stateful workloads tend to be still climbing a bit of a growth curve in the industry. So our data workloads are not primarily based on WCNP. They're VM based. Now that said, there is opportunity to make some progress there, and we are looking at ways to move things into containers that are currently running in VMs which are stateful. The other question you asked is related to how we expose data to third parties and also functionality. Right now we do have in-house, for our own use, a very robust data architecture, and we have followed the sort of domain-oriented data architecture guidance from Martin Fowler. And we have data lakes in which we collect data from all the transactional systems and which we can then use and do use to build models which are then used in our applications. But right now we're not exposing the data directly to customers as a product. That's an interesting direction that's been talked about and may happen at some point, but right now that's internal. What we are exposing to customers is applications. So we're offering our global integrated fulfillment capabilities, our order picking and curbside pickup capabilities, and our cloud powered checkout capabilities to third parties. And this means we're standing up our own internal applications as externally facing SaaS applications which can serve our partners' customers. >> Yeah, of course, Martin Fowler really first introduced to the world Zhamak Dehghani's data mesh concept and this whole idea of data products and domain oriented thinking. Zhamak Dehghani, by the way, is a speaker at our event as well. Last question I had is edge, and how you think about the edge? You know, the stores are an edge. Are you putting resources there that sort of mirror this this triplet model? Or is it better to consolidate things in the cloud? I know there are trade-offs in terms of latency. How are you thinking about that? >> All really good questions. It's a challenging area as you can imagine because edges are subject to disconnection, right? Or reduced connection. So we do place the same architecture at the edge. So WCNP runs at the edge, and an application that's designed to run at WCNP can run at the edge. That said, there are a number of very specific considerations that come up when running at the edge, such as the possibility of disconnection or degraded connectivity. And so one of the challenges we have faced and have grappled with and done a good job of I think is dealing with the fact that applications go offline and come back online and have to reconnect and resynchronize, the sort of online offline capability is something that can be quite challenging. And we have a couple of application architectures that sort of form the two core sets of patterns that we use. One is an offline/online synchronization architecture where we discover that we've come back online, and we understand the differences between the online dataset and the offline dataset and how they have to be reconciled. The other is a message-based architecture. And here in our health and wellness domain, we've developed applications that are queue based. So they're essentially business processes that consist of multiple steps where each step has its own queue. And what that allows us to do is devote whatever bandwidth we do have to those pieces of the process that are most latency sensitive and allow the queue lengths to increase in parts of the process that are not latency sensitive, knowing that they will eventually catch up when the bandwidth is restored. And to put that in a little bit of context, we have fiber lengths to all of our locations, and we have I'll just use a round number, 10-ish thousand locations. It's larger than that, but that's the ballpark, and we have fiber to all of them, but when the fiber is disconnected, and it does get disconnected on a regular basis. In fact, I forget the exact number, but some several dozen locations get disconnected daily just by virtue of the fact that there's construction going on and things are happening in the real world. When the disconnection happens, we're able to fall back to 5G and to Starlink. Starlink is preferred. It's a higher bandwidth. 5G if that fails. But in each of those cases, the bandwidth drops significantly. And so the applications have to be intelligent about throttling back the traffic that isn't essential, so that it can push the essential traffic in those lower bandwidth scenarios. >> So much technology to support this amazing business which started in the early 1960s. Jack, unfortunately, we're out of time. I would love to have you back or some members of your team and drill into how you're using open source, but really thank you so much for explaining the approach that you've taken and participating in SuperCloud2. >> You're very welcome, Dave, and we're happy to come back and talk about other aspects of what we do. For example, we could talk more about the data lakes and the data mesh that we have in place. We could talk more about the directions we might go with serverless. So please look us up again. Happy to chat. >> I'm going to take you up on that, Jack. All right. This is Dave Vellante for John Furrier and the Cube community. Keep it right there for more action from SuperCloud2. (upbeat music)
SUMMARY :
and the Chief Architect for and appreciate the the Walmart Cloud Native Platform? and that is the DevOps Was the real impetus to tap into Sure, and in the course And the way it's configured, and the humans have to the dataset to actually, but people run that machine, if you will, Is it sort of the deployment so that the developers and the specific cloud platform? and that we think are going in the form of data, tools, applications a bit of a growth curve in the industry. and how you think about the edge? and allow the queue lengths to increase for explaining the and the data mesh that we have in place. and the Cube community.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Steve | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Jack Greenfield | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Jack | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
Martin Fowler | PERSON | 0.99+ |
US | LOCATION | 0.99+ |
Zhamak Dehghani | PERSON | 0.99+ |
Today | DATE | 0.99+ |
each | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Starlink | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
two things | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
each step | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
early 1960s | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
a day | QUANTITY | 0.98+ |
GCP | TITLE | 0.97+ |
Azure | TITLE | 0.96+ |
WCNP | TITLE | 0.96+ |
10 milliseconds | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.94+ |
Cloud Spanner | TITLE | 0.94+ |
Linkerd | ORGANIZATION | 0.93+ |
Cube | ORGANIZATION | 0.93+ |
triplet | QUANTITY | 0.92+ |
three cloud providers | QUANTITY | 0.91+ |
two core sets | QUANTITY | 0.88+ |
John Furrier | PERSON | 0.86+ |
one more piece | QUANTITY | 0.86+ |
SuperCloud2 | ORGANIZATION | 0.86+ |
two public clouds | QUANTITY | 0.86+ |
thousand locations | QUANTITY | 0.83+ |
Vice President | PERSON | 0.8+ |
10-ish | QUANTITY | 0.79+ |
WCNP | ORGANIZATION | 0.75+ |
decades | QUANTITY | 0.75+ |
three different major regions | QUANTITY | 0.74+ |
Manu Parbhakar, AWS & Joel Jackson, Red Hat | AWS re:Invent 2022
>>Hello, brilliant humans and welcome back to Las Vegas, Nevada, where we are live from the AWS Reinvent Show floor here with the cube. My name is Savannah Peterson, joined with Dave Valante, and we have a very exciting conversation with you. Two, two companies you may have heard of. We've got AWS and Red Hat in the house. Manu and Joel, thank you so much for being here. Love this little fist bump. Started off, that's right. Before we even got rolling, Manu, you said that you wanted this to be the best segment of, of the cubes airing. We we're doing over a hundred segments, so you're gonna have to bring the heat. >>We're ready. We're did go. Are we ready? Yeah, go. We're ready. Let's bring it on. >>We're ready. All right. I'm, I'm ready. Dave's ready. Let's do it. How's the show going for you guys real quick before we dig in? >>Yeah, I think after Covid, it's really nice to see that we're back into the 2019 level and, you know, people just want to get out, meet people, have that human touch with each other, and I think a lot of trust gets built as a functional that, so it's super amazing to see our partners and customers here at Reedman. Yeah, >>And you've got a few in the house. That's true. Just a few maybe, maybe a couple >>Very few shows can say that, by the way. Yeah, it's maybe a handful. >>I think one of the things we were saying, it's almost like the entire Silicon Valley descended in the expo hall area, so >>Yeah, it's >>For a few different reasons. There's a few different silicon defined. Yeah, yeah, yeah. Don't have strong on for you. So far >>It's, it's, it is amazing. It's the 10th year, right? It's decade, I think I've been to five and it's, it grows every single year. It's the, you have to be here. It's as simple as that. And customers from every single industry are here too. You don't get, a lot of shows have every single industry and almost every single location around the globe. So it's, it's a must, must be >>Here. Well, and the personas evolved, right? I was at reinvent number two. That was my first, and it was all developers, not all, but a lot of developers. And today it's a business mix, really is >>Totally, is a business mix. And I just, I've talked about it a little bit down the show, but the diversity on the show floor, it's the first time I've had to wait in line for the ladies' room at a tech conference. Almost a two decade career. It is, yeah. And it was really refreshing. I'm so impressed. So clearly there's a commitment to community, but also a commitment to diversity. Yeah. And, and it's brilliant to see on the show floor. This is a partnership that is robust and has been around for a little while. Money. Why don't you tell us a little bit about the partnership here? >>Yes. So Red Hand and AWS are best friends, you know, forever together. >>Aw, no wonder we got the fist bumps and all the good vibes coming out. I know, it's great. I love that >>We have a decade of working together. I think the relationship in the first phase was around running rail bundled with E two. Sure. We have about 70,000 customers that are running rail, which are running mission critical workloads such as sap, Oracle databases, bespoke applications across the state of verticals. Now, as more and more enterprise customers are finally, you know, endorsing and adopting public cloud, I think that business is just gonna continue to grow. So a, a lot of progress there. The second titration has been around, you know, developers tearing Red Hat and aws, Hey, listen, we wanna, it's getting competitive. We wanna deliver new features faster, quicker, we want scale and we want resilience. So just entire push towards devs containers. So that's the second chapter with, you know, red Hat OpenShift on aws, which launched as a, a joint manage service in 2021 last year. And I think the third phase, which you're super excited about, is just bringing the ease of consumption, one click deployment, and then having our customers, you know, benefit from the joint committed spend programs together. So, you know, making sure that re and Ansible and JBoss, the entire portfolio of Red Hat products are available on AWS marketplace. So that's the 1, 2, 3, it of our relationship. It's a decade of working together and, you know, best friends are super committed to making sure our customers and partners continue successful. >>Yeah, that he said it, he said it perfectly. 2008, I know you don't like that, but we started with Rel on demand just in 2008 before E two even had a console. So the partnership has been there, like Manu says, for a long time, we got the partnership, we got the products up there now, and we just gotta finalize that, go to market and get that gas on the fire. >>Yeah. So Graviton Outpost, local zones, you lead it into all the new stuff. So that portends, I mean, 2008, we're talking two years after the launch of s3. >>That's right. >>Right. So, and now look, so is this a harbinger of things to come with these new innovations? >>Yeah, I, I would say, you know, the innovation is a key tenant of our partnership, our relationship. So if you look at from a product standpoint, red Hat or Rel was one of the first platforms that made a support for graviton, which is basically 40% better price performance than any other distribution. Then that translated into making sure that Rel is available on all of our regions globally. So this year we launched Switzerland, Spain, India, and Red Hat was available on launch there, support for Nitro support for Outpost Rosa support on Outpost as well. So I think that relationship, that innovation on the product side, that's pretty visible. I think that innovation again then translates into what we are doing on marketplace with one click deployments we spoke about. I think the third aspect of the know innovation is around making sure that we are making our partners and our customers successful. So one of the things that we've done so far is Joe leads a, you know, a black belt team that really goes into each customer opportunity, making sure how can we help you be successful. We launched and you know, we should be able to share that on a link. After this, we launched like a big playlist, which talks about every single use case on how do you get successful and running OpenShift on aws. So that innovation on behalf of our customers partners to make them successful, that's been a key tenant for us together as >>Well. That's right. And that team that Manu is talking about, we're gonna, gonna 10 x that team this year going into January. Our fiscal yield starts in January. Love that. So yeah, we're gonna have a lot of no hiring freeze over here. Nope. No ma'am. No. Yeah, that's right. Yeah. And you know what I love about working with aws and, and, and Manu just said it very, all of that's customer driven. Every single event that we, that he just talked about in that timeline, it's customer driven, right? Customers wanted rail on demand, customers want JBoss up in the cloud, Ansible this week, you know, everything's up there now. So it's just getting that go to market tight and we're gonna, we're gonna get that done. >>So what's the algorithm for customer driven in terms of taking the input? Because if every customers saying, Hey, I this a >>Really similar >>Question right up, right? I, that's what I want. And if you know, 95% of the customers say it, Jay, maybe that's a good idea. >>Yeah, that's right. Trends. But >>Yeah. You know, 30% you might be like, mm, you know, 20%, you know, how do you guys decide when to put gas on the fire? >>No, that, I think, as I mentioned, there are about 70,000 large customers that are running rail on Easy Two, many of these customers are informing our product strategy. So we have, you know, close to about couple of thousand power users. We have customer advisory booths, and these are the, you know, customers are informing us, Hey, let's get all of the Red Hat portfolio and marketplace support for graviton, support for Outpost. Why don't we, why are we not able to dip into the consumption committed spend programs for both Red Hat and aws? That's right. So it's these power users both at the developer level as well as the guys who are actually doing large commercial consumption. They are the ones who are informing the roadmap for both Red Hat and aws. >>But do, do you codify the the feedback? >>Yeah, I'm like, I wanna see the database, >>The, I think it was, I don't know, it was maybe Chasy, maybe it was Besos, that that data beats intuition. So do you take that information and somehow, I mean, it's global, 70,000 customers, right? And they have different weights, different spending patterns, different levels of maturity. Yeah. Do you, how do you codify that and then ultimately make the decision? Yeah, I >>If, I mean, well you, you've got the strategic advisory boards, which are made up of customers and partners and you know, you get, you get a good, you gotta get a good slice of your customer base to get, and you gotta take their feedback and you gotta do something with it, right? That's the, that's the way we do it and codify it at the product level, I'm sure open source. That's, that's basically how we work at the product level, right? The most elegant solution in open source wins. And that's, that's pretty much how we do that at the, >>I would just add, I think it's also just the implicit trust that the two companies had built with each other, working in the trenches, making our customers and partners successful over the last decade. And Alex, give an example. So that manifests itself in context of like, you know, Amazon and Red Hat just published the entire roadmap for OpenShift. What are the new features that are becoming over the next six to nine to 12 months? It's open source available on GitHub. Customers can see, and then they can basically come back and give feedback like, Hey, you know, we want hip compliance. We just launched. That was a big request that was coming from our >>Customers. That is not any process >>Also for Graviton or Nvidia instances. So I I I think it's a, >>Here's the thing, the reason I'm pounding on this is because you guys have a pretty high hit rate, and I think as a >>Customer, mildly successful company >>As, as a customer advocate, the better, you know, if, if you guys make bets that pay off, it's gonna pay off for customers. Right. And because there's a lot of failures in it. Yeah. I mean, let's face it. That's >>Right. And I think, I think you said the key word bets. You place a lot of small bets. Do you have the, the innovation engine to do that? AWS is the perfect place to place those small bets. And then you, you know, pour gas on the fire when, when they take off. >>Yeah, it's a good point. I mean, it's not expensive to experiment. Yeah. >>Especially in the managed service world. Right? >>And I know you love taking things to market and you're a go to market guy. Let's talk gtm, what's got your snow pumped about GTM for 2023? >>We, we are gonna, you know, 10 x the teams that's gonna be focused on these products, right? So we're gonna also come out with a hybrid committed spend program for our customers that meet them where they want to go. So they're coming outta the data center going into a cloud. We're gonna have a nice financial model for them to do that. And that's gonna take a lot of the friction out. >>Yeah. I mean, you've nailed it. I, I think the, the fact that now entire Red Hat portfolio is available on marketplace, you can do it on one click deployment. It's deeply integrated with Amazon services and the most important part that Joel was making now customers can double dip. They can drive benefit from the consumption committed spend programs, both from Red Hat and from aws, which is amazing. Which is a game changer That's right. For many of our large >>Customers. That's right. And that, so we're gonna, we're gonna really go to town on that next year. That's, and all the, all the resources that I have, which are the technology sellers and the sas, you know, the engineers we're growing this team the most out that team. So it's, >>When you say 10 x, how many are you at now? I'm >>Curious to see where you're headed. Tell you, okay. There's not right? Oh no, there's not one. It's triple digit. Yeah, yeah. >>Today. Oh, sweet. Awesome. >>So, and it's a very sizable team. They're actually making sure that each of our customers are successful and then really making sure that, you know, no customer left behind policy. >>And it's a great point that customers love when Amazonians and Red Hats show up, they love it and it's, they want to get more of it, and we're gonna, we're gonna give it to 'em. >>Must feel great to be loved like that. >>Yeah, that's right. Yeah. Yeah. I would say yes. >>Seems like it's safe to say that there's another decade of partnership between your two companies. >>Hope so. That's right. That's the plan. >>Yeah. And I would say also, you know, just the IBM coming into the mix here. Yeah. I, you know, red Hat has informed the way we have turned around our partnership with ibm, essentially we, we signed the strategic collaboration agreement with the company. All of IBM software now runs on Rosa. So that is now also providing a lot of tailwinds both to our rail customers and as well as Rosa customers. And I think it's a very net creative, very positive for our partnership. >>That's right. It's been very positive. Yep. Yeah. >>You see the >>Billboards positive. Yeah, right. Also that, that's great. Great point, Dave. Yep. We have a, we have a new challenge, a new tradition on the cube here at Reinvent where we're, well, it's actually kind of a glamor moment for you, depending on how you leverage it. We're looking for your 32nd hot take your Instagram reel, your sizzle thought leadership, biggest takeaway, most important theme from this year's show. I know you want, right, Joel? I mean, you TM boy, I feel like you can spit the time. >>Yeah. It is all about Rosa for us. It is all in on that, that's the native OpenShift offering on aws and that's, that's the soundbite we're going go to town with. Now, I don't wanna forget all the other products that are in there, but Rosa is a, is a very key push for us this year. >>Fantastic. All right. Manu. >>I think our customers, it's getting super competitive. Our customers want to innovate just a >>Little bit. >>The enterprise customers see the cloud native companies. I wanna do what these guys are doing. I wanna develop features at a fast clip. I wanna scale, I wanna be resilient. And I think that's really the spirit that's coming out. So to Joel's point, you know, move to worlds containers, serverless, DevOps, which was like, you know, aha, something that's happening on the side of an enterprise is not becoming mainstream. The business is demanding it. The, it is becoming the centerpiece in the business strategy. So that's been really like the aha. Big thing that's happening here. >>Yeah. And those architectures are coming together, aren't they? That's correct. Right. You know, VMs and containers, it used to be one architecture and then at the other end of the spectrum is serverless. People thought of those as different things and now it's a single architecture and, and it's kind of right approach for the right job. >>And, and a compliments say to Red Hat, they do an incredible job of hiding that complexity. Yeah. Yes. And making sure that, you know, for example, just like, make it easier for the developers to create value and then, and you know, >>Yeah, that's right. Those, they were previously siloed architectures and >>That's right. OpenShift wanna be place where you wanna run containers or virtual machines. We want that to be this Yeah. Single place. Not, not go bolt on another piece of architecture to just do one or the other. Yeah. >>And hey, the hybrid cloud vision is working for ibm. No question. You know, and it's achievable. Yeah. I mean, I just, I've said unlike, you know, some of the previous, you know, visions on fixing the world with ai, hybrid cloud is actually a real problem that you're attacking and it's showing the results. Agreed. Oh yeah. >>Great. Alright. Last question for you guys. Cause it might be kind of fun, 10 years from now, oh, we're at another, we're sitting here, we all look the same. Time has passed, but we are not aging, which is a part of the new technology that's come out in skincare. That's my, I'm just throwing that out there. Why not? What do you guys hope that you can say about the partnership and, and your continued commitment to community? >>Oh, that's a good question. You go first this time. Yeah. >>I think, you know, the, you know, for looking into the future, you need to look into the past. And Amazon has always been driven by working back from our customers. That's like our key tenant, principle number 1 0 1. >>Couple people have said that on this stage this week. Yeah. >>Yeah. And I think our partnership, I hope over the next decade continues to keep that tenant as a centerpiece. And then whatever comes out of that, I think we, we are gonna be, you know, working through that. >>Yeah. I, I would say this, I think you said that, well, the customer innovation is gonna lead us to wherever that is. And it's, it's, it's gonna be in the cloud for sure. I think we can say that in 10 years. But yeah, anything from, from AI to the quant quantum computing that IBM's really pushing behind that, you know, those are, those are gonna be things that hopefully we show up on a, on a partnership with Manu in 10 years, maybe sooner. >>Well, whatever happens next, we'll certainly be covering it here on the cube. That's right. Thank you both for being here. Joel Manu, fantastic interview. Thanks to see you guys. Yeah, good to see you brought the energy. I think you're definitely ranking high on the top interviews. We >>Love that for >>The day. >>Thank >>My pleasure >>Job, guys. Now that you're competitive at all, and thank you all for tuning in to our live coverage here from AWS Reinvent in Las Vegas, Nevada, with Dave Valante. I'm Savannah Peterson. You're watching The Cube, the leading source for high tech coverage.
SUMMARY :
Manu and Joel, thank you so much for being here. Are we ready? How's the show going for you guys real and, you know, people just want to get out, meet people, have that human touch with each other, And you've got a few in the house. Very few shows can say that, by the way. So far It's the, you have to be here. I was at reinvent number two. And I just, I've talked about it a little bit down the show, but the diversity on the show floor, you know, forever together. I love that you know, benefit from the joint committed spend programs together. 2008, I know you don't like that, but we started So that portends, I mean, 2008, we're talking two years after the launch of s3. harbinger of things to come with these new innovations? Yeah, I, I would say, you know, the innovation is a key tenant of our So it's just getting that go to market tight and we're gonna, we're gonna get that done. And if you know, 95% of the customers say it, Yeah, that's right. how do you guys decide when to put gas on the fire? So we have, you know, close to about couple of thousand power users. So do you take that information and somehow, I mean, it's global, you know, you get, you get a good, you gotta get a good slice of your customer base to get, context of like, you know, Amazon and Red Hat just published the entire roadmap for OpenShift. That is not any process So I I I think it's a, As, as a customer advocate, the better, you know, if, if you guys make bets AWS is the perfect place to place those small bets. I mean, it's not expensive to experiment. Especially in the managed service world. And I know you love taking things to market and you're a go to market guy. We, we are gonna, you know, 10 x the teams that's gonna be focused on these products, Red Hat portfolio is available on marketplace, you can do it on one click deployment. you know, the engineers we're growing this team the most out that team. Curious to see where you're headed. then really making sure that, you know, no customer left behind policy. And it's a great point that customers love when Amazonians and Red Hats show up, I would say yes. That's the plan. I, you know, red Hat has informed the way we have turned around our partnership with ibm, That's right. I mean, you TM boy, I feel like you can spit the time. It is all in on that, that's the native OpenShift offering I think our customers, it's getting super competitive. So to Joel's point, you know, move to worlds containers, and it's kind of right approach for the right job. And making sure that, you know, for example, just like, make it easier for the developers to create value and Yeah, that's right. OpenShift wanna be place where you wanna run containers or virtual machines. I mean, I just, I've said unlike, you know, some of the previous, What do you guys hope that you can say about Yeah. I think, you know, the, you know, Couple people have said that on this stage this week. you know, working through that. you know, those are, those are gonna be things that hopefully we show up on a, on a partnership with Manu Yeah, good to see you brought the energy. Now that you're competitive at all, and thank you all for tuning in to our live coverage here from
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Joel | PERSON | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Dave Valante | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Manu | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Manu Parbhakar | PERSON | 0.99+ |
40% | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Joel Manu | PERSON | 0.99+ |
2021 | DATE | 0.99+ |
Two | QUANTITY | 0.99+ |
January | DATE | 0.99+ |
two companies | QUANTITY | 0.99+ |
two companies | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
Red Hand | ORGANIZATION | 0.99+ |
95% | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
third phase | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
Joel Jackson | PERSON | 0.99+ |
Rosa | ORGANIZATION | 0.99+ |
second chapter | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
2008 | DATE | 0.99+ |
20% | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
30% | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
Outpost | ORGANIZATION | 0.99+ |
red Hat | ORGANIZATION | 0.99+ |
10th year | QUANTITY | 0.99+ |
Jay | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
both | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
first phase | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
ibm | ORGANIZATION | 0.99+ |
aws | ORGANIZATION | 0.99+ |
Joe | PERSON | 0.99+ |
70,000 customers | QUANTITY | 0.99+ |
2023 | DATE | 0.99+ |
one | QUANTITY | 0.99+ |
Las Vegas, Nevada | LOCATION | 0.98+ |
Covid | PERSON | 0.98+ |
first time | QUANTITY | 0.98+ |
nine | QUANTITY | 0.98+ |
12 months | QUANTITY | 0.98+ |
10 | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
this year | DATE | 0.98+ |
about 70,000 customers | QUANTITY | 0.98+ |
Graviton | ORGANIZATION | 0.98+ |
this week | DATE | 0.97+ |
Amazonians | ORGANIZATION | 0.97+ |
third aspect | QUANTITY | 0.97+ |
Spain | LOCATION | 0.97+ |
E two | EVENT | 0.97+ |
each customer | QUANTITY | 0.97+ |
The Cube | TITLE | 0.96+ |
Steve Mullaney, CEO, Aviatrix | AWS re:Invent 2022
(upbeat music) >> You got it, it's theCUBE. We are in Vegas. This is the Cube's live coverage day one of the full event coverage of AWS reInvent '22 from the Venetian Expo Center. Lisa Martin here with Dave Vellante. We love being in Vegas, Dave. >> Well, you know, this is where Super Cloud sort of was born. >> It is. >> Last year, just about a year ago. Steve Mullaney, CEO of of Aviatrix, you know, kind of helped us think it through. And we got some fun stories around. It's happening, but... >> It is happening. We're going to be talking about Super Cloud guys. >> I guess I just did the intro, Steve Mullaney >> You did my intro, don't do it again. >> Sorry I stole that from you, yeah. >> Steve Mullaney, joined just once again, one of our alumni. Steve, great to have you back on the program. >> Thanks for having me back. >> Dave: It's happening. >> It is happening. >> Dave: We talked about a year ago. Net Studio was right there. >> That was two years. Was that year ago, that was a year ago. >> Dave: It was last year. >> Yeah, I leaned over >> What's happening? >> so it's happening. It's happening. You know what, the thing I noticed what's happening now is the maturity of the cloud, right? So, if you think about this whole journey to cloud that has been, what, AWS 12 years. But really over the last few years is when enterprises have really kind of joined that journey. And three or four years ago, and this is why I came out of retirement and went to Aviatrix, was they all said, okay, now we're going to do cloud. You fast forward now three, four years from now, all of a sudden those five-year plans of evacuating the data center, they got one year left, two year left, and they're going, oh crap, we don't have five years anymore. We're, now the maturity's starting to say, we're starting to put more apps into the cloud. We're starting to put business critical apps like SAP into the cloud. This is not just like the low-hanging fruit anymore. So what's happening now is the business criticality, the scale, the maturity. And they're all now starting to hit a lot of limits that have been put into the CSPs that you never used to hit when you didn't have business critical and you didn't have that scale. They were always there. The rocks were always there. Just it was, you never hit 'em. People are starting to hit 'em now. So what's happening now is people are realizing, and I'm going to jump the gun, you asked me for my bumper sticker. The bumper sticker for Aviatrix is, "Good enough is no longer good enough." Now it's funny, it came in a keynote today, but what we see from our customers is it's time to upgrade the native constructs of networking and network security to be enterprise-grade now. It's no longer good enough to just use the native constructs because of a lack of visibility, the lack of controls, the lack of troubleshooting capabilities, all these things. "I now need enterprise grade networking." >> Let me ask you a question 'cause you got a good historical perspective on the industry. When you think about when Maritz was running VMWare. He was like any app, he said basically we're building a software mainframe. And they kind of did that, right? But then they, you know, hit the issue with scale, right? And they can't replicate the cloud. Are there things that we can draw from that experience and apply that to the cloud? What's the same, what's different? >> Oh yeah. So, 1992, do you remember what happened in 1992? I do this, weird German software company called SAP >> Yeah, R3. announced a release as R/3. Which was their first three-tier client-server application of SAP. Before that it ran on mainframes, TCP/IP. Remember that Protocol War? Guess what happened post-1992, everybody goes up like this. Infrastructure completely changes. Cisco, EMC, you name it, builds out these PCE client-server architectures. The WAN changes, MPLS, the campus, everything's home running back to that data center running SAP. That was the last 30 years ago. Great transformation of SAP. They've did it again. It's called S/4Hana. And now it's running and people are switching to S/4Hana and they're moving to the cloud. It's just starting. And that is going to alter how you build infrastructure. And so when you have that, being able to troubleshoot in hours versus minutes is a big deal. This is business critical, millions of dollars. This is not fun and games. So again, back to my, what was good enough for the last three or four years for enterprises no longer good enough, now I'm running business critical apps like SAP, and it's going to completely change infrastructure. That's happening in the cloud right now. And that's obviously a significant seismic shift, but what are some of the barriers that customers have been able to eliminate in order to get there? Or is it just good enough isn't good enough anymore? >> Barriers in terms of, well, I mean >> Lisa: The adoption. Yeah well, I mean, I think it's all the things that they go to cloud is, you know, the complexity, really, it's the agility, right? So the barrier that they have to get over is how do I keep the developer happy because the developer went to the cloud in the first place, why? Swipe the credit card because IT wasn't doing their job, 'cause every time I asked them for something, they said no. So I went around 'em. We need that. That's what they have to overcome in the move to the cloud. That is the obstacle is how do I deliver that visibility, that control, the enterprise, great functionality, but yet give the developer what they want. Because the minute I stop giving them that swipe the card operational model, what do you think they're going to do? They're going to go around me again and I can't, and the enterprise can't have that. >> That's a cultural shift. >> That's the main barrier they've got to overcome. >> Let me ask you another question. Is what we think of as mission critical, the definition changing? I mean, you mentioned SAP, obviously that's mission critical for operations, but you're also seeing new applications being developed in the cloud. >> I would say anything that's, I call business critical, same thing, but it's, business critical is internal to me, like SAP, but also anything customer-facing. That's business critical to me. If that app goes down or it has a problem, I'm not collecting revenue. So, you know, back 30 years ago, we didn't have a lot of customer-facing apps, right? It really was just SAP. I mean there wasn't a heck of a lot of cust- There were customer-facing things. But you didn't have all the digitalization that we have now, like the digital economy, where that's where the real explosion has come, is you think about all the customer-facing applications. And now every enterprise is what? A technology, digital company with a customer-facing and you're trying to get closer and closer to who? The consumer. >> Yeah, self-service. >> Self-service, B2C, everybody wants to do that. Get out of the middle man. And those are business critical applications for people. >> So what's needed under the covers to make all this happen? Give us a little double click on where you guys fit. >> You need consistent architecture. Obviously not just for one cloud, but for any cloud. But even within one cloud, forget multicloud, it gets worst with multicloud. You need a consistent architecture, right? That is automated, that is as code. I can't have the human involved. These are all, this is the API generation, you've got to be able to use automation, Terraform. And all the way from the application development platform you know, through Jenkins and all other software, through CICD pipeline and Terraform, when you, when that developer says, I want infrastructure, it has to go build that infrastructure in real time. And then when it says, I don't need it anymore it's got to take it away. And you cannot have a human involved in that process. That's what's completely changed. And that's what's giving the agility. And that's kind of a cloud model, right? Use software. >> Well, okay, so isn't that what serverless does, right? >> That's part of it. Absolutely. >> But I might still want control sometimes over the runtime if I'm running those mission critical applications. Everything in enterprise is a heterogeneous thing. It's like people, people say, well there's going to, the people going to repatriate back to on-prem, they are not repatriating back to on-prem. >> We were just talking about that, I'm like- >> Steve: It's not going to happen, right? >> It's a myth, it's a myth. >> And there's things that maybe shouldn't have ever gone into the cloud, I get that. Look, do people still have mainframes? Of course. There's certain things that you just, doesn't make sense to move to the new generation. There were things, certain applications that are very static, they weren't dynamic. You know what, keeping it on-prem it's, probably makes sense. So some of those things maybe will go back, but they never should have gone. But we are not repatriating ever, you know, that's not going to happen. >> No I agree. I mean, you know, there was an interesting paper by Andreessen, >> Yeah. >> But, I mean- >> Steve: Yeah it was a little self-serving for some company that need more funding, yeah. You look at the numbers. >> Steve: Yeah. >> It tells the story. It's just not happening. >> No. And the reason is, it's that agility, right? And so that's what people, I would say that what you need to do is, and in order to get that agility, you have to have that consistency. You have to have automation, you have to get these people out of the way. You have to use software, right? So it's that you have that swipe the card operational model for the developers. They don't want to hear the word no. >> Lisa: Right. >> What do you think is going to happen with AWS? Because we heard, I don't know if you heard Selipsky's keynote this morning, but you've probably heard the hallway talk. >> Steve: I did, yeah. >> Okay. You did. So, you know, connecting the dots, you know doubling down on all the primitives, that we expected. We kind of expected more of the higher level stuff, which really didn't see much of that, a little bit. >> Steve: Yeah. So, you know, there's a whole thing about, okay, does the cloud get commoditized? Does it not? I think the secret weapon's the ecosystem, right? Because they're able to sell through with guys like you. Make great margins on that. >> Steve: Yeah, well, yeah. >> What are your thoughts though on the future of AWS? >> IAS is going to get commoditized. So this is the fallacy that a lot of the CSPs have, is they thought that they were going to commoditize enterprise. It never happens that way. What's going to happen is infrastructure as a service, the lower level, which is why you see all the CSPs talking about what? Oracle Cloud, industry cloud. >> Well, sure, absolutely, yeah. >> We got to get to the apps, we got to get to SAP, we got to get to all that, because that's not going to get commoditized, right. But all the infrastructural service where AWS is king that is going to get commoditized, absolutely. >> Okay, so, but historically, you know Cisco's still got 60% plus gross margins. EMC always had good margin. How pure is the lone survivor in Flash? They got 70% gross margins. So infrastructure actually has always been a pretty good business. >> Yeah that's true. But it's a hell of a lot easier, particularly with people like Aviatrix and others that are building these common architectural things that create simplicity and abstract the way the complexities of underneath such that we allow your network to run an AWS, Azure, Google, Oracle, whatever, exactly the same. So it makes it a hell of a lot easier >> Dave: Super cloud. >> to go move. >> But I want to tap your brain because you have a good perspective of this because servers used to be a great margin business too on-prem and now it's not. It's a low margin business 'cause all the margin went to Intel. >> Yeah. But the cloud guys, you know, AWS in particular, makes a ton of dough on servers, so, or compute. So it's going to be interesting to see over time if that gets com- that's why they're going so hard after silicon. >> I think if they can, I think if you can capture the workload. So AWS and everyone else, as another example, this SAP, they call that a gravity workload. You know what gravity workload is? It's a black hole. It drags everything else with it. If you get SAP or Oracle or a mainframe app, it ain't going anywhere. And then what's going to happen is all your other apps are going to follow it. So that's what they're all going to fight for, is type of app. >> You said something earlier about, forget multicloud, for a moment, but, that idea of the super cloud, this abstraction layer, I mean, is that a real business value for customers other than, oh I got all these clouds, I need 'em to work together. You know, from your perspective from Aviatrix perspective, is it an opportunity for you to build on top of that? Or are you just looking at, look, I'm going to do really good work in AWS, in Azure? Now we're making the same experience. >> I hear this every single day from our customers is they look and they say, good enough isn't good enough. I've now hit the point, I'm hitting route limitations. I'm hitting, I'm doing things manually, and that's fine when I don't have that many applications or I don't have mission critical. The dogs are eating the dog food, we're going into the cloud and they're looking and then saying this is not an operational model for me. I've hit the point where I can't keep doing this, I can't throw bodies at this, I need software. And that's the opportunity for us, is they look and they say, I'm doing it in one cloud, but, and there's zero chance I'm going to be able to figure that out in the two or three other clouds. Every enterprise I talk to says multicloud is inevitable. Whether they're in it now, they all know they're going to go, because it's the business units that demand it. It's not the IT teams that demand it, it's the line of business that says, I like GCP for this reason. >> The driver's functionality that they're getting. >> It's the app teams that say, I have this service and GCP's better at it than AWS. >> Yeah, so it's not so much a cost game or the end all coffee mug, right? >> No, no. >> Google does this better than Microsoft, or better than- >> If you asked an IT person, they would rather not have multicloud. They actually tried to fight it. No, why would you want to support four clouds when you could support one right? That's insane. >> Dave and Lisa: Right. If they didn't have a choice and, and so it, the decision was made without them, and actually they weren't even notified until day before. They said, oh, good news, we're going to GCP tomorrow. Well, why wasn't I notified? Well, we're notifying you now. >> Yeah, you would've said, no. >> Steve: This is cloud bottle, let's go. >> Super cloud again. Did you see the Berkeley paper, sky computing I think they call it? Down at Berkeley, yep Dave Linthicum from Deloitte. He's talking about, I think he calls it meta cloud. It's happening. >> Yeah, yeah, yeah. >> It's happening. >> No, and because customers, customers want that. They... >> And talk about some customer example or two that you think really articulates the value of why it's happening and the outcomes that it's generating. >> I mean, I was just talking to Lamb Weston last night. So we had a reception, Lamb Weston, huge, frozen potatoes. They serve like, I dunno, some ungodly percentage of all the french fries to all the fast food. It's unbelievable what they do. Do you know, they have special chemicals they put on the french fries. So when you get your DoorDash, they stay crispy longer. They've invented that patented it. But anyway, it's all these businesses you've never heard of and they do all the, and again, they're moving to SAP or they're actually SAP in the cloud, they're one of the first ones. They did it through Accenture. They're pulling it back off from Accenture. They're not happy with the service they're getting. They're going to use us for their networking and network security because they're going to get that visibility and control back. And they're going to repatriate it back from a managed service and bring it back and run it in-house. And the SAP basis engineers want it to happen because they see the visibility and control that the infrastructure guy's going to get because of us, which leads to, all they care about is uptime and performance. That's it. And they're going to say the infrastructure team's going to lead to better uptime and better performance if it's running on Aviatrix. >> And business performance and uptime, business critical >> That is the business. That is the business. >> It is. So what are some of the things next coming down the pike from Aviatrix? Any secret sauce you can share? >> Lot of secrets. So, two secrets. One, the next thing people really want to do, embedded network security into the network. We've kind of talked about this. You're going to be seeing some things from us. Where does network security belong? In the network. Embedded in the fabric of the network, not as this dumb device called the next-gen firewall that you steer traffic to. It has to be into the fabric of what we do, what we call airspace. You're going to see us talk about that. And then the next thing, back to the maturity of the cloud, as they build out the core, guess what they're doing? It's this thing called edge, Dave, right? And guess what they're going to do? It's not about connecting the cloud to the edge to the cloud with dumb things like SD-WAN, right? Or SaaS. It's actually the other way around. Go into the cloud, turn around, look out at the edge and say, how do I extend the cloud out to the edge, and make it look like a VPC. That's what people are doing. Why, 'cause I want the operational model. I want all the things that I can do in the cloud out at the edge. And everyone knows it's been in networking. I've been in networking for 37 years. He who wins the core does what? Wins the edge, 'cause that's what happens. You do it first in the core and then you want one architecture, one common architecture, one consistent way of doing everything. And that's going to go out to the edge and it's going to look like a VPC from an operational model. >> And Amazon's going to support that, no doubt. >> Yeah, I mean every, you know, every, and then it's just how do you want to go do that? And us as the networking and network security provider, we're getting dragged to the edge by our customer. Because you're my networking provider. And that means, end to end. And they're trying to drag us into on-prem too, yeah. >> Lot's going on, you're going to have to come back- >> Because they want one networking vendor. >> But wait, and you say what? >> We will never do like switches and any of the keep Arista, the Cisco, and all that kind of stuff. But we will start sucking in net flow. We will start doing, from an operational perspective, we will integrate a lot of the things that are happening in on-prem into our- >> No halfway house. >> Copilot. >> No halfway house, no two architectures. But you'll take the data in. >> You want one architecture. >> Yeah. >> Yeah, totally. >> Right play. >> Amazing stuff. >> And he who wins the core, guess what's more strategic to them? What's more strategic on-prem or cloud? Cloud. >> It flipped three years ago. >> Dave: Yeah. >> So he who wins in the clouds going to win everywhere. >> Got it, We'll keep our eyes on that. >> Steve: Cause and effect. >> Thank you so much for joining us. We've got your bumper sticker already. It's been a great pleasure having you on the program. You got to come back, there's so, we've- >> You posting the bumper sticker somewhere? >> Lisa: It's going to be our Instagram. >> Oh really, okay. >> And an Instagram sto- This is new for you guys. Always coming up with new ideas. >> Raising the bar. >> It is, it is. >> Me advance, I mean, come on. >> I love it. >> All right, for our guest Steve Mullaney and Dave Vellante, I'm Lisa Martin. You're watching theCUBE, the leader in live enterprise and emerging tech coverage.
SUMMARY :
This is the Cube's live coverage day one Well, you know, this is where you know, kind of helped We're going to be talking don't do it again. I stole that from you, yeah. Steve, great to have you Dave: We talked about Was that year ago, that was a year ago. We're, now the maturity's starting to say, and apply that to the cloud? 1992, do you remember And that is going to alter in the move to the cloud. That's the main barrier being developed in the cloud. like the digital economy, Get out of the middle man. covers to make all this happen? And all the way from the That's part of it. the people going to into the cloud, I get that. I mean, you know, there You look at the numbers. It tells the story. and in order to get that agility, going to happen with AWS? of the higher level stuff, does the cloud get commoditized? a lot of the CSPs have, that is going to get How pure is the lone survivor in Flash? and abstract the way 'cause all the margin went to Intel. But the cloud guys, you capture the workload. of the super cloud, this And that's the opportunity that they're getting. It's the app teams that say, to support four clouds the decision was made without them, Did you see the Berkeley paper, No, and that you think really that the infrastructure guy's That is the business. coming down the pike from Aviatrix? It's not about connecting the cloud to And Amazon's going to And that means, end to end. Because they want and any of the keep Arista, the Cisco, But you'll take the data in. And he who wins the core, clouds going to win everywhere. You got to come back, there's so, we've- This is new for you guys. the leader in live enterprise
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Steve | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Steve Mullaney | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Vegas | LOCATION | 0.99+ |
Dave | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
60% | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Dave Linthicum | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
37 years | QUANTITY | 0.99+ |
Aviatrix | ORGANIZATION | 0.99+ |
EMC | ORGANIZATION | 0.99+ |
five-year | QUANTITY | 0.99+ |
1992 | DATE | 0.99+ |
Deloitte | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
five years | QUANTITY | 0.99+ |
Last year | DATE | 0.99+ |
Accenture | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
a year ago | DATE | 0.99+ |
Andreessen | PERSON | 0.99+ |
Arista | ORGANIZATION | 0.99+ |
Lamb Weston | ORGANIZATION | 0.99+ |
three | DATE | 0.99+ |
tomorrow | DATE | 0.99+ |
Net Studio | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
One | QUANTITY | 0.98+ |
last year | DATE | 0.98+ |
SAP | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
Lamb Weston | ORGANIZATION | 0.98+ |
last night | DATE | 0.98+ |
Berkeley | ORGANIZATION | 0.98+ |
one cloud | QUANTITY | 0.98+ |
one cloud | QUANTITY | 0.98+ |
two secrets | QUANTITY | 0.97+ |
two architectures | QUANTITY | 0.97+ |
last 30 years ago | DATE | 0.97+ |
12 years | QUANTITY | 0.97+ |
zero chance | QUANTITY | 0.97+ |
three years ago | DATE | 0.97+ |
Venetian Expo Center | LOCATION | 0.97+ |
Subbu Iyer
>> And it'll be the fastest 15 minutes of your day from there. >> In three- >> We go Lisa. >> Wait. >> Yes >> Wait, wait, wait. I'm sorry I didn't pin the right speed. >> Yap, no, no rush. >> There we go. >> The beauty of not being live. >> I think, in the background. >> Fantastic, you all ready to go there, Lisa? >> Yeah. >> We are speeding around the horn and we are coming to you in five, four, three, two. >> Hey everyone, welcome to theCUBE's coverage of AWS re:Invent 2022. Lisa Martin here with you with Subbu Iyer one of our alumni who's now the CEO of Aerospike. Subbu, great to have you on the program. Thank you for joining us. >> Great as always to be on theCUBE Lisa, good to meet you. >> So, you know, every company these days has got to be a data company, whether it's a retailer, a manufacturer, a grocer, a automotive company. But for a lot of companies, data is underutilized yet a huge asset that is value added. Why do you think companies are struggling so much to make data a value added asset? >> Well, you know, we see this across the board. When I talk to customers and prospects there is a desire from the business and from IT actually to leverage data to really fuel newer applications, newer services newer business lines if you will, for companies. I think the struggle is one, I think one the, the plethora of data that is created. Surveys say that over the next three years data is going to be you know by 2025 around 175 zettabytes, right? A hundred and zettabytes of data is going to be created. And that's really a growth of north of 30% year over year. But the more important and the interesting thing is the real time component of that data is actually growing at, you know 35% CAGR. And what enterprises desire is decisions that are made in real time or near real time. And a lot of the challenges that do exist today is that either the infrastructure that enterprises have in place was never built to actually manipulate data in real time. The second is really the ability to actually put something in place which can handle spikes yet be cost efficient to fuel. So you can build for really peak loads, but then it's very expensive to operate that particular service at normal loads. So how do you build something which actually works for you for both users, so to speak. And the last point that we see out there is even if you're able to, you know bring all that data you don't have the processing capability to run through that data. So as a result, most enterprises struggle with one capturing the data, making decisions from it in real time and really operating it at the cost point that they need to operate it at. >> You know, you bring up a great point with respect to real time data access. And I think one of the things that we've learned the last couple of years is that access to real time data it's not a nice to have anymore. It's business critical for organizations in any industry. Talk about that as one of the challenges that organizations are facing. >> Yeah, when we started Aerospike, right? When the company started, it started with the premise that data is going to grow, number one exponentially. Two, when applications open up to the internet there's going to be a flood of users and demands on those applications. And that was true primarily when we started the company in the ad tech vertical. So ad tech was the first vertical where there was a lot of data both on the supply set and the demand side from an inventory of ads that were available. And on the other hand, they had like microseconds or milliseconds in which they could make a decision on which ad to put in front of you and I so that we would click or engage with that particular ad. But over the last three to five years what we've seen is as digitization has actually permeated every industry out there the need to harness data in real time is pretty much present in every industry. Whether that's retail, whether that's financial services telecommunications, e-commerce, gaming and entertainment. Every industry has a desire. One, the innovative companies, the small companies rather are innovating at a pace and standing up new businesses to compete with the larger companies in each of these verticals. And the larger companies don't want to be left behind. So they're standing up their own competing services or getting into new lines of business that really harness and are driven by real time data. So this compelling pressures, one, you know customer experience is paramount and we as customers expect answers in you know an instant, in real time. And on the other hand, the way they make decisions is based on a large data set because you know larger data sets actually propel better decisions. So there's competing pressures here which essentially drive the need one from a business perspective, two from a customer perspective to harness all of this data in real time. So that's what's driving an incessant need to actually make decisions in real or near real time. >> You know, I think one of the things that's been in short supply over the last couple of years is patience. We do expect as consumers whether we're in our business lives our personal lives that we're going to be getting be given information and data that's relevant it's personal to help us make those real time decisions. So having access to real time data is really business critical for organizations across any industries. Talk about some of the main capabilities that modern data applications and data platforms need to have. What are some of the key capabilities of a modern data platform that need to be delivered to meet demanding customer expectations? >> So, you know, going back to your initial question Lisa around why is data really a high value but underutilized or under-leveraged asset? One of the reasons we see is a lot of the data platforms that, you know, some of these applications were built on have been then around for a decade plus. And they were never built for the needs of today, which is really driving a lot of data and driving insight in real time from a lot of data. So there are four major capabilities that we see that are essential ingredients of any modern data platform. One is really the ability to, you know, operate at unlimited scale. So what we mean by that is really the ability to scale from gigabytes to even petabytes without any degradation in performance or latency or throughput. The second is really, you know, predictable performance. So can you actually deliver predictable performance as your data size grows or your throughput grows or your concurrent user on that application of service grows? It's really easy to build an application that operates at low scale or low throughput or low concurrency but performance usually starts degrading as you start scaling one of these attributes. The third thing is the ability to operate and always on globally resilient application. And that requires a really robust data platform that can be up on a five nine basis globally, can support global distribution because a lot of these applications have global users. And the last point is, goes back to my first answer which is, can you operate all of this at a cost point which is not prohibitive but it makes sense from a TCO perspective. 'Cause a lot of times what we see is people make choices of data platforms and as ironically their service or applications become more successful and more users join their journey the revenue starts going up, the user base starts going up but the cost basis starts crossing over the revenue and they're losing money on the service, ironically as the service becomes more popular. So really unlimited scale predictable performance always on a globally resilient basis and low TCO. These are the four essential capabilities of any modern data platform. >> So then talk to me with those as the four main core functionalities of a modern data platform, how does Aerospike deliver that? >> So we were built, as I said from day one to operate at unlimited scale and deliver predictable performance. And then over the years as we work with customers we build this incredible high availability capability which helps us deliver the always on, you know, operations. So we have customers who are who have been on the platform 10 years with no downtime for example, right? So we are talking about an amazing continuum of high availability that we provide for customers who operate these, you know globally resilient services. The key to our innovation here is what we call the hybrid memory architecture. So, you know, going a little bit technically deep here essentially what we built out in our architecture is the ability on each node or each server to treat a bank of SSDs or solid-state devices as essentially extended memory. So you're getting memory performance but you're accessing these SSDs. You're not paying memory prices but you're getting memory performance. As a result of that you can attach a lot more data to each node or each server in a distributed cluster. And when you kind of scale that across basically a distributed cluster you can do with Aerospike the same things at 60 to 80% lower server count. And as a result 60 to 80% lower TCO compared to some of the other options that are available in the market. Then basically, as I said that's the key kind of starting point to the innovation. We lay around capabilities like, you know replication, change data notification, you know synchronous and asynchronous replication. The ability to actually stretch a single cluster across multiple regions. So for example, if you're operating a global service you can have a single Aerospike cluster with one node in San Francisco one node in New York, another one in London and this would be basically seamlessly operating. So that, you know, this is strongly consistent, very few no SQL data platforms are strongly consistent or if they are strongly consistent they will actually suffer performance degradation. And what strongly consistent means is, you know all your data is always available it's guaranteed to be available there is no data lost any time. So in this configuration that I talked about if the node in London goes down your application still continues to operate, right? Your users see no kind of downtime and you know, when London comes up it rejoins the cluster and everything is back to kind of the way it was before, you know London left the cluster so to speak. So the ability to do this globally resilient highly available kind of model is really, really powerful. A lot of our customers actually use that kind of a scenario and we offer other deployment scenarios from a higher availability perspective. So everything starts with HMA or Hybrid Memory Architecture and then we start building a lot of these other capabilities around the platform. And then over the years what our customers have guided us to do is as they're putting together a modern kind of data infrastructure, we don't live in the silo. So Aerospike gets deployed with other technologies like streaming technologies or analytics technologies. So we built connectors into Kafka, Pulsar, so that as you're ingesting data from a variety of data sources you can ingest them at very high ingest speeds and store them persistently into Aerospike. Once the data is in Aerospike you can actually run Spark jobs across that data in a multi-threaded parallel fashion to get really insight from that data at really high throughput and high speed. >> High throughput, high speed, incredibly important especially as today's landscape is increasingly distributed. Data centers, multiple public clouds, Edge, IoT devices, the workforce embracing more and more hybrid these days. How are you helping customers to extract more value from data while also lowering costs? Go into some customer examples 'cause I know you have some great ones. >> Yeah, you know, I think, we have built an amazing set of customers and customers actually use us for some really mission critical applications. So, you know, before I get into specific customer examples let me talk to you about some of kind of the use cases which we see out there. We see a lot of Aerospike being used in fraud detection. We see us being used in recommendations engines we get used in customer data profiles, or customer profiles, Customer 360 stores, you know multiplayer gaming and entertainment. These are kind of the repeated use case, digital payments. We power most of the digital payment systems across the globe. Specific example from a specific example perspective the first one I would love to talk about is PayPal. So if you use PayPal today, then you know when you're actually paying somebody your transaction is, you know being sent through Aerospike to really decide whether this is a fraudulent transaction or not. And when you do that, you know, you and I as a customer are not going to wait around for 10 seconds for PayPal to say yay or nay. We expect, you know, the decision to be made in an instant. So we are powering that fraud detection engine at PayPal. For every transaction that goes through PayPal. Before us, you know, PayPal was missing out on about 2% of their SLAs which was essentially millions of dollars which they were losing because, you know, they were letting transactions go through and taking the risk that it's not a fraudulent transaction. With Aerospike they can now actually get a much better SLA and the data set on which they compute the fraud score has gone up by you know, several factors. So by 30X if you will. So not only has the data size that is powering the fraud engine actually gone up 30X with Aerospike but they're actually making decisions in an instant for, you know, 99.95% of their transactions. So that's- >> And that's what we expect as consumers, right? We want to know that there's fraud detection on the swipe regardless of who we're interacting with. >> Yes, and so that's a really powerful use case and you know, it's a great customer success story. The other one I would talk about is really Wayfair, right, from retail and you know from e-commerce. So everybody knows Wayfair global leader in really in online home furnishings and they use us to power their recommendations engine. And you know it's basically if you're purchasing this, people who bought this also bought these five other things, so on and so forth. They have actually seen their cart size at checkout go up by up to 30%, as a result of actually powering their recommendations engine through Aerospike. And they were able to do this by reducing the server count by 9X. So on one ninth of the servers that were there before Aerospike, they're now powering their recommendations engine and seeing cart size checkout go up by 30%. Really, really powerful in terms of the business outcome and what we are able to, you know, drive at Wayfair. >> Hugely powerful as a business outcome. And that's also what the consumer wants. The consumer is expecting these days to have a very personalized relevant experience that's going to show me if I bought this show me something else that's related to that. We have this expectation that needs to be really fueled by technology. >> Exactly, and you know, another great example you asked about you know, customer stories, Adobe. Who doesn't know Adobe, you know. They're on a mission to deliver the best customer experience that they can. And they're talking about, you know great Customer 360 experience at scale and they're modernizing their entire edge compute infrastructure to support this with Aerospike. Going to Aerospike basically what they have seen is their throughput go up by 70%, their cost has been reduced by 3X. So essentially doing it at one third of the cost while their annual data growth continues at, you know about north of 30%. So not only is their data growing they're able to actually reduce their cost to actually deliver this great customer experience by one third to one third and continue to deliver great Customer 360 experience at scale. Really, really powerful example of how you deliver Customer 360 in a world which is dynamic and you know on a data set which is constantly growing at north of 30% in this case. >> Those are three great examples, PayPal, Wayfair, Adobe, talking about, especially with Wayfair when you talk about increasing their cart checkout sizes but also with Adobe increasing throughput by over 70%. I'm looking at my notes here. While data is growing at 32%, that's something that every organization has to contend with data growth is continuing to scale and scale and scale. >> Yap, I'll give you a fun one here. So, you know, you may not have heard about this company it's called Dream11 and it's a company based out of India but it's a very, you know, it's a fun story because it's the world's largest fantasy sports platform. And you know, India is a nation which is cricket crazy. So you know, when they have their premier league going on and there's millions of users logged onto the Dream11 platform building their fantasy league teams and you know, playing on that particular platform, it has a hundred million users a hundred million plus users on the platform, 5.5 million concurrent users and they have been growing at 30%. So they are considered an amazing success story in terms of what they have accomplished and the way they have architected their platform to operate at scale. And all of that is really powered by Aerospike. Think about that they're able to deliver all of this and support a hundred million users 5.5 million concurrent users all with, you know 99 plus percent of their transactions completing in less than one millisecond. Just incredible success story. Not a brand that is, you know, world renowned but at least you know from what we see out there it's an amazing success story of operating at scale. >> Amazing success story, huge business outcomes. Last question for you as we're almost out of time is talk a little bit about Aerospike AWS the partnership Graviton2 better together. What are you guys doing together there? >> Great partnership. AWS has multiple layers in terms of partnerships. So, you know, we engage with AWS at the executive level. They plan out, really roll out of new instances in partnership with us, making sure that, you know those instance types work well for us. And then we just released support for Aerospike on the Graviton platform and we just announced a benchmark of Aerospike running on Graviton on AWS. And what we see out there is with the benchmark a 1.6X improvement in price performance. And you know about 18% increase in throughput while maintaining a 27% reduction in cost, you know, on Graviton. So this is an amazing story from a price performance perspective, performance per watt for greater energy efficiencies, which basically a lot of our customers are starting to kind of talk to us about leveraging this to further meet their sustainability target. So great story from Aerospike and AWS not just from a partnership perspective on a technology and an executive level, but also in terms of what joint outcomes we are able to deliver for our customers. >> And it sounds like a great sustainability story. I wish we had more time so we would talk about this but thank you so much for talking about the main capabilities of a modern data platform, what's needed, why, and how you guys are delivering that. We appreciate your insights and appreciate your time. >> Thank you very much. I mean, if folks are at re:Invent next week or this week come on and see us at our booth and we are in the data analytics pavilion and you can find us pretty easily. Would love to talk to you. >> Perfect, we'll send them there. Subbu Iyer, thank you so much for joining me on the program today. We appreciate your insights. >> Thank you Lisa. >> I'm Lisa Martin, you're watching theCUBE's coverage of AWS re:Invent 2022. Thanks for watching. >> Clear- >> Clear cutting. >> Nice job, very nice job.
SUMMARY :
the fastest 15 minutes I'm sorry I didn't pin the right speed. and we are coming to you in Subbu, great to have you on the program. Great as always to be on So, you know, every company these days And a lot of the challenges that access to real time data to put in front of you and I and data platforms need to have. One of the reasons we see is So the ability to do How are you helping customers let me talk to you about fraud detection on the swipe and you know, it's a great We have this expectation that needs to be Exactly, and you know, with Wayfair when you talk So you know, when they have What are you guys doing together there? And you know about 18% and how you guys are delivering that. and you can find us pretty easily. for joining me on the program today. of AWS re:Invent 2022.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
60 | QUANTITY | 0.99+ |
London | LOCATION | 0.99+ |
Lisa | PERSON | 0.99+ |
PayPal | ORGANIZATION | 0.99+ |
New York | LOCATION | 0.99+ |
15 minutes | QUANTITY | 0.99+ |
3X | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
Wayfair | ORGANIZATION | 0.99+ |
35% | QUANTITY | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
30% | QUANTITY | 0.99+ |
99.95% | QUANTITY | 0.99+ |
10 seconds | QUANTITY | 0.99+ |
San Francisco | LOCATION | 0.99+ |
30X | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
32% | QUANTITY | 0.99+ |
27% | QUANTITY | 0.99+ |
1.6X | QUANTITY | 0.99+ |
each server | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
Aerospike | ORGANIZATION | 0.99+ |
millions of dollars | QUANTITY | 0.99+ |
India | LOCATION | 0.99+ |
Subbu | PERSON | 0.99+ |
9X | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
99 plus percent | QUANTITY | 0.99+ |
first answer | QUANTITY | 0.99+ |
third thing | QUANTITY | 0.99+ |
less than one millisecond | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
this week | DATE | 0.99+ |
Subbu Iyer | PERSON | 0.99+ |
one third | QUANTITY | 0.99+ |
millions of users | QUANTITY | 0.99+ |
over 70% | QUANTITY | 0.98+ |
both users | QUANTITY | 0.98+ |
Dream11 | ORGANIZATION | 0.98+ |
80% | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
Graviton | TITLE | 0.98+ |
each node | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
three | QUANTITY | 0.98+ |
four | QUANTITY | 0.98+ |
Two | QUANTITY | 0.98+ |
one node | QUANTITY | 0.98+ |
hundred million users | QUANTITY | 0.98+ |
first vertical | QUANTITY | 0.97+ |
about 2% | QUANTITY | 0.97+ |
Aerospike | TITLE | 0.97+ |
single cluster | QUANTITY | 0.96+ |
Thijs Ebbers & Arno Vonk, ING | KubeCon + CloudNativeCon NA 2022
>>Good morning, brilliant humans. Good afternoon or good evening, depending on your time zone. My name is Savannah Peterson and I'm here live with the Cube. We are at CubeCon in Detroit, Michigan. And joining me is my beautiful co-host, Lisa, how you feeling? Afternoon of day three. >>Afternoon day three. We've had such great conversations. We have's been fantastic. The momentum has just been going like this. I love it. >>Yes. You know, sometimes we feel a little low when we're at the end of a conference. Not today. Don't feel that that way at all, which is very exciting. Just like the guests that we have up for you next. Kind of an unexpected player when we think about technology. However, since every company, one of the themes is every company is trying to be a software company. I love that we're talking to I n G. Joining us today is Ty Evers and Arno vk. Welcome to the show gentlemen. Thank >>You very much. Glad to be you. Thank you. >>Yes, it's wonderful. All the way in from Amsterdam. Probably some of the farthest flying folks here for this adventure. Starting off. I forgot what's going on with the shirts guys. You match very well. Tell, tell everyone. >>Well these are our VR code shirts. VR code is basically the player of our company to get people interested as an IT person in banking. Right? Actually, people don't think banking is a good place to work as an IT professional, but actually this, and we are using the OC went with these nice logos to get it attention. >>I love that. So let's actually, let's just talk about that for a second. Why is it such an exciting role to be working in technology at a company like I N G or traditional bank? >>I N G is a challenging environment. That's how do you make an engineer happy, basically give them a problem to solve. So we have lots and lots of problems to solve. So that makes it challenging. But yeah, also rewarding. And you can say a lot of things about banks and with looking at the IT perspective, we are doing amazing things in I and that's what we talked about. Can >>You, can you tell us any of those amazing things or are they secrets? >>Think we talked about last Tuesday at S shift commons conference. Yeah, so we had two, two presentations I presented with my coho sand on my journey over the last three years. So what has IG done? Basically building a secure container hosting platform. Yeah. How do we live a banking cot with cloud native technology and together with our coho young villa presented actually showed it by demo making life and >>Awesome >>In person. So we were not just presenting, >>It's not all smoke and mirrors. It's >>Not smoke and mirror, which we're not presenting our fufu marketing block now. We actually doing it today. And that's what we wanted to share here. >>Well, and as consumers we expect we can access our banking on any device 24 by seven. I wanna be able to do all my transactions in a way that I know is secure. Obviously security's a huge thing there, but talk about I n G Bank aren't always been around for a very long time. Talk about this financial institution as a software company. Really obviously a lot of challenges to solve, a lot of opportunity. But talk about what it's like working for a history and bank that's really now a tech company. >>Yes. It's been really changing as a bank to a tech company. Yeah. We have a lot of developers and operators and we do deliver offer. We OnPrem, we run in the public. So we have a huge engineers and people around to make our software. Yes. And I am responsible for the i Container Ocean platform and we deliver that the name space as a surface and as a real, real secure environment. So our developers, all our developers in, I can request it, but they only get a name space. Yeah, that's very important there. They >>Have >>Resources and all sort of things. Yeah. And it is, they cannot access it. They can only access it by one wifi. So, >>So Lisa and I were chatting before we brought you up here. Name space as a service. This is a newer term for us. Educate us. What does that mean? >>Basically it means we don't give a full cluster to our consumers, right? We only give them basically cpu, memory networking. That's all they need to host application. Everything else we abstract away. And especially in a banking context where compliance is a big thing, you don't need to do compliance for an entire s clusterized developer. It's really saves development time for the colleagues in the bank. It >>Decreases the complexity of projects, which is a huge theme here, especially at scale. I can imagine. I mean, my gosh, you're serving so many different people, it probably saves you time. Let's talk about regulation. What, how challenging is that for you as technologists to balance in all the regulations around banking and FinTech? It's, it's, it's, it's not like some of these kind of wild, wild west industries where we can just go out and play and prototype and do whatever we want. There's a lot of >>Rules. There's a lot of rules. And the problem is you have legislation and you have the real world. Right. And you have to find something in, they're >>Not the same thing. >>You have to find something in between with both parties on the stands and cannot adhere to. Yeah. So the challenge we had, basically we had to wide our, in our own container security standards to prove that the things we were doing were the white things to be in control as a bank because there was no market standard for container security. So basically we took some input from this. So n did a lot of good work. We basically added some things on top to be valid for a bank in Europe. So yeah, that's what we did. And the nice thing is today we take all the boxes we defined back in 2019. >>Hey, so you what it's, I guess, I guess the rules are a little bit easier when you get to help define them. Yep. Yeah. That it feels like a very good strategic call >>And they makes sense. Yeah. Right. Because the hardest problem is try to be compliant for something which doesn't make sense. Right, >>Right. Arnold, talk about, let's double click on namespace as a service. You talked about what that is, but give us a little bit of information on why I N G really believes this is the right approach for this company. >>It's protects for the security that developers doing things they don't shoot. Yeah. They cannot access their store anymore when it is running in production. And that is the most, most important. That is, it is immutable running in our platform. >>Excellent. Talk about both of you. How long have you, have you both been at I n G for a long time? >>I've been with I N G since September, 2001. So that's more than 20 years >>Now. Long time. Ana, what about you? >>Before 2000 already before. >>So both of your comment on that's a long time. Yeah. Talk about the culture of innovation that's at I N G to be able to move at such speed and be groundbreaking in what you're, how you're using technology, what, what's the appetite like at the bank to embrace new and emerging technologies? >>So we are really looking, basically the, the mantra of the bank is to help our customers get a step ahead in life and in business. And we do that by one superior customer service and secondly, sustainability at the heart. So anything which contributes to those targets, you can go to your manager and if you can make goods case why it contributes most of the cases you get some time or some budgets or even some additional colleagues to help you out and give it a try require from a culture perspective required open to trying things out before we reach production. Once you go to production. Yeah. Then we are back to being a bank and you need to take all the boxes to make really sure that we are confident with our customers data and basically we're still a bank but a lot of is possible. >>A lot. It is possible. And there's the customer on the other end who's expecting, like I said earlier, that they can access their data any time that they want, be able to do any transaction they want, making sure the content that's delivered to them is relevant, that it's secure. Obviously with, that's the biggest challenge especially is we think about how many generations are alive today and and those that aren't tech savvy. Yeah. Have challenges with that. Talk about what the bank's dedication is to ensuring from a security perspective that its customers don't have anything to worry about. >>That's always a thin line between security and the user experience. So I n g, like every other bank needs to make choices. Yes. We want the really ease of customers and take the risk that somebody abuses it or do we make it really, really secure and alienate part of our customer base. And that's an ongoing, that's a, that's a a hard, >>It's a trade off. That's >>A line. >>So it's really hard. Interesting part is in Netherlands we had some debates about banks closing down locations, but the moment we introduced our mobile weapon iPads, basically the debates became a lot quieter because a lot of elderly people couldn't work with an iPhone. It turned out they were perfectly fine with a well-designed iPad app to do their banking. Really? >>Okay. >>But that's already learning from like 15 years ago. >>What was the, what was the product roadmap on that? So how, I mean I can imagine you released a mobile app, you're not really thinking that. >>That's basically, I think that was a heavy coincidence. We just, Yeah, okay. Went out to design a very good mobile app. Yeah. And then looking out afterwards at the statistics we say, hey, who was using this way? We've got somebody who's signing on and I dunno the exact age, but it was something like somebody of 90 plus who signed on to use that mobile app. >>Wow. Wow. I mean you really are the five different generations living and working right now. Designing technology. Everybody has to go to the bank whether we are fans of our bank or we're not. Although now I'm thinking about IG as a bank in general. Y'all have a a very good attitude about it. What has kept you at the company for over 20 years? That is we, we see people move around, especially in this technology industry. Yes. Yeah. You know, every two to three years. Sometimes obviously you're in positions of leadership, they're obviously taking good care of you. But I mean multiple decades. Why have you stuck? >>Well first I didn't have the same job in I N D for two decades. Nice. So I went around the infrastructure domain. I did storage initially I did security, I did solution design and in the end I ended up in enterprise architecture. So yeah, it's not like I stuck 20 years in the same role. So every so years >>Go up the ladder but also grow your own skill sets. >>Explore. Yeah. >>So basically I think that's what's every, everybody should be thinking in these days. If you're in a cloud head industry, if you're good at it, you can out quite a nice salary. But it also means that you have some kind of obligation to society to make a difference. And I think, yeah, >>I wouldn't say that everybody feels that way. I >>Need to make a difference with I N G A difference for being more available to our consumers, be more secure to, to our consumers. I, I think that's what's driving me to stick with the company. >>What about you R Now? >>Yes, for me it's very important. Every two, three years are doing new things. I can work with the latest technology so I become really, really innovative so that it is the place to be. >>Yeah. You sort of get that rotation every two to three years with the different tools that you're using. Speaking of or here we're at Cuan, we're talking cloud native, we're talking Kubernetes. Do you think it's possible to, I'm coming back to the regulations. Do you think it's possible to get to banking grade security with cloud native Tech? >>Initially I said we would be at least as secure traditional la but last Tuesday we've proven we can get more secure than situational it. So yeah, definitely. Yes. >>Awesome. I mean, sounds like you proved it to yourself too, which is really saying something. >>Well we actually have Penta results and of course I cannot divulge those, but I about pretty good. >>Can you define, I wanna kind of double book on thanking great security, define what that is, thanking great security and how could other industries aim to Yeah, >>Hit that, that >>Standard. I want security everywhere. Especially my bank. The >>Architecture is zero privilege. So you hear a lot about lease privilege in all the security talks. That's not what you should be aiming for. Zero privilege is what you should be aiming for. And once you're at zero privileged environments, okay, who can leak data because no natural person has access to it. Even if you have somebody invading your infrastructure, there are no privileges. They cannot do privilege escalations. Yeah. So the answer for me is really clear. If you are handling customer data, if you're and customer funds aim for zero privilege architecture, >>What, what are you most excited about next? What's next for you guys? What's next for I n G? What are we gonna be talking about when we're chatting to you Right here? Atan next year or in Amsterdam actually, since we're headed that way in the spring, which is fun. Yes. >>Happy to be your host in Amsterdam. The >>Other way around. We're holding you to that. You've talked about how fun the culture is. Now you're gonna ask, she and I we need, but we need the tee-shirts. We, we obviously need a matching outfit. >>Definitely. We'll arrange some teachers for you as well. Yeah, no, for me, two highlights from this com. The first one was kcp. That can potentially be a paradigm change on how we deal with workloads on Kubernetes. So that's very interesting. I don't know if you see any implementations by next year, but it's definitely something. Looks >>Like we had them on the show as well. Yeah. So it's, it's very fun. I'm sure, I'm sure they'll be very flattered that you just just said. What about you Arnoldo that got you most excited? >>The most important for me was talking to a lot of Asian is other people. What if they thinking how we go forward? So the, the, the community and talk to each other. And also we found those and people how we go forward. >>Yeah, that's been a big thing for us here on the cube and just the energy, the morale. I mean the open source community is so collaborative. It creates an entirely different ethos. Arna. Ty, thank you so much for being here. It's wonderful to have you and hear what I n g is doing in the technology space. Lisa, always a pleasure to co-host with you. Of course. And thank you Cube fans for hanging out with us here on day three of Cuban Live from Detroit, Michigan. My name is Savannah Peterson and we'll see you up next for a great chat coming soon.
SUMMARY :
And joining me is my beautiful co-host, Lisa, how you feeling? I love it. Just like the guests that we have up for you next. Glad to be you. I forgot what's going on with the shirts guys. VR code is basically the player of our company So let's actually, let's just talk about that for a second. So we have lots and lots of problems to solve. How do we live a banking cot with cloud native technology and together So we were not just presenting, It's not all smoke and mirrors. And that's what we wanted to share here. Well, and as consumers we expect we can access our banking on any device 24 So we have a huge engineers and people around to And it is, they cannot access it. So Lisa and I were chatting before we brought you up here. Basically it means we don't give a full cluster to our consumers, right? What, how challenging is that for you as technologists And the problem is you have legislation and So the challenge we had, basically we had to wide our, in our own container security standards to prove Hey, so you what it's, I guess, I guess the rules are a little bit easier when you get to help define them. Because the hardest problem is try to be compliant for something You talked about what that is, And that is the most, most important. Talk about both of you. So that's more than 20 years Ana, what about you? So both of your comment on that's a long time. of the cases you get some time or some budgets or even some additional colleagues to help you out and making sure the content that's delivered to them is relevant, that it's secure. abuses it or do we make it really, really secure and alienate part of our customer It's a trade off. but the moment we introduced our mobile weapon iPads, basically the debates became a So how, I mean I can imagine you released a mobile app, And then looking out afterwards at the statistics we say, What has kept you at the company for over 20 years? I did solution design and in the end I ended up in enterprise architecture. Yeah. that you have some kind of obligation to society to make a difference. I wouldn't say that everybody feels that way. Need to make a difference with I N G A difference for being more available to our consumers, technology so I become really, really innovative so that it is the place to be. Do you think it's possible to get to we can get more secure than situational it. I mean, sounds like you proved it to yourself too, which is really saying something. I want security everywhere. So you hear a lot about lease privilege in all the security talks. What are we gonna be talking about when we're chatting to you Right here? Happy to be your host in Amsterdam. We're holding you to that. I don't know if you see any implementations by What about you Arnoldo that got you most excited? And also we And thank you Cube fans for hanging out with us here on day three of Cuban Live from Detroit,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa | PERSON | 0.99+ |
Amsterdam | LOCATION | 0.99+ |
2019 | DATE | 0.99+ |
Ana | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Netherlands | LOCATION | 0.99+ |
Arnold | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
20 years | QUANTITY | 0.99+ |
September, 2001 | DATE | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
both | QUANTITY | 0.99+ |
I N G | ORGANIZATION | 0.99+ |
iPads | COMMERCIAL_ITEM | 0.99+ |
two decades | QUANTITY | 0.99+ |
three years | QUANTITY | 0.99+ |
iPad | COMMERCIAL_ITEM | 0.99+ |
Detroit, Michigan | LOCATION | 0.99+ |
Detroit, Michigan | LOCATION | 0.99+ |
today | DATE | 0.99+ |
next year | DATE | 0.99+ |
KubeCon | EVENT | 0.99+ |
Arno Vonk | PERSON | 0.99+ |
both parties | QUANTITY | 0.99+ |
IG | ORGANIZATION | 0.99+ |
more than 20 years | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
last Tuesday | DATE | 0.99+ |
over 20 years | QUANTITY | 0.98+ |
I n G | ORGANIZATION | 0.98+ |
Thijs Ebbers | PERSON | 0.98+ |
15 years ago | DATE | 0.97+ |
CloudNativeCon | EVENT | 0.97+ |
seven | QUANTITY | 0.97+ |
Cuan | ORGANIZATION | 0.97+ |
first | QUANTITY | 0.97+ |
90 plus | QUANTITY | 0.96+ |
Cube | ORGANIZATION | 0.96+ |
Zero privilege | QUANTITY | 0.95+ |
Penta | ORGANIZATION | 0.94+ |
Arna | PERSON | 0.94+ |
first one | QUANTITY | 0.93+ |
zero privilege | QUANTITY | 0.93+ |
one wifi | QUANTITY | 0.92+ |
Kubernetes | ORGANIZATION | 0.92+ |
2000 | DATE | 0.92+ |
Arnoldo | PERSON | 0.92+ |
OnPrem | ORGANIZATION | 0.92+ |
two highlights | QUANTITY | 0.92+ |
day three | QUANTITY | 0.91+ |
five different generations | QUANTITY | 0.9+ |
ING | ORGANIZATION | 0.9+ |
24 | QUANTITY | 0.89+ |
CubeCon | ORGANIZATION | 0.88+ |
G Bank | ORGANIZATION | 0.87+ |
zero privilege architecture | QUANTITY | 0.86+ |
secondly | QUANTITY | 0.86+ |
Atan | LOCATION | 0.85+ |
two presentations | QUANTITY | 0.83+ |
S shift commons conference | EVENT | 0.82+ |
NA 2022 | EVENT | 0.82+ |
zero privileged | QUANTITY | 0.81+ |
every two | QUANTITY | 0.81+ |
last three years | DATE | 0.79+ |
double | QUANTITY | 0.77+ |
Ty Evers | ORGANIZATION | 0.76+ |
device | QUANTITY | 0.72+ |
Afternoon | DATE | 0.72+ |
Cuban Live | EVENT | 0.7+ |
a second | QUANTITY | 0.69+ |
Ty | PERSON | 0.68+ |
three | QUANTITY | 0.65+ |
every | QUANTITY | 0.57+ |
i Container Ocean | ORGANIZATION | 0.56+ |
Afternoon of day | DATE | 0.54+ |
Kubernetes | TITLE | 0.52+ |
Drew Nielsen, Teleport | KubeCon + CloudNativeCon NA 2022
>>Good afternoon, friends. My name is Savannah Peterson here in the Cube Studios live from Detroit, Michigan, where we're at Cuban and Cloud Native Foundation, Cloud Native Con all week. Our last interview of the day served me a real treat and one that I wasn't expecting. It turns out that I am in the presence of two caddies. It's a literal episode of Caddy Shack up here on Cube. John Furrier. I don't think the audience knows that you were a caddy. Tell us about your caddy days. >>I used to caddy when I was a kid at the local country club every weekend. This is amazing. Double loops every weekend. Make some bang, two bags on each shoulder. Caddying for the members where you're going. Now I'm >>On show. Just, just really impressive >>Now. Now I'm caddying for the cube where I caddy all this great content out to the audience. >>He's carrying the story of emerging brands and established companies on their cloud journey. I love it. John, well played. I don't wanna waste any more of this really wonderful individual's time, but since we now have a new trend of talking about everyone's Twitter handle here on the cube, this may be my favorite one of the day, if not Q4 so far. Drew, not reply. AKA Drew ne Drew Nielsen, excuse me, there is here with us from Teleport. Drew, thanks so much for being here. >>Oh, thanks for having me. It's great to be here. >>And so you were a caddy on a whole different level. Can you tell us >>About that? Yeah, so I was in university and I got tired after two years and didn't have a car in LA and met a pro golfer at a golf course and took two years off and traveled around caddying for him and tried to get 'em through Q School. >>This is, this is fantastic. So if you're in school and your parents are telling you to continue going to school, know that you can drop out and be a caddy and still be a very successful television personality. Like both of the gentlemen at some point. >>Well, I never said my parents like >>That decision, but we'll keep our day jobs. Yeah, exactly. And one of them is Cloud Native Security. The hottest topic here at the show. Yep. I want to get into it. You guys are doing some really cool things. Are we? We hear Zero Trust, you know, ransomware and we even, I even talked with the CEO of Dockets morning about container security issues. Sure. There's a lot going on. So you guys are in the middle of teleport. You guys have a unique solution. Tell us what you guys got going on. What do you guys do? What's the solution and what's the problem you solve? >>So Teleport is the first and only identity native infrastructure access solution in the market. So breaking that down, what that really means is identity native being the combination of secret list, getting rid of passwords, Pam Vaults, Key Vaults, Yeah. Passwords written down. Basically the number one source of breach. And 50 to 80% of breaches, depending on whose numbers you want to believe are how organizations get hacked. >>But it's not password 1 23 isn't protecting >>Cisco >>Right >>Now. Well, if you think about when you're securing infrastructure and the second component being zero trust, which assumes the network is completely insecure, right? But everything is validated. Resource to resource security is validated, You know, it assumes work from anywhere. It assumes the security comes back to that resource. And we take the combination of those two into identity, native access where we cryptographically ev, validate identity, but more importantly, we make an absolutely frictionless experience. So engineers can access infrastructure from anywhere at any time. >>I'm just flashing on my roommates, checking their little code, changing Bob login, you know, dongle essentially, and how frustrating that always was. I mean, talk about interrupting workflow was something that's obviously necessary, but >>Well, I mean, talk about frustration if I'm an engineer. Yeah, absolutely. You know, back in the day when you had these three tier monolithic applications, it was kind of simple. But now as you've got modern application development environments Yeah, multi-cloud, hybrid cloud, whatever marketing term around how you talk about this, expanding sort of disparate infrastructure. Engineers are sitting there going from system to system to machine to database to application. I mean, not even a conversation on Kubernetes yet. Yeah. And it's just, you know, every time you pull an engineer or a developer to go to a vault to pull something out, you're pulling them out for 10 minutes. Now, applications today have hundreds of systems, hundreds of microservices. I mean 30 of these a day and nine minutes, 270 minutes times 60. And they also >>Do the math. Well, there's not only that, there's also the breach from manual error. I forgot to change the password. What is that password? I left it open, I left it on >>Cognitive load. >>I mean, it's the manual piece. But even think about it, TR security has to be transparent and engineers are really smart people. And I've talked to a number of organizations who are like, yeah, we've tried to implement security solutions and they fail. Why? They're too disruptive. They're not transparent. And engineers will work their way around them. They'll write it down, they'll do a workaround, they'll backdoor it something. >>All right. So talk about how it works. But I, I mean, I'm getting the big picture here. I love this. Breaking down the silos, making engineers lives easier, more productive. Clearly the theme, everyone they want, they be gonna need. Whoever does that will win it all. How's it work? I mean, you deploying something, is it code, is it in line? It's, >>It's two binaries that you download and really it starts with the core being the identity native access proxy. Okay. So that proxy, I mean, if you look at like the zero trust principles, it all starts with a proxy. Everything connects into that proxy where all the access is gated, it's validated. And you know, from there we have an authorization engine. So we will be the single source of truth for all access across your entire infrastructure. So we bring machines, engineers, databases, applications, Kubernetes, Linux, Windows, we don't care. And we basically take that into a single architecture and single access platform that essentially secures your entire infrastructure. But more importantly, you can do audit. So for all of the organizations that are dealing with FedRAMP, pci, hipaa, we have a complete audit trail down to a YouTube style playback. >>Oh, interesting. We're we're California and ccpa. >>Oh, gdpr. >>Yeah, exactly. It, it, it's, it's a whole shebang. So I, I love, and John, maybe you've heard this term a lot more than I have, but identity native is relatively new to me as as a term. And I suspect you have a very distinct way of defining identity. How do you guys define identity internally? >>So identity is something that is cryptographically validated. It is something you have. So it's not enough. If you look at, you know, credentials today, everyone's like, Oh, I log into my computer, but that's my identity. No, it's not. Right. Those are attributes. Those are something that is secret for a period of time until you write it down. But I can't change my fingerprint. Right. And now I >>Was just >>Thinking of, well no, perfect case in point with touch ID on your meth there. Yeah. It's like when we deliver that cryptographically validated identity, we use these secure modules in like modern laptops or servers. Yeah. To store that identity so that even if you're sitting in front of your computer, you can't get to it. But more importantly, if somebody were to take that and try to be you and try to log in with your fingerprint, it's >>Not, I'm not gonna lie, I love the apple finger thing, you know, it's like, you know, space recognition, like it's really awesome. >>It save me a lot of time. I mean, even when you go through customs and they do the face scan now it actually knows who you are, which is pretty wild in the last time you wanna provide ones. But it just shifted over like maybe three months ago. Well, >>As long as no one chops your finger off like they do in the James Bond movies. >>I mean, we try and keep it a light and fluffy here on the queue, but you know, do a finger teams, we can talk about that >>Too. >>Gabby, I was thinking more minority report, >>But you >>Knows that's exactly what I, what I think of >>Hit that one outta bounds. So I gotta ask, because you said you're targeting engineers, not IT departments. What's, is that, because I in your mind it is now the engineers or what's the, is always the solution more >>Targeted? Well, if you really look at who's dealing with infrastructure on a day-to-day basis, those are DevOps individuals. Those are infrastructure teams, Those are site reliability engineering. And when it, they're the ones who are not only managing the infrastructure, but they're also dealing with the code on it and everything else. And for us, that is who is our primary customer and that's who's doing >>It. What's the biggest problem that you're solving in this use case? Because you guys are nailing it. What's the problem that your identity native solution solves? >>You know, right out of the backs we remove the number one source of breach. And that is taking passwords, secrets and, and keys off the board. That deals with most of the problem right there. But there are really two problems that organizations face. One is scaling. So as you scale, you get more secrets, you get more keys, you get all these things that is all increasing your attack vector in real time. Oh >>Yeah. Across teams locations. I can't even >>Take your pick. Yeah, it's across clouds, right? Any of it >>On-prem doesn't. >>Yeah. Any of it. We, and we allow you to scale, but do it securely and the security is transparent and your engineers will absolutely love it. What's the most important thing about this product Engineers. Absolutely. >>What are they saying? What are some of those examples? Anecdotally, pull boats out from engineering. >>You're too, we should have invent, we should have invented this ourselves. Or you know, we have run into a lot of customers who have tried to home brew this and they're like, you know, we spend an in nor not of hours on it >>And IT or they got legacy from like Microsoft or other solutions. >>Sure, yeah. Any, but a lot of 'em is just like, I wish I had done it myself. Or you know, this is what security should be. >>It makes so much sense and it gives that the team such a peace of mind. I mean, you never know when a breach is gonna come, especially >>It's peace of mind. But I think for engineers, a lot of times it deals with the security problem. Yeah. Takes it off the table so they can do their jobs. Yeah. With zero friction. Yeah. And you know, it's all about speed. It's all about velocity. You know, go fast, go fast, go fast. And that's what we enable >>Some of the benefits to them is they get to save time, focus more on, on task that they need to work on. >>Exactly. >>And get the >>Job done. And on top of it, they answer the audit and compliance mail every time it comes. >>Yeah. Why are people huge? Honestly, why are people doing this? Because, I mean, identity is just such an hard nut to crack. Everyone's got their silos, Vendors having clouds have 'em. Identity is the most fragmented thing on >>The planet. And it has been fragmented ever since my first RSA conference. >>I know. So will we ever get this do over? Is there a driver? Is there a market force? Is this the time? >>I think the move to modern applications and to multi-cloud is driving this because as those application stacks get more verticalized, you just, you cannot deal with the productivity >>Here. And of course the next big thing is super cloud and that's coming fast. Savannah, you know, You know that's Rocket. >>John is gonna be the thought leader and keyword leader of the word super cloud. >>Super Cloud is enabling super services as the cloud cast. Brian Gracely pointed out on his Sunday podcast of which if that happens, Super Cloud will enable super apps in a new architectural >>List. Please don't, and it'll be super, just don't. >>Okay. Right. So what are you guys up to next? What's the big hot spot for the company? What are you guys doing? What are you guys, What's the idea guys hiring? You put the plug in. >>You know, right now we are focused on delivering the best identity, native access platform that we can. And we will continue to support our customers that want to use Kubernetes, that want to use any different type of infrastructure. Whether that's Linux, Windows applications or databases. Wherever they are. >>Are, are your customers all of a similar DNA or are you >>No, they're all over the map. They range everything from tech companies to financial services to, you know, fractional property. >>You seem like someone everyone would need. >>Absolutely. >>And I'm not just saying that to be a really clean endorsement from the Cube, but >>If you were doing DevOps Yeah. And any type of forward-leaning shift, left engineering, you need us because we are basically making security as code a reality across your entire infrastructure. >>Love this. What about the team dna? Are you in a scale growth stage right now? What's going on? Absolutely. Sounds I was gonna say, but I feel like you would have >>To be. Yeah, we're doing, we're, we have a very positive outlook and you know, even though the economic time is what it is, we're doing very well meeting. >>How's the location? Where's the location of the headquarters now? With remote work is pretty much virtual. >>Probably. We're based in downtown Oakland, California. >>Woohoo. Bay area representing on this stage right now. >>Nice. Yeah, we have a beautiful office right in downtown Oakland and yeah, it's been great. Awesome. >>Love that. And are you hiring right now? I bet people might be. I feel like some of our cube watchers are here waiting to figure out their next big play. So love to hear that. Absolutely love to hear that. Besides Drew, not reply, if people want to join your team or say hello to you and tell you how brilliant you looked up here, or ask about your caddy days and maybe venture a guest to who that golfer may have been that you were CAD Inc. For, what are the best ways for them to get in touch with you? >>You can find me on LinkedIn. >>Great. Fantastic. John, anything else >>From you? Yeah, I mean, I just think security is paramount. This is just another example of where the innovation has to kind of break through without good identity, everything could cripple. Then you start getting into the silos and you can start getting into, you know, tracking it. You got error user errors, you got, you know, one of the biggest security risks. People just leave systems open, they don't even know it's there. So like, I mean this is just, just identity is the critical linchpin to, to solve for in security to me. And that's totally >>Agree. We even have a lot of customers who use us just to access basic cloud consoles. Yeah. >>So I was actually just gonna drive there a little bit because I think that, I'm curious, it feels like a solution for obviously complex systems and stacks, but given the utility and what sounds like an extreme ease of use, I would imagine people use this for day-to-day stuff within their, >>We have customers who use it to access their AWS consoles. We have customers who use it to access Grafana dashboards. You know, for, since we're sitting here at coupon accessing a Lens Rancher, all of the amazing DevOps tools that are out there. >>Well, I mean true. I mean, you think about all the reasons why people don't adopt this new federated approach or is because the IT guys did it and the world we're moving into, the developers are in charge. And so we're seeing the trend where developers are taking the DevOps and the data and the security teams are now starting to reset the guardrails. What's your >>Reaction to that? Well, you know, I would say that >>Over the top, >>Well I would say that you know, your DevOps teams and your infrastructure teams and your engineers, they are the new king makers. Yeah. Straight up. Full stop. >>You heard it first folks. >>And that's >>A headline right >>There. That is a headline. I mean, they are the new king makers and, but they are being forced to do it as securely as possible. And our job is really to make that as easy and as frictionless as possible. >>Awesome. >>And it sounds like you're absolutely nailing it. Drew, thank you so much for being on the show. Thanks for having today. This has been an absolute pleasure, John, as usual a joy. And thank all of you for tuning in to the Cube Live here at CU Con from Detroit, Michigan. We look forward to catching you for day two tomorrow.
SUMMARY :
I don't think the audience knows that you were a caddy. the members where you're going. Just, just really impressive He's carrying the story of emerging brands and established companies on It's great to be here. And so you were a caddy on a whole different level. Yeah, so I was in university and I got tired after two years and didn't have to school, know that you can drop out and be a caddy and still be a very successful television personality. What's the solution and what's the problem you solve? And 50 to 80% of breaches, depending on whose numbers you want to believe are how organizations It assumes the security comes back to that resource. you know, dongle essentially, and how frustrating that always was. You know, back in the day when you had these three tier I forgot to change I mean, it's the manual piece. I mean, you deploying something, is it code, is it in line? And you know, from there we have an authorization engine. We're we're California and ccpa. And I suspect you have a very distinct way of that is secret for a period of time until you write it down. try to be you and try to log in with your fingerprint, it's Not, I'm not gonna lie, I love the apple finger thing, you know, it's like, you know, space recognition, I mean, even when you go through customs and they do the face scan now So I gotta ask, because you said you're targeting Well, if you really look at who's dealing with infrastructure on a day-to-day basis, those are DevOps individuals. Because you guys are nailing it. So as you scale, you get more secrets, you get more keys, I can't even Take your pick. We, and we allow you to scale, but do it securely What are they saying? they're like, you know, we spend an in nor not of hours on it Or you know, you never know when a breach is gonna come, especially And you know, it's all about speed. And on top of it, they answer the audit and compliance mail every time it comes. Identity is the most fragmented thing on And it has been fragmented ever since my first RSA conference. I know. Savannah, you know, Super Cloud is enabling super services as the cloud cast. So what are you guys up to next? And we will continue to support our customers that want to use Kubernetes, you know, fractional property. If you were doing DevOps Yeah. Sounds I was gonna say, but I feel like you would have Yeah, we're doing, we're, we have a very positive outlook and you know, How's the location? We're based in downtown Oakland, California. Bay area representing on this stage right now. it's been great. And are you hiring right now? John, anything else Then you start getting into the silos and you can start getting into, you know, tracking it. We even have a lot of customers who use us just to access basic cloud consoles. a Lens Rancher, all of the amazing DevOps tools that are out there. I mean, you think about all the reasons why people don't adopt this Well I would say that you know, your DevOps teams and your infrastructure teams and your engineers, I mean, they are the new king makers and, but they are being forced to We look forward to catching you for day
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Savannah Peterson | PERSON | 0.99+ |
30 | QUANTITY | 0.99+ |
Drew | PERSON | 0.99+ |
10 minutes | QUANTITY | 0.99+ |
Brian Gracely | PERSON | 0.99+ |
John | PERSON | 0.99+ |
LA | LOCATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Drew Nielsen | PERSON | 0.99+ |
two binaries | QUANTITY | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
270 minutes | QUANTITY | 0.99+ |
50 | QUANTITY | 0.99+ |
Savannah | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
two problems | QUANTITY | 0.99+ |
Detroit, Michigan | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
Sunday | DATE | 0.99+ |
two years | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
second component | QUANTITY | 0.99+ |
Zero Trust | ORGANIZATION | 0.99+ |
Teleport | ORGANIZATION | 0.99+ |
Windows | TITLE | 0.99+ |
ORGANIZATION | 0.99+ | |
three tier | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
Cloud Native Foundation | ORGANIZATION | 0.99+ |
YouTube | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
California | LOCATION | 0.99+ |
tomorrow | DATE | 0.98+ |
two bags | QUANTITY | 0.98+ |
Linux | TITLE | 0.98+ |
One | QUANTITY | 0.98+ |
80% | QUANTITY | 0.98+ |
three months ago | DATE | 0.98+ |
FedRAMP | ORGANIZATION | 0.98+ |
day two | QUANTITY | 0.98+ |
KubeCon | EVENT | 0.98+ |
CloudNativeCon | EVENT | 0.97+ |
Super Cloud | TITLE | 0.97+ |
Gabby | PERSON | 0.96+ |
nine minutes | QUANTITY | 0.96+ |
Cube Studios | ORGANIZATION | 0.95+ |
a day | QUANTITY | 0.95+ |
CU Con | EVENT | 0.95+ |
Double | QUANTITY | 0.94+ |
ORGANIZATION | 0.94+ | |
zero friction | QUANTITY | 0.94+ |
Bob | PERSON | 0.93+ |
Cube | ORGANIZATION | 0.92+ |
Caddy Shack | TITLE | 0.92+ |
Q School | ORGANIZATION | 0.91+ |
single access platform | QUANTITY | 0.91+ |
zero trust | QUANTITY | 0.89+ |
single architecture | QUANTITY | 0.89+ |
60 | QUANTITY | 0.88+ |
downtown Oakland, California | LOCATION | 0.88+ |
teleport | ORGANIZATION | 0.87+ |
Kubernetes | TITLE | 0.87+ |
two caddies | QUANTITY | 0.87+ |
pci | ORGANIZATION | 0.86+ |
each shoulder | QUANTITY | 0.85+ |
Cuban | ORGANIZATION | 0.85+ |
single source | QUANTITY | 0.85+ |
hundreds of microservices | QUANTITY | 0.84+ |
zero trust | QUANTITY | 0.83+ |
Dockets | ORGANIZATION | 0.83+ |
NA 2022 | EVENT | 0.82+ |
CAD Inc. | ORGANIZATION | 0.81+ |
Bay | LOCATION | 0.8+ |
one source | QUANTITY | 0.78+ |
RSA conference | EVENT | 0.78+ |
hundreds of systems | QUANTITY | 0.77+ |
Cloud Native | EVENT | 0.76+ |
Jason Collier, AMD | VMware Explore 2022
(upbeat music) >> Welcome back to San Francisco, "theCUBE" is live, our day two coverage of VMware Explore 2022 continues. Lisa Martin with Dave Nicholson. Dave and I are pleased to welcome Jason Collier, principal member of technical staff at AMD to the program. Jason, it's great to have you. >> Thank you, it's great to be here. >> So what's going on at AMD? I hear you have some juicy stuff to talk about. >> Oh, we've got a ton of juicy stuff to talk about. Clearly the Project Monterey announcement was big for us, so we've got that to talk about. Another thing that I really wanted to talk about was a tool that we created and we call it, it's the VMware Architecture Migration Tool, call it VAMT for short. It's a tool that we created and we worked together with VMware and some of their professional services crew to actually develop this tool. And it is also an open source based tool. And really the primary purpose is to easily enable you to move from one CPU architecture to another CPU architecture, and do that in a cold migration fashion. >> So we're probably not talking about CPUs from Tandy, Radio Shack systems, likely this would be what we might refer to as other X86 systems. >> Other X86 systems is a good way to refer to it. >> So it's interesting timing for the development and the release of a tool like this, because in this sort of X86 universe, there are players who have been delayed in terms of delivering their next gen stuff. My understanding is AMD has been public with the idea that they're on track for by the end of the year, Genoa, next gen architecture. So can you imagine a situation where someone has an existing set of infrastructure and they're like, hey, you know what I want to get on board, the AMD train, is this something they can use from the VMware environment? >> Absolutely, and when you think about- >> Tell us exactly what that would look like, walk us through 100 servers, VMware, 1000 VMs, just to make the math easy. What do you do? How does it work? >> So one, there's several things that the tool can do, we actually went through, the design process was quite extensive on this. And we went through all of the planning phases that you need to go through to do these VM migrations. Now this has to be a cold migration, it's not a live migration. You can't do that between the CPU architectures. But what we do is you create a list of all of the virtual machines that you want to migrate. So we take this CSV file, we import this CSV file, and we ask for things like, okay, what's the name? Where do you want to migrate it to? So from one cluster to another, what do you want to migrate it to? What are the networks that you want to move it to? And then the storage platform. So we can move storage, it could either be shared storage, or we could move say from VSAN to VSAN, however you want to set it up. So it will do those storage migrations as well. And then what happens is it's actually going to go through, it's going to shut down the VM, it's going to take a snapshot, it is going to then basically move the compute and/or storage resources over. And once it does that, it's going to power 'em back up. And it's going to check, we've got some validation tools, where it's going to make sure VM Tools comes back up where everything is copacetic, it didn't blue screen or anything like that. And once it comes back up, then everything's good, it moves onto the next one. Now a couple of things that we've got feature wise, we built into it. You can parallelize these tasks. So you can say, how many of these machines do you want to do at any given time? So it could be, say 10 machines, 50 machines, 100 machines at a time, that you want to go through and do this move. Now, if it did blue screen, it will actually roll it back to that snapshot on the origin cluster. So that there is some protection on that. A couple other things that are actually in there are things like audit tracking. So we do full audit logging on this stuff, we take a snapshot, there's basically kind of an audit trail of what happens. There's also full logging, SYS logging, and then also we'll do email reporting. So you can say, run this and then shoot me a report when this is over. Now, one other cool thing is you can also actually define a change window. So I don't want to do this in the middle of the afternoon on a Tuesday. So I want to do this later at night, over the weekend, you can actually just queue this up, set it, schedule it, it'll run. You can also define how long you want that change window to be. And what it'll do, it'll do as many as it can, then it'll effectively stop, finish up, clean up the tasks and then send you a report on what all was successfully moved. >> Okay, I'm going to go down the rabbit hole a little bit on this, 'cause I think it's important. And if I say something incorrect, you correct me. >> No problem. >> In terms of my technical understanding. >> I got you. >> So you've got a VM, essentially a virtual machine typically will consist of an entire operating system within that virtual machine. So there's a construct that containerizes, if you will, the operating system, what is the difference, where is the difference in the instruction set? Where does it lie? Is it in the OS' interaction with the CPU or is it between the construct that is the sort of wrapper around the VM that is the difference? >> It's really primarily the OS, right? And we've not really had too many issues doing this and most of the time, what is going to happen, that OS is going to boot up, it's going to recognize the architecture that it's on, it's going to see the underlying architecture, and boot up. All the major operating systems that we test worked fine. I mean, typically they're going to work on all the X86 platforms. But there might be instruction sets that are kind of enabled in one architecture that may not be in another architecture. >> And you're looking for that during this process. >> Well usually the OS itself is going to kind of detect that. So if it pops up, the one thing that is kind of a caution that you need to look for. If you've got an application that's explicitly using an instruction set that's on one CPU vendor and not the other CPU vendor. That's the one thing where you're probably going to see some application differences. That said, it'll probably be compatible, but you may not get that instruction set advantage in it. >> But this tool remediates against that. >> Yeah, and what we do, we're actually using VM Tools itself to go through and validate a lot of those components. So we'll look and make sure VM Tools is enabled in the first place, on the source system. And then when it gets to the destination system, we also look at VM Tools to see what is and what is not enabled. >> Okay, I'm going to put you on the spot here. What's the zinger, where doesn't it work? You already said cold, we understand, you can schedule for cold migrations, that's not a zinger. What's the zinger, where doesn't it work? >> It doesn't work like, live migrations just don't work. >> No live, okay, okay, no live. What about something else? What's the oh, you've got that version, you've got that version of X86 architecture, it-won't work, anything? >> A majority of those cases work, where it would fail, where it's going to kick back and say, hey, VM Tools is not installed. So where you would see this is if you're running a virtual appliance from some vendor, like insert vendor here that say, got a firewall, or got something like that, and they don't have VM Tools enabled. It's going to fail it out of the gate, and say, hey, VM Tools is not on this, you might want to manually do it. >> But you can figure out how to fix that? >> You can figure out how to do that. You can also, and there's a flag in there, so in kind of the options that you give it, you say, ignore VM Tools, don't care, move it anyway. So if you've got less, some VMs that are in there, but they're not a priority VM, then it's going to migrate just fine. >> Got It. >> Can you elaborate a little bit on the joint development work that AMD and VMware are doing together and the value in it for customers? >> Yeah, so it's one of those things we worked with VMware to basically produce this open source tool. So we did a lot of the core component and design and we actually engaged VMware Professional Services. And a big shout out to Austin Browder. He helped us a ton in this project specifically. And we basically worked, we created this, kind of co-designed, what it was going to look like. And then jointly worked together on the coding, of pulling this thing together. And then after that, and this is actually posted up on VMware's public repos now in GitHub. So you can go to GitHub, you can go to the VMware samples code, and you can download this thing that we've created. And it's really built to help ease migrations from one architecture to another. So if you're looking for a big data center move and you got a bunch of VMs to move. I mean, even if it's same architecture to same architecture, it's definitely going to ease the pain of going through and doing a migration of, it's one thing when you're doing 10 machines, but when you're doing 10,000 virtual machines, that's a different story. It gets to be quite operationally inefficient. >> I lose track after three. >> Yeah. >> So I'm good for three, not four. >> I was going to ask you what your target market segment is here. Expand on that a little bit and talk to me about who you're working with and those organizations. >> So really this is targeted toward organizations that have large deployments in enterprise, but also I think this is a big play with channel partners as well. So folks out there in the channel that are doing these migrations and they do a lot of these, when you're thinking about the small and mid-size organizations, it's a great fit for that. Especially if they're kind of doing that upgrade, the lift and shift upgrade, from here's where you've been five to seven years on an architecture and you want to move to a new architecture. This is really going to help. And this is not a point and click GUI kind of thing. It's command line driven, it's using PowerShell, we're using PowerCLI to do the majority of this work. And for channel partners, this is an excellent opportunity to put the value and the value add and VAR, And there's a lot of opportunity for, I think, channel partners to really go and take this. And once again, being open source. We expect this to be extensible, we want the community to contribute and put back into this to basically help grow it and make it a more useful tool for doing these cold migrations between CPU architectures. >> Have you seen any in the last couple of years of dynamics, obviously across the world, any industries in particular that are really leading edge for what you guys are doing? >> Yeah, that's really, really interesting. I mean, we've seen it, it's honestly been a very horizontal problem, pretty much across all vertical markets. I mean, we've seen it in financial services, we've seen it in, honestly, pretty much across the board. Manufacturing, financial services, healthcare, we have seen kind of a strong interest in that. And then also we we've actually taken this and presented this to some of our channel partners as well. And there's been a lot of interest in it. I think we presented it to about 30 different channel partners, a couple of weeks back about this. And I got contact from 30 different channel partners that said they're interested in basically helping us work on it. >> Tagging on to Lisa's question, do you have visibility into the AMD thought process around the timing of your next gen release versus others that are competitors in the marketplace? How you might leverage that in terms of programs where partners are going out and saying, hey, perfect time, you need a refresh, perfect time to look at AMD, if you haven't looked at them recently. Do you have any insight into that in what's going on? I know you're focused on this area. But what are your thoughts on, well, what's the buzz? What's the buzz inside AMD on that? >> Well, when you look overall, if you look at the Gartner Hype Cycle, when VMware was being broadly adopted, when VMware was being broadly adopted, I'm going to be blunt, and I'm going to be honest right here, AMD didn't have a horse in the race. And the majority of those VMware deployments we see are not running on AMD. Now that said, there's an extreme interest in the fact that we've got these very cored in systems that are now coming up on, now you're at that five to seven year refresh window of pulling in new hardware. And we have extremely attractive hardware when it comes to running virtualized workloads. The test cluster that I'm running at home, I've got that five to seven year old gear, and I've got some of the, even just the Milan systems that we've got. And I've got three nodes of another architecture going onto AMD. And when I got these three nodes completely maxed to the number of VMs that I can run on 'em, I'm at a quarter of the capacity of what I'm putting on the new stuff. So what you get is, I mean, we worked the numbers, and it's definitely, it's like a 30% decrease in the amount of resources that you need. >> That's a compelling number. >> It's a compelling number. >> 5%, 10%, nobody's going to do anything for that. You talk 30%. >> 30%. It's meaningful, it's meaningful. Now you you're out of Austin, right? >> Yes. >> So first thing I thought of when you talk about running clusters in your home is the cost of electricity, but you're okay. >> I'm okay. >> You don't live here, you don't live here, you don't need to worry about that. >> I'm okay. >> Do you have a favorite customer example that you think really articulates the value of AMD when you're in customer conversations and they go, why AMD and you hit back with this? >> Yeah. Actually it's funny because I had a conversation like that last night, kind of random person I met later on in the evening. We were going through this discussion and they were facing exactly this problem. They had that five to seven year infrastructure. It's funny, because the guy was a gamer too, and he's like, man, I've always been a big AMD fan, I love the CPUs all the way since back in basically the Opterons and Athlons right. He's like, I've always loved the AMD systems, loved the graphics cards. And now with what we're doing with Ryzen and all that stuff. He's always been a big AMD fan. He's like, and I'm going through doing my infrastructure refresh. And I told him, I'm just like, well, hey, talk to your VAR and have 'em plug some AMD SKUs in there from the Dells, HPs and Lenovos. And then we've got this tool to basically help make that migration easier on you. And so once we had that discussion and it was great, then he swung by the booth today and I was able to just go over, hey, this is the tool, this is how you use it, here's all the info. Call me if you need any help. >> Yeah, when we were talking earlier, we learned that you were at Scale. So what are you liking about AMD? How does that relate? >> The funny thing is this is actually the first time in my career that I've actually had a job where I didn't work for myself. I've been doing venture backed startups the last 25 years and we've raised couple hundred million dollars worth of investment over the years. And so one, I figured, here I am going to AMD, a larger corporation. I'm just like, am I going to be able to make it a year? And I have been here longer than a year and I absolutely love it. The culture at AMD is amazing. We still have that really, I mean, almost it's like that underdog mentality within the organization. And the team that I'm working with is a phenomenal team. And it's actually, our EVP and our Corp VP, were actually my executive sponsors, we were at a prior company. They were one of my executive sponsors when I was at Scale. And so my now VP boss calls me up and says, hey, I'm putting a band together, are you interested? And I was kind of enjoying a semi-retirement lifestyle. And then I'm just like, man, because it's you, yes, I am interested. And the group that we're in, the work that we're doing, the way that we're really focusing on forward looking things that are affecting the data center, what's going to be the data center like three to five years from now. It's exciting, and I am having a blast, I'm having the time of my life. I absolutely love it. >> Well, that relationship and the trust that you will have with each other, that bleeds into the customer conversations, the partner conversations, the employee conversations, it's all inextricably linked. >> Yes it is. >> And we want to know, you said three to five years out, like what? Like what? Just general futurist stuff, where do you think this is going. >> Well, it's interesting. >> So moon collides with the earth in 2025, we already know that. >> So we dialed this back to the Pensando acquisition. When you look at the Pensando acquisition and you look at basically where data centers are today, but then you look at where basically the big hyperscalers are. You look at an AWS, you look at their architecture, you specifically wrap Nitro around that, that's a very different architecture than what's being run in the data center. And when you look at what Pensando does, that's a lot of starting to bring what these real clouds out there, what these big hyperscalers are running into the grasps of the data center. And so I think you're going to see a fundamental shift. The next 10 years are going to be exciting because the way you look at a data center now, when you think of what CPUs do, what shared storage, how the networking is all set up, it ain't going to look the same. >> Okay, so the competing vision with that, to play devil's advocate, would be DPUs are kind of expensive. Why don't we just use NICs, give 'em some more bandwidth, and use the cheapest stuff. That's the competing vision. >> That could be. >> Or the alternative vision, and I imagine everything else we've experienced in our careers, they will run in parallel paths, fit for function. >> Well, parallel paths always exist, right? Otherwise, 'cause you know how many times you've heard mainframe's dead, tape's dead, spinning disk is dead. None of 'em dead, right? The reality is you get to a point within an industry where it basically goes from instead of a growth curve like that, it goes to a growth curve of like that, it's pretty flat. So from a revenue growth perspective, I don't think you're going to see the revenue growth there. I think you're going to see the revenue growth in DPUs. And when you actually take, they may be expensive now, but you look at what Monterey's doing and you look at the way that those DPUs are getting integrated in at the OEM level. It's going to be a part of it. You're going to order your VxRail and VSAN style boxes, they're going to come with them. It's going to be an integrated component. Because when you start to offload things off the CPU, you've driven your overall utilization up. When you don't have to process NSX on basically the X86, you've just freed up cores and a considerable amount of them. And you've also moved that to where there's a more intelligent place for that pack to be processed right, out here on this edge. 'Cause you know what, that might not need to go into the host bus at all. So you have just alleviated any transfers over a PCI bus, over the PCI lanes, into DRAM, all of these components, when you're like, but all to come with, oh, that bit needs to be on this other machine. So now it's coming in and it's making that decision there. And then you take and integrate that into things like the Aruba Smart Switch, that's running the Pensando technology. So now you got top of rack that is already making those intelligent routing decisions on where packets really need to go. >> Jason, thank you so much for joining us. I know you guys could keep talking. >> No, I was going to say, you're going to have to come back. You're going to have to come back. >> We've just started to peel the layers of the onion, but we really appreciate you coming by the show, talking about what AMD and VMware are doing, what you're enabling customers to achieve. Sounds like there's a lot of tailwind behind you. That's awesome. >> Yeah. >> Great stuff, thank you. >> It's a great time to be at AMD, I can tell you that. >> Oh, that's good to hear, we like it. Well, thank you again for joining us, we appreciate it. For our guest and Dave Nicholson, I'm Lisa Martin. You're watching "theCUBE Live" from San Francisco, VMware Explore 2022. We'll be back with our next guest in just a minute. (upbeat music)
SUMMARY :
Jason, it's great to have you. I hear you have some to easily enable you to move So we're probably good way to refer to it. and the release of a tool like this, 1000 VMs, just to make the math easy. And it's going to check, we've Okay, I'm going to In terms of my that is the sort of wrapper and most of the time, that during this process. that you need to look for. in the first place, on the source system. What's the zinger, where doesn't it work? It doesn't work like, live What's the oh, you've got that version, So where you would see options that you give it, And a big shout out to Austin Browder. I was going to ask you what and the value add and VAR, and presented this to some of competitors in the marketplace? in the amount of resources that you need. nobody's going to do anything for that. Now you you're out of Austin, right? is the cost of electricity, you don't live here, you don't They had that five to So what are you liking about AMD? that are affecting the data center, Well, that relationship and the trust where do you think this is going. we already know that. because the way you look Okay, so the competing Or the alternative vision, And when you actually take, I know you guys could keep talking. You're going to have to come back. peel the layers of the onion, to be at AMD, I can tell you that. Oh, that's good to hear, we like it.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Jason Collier | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
50 machines | QUANTITY | 0.99+ |
10 machines | QUANTITY | 0.99+ |
Jason | PERSON | 0.99+ |
10 machines | QUANTITY | 0.99+ |
100 machines | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Austin | LOCATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
five | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
100 servers | QUANTITY | 0.99+ |
seven year | QUANTITY | 0.99+ |
theCUBE Live | TITLE | 0.99+ |
10,000 virtual machines | QUANTITY | 0.99+ |
Lenovos | ORGANIZATION | 0.99+ |
30% | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
four | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
10% | QUANTITY | 0.99+ |
30 different channel partners | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
earth | LOCATION | 0.99+ |
5% | QUANTITY | 0.99+ |
1000 VMs | QUANTITY | 0.99+ |
Dells | ORGANIZATION | 0.99+ |
GitHub | ORGANIZATION | 0.99+ |
seven years | QUANTITY | 0.98+ |
Austin Browder | PERSON | 0.98+ |
a year | QUANTITY | 0.98+ |
Tandy | ORGANIZATION | 0.98+ |
Radio Shack | ORGANIZATION | 0.98+ |
VMware | ORGANIZATION | 0.98+ |
Monterey | ORGANIZATION | 0.98+ |
today | DATE | 0.97+ |
HPs | ORGANIZATION | 0.97+ |
first time | QUANTITY | 0.97+ |
Tuesday | DATE | 0.97+ |
Scale | ORGANIZATION | 0.97+ |
VM Tools | TITLE | 0.97+ |
one thing | QUANTITY | 0.96+ |
last night | DATE | 0.96+ |
about 30 different channel partners | QUANTITY | 0.95+ |
first | QUANTITY | 0.95+ |
Athlons | COMMERCIAL_ITEM | 0.95+ |
VxRail | COMMERCIAL_ITEM | 0.95+ |
X86 | TITLE | 0.94+ |
Pensando | ORGANIZATION | 0.94+ |
VMware Explore 2022 | TITLE | 0.94+ |
Ryzen | COMMERCIAL_ITEM | 0.94+ |
five years | QUANTITY | 0.93+ |
Brent Meadows, Expedient & Bryan Smith, Expedient | VMware Explore 2022
(upbeat music) >> Hey everyone. Welcome to theCUBE's coverage of VMware Explore 2022. We are at Moscone West. Lisa Martin and Dave Nicholson here. Excited, really excited, whereas they were saying in the VMware keynote, pumped and jacked and jazzed to be back in-person with a lot of folks here. Keynote with standing room only. We've just come from that. We've got a couple of guests here from Expedient, going to unpack their relationship with VMware. Please welcome Brian Smith, the Senior Vice President and Chief Strategy Officer at Expedient. And Brent Meadows, the Vice President of Advanced Solution Architecture at Expedient. Guys it's great to have you on the program. >> Appreciate it bringing us on. >> Yep, welcome. >> Isn't it great to be back in person? >> It is phenomenal to be back. >> So let's talk about obviously three years since the last, what was called VMworld, so many dynamics in the market. Talk to us about what's going on at Expedient, we want to dig into Cloud Different, but kind of give us a lay of the land of what's going on and then we're going to uncrack the VMware partnership as well. >> Sure, so Expedient we're a full stack cloud service provider. So we have physical data centers that we run and then have VMware-based cloud and we've seen a huge shift from the client perspective during the pandemic in how they've really responded from everything pre-pandemic was very focused with Cloud First and trying to go that route only with hyper scaler. And there's been a big evolution with how people have to change how they think about their transformation to get the end result they're looking for. >> Talk about Cloud Different and what it's helping customers to achieve as everyone's in this accelerated transformation. >> Yeah. So, Cloud Different is something that Expedient branded. It's really about how the transformation works. And traditionally, companies thought about doing their transformation, at first they kept everything in house that they were doing and they started building their new applications out into a hyper scale cloud. And what that really is like is, a good analogy would be, it's like living in a house while you're renovating it. And I know what that's like from my relationship versus if you build a new house, or move to a new property that's completed already. And that's really the difference in that experience from a Cloud Different approach from transformation is you think of all the things that you have internally, and there's a lot of technical debt there, and that's a lot of weight that you're carrying when you're trying to do that transformation. So if you kind of flip that around and instead look to make that transformation and move all that technical debt into a cloud that's already built to run those same types of applications, a VMware-based cloud, now you can remove all of that noise, move into a curated stack of technology and everything just works. It has the security in place, your teams know how to run it, and then you can take that time you really reclaim and apply that towards new applications and new things that are strategic to the business. >> That's really critical, Brent, to get folks in the IT organization across the business, really focused on strategic initiatives rather than a lot of the mundane tasks that they just don't have time for. Brent, what are you hearing in the last couple of years with the dynamics we talked about, what are you hearing from the customer? >> Right. So, one of the big things and the challenges in the current dynamic is kind of that staffing part. So as people have built their infrastructure over the years, there's a lot of tribal knowledge that's been created during that process and every day more and more of that knowledge is walking out the door. So taking some of that technical debt that Brian mentioned and kind of removing that so you don't have to have all that tribal knowledge, really standardizing on the foundational infrastructure pieces, allows them to make that transition and not have to carry that technical debt along with them as they make their digital transformations. >> We heard a lot this morning in the keynote guys about customers going, most of them still being in cloud chaos, but VMware wanting them to get to cloud smart. What does that mean, Brian, from Expedient's perspective? What does cloud smart look like to Expedient and its customers? >> Yeah, we completely agree with that message. And it's something we've been preaching for a couple years in part of that Cloud Different story. And it's really about having a consistent wrapper across all of your environments. It doesn't matter if it's things that you're running on-premises that's legacy to things that are in a VMware-based cloud, like an Expedient cloud or things that are in a hyper scale, but having one consistent security, one consistent automation, one consistent cost management, really gives you the governance so that you can get the value out of cloud that you are hoping for and remove a lot of the noise and think less about the technology and more about what the business is getting out of the technology. >> So what does that look like as a practical matter? I imagine you have customers whose on-premises VMware environments look different than what you've created within Expedient data centers. I'm thinking of things like the level of adoption of NSX, how well a customer may embrace VSAN on-prem as an example. Is part of this transmogrification into your data center, kind of nudging people to adopt frameworks that are really necessary for success in the future? >> It's less of a nudge because a lot of times as a service provider, we don't talk about the technology, we talk more about the outcome. So the nice thing with VMware is we can move that same virtual machine or that container into the platform and the client doesn't always know exactly what's underneath because we have that standardized VMware stack and it just works. And that's part of the beauty of the process. I dunno if you want to talk about a specific client or... >> Yeah, so one of the ones we worked with is Bob Evans Foods. So they were in that transformation stage of refreshing, not only their office space and their data center, but also their VMware environment. So we helped them go through and first thing is looking at their existing environment, figuring out what they currently have, because you can't really make a good decision of what you need to change until you know where you're starting from. So we worked with them through that process, completely evacuated their data center. And from a business perspective, what that allowed them to do as well is have more flexibility in the choice of their next corporate office, because they didn't have to have a data center attached to it. So just from that data center perspective, we gave them some flexibility there. But then from an operations perspective, really standardize that process, offloaded some of those menial tasks that you mentioned earlier, and allow them to really look more towards business-driving projects, instead of just trying to keep those lights on, keeping the backups running, et cetera. >> Brian, question for you, here we are, the theme of the event is "The Center of the Multi-cloud Universe" which seems like a Marvel movie, I haven't seen any new superheroes yet, but I suspect there might be some here. But as customers end up and land in multi-cloud by default not by strategy, how does Expedient and VMware help them actually take the environment that they have and make it strategic so that the business can achieve the outcomes, improving revenue, finding new revenue streams, new products, new routes to market to delight those customers. How do you turn that kind of cloud chaos into a strategy? >> Yeah. I'd say there's a couple different components. One is really time. How can you give them time back for things that are creating noise and aren't really strategic to the business? And so if you can give that time back, that's the first way that you can really impact the business. And the second is through that standardization, but also a lot of times when people think of that new standard, they're only thinking if you're building from scratch. And what VMware has really helped is by taking those existing workloads and giving a standard that works for those applications and what you're building new and brings those together under a common platform and so had a really significant impact to the speed that somebody can get to that cloud operating model, that used to be a multi-year process and most of our clients can go from really everything or almost everything on-prem and a little bit in a cloud to a complete cloud operating model, on average, in four to six months. >> Wow! >> So if I have an on-premises environment and some of my workloads are running in a VMware context, VMware would make the pitch in an agnostic way that, "Well, you can go and deploy that "on top of a stack of infrastructure "and anybody and anywhere now." Why do customers come to you instead of saying, "Oh, we'll go to "pick your flavor of hyper scale cloud provider." What's kind of your superpower? You've mentioned a couple of things, but really hone it in on, why would someone want to go to Expedient? >> Yeah. In a single word, service. I mean, we have a 99% client retention rate and have for well over a decade. So it's really that expertise that wraps around all the different technology so that you're not worried about what's happening and you're not worried about trying to keep the lights on and doing the firefighting. You're really focused on the business. And the other way to, I guess another analogy is, if you think about a lot of the technology and the way people go to cloud, it's like if you got a set of Legos without the box or the instructions. So you can build stuff, it could be cool, but you're not going to get to that end state-- >> Hold on. That's how Legos used to work. Just maybe you're too young to remember a time-- >> You see their sales go up because now you buy a different set for this-- >> I build those sets with my son, but I do it grudgingly. >> Do you ever step on one? >> Of course I do. >> Yeah, there's some pain involved. Same thing happens in the transformation. So when they're buying services from an Expedient, you're buying that box set where you have a picture of what your outcome's going to be, the instructions are there. So you also have confidence that you're going to get to the end outcome much faster than you would if you're trying to assemble everything yourself. (David laughing) >> In my mind, I'm imagining the things that I built with Lego, before there were instructions. >> No death star? >> No. Nothing close with the death star. Definitely something that you would not want your information technology to depend upon. >> Got it. >> Brent, we've seen obviously, it seems like every customer these days, regardless of industry has a cloud first initiative. They have competitors in the rear view mirror who are, if they're able to be more agile and faster to market, are potential huge competitive threat. As we see the rise of multi-cloud in the last 12 months, there's also been a lot of increased analyst coverage for alternate specialty hybrid cloud. Talk to us about, Expedient was in the recent Gartner market guide for specialty cloud. How are these related? What's driving this constant change out in the customer marketplace? >> Sure. So a lot of that agility that clients are getting and trying to do that digital transformation or refactor their applications requires a lot of effort from the developers and the internal IT practitioners. So by moving to a model with an enterprise kind of like Expedient, that allows them to get a consistent foundational level for those technical debt, the 'traditional workloads' where they can start focusing their efforts more on that refactoring of their applications, to get that agility, to get the flexibility, to get the market advantage of time to market with their new refactored applications. That takes them much faster to market, allows them to get ahead of those competitors, if they're not already ahead of them, get further ahead of them or catch up the ones that may have already made that transition. >> And I would add that the analyst coverage you've seen in the last 9 to 12 months, really accelerate for our type of cloud because before everything was hyper scale, everything's going to be hyper scale and they realized that companies have been trying to go to the cloud really for over a decade, really 15 years, that digital transformation, but most companies, when you look at the analysts say they're about 30% there, they've hit a plateau. So they need to look at a different way to approach that. And they're realizing that a VMware-based cloud or the specialty cloud providers give a different mode of cloud. Because you had of a pendulum that everything was on-premises, everything swung to cloud first and then it swung to multi-cloud, which meant multiple hyper scale providers and now it's really landing at that equilibrium where you have different modes of cloud. So it's similar like if you want to travel the world, you don't use one mode of transportation to get from one continent to the other. You have to use different modes. Same thing to get all the way to that cloud transformation, you need to use different modes of cloud, an enterprise cloud, a hyper scale cloud, working them together with that common management plan. >> And with that said Brian, where have customer conversations gone in the last couple of years? Obviously this has got to be an executive level, maybe even a board level conversation. Talk to us about how your customer conversations have changed. Have the stakeholders changed? Has things gone up to stack? >> Yeah. The business is much more involved than what it's been in the past and some of the drivers, even through the pandemic, as people reevaluate office space, a lot of times data centers were part of the same building. Or they were added into a review that nobody ever asked, "Well, why are you only using 20% of your data center?" So now that conversation is very active and they're reevaluating that and then the conversation shifts to "Where's the best place?" And that's a lot of, the conference also talks about the best place for your application for the workload in the right location. >> My role here is to dive down into the weeds constantly to stay away from business outcomes and things like that. But somewhere in the middle there's this question of how what you provide is consumed. So fair to assume that often people are moving from CapEx model to an OPEX model where they're consuming by the glass, by the drink. What does that mean organizationally for your customers? And do you help them work through that journey, reorganizing their internal organization to take advantage of cloud? Is that something that Expedient is a part of, or do you have partners that help them through that? How does that work? >> Yeah. There's some unique things that an enterprise doesn't understand when they think about what they've done on-prem versus a service provider is. There's whole models that they can purchase with us in consumption, not just the physical hardware, but licensing as well. Do you want to talk about how clients actually step in and start to do that evaluation? >> Sure. So it really kind of starts on the front end of evaluating what they have. So going through an assessment process, because traditionally, if you have a big data center full of hardware, you've already paid for it. So as you're deploying new workloads, it's "free to deploy." But when you go to that cloud operating model, you're paying for each drink that you're taking. So we want to make sure that as they're going into that cloud operating model, that they are right sized on the front end. They're not over-provisioned on anything that they're going to just waste money and resources on after they make that transition. So it's really about giving them great data on the front end, doing all that collection from a foundational level, from a infrastructure level, but also from a business and IT operations perspective and figuring out where they're spending, not just their money, but also their time and effort and helping them streamline and simplify those IT operations. >> Let's talk about one of the other elephants in the room and that is the remote hybrid workforce. Obviously it's been two and a half years, which is hard to believe. I think I'm one of the only people that hates working from home. Most people, do you too? Okay, good. Thank you, we're normal. >> Absolutely. (Lisa laughing) But VMware was talking about desktop as a service, there was so much change and quick temporary platform set up to accommodate offsite workers during the pandemic. What are some of the experiences that your clients are having and how is Expedient plus VMware helping businesses adapt and really create them the right hybrid model for them going forward? >> Sure. So as part of being that full sack cloud service provider, desktop in that remote user has to be part of that consideration. And one of the biggest things we saw with the pandemic was people stood up what we call pandemic VDI, very temporary solutions. And you saw the news articles that they said, "We did it in 10 days." And how many big transformational events do people plan and execute in 10 days that transform their workforce? So now they're having to come back and say, "Okay, what's the right way to deploy it?" And do you want to talk about some of the specifics of what we're seeing in the adjustments that they're doing? >> Sure. So it is, when you look at it from the end user perspective, it's how they're operating, how they're getting their tools through their day to day job, but it's also the IT administrators that are having to provide that service to the end users. So it's really kind of across the board, it's affecting everyone. So it's really kind of going through and helping them figure out how they're going to support their users going forward. So we've spun up things like VMware desktop as a service providing that multi-tenant ability to consume on a per desktop basis, but then we've also wrapped around with a lot of security features. So one of the big things is as people are going and distributing where they're working from, that data and access to data is also opened up to those locations. So putting those protections in place to be able to protect the environment and then be able, if something does get in, to be able to detect what's going on. And then of course, with a lot of the other components, being able to recover those environments. So building the desktops, the end user access into the disaster recovery plans. >> And talk more, a little bit Brent, about the security aspect. We've seen the threat landscape change dramatically in the last couple of years, ransomware is a household word. I'm pretty sure even my mom knows what that means, to some degree. Where is that in customer conversations? I can imagine in certain industries like financial services and healthcare with PII, it's absolutely critical to ensure that that data is, they know where it is. It's protected and it's recoverable, 'cause everyone's talking about cyber resilience these days. >> Right. And if it's not conversation 1, it's conversation 1A. So it's really kind of core to everything that we do when we're talking to clients. It's whether it's production DR or the desktops, is building that security in place to help them build their security practice up. So when you think about it, it's doing it at layers. So starting with things like more advanced antivirus to see what's actually going on the desktop and then kind of layering above there. So even up to micro-segmentation, where you can envelop each individual desktop in their own quasi network, so that they're only allowed kind of that zero trust model where, Hey, if you can get to a file share, that's the only place you should be going or do I need web apps to get my day to day job done, but really restricting that access and making sure that everything is more good traffic versus unknown traffic. >> Yeah. >> And also on the, you asked about the clouds smarter earlier. And you can really weave the desktop into that because when you're thinking of your production compute environment and your remote desktop environment, and now you can actually share storage together, you can share security together and you start to get economies of scale across those different environments as well. >> So as we are in August, I think still yeah, 2022, barely for a couple more days, lot of change going on at VMware. Expedient has been VMware America's partner of the year before. Talk to us about some of the things that you think from a strategic perspective are next for the partnership. >> That it's definitely the multi-cloud world is here. And it's how we can go deeper, how we're going to see that really mature. You know, one of the things that we've actually done together this year was we worked on a project and evaluated over 30 different companies of what they spend on IT. Everything from the physical data center to the entire stack, to people and actually build a cloud transformation calculator that allows you to compare strategies, so that if you look at Strategy A over a five year period, doing your current transformation, versus that Cloud Different approach, it can actually help quantify the number of hours difference that you can get, the total cost of ownership and the speed that you can get there. So it's things like that that help people make easier decisions and simplify information are going to be part of it. But without a doubt, it's going to be how you can have that wrapper across all of your different environments that really delivers that cloud-like environment that panacea people have been looking for. >> Yeah. That panacea, that seems like it's critical for every organization to achieve. Last question for you. When customers come to you, when they've hit that plateau. They come to Expedient saying, "Guys, with VMware, help us accelerate past this. "We don't have the time, we need to get this done quickly." How do you advise them to move forward? >> Sure. So it goes back to that, what's causing them to hit that plateau? Is it more on the development side of things? Is it the infrastructure teams, not being able to respond fast enough to the developers? And really putting a plan in place to really get rid of those plateaus. It could be getting rid of the technical debt. It could be changing the IT operations and kind of that, the way that they're looking at a cloud transformation model, to help them kind of get accelerated and get them back on the right path. >> Back on the right path. I think we all want to get back on the right path. Guys, thank you so much for joining David and me on theCUBE today, talking about Expedient Cloud Different, what you're seeing in the marketplace, and how Expedient and VMware are helping customers to succeed. We appreciate your time. >> Yep. >> Thanks for having us. >> For our guests and Dave Nicholson, I'm Lisa Martin. You're watching theCUBE live from VMware Explorer '22, stick around, Dave and I will be back shortly with our next guest. (gentle upbeat music)
SUMMARY :
And Brent Meadows, the Vice President the land of what's going on to get the end result they're looking for. and what it's helping customers to achieve and instead look to in the last couple of years and kind of removing that to get to cloud smart. so that you can get the value out of cloud kind of nudging people to adopt frameworks or that container into the platform and allow them to really look more towards so that the business can that you can really impact the business. Why do customers come to and the way people go to cloud, Just maybe you're too I build those sets with my son, So you also have confidence I'm imagining the things that you would not want agile and faster to market, that allows them to get a and then it swung to multi-cloud, in the last couple of years? and some of the drivers, So fair to assume that and start to do that evaluation? that they're going to just and that is the remote hybrid workforce. What are some of the experiences And one of the biggest things that service to the end users. in the last couple of years, that's the only place you should be going and now you can actually that you think from a and the speed that you can get there. "We don't have the time, we of the technical debt. Back on the right path. with our next guest.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Brian | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Brent Meadows | PERSON | 0.99+ |
Brian Smith | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
August | DATE | 0.99+ |
20% | QUANTITY | 0.99+ |
Brent | PERSON | 0.99+ |
Expedient | ORGANIZATION | 0.99+ |
15 years | QUANTITY | 0.99+ |
99% | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Lego | ORGANIZATION | 0.99+ |
Bryan Smith | PERSON | 0.99+ |
two and a half years | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
four | QUANTITY | 0.99+ |
2022 | DATE | 0.99+ |
three years | QUANTITY | 0.99+ |
each drink | QUANTITY | 0.99+ |
VMware America | ORGANIZATION | 0.99+ |
10 days | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
first initiative | QUANTITY | 0.98+ |
six months | QUANTITY | 0.98+ |
Legos | ORGANIZATION | 0.98+ |
Gartner | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
pandemic | EVENT | 0.98+ |
VMworld | ORGANIZATION | 0.98+ |
today | DATE | 0.97+ |
over a decade | QUANTITY | 0.97+ |
about 30% | QUANTITY | 0.97+ |
this year | DATE | 0.96+ |
first thing | QUANTITY | 0.96+ |
over 30 different companies | QUANTITY | 0.96+ |
one continent | QUANTITY | 0.95+ |
single word | QUANTITY | 0.95+ |
Moscone West | LOCATION | 0.95+ |
first way | QUANTITY | 0.94+ |
Marvel | ORGANIZATION | 0.94+ |
each individual desktop | QUANTITY | 0.94+ |
Bob Evans Foods | ORGANIZATION | 0.94+ |
VMware | TITLE | 0.93+ |
The Center of the Multi | TITLE | 0.93+ |
this morning | DATE | 0.92+ |
Lisa | PERSON | 0.91+ |
last couple of years | DATE | 0.9+ |
VMware Explore 2022 | TITLE | 0.89+ |
12 months | QUANTITY | 0.89+ |
theCUBE | ORGANIZATION | 0.88+ |
last 12 months | DATE | 0.88+ |
Breaking Analysis: How Lake Houses aim to be the Modern Data Analytics Platform
from the cube studios in palo alto in boston bringing you data driven insights from the cube and etr this is breaking analysis with dave vellante earnings season has shown a conflicting mix of signals for software companies well virtually all firms are expressing caution over so-called macro headwinds we're talking about ukraine inflation interest rates europe fx headwinds supply chain just overall i.t spend mongodb along with a few other names appeared more sanguine thanks to a beat in the recent quarter and a cautious but upbeat outlook for the near term hello and welcome to this week's wikibon cube insights powered by etr in this breaking analysis ahead of mongodb world 2022 we drill into mongo's business and what etr survey data tells us in the context of overall demand and the patterns that we're seeing from other software companies and we're seeing some distinctly different results from major firms these days we'll talk more about [Â __Â ] in this session which beat eps by 30 cents in revenue by more than 18 million dollars salesforce had a great quarter and its diversified portfolio is paying off as seen by the stocks noticeable uptick post earnings uipath which had been really beaten down prior to this quarter it's brought in a new co-ceo and it's business is showing a nice rebound with a small three cent eps beat and a nearly 20 million dollar top line beat crowdstrike is showing strength as well meanwhile managements at microsoft workday and snowflake expressed greater caution about the macroeconomic climate and especially on investors minds his concern about consumption pricing models snowflake in particular which had a small top-line beat cited softness and effects from reduced consumption especially from certain consumer-facing customers which has analysts digging more deeply into the predictability of their models in fact barclays analyst ramo lenchow published an especially thoughtful piece on this topic concluding that [Â __Â ] was less susceptible to consumption headwinds than for example snowflake essentially for a few reasons one because atlas mongo's cloud managed service which is the consumption model comprises only about 60 percent of mongo's revenue second is the premise that [Â __Â ] is supporting core operational applications that can't be easily dialed down or turned off and three that snowflake customers it sounds like has a more concentrated customer base and due to that fact there's a preponderance of its revenue is consumption driven and would be more sensitive to swings in these consumption patterns now i'll say this first consumption pricing models are here to stay and the much preferred model for customers is consumption the appeal of consumption is i can actually dial down turn off if i need to and stop spending for a while which happened or at least happened to a certain extent this quarter for certain companies but to the point about [Â __Â ] supporting core applications i do believe that over time you're going to see the increased emergence of data products that will become core monetization drivers in snowflake along with other data platforms is going to feed those data products and services and become over time maybe less susceptible and less sensitive to these consumption patterns it'll always be there but i think increasingly it's going to be tied to operational revenue last two points here in this slide software evaluations have reverted to their historical mean which is a good thing in our view we've taken some air out of the bubble and returned to more normalized valuations was really predicted and looked forward to look we're still in a lousy market for stocks it's really a bear market for tech the market tends to be at least six months ahead of the economy and often not always but often is a good predictor we've had some tough compares relative to the pandemic days in tech and we'll be watching next quarter very closely because the macro headwinds have now been firmly inserted into the guidance of software companies okay let's have a look at how certain names have performed relative to a software index benchmark so far this year here's a year-to-date chart comparing microsoft salesforce [Â __Â ] and snowflake to the igv software heavy etf which is shown in the darker blue line which by the way it does not own the ctf does not own snowflake or [Â __Â ] you can see that these big super caps have fared pretty well whereas [Â __Â ] and especially snowflake those higher growth companies have been much more negatively impacted year to date from a stock price standpoint now let's move on let's take a financial snapshot of [Â __Â ] and put it next to snowflake so we can compare these two higher growth names what we've done here in this chart has taken the most recent quarters revenue and multiplied it by 4x to get a revenue run rate and we've parenthetically added a projection for the full year revenue [Â __Â ] as you see will do north of a billion dollars in revenue while snowflake will begin to approach three billion dollars 2.7 and run right through that that four quarter run rate that they just had last quarter and you can see snowflake is growing faster than [Â __Â ] at 85 percent this past quarter and we took now these most of these profit of these next profitability ratios off the current quarter with one exception both companies have high gross margins of course you'd expect that but as we've discussed not as high as some traditional software companies in part because of their cloud costs but also you know their maturity or lack thereof both [Â __Â ] and snowflake because they are in high growth mode have thin operating margins they spend nearly half or more than half of their revenue on growth that's the sg a line mostly the s the sales and marketing is really where they're spending money uh and and they're specialists so they spend a fair amount of their revenue on r d but maybe not as high as you might think but a pretty hefty percentage the free cash flow as a percentage of revenue line we calculated off the full year projections because there was a kind of an anomaly this quarter in the in the snowflake numbers and you can see snowflakes free cash flow uh which again was abnormally high this quarter is going to settle in around 16 this year versus mongo's six percent so strong focus by snowflake on free cash flow and its management snowflake is about four billion dollars in cash and marketable securities on its balance sheet with little or no debt whereas [Â __Â ] has about two billion dollars on its balance sheet with a little bit of longer term debt and you can see snowflakes market cap is about double that of mongos so you're paying for higher growth with snowflake you're paying for the slootman scarpelli execution engine the expectation there a stronger balance sheet etc but snowflake is well off its roughly 100 billion evaluation which it touched during the peak days of tech during the pandemic and just that as an aside [Â __Â ] has around 33 000 customers about five times the number of customers snowflake has so a bit of a different customer mix and concentration but both companies in our view have no lack of market in terms of tam okay now let's dig a little deeper into mongo's business and bring in some etr data this colorful chart shows the breakdown of mongo's net score net score is etr's proprietary methodology that measures the percent of customers in the etr survey that are adding the platform new that's the lime green at nine percent existing customers that are spending six percent or more on the platform that's the forest green at 37 spending flat that's the gray at 46 percent decreasing spend that's the pinkish at around 5 and churning that's only 3 that's the bright red for [Â __Â ] subtract the red from the greens and you net out to a 38 which is a very solid net score figure note this is a survey of 1500 or so organizations and it includes 150 mongodb customers which includes by the way 68 global 2000 customers and they show a spending velocity or a net score of 44 so notably higher among the larger clients and while it's a smaller sample only 27 emea's net score for [Â __Â ] is 33 now that's down from 60 last quarter note that [Â __Â ] cited softness in its european business on its earning calls so that aligns to the gtr data okay now let's plot [Â __Â ] relative to some other data platforms these don't all necessarily compete head to head with [Â __Â ] but they are in data and database platforms in the etr data set and that's what this chart shows it's an xy graph with net score or as we say spending momentum on the vertical axis and overlap or presence or pervasiveness in the data set on the horizontal axis see that red dotted line there at 40 that indicates an elevated level of spending anything above that is highly elevated we've highlighted [Â __Â ] in that red box which is very close to that 40 percent line it has a pretty strong presence on the x-axis right there with gcp snowflake as we've reported has come down to earth but still well elevated again that aligns with the earnings releases uh aws and microsoft they have many data platforms especially aws so their plot position reflects their broad portfolio massive size on the x-axis um that's the presence and and very impressive on the vertical axis so despite that size they have strong spending momentum and you can see the pack of others including cockroach small on the verdict on the horizontal but elevated on the vertical couch base is creeping up since its ipo redis maria db which was launched the day that oracle bought sun and and got my sequel and some legacy platforms including the leader in database oracle as well as ibm and teradata's both cloud and on-prem platforms now one interesting side note here is on mongo's earning call it clearly cited the advantages of its increasingly all-in-one approach relative to others that offer a portfolio of bespoke or what we some sometimes call horses for courses databases [Â __Â ] cited the advantages of its simplicity and lower costs as it adds more and more functionality this is an argument often made by oracle and they often target aws as the company with too many databases and of course [Â __Â ] makes that argument uh as well but they also make the argument that oracle they don't necessarily call them out but they talk about traditional relational databases of course they're talking about oracle and others they say that's more complex less flexible and less appealing to developers than is [Â __Â ] now oracle of course would retur we retort saying hey we now support a mongodb api so why go anywhere else we're the most robust and the best for mission critical but this gives credence to the fact that if oracle is trying to capture business by offering a [Â __Â ] api for example that [Â __Â ] must be doing something right okay let's look at why they buy [Â __Â ] here's an etr chart that addresses that question it's it's mongo's feature breadth is the number one reason lower cost or better roi is number two integrations and stack alignment is third and mongo's technology lead is fourth those four kind of stand out with notice on the right hand side security and vision much lower there in the right that doesn't necessarily mean that [Â __Â ] doesn't have good security and and good vision although it has been cited uh security concerns um and and so we keep an eye on that but look [Â __Â ] has a document database it's become a viable alternative to traditional relational databases meaning you have much more flexibility over your schema um and in fact you know it's kind of schema-less you can pretty much put anything into a document database uh developers seem to love it generally it's fair to say mongo's architecture would favor consistency over availability because it uses a single master architecture as a primary and you can create secondary nodes in the event of a primary failure but you got to think about that and how to architect availability into the platform and got to consider recovery more carefully now now no schema means it's not a tables and rows structure and you can again shove anything you want into the database but you got to think about how to optimize performance um on queries now [Â __Â ] has been hard at work evolving the platform from the early days when you go back and look at its roadmap it's been you know started as a document database purely it added graph processing time series it's made search you know much much easier and more fundamental it's added atlas that fully managed cloud database uh service which we said now comprises 60 of its revenue it's you know kubernetes integrations and kind of the modern microservices stack and dozens and dozens and dozens of other features mongo's done a really fine job we think of creating a leading database platform today that is loved by customers loved by developers and is highly functional and next week the cube will be at mongodb world and we'll be looking for some of these items that we're showing here and this this chart this always going to be main focus on developers [Â __Â ] prides itself on being a developer friendly platform we're going to look for new features especially around security and governance and simplification of configurations and cluster management [Â __Â ] is likely going to continue to advance its all-in-one appeal and add more capabilities that reduce the need to to spin up bespoke platforms and we would expect enhance enhancements to atlas further enhancements there is atlas really is the future you know maybe adding you know more cloud native features and integrations and perhaps simplified ways to migrate to the cloud to atlas and improve access to data sources generally making the lives of developers and data analysts easier that's going to be we think a big theme at the event so these are the main things that we'll be scoping out at the event so please stop by if you're in new york city new york city at mongodb world or tune in to thecube.net okay that's it for today thanks to my colleagues stephanie chan who helps research breaking analysis from time to time alex meyerson is on production as today is as is andrew frick sarah kenney steve conte conte anderson hill and the entire team in palo alto thank you kristen martin and cheryl knight helped get the word out and rob hof is our editor-in-chief over there at siliconangle remember all these episodes are available as podcasts wherever you listen just search breaking analysis podcast we do publish each week on wikibon.com and siliconangle.com want to reach me email me david.velante siliconangle.com or dm me at divalante or a comment on my linkedin post and please do check out etr.ai for the best survey data in the enterprise tech business this is dave vellante for the cube insights powered by etr thanks for watching see you next time [Music] you
SUMMARY :
into the platform and got to consider
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
nine percent | QUANTITY | 0.99+ |
30 cents | QUANTITY | 0.99+ |
six percent | QUANTITY | 0.99+ |
46 percent | QUANTITY | 0.99+ |
ramo lenchow | PERSON | 0.99+ |
new york | LOCATION | 0.99+ |
next week | DATE | 0.99+ |
thecube.net | OTHER | 0.99+ |
85 percent | QUANTITY | 0.99+ |
microsoft | ORGANIZATION | 0.99+ |
40 percent | QUANTITY | 0.99+ |
six percent | QUANTITY | 0.99+ |
cheryl | PERSON | 0.99+ |
andrew frick | PERSON | 0.99+ |
three billion dollars | QUANTITY | 0.99+ |
more than 18 million dollars | QUANTITY | 0.99+ |
dave vellante | PERSON | 0.99+ |
oracle | ORGANIZATION | 0.99+ |
this year | DATE | 0.99+ |
stephanie chan | PERSON | 0.99+ |
alex meyerson | PERSON | 0.99+ |
next quarter | DATE | 0.99+ |
37 | QUANTITY | 0.99+ |
44 | QUANTITY | 0.99+ |
last quarter | DATE | 0.99+ |
boston | LOCATION | 0.99+ |
60 | QUANTITY | 0.99+ |
both companies | QUANTITY | 0.99+ |
38 | QUANTITY | 0.99+ |
david.velante | OTHER | 0.99+ |
today | DATE | 0.99+ |
about two billion dollars | QUANTITY | 0.99+ |
dozens | QUANTITY | 0.98+ |
about four billion dollars | QUANTITY | 0.98+ |
rob hof | PERSON | 0.98+ |
33 | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
each week | QUANTITY | 0.98+ |
around 33 000 customers | QUANTITY | 0.98+ |
27 | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
4x | QUANTITY | 0.97+ |
150 mongodb customers | QUANTITY | 0.97+ |
three | QUANTITY | 0.97+ |
more than half | QUANTITY | 0.97+ |
fourth | QUANTITY | 0.96+ |
aws | ORGANIZATION | 0.96+ |
this week | DATE | 0.96+ |
both | QUANTITY | 0.96+ |
nearly 20 million dollar | QUANTITY | 0.96+ |
anderson hill | PERSON | 0.96+ |
2022 | DATE | 0.95+ |
palo alto | ORGANIZATION | 0.94+ |
mongo | ORGANIZATION | 0.94+ |
sarah kenney | PERSON | 0.94+ |
kristen martin | PERSON | 0.93+ |
about 60 percent | QUANTITY | 0.93+ |
one | QUANTITY | 0.93+ |
40 | QUANTITY | 0.93+ |
one exception | QUANTITY | 0.93+ |
2.7 | QUANTITY | 0.93+ |
third | QUANTITY | 0.93+ |
four | QUANTITY | 0.93+ |
atlas | TITLE | 0.92+ |
two higher growth | QUANTITY | 0.92+ |
about five times | QUANTITY | 0.92+ |
3 | QUANTITY | 0.91+ |
etr | ORGANIZATION | 0.91+ |
pandemic | EVENT | 0.91+ |
atlas mongo | ORGANIZATION | 0.91+ |
this quarter | DATE | 0.9+ |
ukraine | LOCATION | 0.9+ |
siliconangle.com | OTHER | 0.89+ |
2000 customers | QUANTITY | 0.88+ |
palo | LOCATION | 0.88+ |
around 5 | QUANTITY | 0.87+ |
Prakash Darji, Pure Storage | At Your Storage Service
(bright intro music) >> The cloud has popularized many useful concepts in the past decade, working backwards from the customer to pizza teams and DevOps mindset, the shared responsibility model, and security, of course, the shift from CapEx to OPEX, and as a service consumption models. The last item is what we're here to talk about today. Pay for consumption is attractive because you're not over provisioning, at least not the way you used to. You'd have to buy for peak capacity events, but there are always two sides to every story, and while pay for use more closely ties IT consumption to business value, procurement teams don't always love the uncertainty of the cloud bill each month, but consumption pricing and as a service models are here to stay in software and hardware. Hello, I'm Dave Vellante and welcome to "At Your Storage Service" made possible by Pure Storage, and with me is Prakash Darji who's the General Manager of the Digital Experience Business Unit at Pure. Prakash, welcome to the program. >> Thanks Dave. Thanks for having me. >> You bet. Okay, we've seen this shift to as a service, the as a service economy, subscription models, and this as a service movement have gained real momentum. It's clear over the past several years. What's driving this shift? Is it pressure from investors and technology companies that are chasing the all important ARR, their annual recurring revenue stream? Is it customer driven? Give us your insights. >> Well, look, I think we'll do some definitional stuff first. I think we often mix the definition of a subscription and a service, but, you know, subscription is, "Hey, I can go for a pay up front or pay as I go." Service is more about, "How do I not buy something just by the outcome?" So, you know, the concept of delivering storage as a service means, what do you want in storage, performance, capacity, availability? Like that's what you want. Well, how do you get that without having to worry about the labor of planning, capacity management? Those labor elements are what's driving it. So I think in the world where you have to do more with less and in a world where security becomes increasingly important where standardization will allow you to secure your landscape against ransomware and those types of things, those trends are driving the SaaSification of storage, and the only way to deliver that is storage as a service. >> So that's good. You maybe thinking about it differently than some of the other companies that I talked to, but so you've made inroads here, pretty big inroads actually, and changed the thinking in enterprise data storage with a huge emphasis on simplicity. That's really Pure's raison d'etre. How does storage as a service fit in to your innovation agenda overall? >> Well, our innovation agenda started, as you mentioned, with the simplicity, you know, a decade ago with the Evergreen Architecture. That architecture was beyond the box. How do you go ahead and say, "I can improve performance or capacity as I need it." Well, that's a foundational element to deliver a service because once you have that technology, you can say, "Oh, you know what? You've subscribed to this performance level. You want to raise your performance level and yes, that'll be a higher dollar per gig or dollar per terabyte," but how do you do that without a data migration? How do you do that with a non-disruptive service change? How do you do that with a delivery via software update? Those elements of non-disruptive updates, when you think SaaS, Salesforce, you don't know when Salesforce doesn't update. You don't know when they're increasing something, adding a new capability. It just shows up. It's not a disruptive event. So to drive that standardization and SaaSification in service delivery, you need to keep that simplicity of delivery first and foremost, and you can't allow, like, if the goal was, "I want to change from this service here to that service here," and a person needed to show up and do a day data migration, that's kind of useless. You've broken the experience of flexibility for a customer. >> Okay, so I like the Salesforce analogy, but I want to jump out do a little side for a second. So I've got to make some commitment to Pure, right? Some baseline commitment, and if I do, then I can dial up and then pay for what I use, and I can dial it down, correct? >> Correct. >> Okay. I can't do that with Salesforce, all right? I could dial up, but then I'm stuck with those licenses. So you have a better model in Salesforce, I would argue. Okay. >> Yeah. I would agree with that. >> Okay, so, and I got to pay for everything up front. Anyway, let's go back to I was kind of pushing at you a little bit at my upfront, you know, about, you know, the ARR model, the all-important, you know, financial metric, but let's talk from the customer standpoint. What are the benefits of consuming storage as a service from your customer's perspective? >> Well, one is when you start your storage journey, do you really know what you need? And I would argue, most of the time, people are guessing, right? It's like, "Well, I think I need this. This is the performance I think I need or this is the capacity I think I need," and, you know, with the scientific method, you actually deploy something, and you're like, "Do I need more? Do I need less?" You find out as you're deploying. So in a storage as a service world, when you have the ability to move up performance levels or move out capacity levels, and you have that flexibility, then you have the ability to just to meet demand as you deploy, and that's the most important element of meeting business needs today. The applications you deploy are not in your control when you're providing storage to your end consumers. >> Yeah. >> They're going to want different levels of storage. They're going to want different performance thresholds. That's kind of a pay, you know, pay for performance type culture, right? You can use HR analogy for it. You pay for performance. You want top talent, you pay for it. You want top storage performance, you pay for it. You don't, you can pay less, and you can actually get lower performance tiers. Not everything is a tier one application, and you need the ability to deploy it, but when you start, how do you know the way your end customers are going to be consuming or do you need a dictated upfront? 'Cause that's infrastructure dictating business inflexibility, and you never want to be in that position. >> I got another analogy for you. It's like, you know, we do a lot of hosting at our home and you know, like Thanksgiving, right? And you go to the liquor store and say, "Okay, what should I get, should we get red wine? We got to go white wine, we got to get some beer. Should I get bubbles? Yeah, I get some bubbles." 'Cause you don't know what people are going to have, and so you over provision everything and then there's a run on bubbles and you're like, "Ah, we're running the bubbles," so you just over buy, but there's a liquor store that actually will take it back. So I got to do business with those guys every time 'cause it's way more flexible. I can dial up capacity or I can dial up performance, and dial it back down if I don't use it. >> Where you're going to be drinking a lot more the next few weeks. >> Yeah, exactly, like which is the last thing you want. Okay, so let's talk about how Pure kind of meets this as a service demand. You've touched upon your differentiators from others in the market. You know, love to hear about the momentum. What are you seeing out there? >> Yeah, look, our business is growing well largely built on, you know, what customers need. Specifically, where the market is at today is there's a set of folks that are interested in the financial transformation of CapEx to OPEX. Like that definitely exists in the industry around, "How do I get a paper use model?" The next kind of more advanced customer is interested in, "How do I go ahead and remove labor to deliver storage?" And a service gets you there on top of a subscription. The most sophisticated customer says, "How do I separate storage production with consumption and production of storage?" Being a storage producer should be about standardization so I could do policy based management. Why is that important? You know, coming back to some of the things I said earlier, in the world where ransomware attacks are common, you need the standardized security policies. Linux has new vulnerabilities every other day, like find two, three critical vulnerabilities a week. How do you stay on top of it? The complexity of staying on top of it should be, "Look, let's standardize and make it a vendor problem, and assume the vendor's going to deliver this to me." So that standardization allows you to have business policies that allow you to stay current and modern. I would argue in, you know, the traditional storage and appliance world, you buy something and the day after you buy it, it's worthless. It's like driving a car off a lot, right? The very next day, the car's not worth what it was when you bought it. Storage is the same way. So how do you ensure that your storage stays current? How do you ensure that it gets a like a fine line that gets better with age? Well, if you're not buying storage and you're buying a performance SLA, it's up to the vendor to meet that SLA so it actually never gets worse over time. This is the way you modernize technology and avoid technology debt as a customer. >> Yeah, I mean, just even though words you're using and the way you're thinking about this precaution, I think are different, and I love the concept of essentially taking my labor cost and transferring them to Pure's R&D. I mean, that's essentially what you're talking about here. So let's stick with the tech for a minute. What do you see as new or emerging technologies that are helping accelerate this shift toward the as a service economy? >> Well, the first thing is I always tell people you can't deliver a service without monitoring because if you can't monitor something, how you're going to know whether you're meeting your service level obligation, right? So everything starts with data monitoring. The next step layering on the technology differentiation is if you need to deliver a service level obligation on top of that data monitoring, you need the ability to flexibly meet whatever performance obligations you have in a tight time window. So supply chain and being able to deliver anywhere becomes important. So if you use the analogy today of how Tesla works or a IoT system works, you have a SaaS management that actually provides instructions that pushes those instructions and policies to the edge. In Tesla's case, that happens to be the car. It'll push software updates to the car. It'll push new map updates to the car, but the car is running independently. It's not like if the car becomes disconnected from the internet, it's going to crash and drive you off the road. In the same way, what if you think about storage as something that needs to be wherever your application is? So people think about cloud as a destination. I think that's a fallacy. You have to think about the world in the view of an application. An application needs data, and that data needs to sit in storage wherever that application sits. So for us, the storage system is just an edge device. It can be sitting in your data center it can be sitting in a Equinix. It can be sitting in hosted and MSP can run it. It can even be sitting in the public cloud, but how do you have central monitoring and central management where you can push policies to update all those devices, very similar to an IOT system? So the technology advantage of doing that means that you can operate anywhere and ensure you have a consistent set of policies, a consistent set of protection, a consistent set of, you know, prevention against ransomware attack regardless of your application, regardless of, you know, where it sits, regardless of what content in it you're on. That approach is very similar to the way the IoT industry has been updating and monitoring edge devices, nest thermostats, you know, Tesla cars, those types of things. That's the thinking that needs to come to storage, and that's the foundation on which we built Pure as a service. >> So that implies or, at least I infer, that you've, obviously, got control of the experience on prem, but you're extending that into AWS, Google, Azure, which suggests to me that you have to hide the underlying complexity of the primitives and APIs in that world, and then eventually, actually today, 'cause you're treating everything like the edge out to the edge, you know, maybe mini Pure at some point in time, but so I call that super cloud, that abstraction layer that floats above all the clouds on-prem and adds that layer of value and is a singular experience, what you're talking about pushing, you know, policy throughout. Is that the right way to think about it? And how does this impact the ability to deliver true storage as a service? >> Oh, that's absolutely the right way of thinking about it. The things that you think about from an abstraction kind of fall in three buckets. First, you need management. So how do you ensure consistent management experience, creating volumes, deleting volumes, creating buckets, creating files, creating directories like management of objects and create a consistent API across the entire landscape? The second one is monitoring. How do you measure utilization and performance obligations or capacity obligations or, you know, policy violations, wherever you're at? And then the third one is more of a business one, which is procurement because you can't do it independent of procurement, meaning what happens when you run out? Do you need to increase your reserve commits? Do you want to go on demand? How do you integrate it into company's procurement models such that you can say, "I can use what I need," and any, it's not like every change order is a request of procurement. That's going to break an as a service delivery model. So to get embedded in a customer's landscape where they don't have to worry about storage, you have to provide that consistency on management, monitoring, and procurement across the tech, and yes, this is deep technology problems, whether it's running our storage on AWS or Azure or running it on prem or, you know, at some point in the future, maybe even, you know, Pure mini at the edge, right? So, you know, all of those things are tied to our Pure as a service delivery. >> Yeah, technically, non-trivial, but hey, you guys are on it. Well, we got to leave it there, Prakash. Thank you, great stuff, really appreciate your time. >> All right, thanks for having me, man. >> You're very welcome. Okay, in a moment, Steve McDowell. For more insights and strategies, he's going to give us the analyst perspective on as a service. You're watching "theCUBE", the leader in high tech enterprise coverage. (bright outro music)
SUMMARY :
at least not the way you used to. Thanks for having me. that are chasing the all important ARR, So, you know, the concept and changed the thinking and you can't allow, So I've got to make some So you have a better model I would agree with that. the all-important, you and you have that flexibility, and you need the ability to deploy it, and you know, like Thanksgiving, right? a lot more the next few weeks. like which is the last thing you want. This is the way you modernize technology and the way you're thinking and ensure you have a out to the edge, you know, such that you can say, but hey, you guys are on it. the leader in high tech
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Steve McDowell | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Tesla | ORGANIZATION | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
OPEX | ORGANIZATION | 0.99+ |
Prakash Darji | PERSON | 0.99+ |
two sides | QUANTITY | 0.99+ |
Prakash | PERSON | 0.99+ |
First | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
AWS | ORGANIZATION | 0.99+ |
CapEx | ORGANIZATION | 0.99+ |
second one | QUANTITY | 0.99+ |
first thing | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
third one | QUANTITY | 0.98+ |
each month | QUANTITY | 0.97+ |
Thanksgiving | EVENT | 0.97+ |
Equinix | ORGANIZATION | 0.96+ |
a decade ago | DATE | 0.95+ |
Pure mini | COMMERCIAL_ITEM | 0.94+ |
three critical vulnerabilities | QUANTITY | 0.93+ |
Salesforce | ORGANIZATION | 0.93+ |
next day | DATE | 0.91+ |
past decade | DATE | 0.91+ |
Linux | TITLE | 0.89+ |
a week | QUANTITY | 0.88+ |
first | QUANTITY | 0.85+ |
Pure. Prakash | ORGANIZATION | 0.84+ |
Azure | ORGANIZATION | 0.83+ |
Evergreen Architecture | ORGANIZATION | 0.79+ |
Azure | TITLE | 0.78+ |
Pure | COMMERCIAL_ITEM | 0.77+ |
next few weeks | DATE | 0.74+ |
years | DATE | 0.72+ |
one | QUANTITY | 0.69+ |
past | DATE | 0.65+ |
Pure | ORGANIZATION | 0.61+ |
a day | QUANTITY | 0.61+ |
SaaS | TITLE | 0.55+ |
Salesforce | TITLE | 0.51+ |
second | QUANTITY | 0.48+ |
theCUBE | TITLE | 0.45+ |
Robert Belson, Verizon | Red Hat Summit 2022
>> Welcome back to the Seaport in Boston and this is theCUBE's coverage of Red Hat Summit 2022. I'm Dave Vellante with my co-host Paul Gillin. Rob Belson is here as the Developer Relations Lead at Verizon. Robbie great to see you. Thanks for coming on theCUBE. >> Thanks for having me. >> So Verizon and developer relations. Talk about your role there. Really interesting. >> Absolutely. If you think about our mobile edge computing portfolio in Verizon 5G Edge, suddenly the developer is a more important persona than ever for actually adopting the cloud itself and adopting the mobile edge. So the question then quickly became how do we go after developers and how do we tell stories that ultimately resonate with them? And so my role has been spearheading our developer relations and experience efforts, which is all about meeting developers in the channels where they actually are, building content that resonates with them. Building out architectures that showcase how do you actually use the technology in the wild? And then ultimately creating automation assets that make their lives easier in deploying to the mobile edge. >> So, you know, telcos get a bad rap, when you're thinking it's amazing what you guys do. You put out all this capital infrastructure, big outlays. You know, we use our phones to drop a call. People like, "Ah, freaking Verizon." But it's amazing what we can actually do too. You think about the pandemic, the shift that the telcos had to go through to landlines to support home, never missed a beat. And yet at the same time you're providing all this infrastructure for people to come over the top, the cost forbid is going down, right? Your cost are going up and yet now we're doing this big 5G buildup. So I feel like there's a renaissance about to occur in edge computing that the telcos are going to lead new forms of monetization new value that you're going to be able to add, new services, new applications. The future's got to be exciting for you guys and it's going to be developer-led, isn't it? >> Absolutely. I mean it's been such an exciting time to be a part of our mobile edge computing portfolio. If you think back to late 2019 we were really asking the question with the advent of high speed 5G mobile networks, how can you drive more immersive experiences from the cloud in a cloud native way without compromising on the tools you know and love? And that's ultimately what caused us to really work with the likes of AWS and others to think about what does a mobile edge computing portfolio look like? So we started with 5G Edge with AWS Wavelength. So taking the compute and storage services you know and love in AWS and bringing it to the edge of our 4G and 5G networks. But then we start to think, well, wait a minute. Why stop at public networks? Let's think about private networks. How can we bring the cloud and private networks together? So you turn back to late 2021 we announced Verizon 5G Edge with AWS Outposts but we didn't even stop there. We said, "Well, interest's cool, but what about network APIs? We've been talking about the ability and the programmability of the 5G network but what does that actually look like to the developers? And one great example is our Edge Discovery Service. So you think about the proliferation of the edge 17 Wavelength Zones today in the US. Well, what edge is the right edge? You think about maybe the airline industry if the closest exit might be behind you absolutely applies to service discovery. So we've built an API that helps answer that seemingly basic question but is the fundamental building block for everything to workload orchestration, workload distribution. A basic network building block has become so important to some of these new sources of revenue streams, as we mentioned, but also the ability to disintermediate that purpose built hardware. You think about the future of autonomous mobile robots either ground and aerial robotics. Well, you want to make those devices as cheap as possible but you don't want to compromise on performance. And that mobile edge layer is going to become so critical for that connectivity, but also the compute itself. >> So I just kind of gave my little narrative up front about telco, but that purpose built hardware that you're talking about is exceedingly reliable. I mean, it's hardened, it's fossilized and so now as you just disaggregate that and go to a more programmable infrastructure, how are you able to and what gives you confidence that you're going to be able to maintain that reliability that I joke about? Oh, but it's so reliable. The network has amazing reliability. How are you able to maintain that? Is that just the pace of technology is now caught up, I wonder if you can explain that? >> I think it's really cool as I see reliability and sort of geo distribution as inextricably linked. So in a world where to get that best in class latency you needed to go to one place and one place only. Well, now you're creating some form of single source of failure whether it's the power, whether it's the compute itself, whether it's the networking, but with a more geo distributed footprint, particularly in the mobile edge more choices for where to deliver that immersive experience you're naturally driving an increase in reliability. But again, infra alone it's not going to do the job. You need the network APIs. So it's the convergence of the cloud and network and infra and the automation behind it that's been incredibly powerful. And as a great example, the work we've been doing in hybrid MEC the ability to converge within one single architecture, the private network, the public network, the AWS Outposts, the AWS Wavelength all in one has been such a fantastic journey and Red Hat has been a really important part in that journey. >> From the perspective of the developer when they're building a full cloud to edge application, where does Verizon pick up? Where do they start working primarily with you versus with their cloud provider? >> Absolutely. And I think you touched on a really important point. I think when you often think about the edge it's thought of as an either, or. Is it the edge? Is it the cloud? Is it both? It's an and I can't emphasize that enough. What we've seen from the customers greenfield or otherwise it's about extending an application or services that were never intended to live at the edge, to the edge itself, to deliver a more performant experience. And for certain control plane operations, metadata, backend operations analytics that can absolutely stay in the cloud itself. And so our role is to be a trusted partner in some of our enterprise customers' journeys. Of course, they can lean on the cloud provider in select cases, but we're an absolutely critical mode of support as you think about what are those architectures? How do you integrate the network APIs? And through our developer relations efforts, we've put a major role in helping to shape what those patterns really look like in the wild. >> When they're developing for 5G I mean, the availability of 5G of particularly you know, the high bandwidth 5G is pretty spotty right now. Mostly urban areas. How should they be thinking in the future developing an application roll out two years from now about where 5G will be at that point? >> Absolutely. I think one of the most important things in this case is the interoperability of our edge computing portfolio with both 4G and 5G. Whenever somebody asks me about the performance of 5G they ask how fast? Or for edge computing. It's always about benchmark. It's not an absolute value. It's always about benchmarking the performance to that next best alternative. What were you going to get if you didn't have edge computing in your back pocket? And so along that line of thought having the option to go either through 4G or 5G, having a mobile edge computing portfolio that works for both modes of connectivity even CAN-AM IoT is incredibly powerful. >> So it sounds like 4G is going to be with us for quite a while still? >> And I think it's an important part of the architecture. >> Yeah. >> Robert, tell us about the developer that's building these applications. Where does that individual come from? What's their persona? >> Oh, boy I think there's a number of different personas and flavors. I've seen everything from the startup in the back of a garage working hard to try to figure out what could I do for a next generation media and entertainment experience but also large enterprises. And I think a great area where we saw this was our 5G Edge Computing Challenge that we hosted last year. Believe it or not 100 submissions from over 22 countries, all building on Verizon 5G Edge. It was so exciting to see because so many different use cases across public safety, healthcare, media and entertainment. And what we found was that education is so important. A lot of developers have great ideas but if you don't understand the fundamentals of the infrastructure you get bogged down in networking and setting up your environment. And that's why we think that developer education is so important. We want to make it easy and in fact, the 5G Edge portfolio was designed in such a way that we'll abstract the complexities of the network away so you can focus on building your application and that's such a central theme and focus for how we approach the development. >> So what kind of services are you exposing via APIs? >> Absolutely, so first and foremost, as you think about 5G Edge with say AWS Wavelength, the infra there are APIs that are exposed by AWS to launch your infra, to patch your infrastructure, to automate your infrastructure. Specifically that Verizon has developed that's our network APIs. And a great example is our Edge Discovery Service. So think of this as like a service registry you've launched an application in all 17 edge zones. You would take that information, you would send it via API to the Edge Discovery Service so that for any mobile client say, you wake up one morning in Boston, you can ask the API or query, "Hey, what's the closest edge zone?" DNS isn't going to be able to figure it out. You need knowledge of the actual topology of the mobile network itself. So the API will answer. Let's say you take a little road trip 1,000 miles south to say Miami, Florida you ask that question again. It could change. So that's the workflow and how you would use the network API today. >> How'd you get into this? You're an engineer it's obvious how'd you stumble into this role? >> Well, yeah, I have a background in networks and distributed systems so I always knew I wanted to stay in the cloud somewhere. And there was a really unique opportunity at Verizon as the portfolio was being developed to really think about what this developer community looked like. And we built this all from scratch. If you look at say our Verizon 5G Edge Blog we launched it just along the timing of the actual GA of Wavelength. You look at our developer newsletter also around the time of the launch of Wavelength. So we've done a lot in such a short period and it's all been sort of organic, interacting with developers, working backwards from the customer. And so it's been a fairly new, but incredibly exciting journey. >> How will your data, architecture, data flow what will that look like in the future? How will that be different than it is sort of historically? >> When I think about customer workloads real time data architecture is an incredibly difficult thing to do. When you overlay the edge, admittedly, it gets more complicated. More places that produce the data, more places that consume data. How do you reconcile all of these environments? Maintain consistency? This is absolutely something we've been working on with the ecosystem at large. We're not going to solve this alone. We've looked at architecture patterns that we think are successful. And some of the things that we found that we believe are pretty cool this idea of taking that embedded mobile database, virtualizing it to the edge, even making it multi-tenant. And then you're producing data to one single source and simplifying how you organize and share data because all of the data being produced to that one location will be relevant to that topology. So Boston, as an example, Boston data being produced to that edge zone will only service Boston clients. So having a geo distributed footprint really does help data architectures, but at the core of all of this database, architectures, you need a compute environment that actually makes sense. That's performant, that's reliable. That's easy to use that you understand how to manage and that the edge doesn't make it any more difficult to manage. >> So are you building that? >> That's exactly what we're doing. So here at Red Hat Summit we've had the unique opportunity to continue to collaborate with our partners at Red Hat to think about how you actually use OpenShift in the context of hybrid MEC. So what have done is we've used OpenShift as is to extend what already exists to some of these new edge zones without adding in an additional layer of complexity that was unmanageable. >> So you use OpenShift so you don't have to cobble this together on your own as a full development environment and that's the role really that OpenShift plays here? >> That's exactly right. And we presented pieces of this at our re:Invent this past year and what we basically did is we said the edge needs to be inextricably linked with the cloud. And you want to be able to manage it from some seamless central pane of glass and using that OpenShift console is a great way. So what we did is we wanted to show a really geo-distributed footprint in action. We started with a Wavelength zone in Boston, the region in Northern Virginia, an outpost in the Texas area. We cobbled it all together in one cluster. So you had a whole compute mesh separated by thousands of miles all within a single cluster, single pane of glass. We take that and are starting to expand on the complexity of these architectures to overlay the network APIs we mentioned, to overlay multi-region support. So when we say you can use all 17 zones at once you actually can. >> So you've been talking about Wavelength and Outposts which are AWS products, but Microsoft and Google both have their distributed architectures as well. Where do you stand with those? Will you support those? Are you working with them? >> That's a great question. We have made announcements with Microsoft and Google but today I focus a lot on the work we do with AWS Wavelength and Outposts and continuing to work backwards from the customer and ultimately meet their needs. >> Yeah I mean, you got to start with an environment that the developers know that obviously a great developer community, you know, you see it at re:Invent. What was the reaction at re:Invent when you showed this from a developer community? >> Absolutely. Developers are excited and I think the best part is we're not the only ones talking about Wavelength not even AWS are the only ones talking about Wavelength. And to me from a developer ecosystem perspective that's when you know it's working. When you're not the one telling the best stories when others are evangelizing the power of your technology on your behalf that's when the ecosystem's starting to pick up. >> Speaking of making a bet on Outposts you know, it's somewhat limited today. I'll say it it's limited today in terms of we think it supports RDS and there's a few storage players. Is it your expectation that Outposts is going to be this essentially the cloud environment on your premises is that? >> That's a great question. I see it more as we want to expand customer choice more than ever and ultimately let the developers and architects decide. That's why I'm so bullish on this idea of hybrid MEC. Let's provide all of the options the most complicated geo distributed hybrid deployment you can imagine and automate it, make it easy. That way if you want to take away components of this architecture all you're doing is simplifying something that's already automated and fairly simple to begin with. So start with the largest problem to solve and then provide customers choice for what exactly meets their requirements their SLAs, their footprint, their network and work backwards from the customer. >> Exciting times ahead. Rob, thanks so much for coming on theCUBE. It's great to have you. >> Appreciate it, thanks for your time. >> Good luck. All right, thank you for watching. Keep it right there. This is Dave Vellante for Paul Gillin. We're live at Red Hat Summit 2022 from the Seaport in Boston. We'll be right back.
SUMMARY :
as the Developer So Verizon and developer relations. and adopting the mobile edge. that the telcos are going to if the closest exit might be behind you Is that just the pace of in hybrid MEC the ability to converge And I think you touched on I mean, the availability having the option to go part of the architecture. Where does that individual come from? of the infrastructure you get bogged down So that's the workflow of the actual GA of Wavelength. and that the edge doesn't make it any more to think about how you We take that and are starting to expand Where do you stand with those? and continuing to work that the developers know that's when you know it's working. Outposts is going to be and fairly simple to begin with. It's great to have you. from the Seaport in Boston.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul Gillin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Verizon | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Microsoft | ORGANIZATION | 0.99+ |
Rob Belson | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Robert Belson | PERSON | 0.99+ |
Robert | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
Rob | PERSON | 0.99+ |
US | LOCATION | 0.99+ |
Texas | LOCATION | 0.99+ |
100 submissions | QUANTITY | 0.99+ |
telco | ORGANIZATION | 0.99+ |
Seaport | LOCATION | 0.99+ |
late 2019 | DATE | 0.99+ |
OpenShift | TITLE | 0.99+ |
late 2021 | DATE | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Northern Virginia | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
1,000 miles | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
thousands of miles | QUANTITY | 0.99+ |
Red Hat Summit 2022 | EVENT | 0.99+ |
pandemic | EVENT | 0.98+ |
one | QUANTITY | 0.98+ |
one cluster | QUANTITY | 0.98+ |
Robbie | PERSON | 0.98+ |
one place | QUANTITY | 0.98+ |
17 zones | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
one morning | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
single cluster | QUANTITY | 0.97+ |
both modes | QUANTITY | 0.97+ |
over 22 countries | QUANTITY | 0.96+ |
one place | QUANTITY | 0.96+ |
theCUBE | ORGANIZATION | 0.93+ |
Red Hat Summit | EVENT | 0.92+ |
telcos | ORGANIZATION | 0.92+ |
Edge Discovery Service | TITLE | 0.91+ |
5G Edge Computing Challenge | EVENT | 0.91+ |
Wavelength | TITLE | 0.89+ |
one location | QUANTITY | 0.89+ |
single pane | QUANTITY | 0.88+ |
one single source | QUANTITY | 0.85+ |
two years | QUANTITY | 0.83+ |
one single architecture | QUANTITY | 0.83+ |
Miami, Florida | LOCATION | 0.82+ |
single source | QUANTITY | 0.81+ |
Outposts | ORGANIZATION | 0.8+ |
past year | DATE | 0.78+ |
G37 Paul Duffy
(bright upbeat music) >> Okay, welcome back everyone to the live CUBE coverage here in Las Vegas for in-person AWS re:Invent 2021. I'm John Furrier host of theCUBE two sets, live wall to wall coverage, all scopes of the hybrid events. Well, great stuff online. That was too much information to consume, but ultimately as usual, great show of new innovation for startups and for large enterprises. We've got a great guest, Paul Duffy head of startups Solutions Architecture for North America for Amazon Web Services. Paul, thanks for coming on. Appreciate it. >> Hi John, good to be here. >> So we saw you last night, we were chatting kind of about the show in general, but also about start ups. Everyone knows I'm a big startup fan and big founder myself, and we talk, I'm pro startups, everyone loves startups. Amazon, the first real customers were developers doing startups. And we know the big unicorns out there now all started on AWS. So Amazon was like a dream for the startup because before Amazon, you had to provision the server, you put in the Colo, you need a system administrator, welcome to EC2. Goodness is there, the rest is history. >> Yeah. >> The legacy and the startups is pretty deep. >> Yeah, you made the right point. I've done it myself. I co-founded a startup in about 2007, 2008. And before we even knew whether we had any kind of product market fit, we were racking the servers and doing all that kind of stuff. So yeah, completely changed it. >> And it's hard too with the new technology now finding someone to actually, I remember when we stood with our first Hadoop and we ran a solar search engine. I couldn't even find anyone to manage it. Because if you knew Hadoop back then, you were working at Facebook or Hyperscaler. So you guys have all this technology coming out, so provisioning and doing the heavy lifting for start is a huge win. That's kind of known, everyone knows that. So that's cool. What are you guys doing now because now you've got large enterprises trying to beat like startups. You got startups coming in with huge white spaces out there in the market. Jerry Chen from Greylock, and it was only yesterday we talked extensively about the net new opportunities in the Cloud that are out there. And now you see companies like Goldman Sachs have super cloud. So there's tons of growth. >> Paul: Yeah. >> Take us through the white space. How do you guys see startups taking advantage of AWS to a whole another level. >> And I think it's very interesting when you look at how things have changed in those kind of 15 years. The old world's horrible, you had to do all this provisioning. And then with AWS, Adam Szalecki was talking in his keynote on the first day of the event where people used to think it was just good for startups. Now for startups, it was this kind of obvious thing because they didn't have any legacy, they didn't have any data centers, they didn't have necessarily a large team and be able to do this thing with no commitment. Spin up a server with an API call was really the revolutionary thing. In that time, 15 years later, startups still have the same kind of urgency. They're constrained by time, they're constrained by money, they're constrained by the engineering talent they have. When you hear some of the announcements this week, or you look what is kind of the building blocks available to those startups. That I think is where it's become revolutionary. So you take a startup in 2011, 2012, and they were trying to build something maybe they were trying to do image recognition on forms for example, and they could build that. But they had to build the whole thing in the cloud. We had infrastructure, we had database stuff, but they would have to do all of the kind of the stuff on top of that. Now you look at some of the kind of the AIML services we have things like Textract, and they could just take that service off the shelf. We've got one startup in Canada called Chisel AI. They're trying to disrupt the insurance industry, and they could just use these services like text extracts to just accelerate them getting into that product market fit instead of having to do this undifferentiated (indistinct). >> Paul, we talk about, I remember back in the day when Web Services and service oriented architecture, building blocks, decoupling APIs, all that's now so real and so excellent, but you brought up a great point, Glue layers had to be built. Now you have with the scale of Amazon Web Services, things we're learning from other companies. It reminds me of the open source vibe where you stand on the shoulders of others to get success. And there's a lot of new things coming out that startups don't have to do because startup before then did. This is like a new, cool thing. It's a whole nother level. >> Yeah, and I think it's a real standing on the shoulders of giants kind of thing. And if you just unpick, like in Verna's announcement this morning, his key to this one, he was talking about the Amplify Studio kind of stuff. And if you think about the before and after for that, front-end developers have had to do this stuff for a long period of time. And in the before version, they would have to do all that kind of integration work, which isn't really what they want to spend that time doing. And now they've kind of got that headstart. Andy Jassy famously would say, when he talked about building AWS, that there is no compression algorithm for experience. I like to kind of misuse that phrase for what we try to do for startups is provide these compression algorithms. So instead of having say, hire a larger engineering team to just do this kind of crafty stuff, they can just take the thing and kind of get from naught to 60 (indistinct). >> Gives some examples today of where this is playing out in real time. What kinds of new compression algorithms can startups leverage that they couldn't get before what's new that's available? >> I think you see it across all parts of the stack. I mean, you could just take it out of a database thing, like in the old days, if you wanted to start, and you had the dream that every startup has, of getting to kind of hyper scale where things bursting that seems is the problem. If you wanted to do that in the database layer back in the day, you would probably have to provision most of that database stuff yourself. And then when you get to some kind of limiting factor, you've got to do that work where all you're really wanting to do is try and add more features to your application. Or whether you've got services like Aurora where that will do all of that kind of scaling from a storage point of view. And it gives that startup the way to stand on the shoulders of giants, all the same kind of thing. You want to do some kind of identity, say you're doing a kind of a dog walking marketplace or something like that. So one of the things that you need to do for the kind of the payments thing is some kind of identity verification. In the old days, you would have to have gone pulled all those premises together to do the stuff that would look at people's ID and so on. Now, people can take things like Textracts for example, to look at those forms and do that kind of stuff. And you can kind of pick that story in all of these different stream lines whether it's compute stuff, whether it's database, whether it's high-level AIML stuff, whether it's stuff like amplify, which just massively compresses that timeframe for the startup. >> So, first of all, I'm totally loving this 'cause this is just an example of how evolution works. But if I'm a startup, one of the big things I would think about, and you're a founder, you know this, opportunity recognition is one thing, opportunity capture is another. So moving fast is what nimble startups do. Maybe there's a little bit of technical debt. There maybe a little bit of model debt, but they can get beach head quickly. Startups can move fast, that's the benefit. So where do I learn if I'm a startup founder about where all these pieces are? Is there a place that you guys are providing? Is there use cases where founders can just come in and get the best of the best composable cloud? How do I stand up something quickly to get going that I could regain and refactor later, but not take on too much technical debt or just actually have new building blocks. Where are all these tools? >> I'm really glad you asked that one. So, I mean, first startups is the core of what everyone in my team does. And most of the people we hire, well, they all have a passion for startups. Some have been former founders, some have been former CTOs, some have come to the passion from a different kind of thing. And they understand the needs of startups. And when you started to talk about technical debt, one of the balances that startups have always got to get right, is you're not building for 10 years down the line. You're building to get yourself often to the next milestone to get the next set of customers, for example. And so we're not trying to do the sort of the perfect anonymity of good things. >> I (indistinct) conception of startups. You don't need that, you just got to get the marketplace. >> Yeah, and how we try to do that is we've got a program called Activate and Activate gives startup founders either things like AWS credits up to a hundred thousand dollars in credits. It gives them other technical capabilities as well. So we have a part of the console, the management console called the Activate Console people can go there. And again, if you're trying to build a backend API, there is something that is built on AWS capability to be launched recently that basically says here's some templatized stuff for you to go from kind of naught to 60 and that kind of thing. So you don't have to spend time searching the web. And for us, we're taking that because we've been there before with a bunch of other startups, so we're trying to help. >> Okay, so how do you guys, I mean, a zillion startups, I mean, you and I could be in a coffee shop somewhere, hey, let's do a startup. Do I get access, does everyone gets access to this program that you have? Or is it an elite thing? Is there a criteria? Is it just, you guys are just out there fostering and evangelizing brilliant tools. Is there a program? How do you guys- >> It's a program. >> How do you guys vet startup's, is there? >> It's a program. It has different levels in terms of benefits. So at the core of it it's open to anybody. So if you were a bootstrap startup tomorrow, or today, you can go to the Activate website and you can sign up for that self-starting tier. What we also do is we have an extensive set of connections with the community, so T1 accelerators and incubators, venture capital firms, the kind of places where startups are going to build and via the relationships with those folks. If you're in one, if you've kind of got investment from a top tier VC firm for example, you may be eligible for a hundred thousand dollars of credit. So some of it depends on where the stock is up, but the overall program is open to all. And a chunk of the stuff we talked about like the guidance that's there for everybody. >> It's free, that's free and that's cool. That's good learning, so yeah. And then they get the free training. What's the coolest thing that you're doing right now that startups should know about around obviously the passionate start ups. I know for a fact at 80%, I can say that I've heard Andy and Adam both say that it's not just enterprising, well, they still love the startups. That's their bread and butter too. >> Yeah, well, (indistinct) I think it's amazing that someone, we were talking about the keynote you see some of these large customers in Adam's keynote to people like United Airlines, very, very large successful enterprise. And if you just look around this show, there's a lot of startups just on this expert floor that we are now. And when I look at these announcements, to me, the thing that just gets me excited and keeps me staying doing this job is all of these little capabilities make it in the environment right now with a good funding environment and all of these technical building blocks that instead of having to take a few, your basic compute and storage, once you have all of these higher and higher levels things, you know the serverless stuff that was announced in Adam's keynotes early, which is just making it easy. Because if you're a founder, you have an idea, you know the thing that you want to disrupt. And we're letting people do that in different ways. I'll pick one start up that I find really exciting to talk to. It's called Study. It's run by a guy called Zack Kansa. And he started that start up relatively recently. Now, if you started 15 years ago, you were going to use EC2 instances building on the cloud, but you were still using compute instances. Zack is really opinionated and a kind of a technology visionary in this sense that he takes this serverless approach. And when you talk to him about how he's building, it's almost this attitude of, if I've had to spin up a server, I've kind of failed in some way, or it's not the right kind of thing. Why would we do that? Because we can build with these completely different kinds of architectures. What was revolutionary 15 years ago, and it's like, okay, you can launch it and serve with an API, and you're going to pay by the hour. But now when you look at how Zack's building, you're not even launching a server and you're paying by the millions. >> So this is a huge history lesson slash important point. Back 15 years ago, you had your alternative to Amazon was provisioning, which is expensive, time consuming, lagging, and probably causes people to give up, frankly. Now you get that in the cloud either you're on your own custom domain. I remember EC2 before they had custom domains. It was so early. But now it's about infrastructures code. Okay, so again, evolution, great time to market, buy what you need in the cloud. And Adam talked about that. Now it's true infrastructure is code. So the smart savvy architects are saying, Hey, I'm just going to program. If I'm spinning up servers, that means that's a low level primitive that should be automated. >> Right. >> That's the new mindset. >> Yeah, that's why the fun thing about being in this industry is in just in the time that I've worked at AWS, since about 2011, this stuff has changed so much. And what was state of the art then? And if you take, it's funny, when you look at some of the startups that have grown with AWS, like whether it's Airbnb, Stripe, Slack and so on. If you look at how they built in 2011, because sometimes new startups will say, oh, we want to go and talk to this kind of unicorn and see how they built. And if you actually talked to the unicorn, some of them would say, we wouldn't build it this way anymore. We would do the kind of stuff that Zack and the folks studied are doing right now, because it's totally different (indistinct). >> And the one thing that's consistent from then to now is only one thing, it has nothing to do with the tech, it's speed. Remember rails front end with some backend Mongo, you're up on EC2, you've got an app, in a week, hackathon. Weekend- >> I'm not tying that time thing, that just goes, it gets smaller and smaller. Like the amplify thing that Verna was talking about this morning. You could've gone back 15 years, it's like, okay, this is how much work the developer would have to do. You could go back a couple of years and it's like, they still have this much work to do. And now this morning, it's like, they've just accelerated them to that kind of thing. >> We'll end on giving Jerry Chan a plug in our chat yesterday. We put the playbook out there for startups. You got to raise your focus on the beach head and solve the problem you got in front of you, and then sequence two adjacent positions, refactor in the cloud. Take that approach. You don't have to boil the ocean over right away. You get in the market, get in and get automating kind of the new playbook. It's just, make everything work for you. Not use the modern. >> Yeah, and the thing for me, that one line, I can't remember it was Paul Gray, or somehow that I stole it from, but he's just encouraging these startups to be appropriately lazy. Like let us do the hard work. Let us do the undifferentiated heavy lifting so people can come up with these super cool ideas. >> Yeah, just plugging the talent, plugging the developer. You got a modern application. Paul, thank you for coming on theCUBE, I appreciate it. >> Thank you. >> Head of Startup Solution Architecture North America, Amazon Web Services is going to continue to birth more startups that will be unicorns and decacorns now. Don't forget the decacorns. Okay, we're here at theCUBE bringing you all the action. I'm John Furrier, theCUBE. You're watching the Leader in Global Tech Coverage. We'll be right back. (bright upbeat music)
SUMMARY :
all scopes of the hybrid events. So we saw you last night, The legacy and the and doing all that kind of stuff. And now you see companies How do you guys see startups all of the kind of the stuff that startups don't have to do And if you just unpick, can startups leverage that So one of the things that you need to do and get the best of the And most of the people we hire, you just got to get the marketplace. So you don't have to spend to this program that you have? So at the core of it it's open to anybody. What's the coolest thing And if you just look around this show, Now you get that in the cloud And if you actually talked to the unicorn, And the one thing that's Like the amplify thing that Verna kind of the new playbook. Yeah, and the thing for me, Yeah, just plugging the bringing you all the action.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
2011 | DATE | 0.99+ |
Paul Duffy | PERSON | 0.99+ |
Canada | LOCATION | 0.99+ |
Andy | PERSON | 0.99+ |
Jerry Chan | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Adam Szalecki | PERSON | 0.99+ |
Paul | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
10 years | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
Jerry Chen | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Goldman Sachs | ORGANIZATION | 0.99+ |
United Airlines | ORGANIZATION | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
today | DATE | 0.99+ |
80% | QUANTITY | 0.99+ |
Paul Gray | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
millions | QUANTITY | 0.99+ |
2008 | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
2012 | DATE | 0.99+ |
Zack | PERSON | 0.99+ |
Hyperscaler | ORGANIZATION | 0.98+ |
tomorrow | DATE | 0.98+ |
15 years | QUANTITY | 0.98+ |
EC2 | TITLE | 0.98+ |
2007 | DATE | 0.98+ |
15 years later | DATE | 0.98+ |
two sets | QUANTITY | 0.98+ |
this week | DATE | 0.97+ |
15 years ago | DATE | 0.97+ |
first day | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
Hadoop | TITLE | 0.97+ |
60 | QUANTITY | 0.97+ |
Solutions Architecture | ORGANIZATION | 0.96+ |
this morning | DATE | 0.96+ |
both | QUANTITY | 0.96+ |
hundred thousand dollars | QUANTITY | 0.95+ |
last night | DATE | 0.95+ |
Airbnb | ORGANIZATION | 0.95+ |
two adjacent positions | QUANTITY | 0.94+ |
theCUBE | ORGANIZATION | 0.93+ |
Greylock | ORGANIZATION | 0.93+ |
one thing | QUANTITY | 0.93+ |
Colo | LOCATION | 0.93+ |
Stripe | ORGANIZATION | 0.91+ |
T1 | ORGANIZATION | 0.9+ |
one line | QUANTITY | 0.88+ |
up to a hundred thousand dollars | QUANTITY | 0.87+ |
Mongo | ORGANIZATION | 0.86+ |
Chisel AI | ORGANIZATION | 0.85+ |
Steve Mullaney, Aviatrix | AWS re:Invent 2021
(bright music) >> Welcome back to AWS re:Invent. You're watching theCUBE. And we're here with Steve Mullaney, who is the president and CEO of Aviatrix. Steve, I got to tell ya, great to see you man. >> We started the whole pandemic, last show we did was with you guys. >> Steve: Don't say we started, we didn't start it. (steve chuckles) >> Right, we kicked it off (all cross talking) >> It's going to be great. >> Our virtual coverage, that hybrid coverage that we did, how ironic? >> Steve: Yeah, was as the world was shutting down. >> So, great to see you face to face. >> Steve: Great to see you too. >> Wow, so you're two years in? >> Steve: Two and a half years yeah. >> Started, the company was standing start $2 billion valuation, raised a bunch of dough. >> Steve: Yeah. >> That's good, you got to feel good about that. >> We were 38 people, two and a half years ago, we're now 400. We had a couple million in ARR, we're now going to be over a 100 million next year, next calendar year, so significant growth. We just raised $200 million, three months ago at a $2 billion valuation. Now have 550 customers, 54 of them are fortune 500, when I started two and a half years ago, we didn't have any fortune 500s, we had probably about a 100 customers. So, massive growth, big growth (indistinct). >> Awesome, I got to ask you, I love to ask CEO's, entrepreneurs, how did you know when to scale? >> You just know it, when you see it. (indistinct) Yeah, there's no formula, you just know it and what you look for is that point where you say, okay, we've now proven the model and until you do that you minimize things and we actually just went through this. We had 12 sales teams, four months ago, we now have 50. 50, five zero and it's that step function as a company, you don't want to linearly grow 'cause you want to hold until you say, it's happening. And then once you say it's happening, okay, the dogs are eating the dog food, this is good then you flip the other way, and then you say, let's grow as fast as we possibly can and that's kind of the mode we're in right now. >> Okay, You've... >> You just know it when you see it. >> Other piece of that is how fast do you scale? And now you're sort of doing that step function as your going. >> Steve: We are going as fast as we possibly can. >> Wow, that's awesome, congratulations and I know you've got to long way to go. So okay, let's talk about the big trends that you're seeing that Aviatrix has taken advantage of, maybe explain a little bit about what you guys do. >> Yeah. So we are, what I like to call Multi- Cloud Native Networking and Network Security. So, if you think of... >> David: What is multicloud native? You got to explain that. >> I got to to explain that. Here's what's happened, it's happening and what I mean by it's happening is, enterprises at two and a half years ago, this is why I joined Aviatrix, all decided for the first time, we mean it now, we are going into Cloud 'cause before that they were just mouthing it. And they said, "We're going into the Cloud." And oh by the way, I knew two and a half years ago of course it was going to be multicloud, 'cause enterprises run workloads where they run best. That's what they do, it's sometimes it's AWS, sometimes it's ads or sometimes it's Google, it's of course going to be multicloud. And so from an enterprise perspective, they love the DevOps, they love the simplicity, the automation, the infrastructure is code, the Terraform, that Cloud operational model, because this is a business transformation, moving to Cloud is not a technology transformation it's the business. It's the CEO saying we are digitizing we have an existential threat to the survival of our company, I want to grow a market share, I want to be more competitive, we're doing this, stop laying across the tracks technology people, will run you over, we're doing this. And so when they do that as an enterprise, I'm BNY Mellon, I'm United Airlines, you name it, your favorite enterprise. I need the visibility and control from a networking and network security perspective like I used to have on-prem. Now I'm not going to do it in the horrible complex operational model the Cisco 1994 data center, do not bring that crap into my wonderful Cloud, so that ain't happening but, all I get from the Native constructs, I don't get enough of that visibility and control, it's a little bit of a black box, I don't get that. So where do I get the best of the Cloud from an operational model, but yet with the visibility and control that I need, that I used to have on-prem from networking network security, that's Aviatrix. And that's where people find us and so from a networking and network security, so that's why I call it multicloud Native because what we do is, create a layer basically an abstraction layer above all the different Clouds, we create one architecture for networking and network security with advanced services not basic services that run on AWS, Azure, Google, Oracle, Ali Cloud, Top Secret Clouds, GovClouds, you name it. And now the customer has one architecture, which is what enterprises want, I want one network, I want one network security architecture, not AWS Native, Azure Native, Google Native. >> David: Right. >> We leverage those native constructs, abstract it, and then provide a single common architecture with demand services, irrespective of what Cloud you're on. >> Dave, I've been saying this for a couple of years now, that Cloud Native... >> Does that make sense Dave? >> Absolutely. >> That abstraction layer, right? And I said, "The guys who do this, who figure this out are going to make a lot of dough." >> Yeah. >> Snowflakes obviously doing it. >> Yeah. >> You guys are doing it, it's the future. >> Yeah. >> And it's really an obvious construct when you look back at the world of call it Legacy IT for a moment... >> Steve: Yeah. >> Because did we have different networks to hookup different things in a data center? >> No, one network. >> One network of course. I don't care if the physical stack comes from Dell, HP or IBM. >> Steve: That's right, I want an attraction layer above that, yeah. >> Exactly. >> So the other thing that happens is, everybody and you'll understand this from being at Oracle, everybody wants to forget about the network. Network security, it's down in the bowels, it's like plumbing, electricity, it's just, it has to be there but people want to forget about it and so you see Datadog, you see Snowflake, you see HashiCorp going IPO in early December. Guess what? That next layer underneath that, I call it the horsemen of the multicloud infrastructure is networking and network security, that's going to be Aviatrix. >> Well, you guys make some announcements recently in that space, every company is a security company but you're really deep into it. >> Well, that's the interesting thing about it. So I said multicloud Native Networking and Network Security, it's integrated, so guess where network security is going to be done in the Cloud? In the network. >> David: Network. >> Yeah in the network. >> What a strange concept but guess what on-prem it's not, you deflect traffic to this thing called a firewall. Well, why was that? I was at Synoptics, I was at Cisco 'cause we didn't care about network security, so that's why firewall companies existed. >> Dave: Right. >> It should be integrated into the infrastructure. So now in the Cloud, your security posture is way worse than it was on-prem. You're connected to the internet by default so guess what? You want your network to do network security, so we announced two things in security; one, we're now a security competency partner for AWS, they do not give that out lightly. We were networks competency four years ago, we're now network security competency. One of the few that are both, they don't do that, that took us nine months of working with them to get there. And they only do that for the people that really are delivering value. And then what we just announced what we call, 'ThreatIQ with ThreatGuard.' So again, built into the network because we are the network, we understand the traffic, we're the control plane and the data plane, we see all traffic. We integrate into the network, we subscribe to threat databases, public databases, where we see what are the malicious IPS. If we have any traffic anywhere in your overall, and this is multicloud, not just AWS, every single Cloud, if we see that malicious traffic going some into IP guess what? It's probably BIT Mining, Bitcoin, crypto mining, it's probably some sort of data ex filtration. It could be some tour thing that you're connected to, whatever it is, you should not have traffic going. And so we do two things we alert and we show you where that all is and then with ThreatGuard, we actually will do a firewall rule right at that gateway, at that point that it's going out and immediately gone. >> You'll take the action. >> We'll take the action. >> Okay. >> And so every single customer, Dave and David, that we've shown this new capability to, it lights up like a Christmas tree. >> Yeah al bet. Okay, but now you've made some controversial statements... >> Steve: Which time? >> Okay, so you said Cisco, I think VMware... >> Dave: He's writing them down. >> I know but I can back it up. >> I think you said the risk, Cisco, VMware and Arista, they're not even in the Cloud conversation now. Arista, Jayshree Ullal is a business hero of mine, so I don't want to... >> Steve: Yeah, mine too. >> I don't want to interrogate her, she's awesome. >> Steve: Yeah. >> But what do you mean by that? Because can't Cisco come at this from their networking perspective and security and bring that in? What do you mean by they're not in the Cloud conversation? >> They're not in the conversation. >> David: Okay, defend that. >> And the reason is they were about four years ago. So when you're four years ago, you're moving into the Cloud, what's the first thing you do? I'm going to grab my CSR and I'm going to try to jam it in the Cloud. Guess what? The CSR doesn't even know it's in the Cloud, it's looking for ports, right? And so what happens is the operational model is horrendous, so all the Cloud people, it just is like oil and water, so they go, oh, that was horrendous. So no one's doing that, so what happens in the Cloud is they realize the number one thing is the Cloud operational model. I need that simplicity, I have to be a single Terraform provider, infrastructure is code. Where do I put my box with my wires? That's what the on-prem hardware people think. >> David: The selling ports your saying? >> The selling boxes. >> David: Yeah. >> And so they'll say, "Oh, we got us software version of it, it runs as a VM, it has no idea it's in the Cloud." It is not Cloud Native, I call that Cloud naive, they don't understand so then the model doesn't work. And so then they say, "Okay, I'm not going to do that." Then the only other thing they can do, is they look at the Cloud providers themselves and they say, "All right, I'm going to use Native constructs, what do you got?" And what happens basically is the Cloud providers say, "Well, we do everything and anything you'll ever need and networking and network security." And the customers, "Oh my God, it's fantastic." Then they try to use it and what they realize is you get very basic level services, and you get no visibility and control because they're a black box, you don't get to go in. How about troubleshooting, Packet Captures, simple things? How about security controls, performance traffic engineering, performance controls, visibility nothing, right? And so then they go, "Oh shit, I'm an enterprise, I'm not just some DevOps Danny three years ago, who was just spinning up workloads and didn't care about security." No, that was the Cloud three years ago. This is now United, BNY, Nike. This is like elite of elite. So when my VC was here, he said, "It's happening." That's what he meant, it's happening. Meaning enterprises, the dogs are eating the dog food and they need visibility and control, they cannot get it from the Cloud providers. >> It's happening in early days Dave. >> So Steve, we're going to stipulate that you can't jam this stuff into Cloud, but those dinosaurs are real and they're there. Explain how you... >> Steve: Well you called them dinosaurs not me but they're roaming the earth and they're going to run out of food pretty soon. (all laughing) The comet hit the earth. >> Hey, they're going to go down fighting. (all laughing) >> But the dinosaurs didn't all die the day after the comet hit the earth... >> Steve: That's right. >> They took awhile. >> Steve: They took a while. >> So, how are you going to saddle them up? That's the question because you're... >> Steve: It's over there walking dead, I don't need to do anything. >> Is it the captain Kirk to con, let them die. >> Steve: Yeah. >> Because you're in the Cloud, you're multicloud... >> Steve: Yeah. >> That's great, but 80% of my IT still on-prem and I still have Cisco switches. Isn't that just not your market or? >> When IBM and DEC did we have to do anything with IBM and DEC in the 90s, early 90s, when we created BC client server, IP architectures? No, they weren't in the conversation. >> David: Yeah. >> So, we dint compete with them, just like whatever they do on-prem, keep doing it, I wish you the best. >> But you need to integrate with them and play with them. >> Steve: No. >> Not at all? >> No, no we integrate, here is the thing that's going to happen, so to the on-prem people, it's all point of reference. They look at Cloud as off-prem, I'm going to take my operational model on-prem and I'm going to push it into the Cloud. And if I push it into multiple Clouds, they're going to call that multicloud, see we are multicloud. You're pushing your operational model into the Cloud. What's happening is Cloud has won, it won two and a half years ago with every enterprise. It's like a rock in the water. And what's going to happen is that operational model is moving out to the edge, it's moving to the branch, it's moving to the data center and it's moving into edge computing. That's what's happening... >> So outpost, so I put an outpost in my data center... >> Outpost looks like... >> Is that Aviatrix? >> Absolutely, we're going to get dragged with that... >> Dave: Okay, alright. >> Because we're the networking and network security provider, and as the company pushes out, that operational model is going to move out, not the existing on-prem OT, IT branch office then pushing in. And so, what's happening is you're coming at it from the wrong perspective. And this wave is just going to push over and so I'm just following behind this wave of AWS and Azure and Google. >> Here's the thing, you can do this and you don't have a bunch of legacy deductible debt... >> Steve: Yeah. >> So you can be Cloud Native, multicloud native, I think you called it? >> Steve: Yeah, yeah. >> I love it, you're building castles on the sand. >> Steve: Yeah. >> Jerry Chen's thing. >> Steve: Yeah. >> Now, the thing is, today's executives, they're not as naive as Ken Olsen, UNIX as, "Snake oil," who would need a PC, so they're not in denial. >> They're probably not in denial, yeah. >> Right, and so they have some resources, so the problem is they can't move as fast as you can. So, you're going to do really well. >> Steve: Yeah. >> I think they'll eventually get there Steve, but you're going to be, I don't know how many, four or five years ahead, that's a nice lead. >> That's a bet I'll take any day. >> David: Then what you don't think they'll ever get there? >> No, 10 years. (steve laughing) >> Okay, but they're not going out of business. >> No, I didn't say that. >> I know you didn't. >> What they're doing, I wish them all the best. >> Because a lot of their customers move... >> I don't compete with them. >> Yeah. We were out of time. >> Yeah. >> What did you mean by AWS is like Sandals? You mean like cool like Sandals? >> Steve: Oh, no, no, no. I don't want to... >> You mean like the vacation place? >> Have you ever been to Sandals? >> I never done it. What do you mean by that? >> There coming, there coming. Which version of sandals (indistinct)? (people cross talking) >> This is for an enterprise by the way, and look, Sandals is great for a lot of people but if you're a Cloud provider, you have to provide the common set of services for the masses because you need to make money. And oh, by the way, when you go to Sandals, go try it, like get a bottle of wine, they say, "We got red wine or white wine?" "Oh, great, what kind of red wine?" "No, red wine and it's in a box." And they hope that you won't know the difference. The problem is some people in enterprises want Four Seasons, so they want to be able to swipe the card and get a good bottle of wine. And so that's the thing with the Cloud, but the Cloud can't offer up a 200 bottle of wine to everybody. My mom loves box wine, so give her box wine. Where ISBs like us come in, is great but complimentary to the Cloud provider for that person who wants that nice bottle of wine because if AWS had to provide all this level of functionality for everybody, their instant sizes would be too big, >> Too much cost for that. (people cross talking) You're right on. And as long as you can innovate fast and stay ahead of that and keep adding value... >> Well, here's the thing, they're not going to do it for multicloud either though. >> David: I wouldn't trust them to do it with multicloud. >> No. >> David: I wouldn't. >> No enterprise would and I don't think they would ever do it anyway. >> That makes sense. Steve, we've got to go man. You're awesome, love to have you on theCUBE, come back anytime. >> Awesome, thank you. >> All right, keep it right there everybody. You're watching theCUBE, the leader in enterprise tech coverage. (bright music)
SUMMARY :
great to see you man. last show we did was with you guys. Steve: Don't say we Steve: Yeah, was as the Started, the company was standing start That's good, you got we didn't have any fortune 500s, and that's kind of the is how fast do you scale? Steve: We are going as So okay, let's talk about the big trends So, if you think of... You got to explain that. It's the CEO saying we are digitizing and then provide a single for a couple of years now, And I said, "The guys who do this, when you look back at the world of call it I don't care if the physical stack I want an attraction and so you see Datadog, you see Snowflake, Well, you guys make Well, that's the you deflect traffic to this and we show you where that all is And so every single Okay, but now you've made some Okay, so you said I think you said the risk, I don't want to interrogate And the reason is they and you get no visibility and control that you can't jam this stuff into Cloud, and they're going to run Hey, they're going to go down fighting. But the dinosaurs didn't all die That's the question because you're... I don't need to do anything. Is it the captain Kirk Because you're in the and I still have Cisco switches. When IBM and DEC did I wish you the best. But you need to integrate with them here is the thing that's going to happen, So outpost, so I put an to get dragged with that... and as the company pushes out, Here's the thing, you can do this building castles on the sand. Now, the thing is, today's executives, so the problem is they can't I don't know how many, No, 10 years. Okay, but they're not What they're doing, I Because a lot of Yeah. I don't want to... do you mean by that? (people cross talking) And so that's the thing with the Cloud, And as long as you can innovate Well, here's the thing, them to do it with multicloud. and I don't think they to have you on theCUBE, the leader in enterprise tech coverage.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Steve | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Steve Mullaney | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
HP | ORGANIZATION | 0.99+ |
Aviatrix | ORGANIZATION | 0.99+ |
DEC | ORGANIZATION | 0.99+ |
Ken Olsen | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Nike | ORGANIZATION | 0.99+ |
nine months | QUANTITY | 0.99+ |
$2 billion | QUANTITY | 0.99+ |
12 sales teams | QUANTITY | 0.99+ |
Jerry Chen | PERSON | 0.99+ |
Jayshree Ullal | PERSON | 0.99+ |
BNY | ORGANIZATION | 0.99+ |
80% | QUANTITY | 0.99+ |
$200 million | QUANTITY | 0.99+ |
Synoptics | ORGANIZATION | 0.99+ |
United Airlines | ORGANIZATION | 0.99+ |
two years | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
38 people | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
Danny | PERSON | 0.99+ |
550 customers | QUANTITY | 0.99+ |
90s | DATE | 0.99+ |
BNY Mellon | ORGANIZATION | 0.99+ |
three years ago | DATE | 0.99+ |