Emilia A'Bell Platform9
(Gentle music) >> Hello and welcome to the Cube here in Palo Alto, California. I'm John Furrier here, joined by Platform nine, Amelia Bell the Chief Revenue Officer, really digging into the conversation around Kubernetes Cloud native and the journey this next generation cloud. Amelia, thanks for coming in and joining me today. >> Thank you, thank you. Great pleasure to be here. >> So, CRO, chief Revenue Officer. So you're mainly in charge of serving the customers, making sure they're they're happy with the solution you guys have. >> That's right. >> And this market must be pretty exciting. >> Oh, it's very exciting and we are seeing a lot of new use cases coming up all the time. So part of my job is to obtain new customers but then of course, service our existing customers and then there's a constant evolution. Nothing is standing still right now. >> We've had all your co-founders on, on the show here and we've kind of talked about the trends and where you guys have come from, where you guys are going now. And it's interesting, if you look at the cloud native market, the scale is still huge. You seeing now this next wave of AI coming on, which I call that's the real web three in my mind in terms of like the next experiences really still points to data infrastructure scale. These next gen apps are coming. And so that's being built on the previous generation of DevSecOps. >> Right >> And so a lot of enterprises are having to grow up really, really fast >> Right. >> And figure out, okay, I got to have scale I got large scale data, I got horizontal scalability I got to apply machine learning now the new software engineering practice. And then, oh, by the way I got the Kubernetes clusters I got to manage >> Right. >> I got what's containers weather, the security problems. This is a really complicated but important area of build out right now in the marketplace. >> Right. What are you seeing? >> So it's, it's really important that the infrastructure is not the hindrance in these cases. And we, one of our customers is in fact a large AI company and we, I met with them yesterday and asked them, you know, why are you giving that to us? You've got really smart engineers. They can run and create the infrastructure, you know in a custom way that you want it. And they said, we've got to be core to our business. There's plenty of work to do just on delivering the AI capabilities, and there's plenty of work to do. We can't get bogged down in the infrastructure. We don't want to have people running the engine we want them driving the car. We want them creating value on top of that. so they can't have the infrastructure being the bottleneck for them. >> It's interesting, the AI companies, that's their value proposition to their customers is that they don't want the technical talent. >> Right. >> Working on, you know, non-differentiated heavy lifting things. >> Right. >> And automate those and scale it up. Can you talk about the problem that you guys are solving? Because there's a lot going on here. >> Yeah. >> You can look at all aspects of the DevOps scale. There's a lot of little problems, some big problems. What are you guys focusing on? What's the bullseye for Platform known? >> Okay, so the bullseye is that Kubernetes infrastructure is really hard, right? It's really hard to create and run. So we introduce a time to market efficiency, let's get this up and running and let's get you into production and and producing results for your customers fast. But at the same time, let's reduce your cost and complexity and increase reliability. So, >> And what are some of the things that they're having problems with that are breaking? Is it more of updates on code? Is it size of the, I mean clusters they have, what what is it more operational? What are the, what are some of the things that are that kind of get them to call you guys up? What's the main thing? >> It's the operations. It's all operations. So what, what happens is that if you have a look at Kubernetes platform it's made up of many, many components. And that's where it gets complex. It's not just Kubernetes. There's load balances, networking, there's observability. All these things have to operate together. And all the piece parts have to be upgraded and maintained. The integrations need to work, you need to have probes into the system to predict where problems can be coming. So the operational part of it is complex. So you need to be observing not only your clusters in the health of the clusters and the nodes and so on but the health of the platform itself. >> We're going to get Peter Frey in on here after I talk about some of the technical issues on deployments. But what's the, what's the big decision for the customer? Because there's kind of, there's two schools of thought. One is, I'm going to build my own and have my team build it or I'm going to go with a partner >> Right. >> Say platform nine, what's the trade offs there? Because it seems to me that, that there's a there's a certain area of where it's core competency but I can outsource it or partner with it and, and work with platform nine versus trying to take it all on internally >> Right. >> Of which requires more costs. So there's a, there's a line where you kind of like figure out that customers have to figure out that, that piece >> Right >> What do, what's your view on that? Because I'm hearing that more people are saying, hey I want to, I want to focus my people on solutions. The app side, not so much the ops >> Right. >> What's the trade off? How do you talk about? >> It's a really interesting question because most companies think they have two options. It's either a DIY option and they love that engineers love playing with the new and on the latest. And then they think the other option is going to cloud, public cloud and have it semi managed by them. And you get very different out of those. So in the DIY you get flexibility coz you get to choose your infrastructure but then you've got all the complexities of the DIY piece. You've got to not only choose all your components but you've got to keep them working. Now if you go to public cloud option, you lose flexibility because a lot of those choices are made for you but you gain agility because quite frankly it's really easy to spin up clusters. So what we are, is that in the middle we bring the agility and the flexibility because we bring the control plane that allows you to spin up clusters and and lifecycle manage them very quickly. So the agility's there but you can do it on the infrastructure of your choice. And in the DIY culture, one of the hardest things to do actually is to convince them they don't have to do it themselves. They can focus on higher value activities, which are more focused on delivering outcomes to their customers. >> So you provide the solution that allows them to feel like they're billing it themselves. >> Correct. >> And get these scale and speed and the efficiencies of the op side. So it's kind of the best of both worlds. It's not a full outsource. >> Right, right. >> You're bringing them in to make their jobs easier >> Right, That's right. So they get choices. >> Yeah. >> We, we, they get choices on how they build it and then we run and operate it for them. But they, they have all the observability. The benefit is that if we are managing their operations and most of our customers choose the managed operations piece of it, then they don't. If something goes wrong, we fix that and they, they they get told, oh, by the way, you had a problem. We've dealt with it. But in the other model is they've got to create all that observability themselves and they've got to get ahead of the issues themselves, and then they've got to raise tickets to whoever they need to raise tickets to. Whereas we have things like auto ticket generation and so on where, look, just drive the car let us worry about the engine and all of that. Let us deal with that. And you can choose whatever you want about the engine but let us manage it for you. So >> What do you, what do you say to folks out there that are may have a need for platform nine? What's the signals inside their company that they should be calling you guys up and, and leaning in with platform nine? >> Right. >> Is it more sprawl on on clusters? Is it more errors? Is it more tickets? Is it more hassle? What are some of the signs? If someone's watching this say, hey I have, I have an issue with this. >> I would say, if there's operational inefficiencies you can't get things to market fast enough because you are building this and it's just taking too long you're spending way too much time operationally on the infrastructure, then you are, you are not using your resources where they should best be used. And, and that is delivering services to the customer. >> Ed me Hora on for International Women's Day. And she was talking about how they love to solve complex problems on the engineering team at Platform nine. It's going to get pretty complex with the edge emerging >> Indeed >> and cloud native on-premises distributed computing. >> Indeed. >> essentially is what it is. That's kind of the core DNA of the team. >> Yeah. >> What, how does that translate to the customers? Because IT seems to be, okay, I have virtual machines were great, now I got to scale up and and convert over a transform to containers, Kubernetes >> Right. >> And then large scale app, app applications. >> Right, so when it comes to Edge it gets complex pretty fast because it's highly distributed. So how do you have standardization and governance across all the different edge locations? So what we bring into play is an ability to, um, at each edge, location eh, provision from bare metal up all the way up to the application. So let's say you have thousands of stores and you want to modernize those stores, you know rather than having a server being sent somewhere to have an image loaded up and then sent that and then you've got to send a technical guide to the store and you've got to implement it all there. Forget all that. That's just, that's just a ridiculous waste of time. So what we've done is we've created the ability where the server can just be sent to the store. You can get your barista or your chef just to plug it in, right? You don't need to send any technical person over there. As long as we have access to it, we get access to it and we provision the whole thing from bare metal up and then we can maintain it according to the standards that are needed and upgrade accordingly. And that gives standardization across all your stores or edge locations or 5G towers or whatever it is, distribution centers. And we can create nice governance and good standardization which allows them to innovate fast as well. >> So this is a real opportunity for you guys. >> Yeah. >> This is an advantage from your expertise. >> Yes. >> The edge piece, dropping in a box, self-provisioning. >> That's right. So yeah. >> Can people do that? What's the, >> No, actually it, it's, it's very difficult to do. I I, from my understanding, we're the only people that can provision it from bare metal up, right? So if anyone has a different story, I'd love to hear about that. But that's my understanding today. >> That's a good value purpose. So talk about the value of the customer. What kind of scope do you got? Can you scope some of the customer environments you have from >> Sure. >> From, you know, small to the large, how give us an idea of the order of magnitude of the >> Yeah, so, so small customers may have 20 clusters or something like that. 20 nodes, I beg your pardon. Our large customers, like we're we are scaling one particular distributed environment from 2200 nodes to 10,000 nodes by the end of this year and 26,000 nodes next year. We have another customer that's scaling up to 10,000 nodes this year as well. So we have some very large scale, but some smaller ones too. And we're, we're happy to work with either end. >> Okay, so pretend I'm a customer. I'm really, I got pain and Kubernetes like I want to, I can't hire enough people. I want to have my all focus. What's the pitch? >> Okay. So skill shortage is something that that everyone is facing right now. And if, if you've got skill shortage it's going to be really hard to hire if you are competing against really, you know, high salary you know, offering companies that are out there. So the pitch is, let us do it for you. We have, we have a team of excellent probably the best Kubernetes engineers on the planet. We will create your environment for you. We will get it up and running. We will allow you to, you know, run your applica, just consume the platform, we'll run it for you. We'll have SLAs and up times guaranteed and you can just focus on delivering the software and the value needed to your customers. >> What are some of the testimonials that you get from people? Just anecdotally, what do they say? Oh my god, you guys save. >> Yeah. >> Our butts. >> Yeah. >> This is amazing. We just shipped our code out much faster. >> Yeah. >> What are some of the things that you hear? >> So, so the number one thing I hear is it just works right? It's, we don't have to worry about it, it just works. So that, that's a really great feedback that we get. The other thing I hear is if we do have issues that your team are amazing, they they fix things, they're proactive, you know, they're we really enjoy working with you. So from, from that perspective, that's great. But the other side of it is we hear things like if we were to do that ourselves we would've taken six to 12 months to build that. And you guys have just saved us six to 12 months. The other thing that we hear is with the same two engineers we started on, you know, a hundred nodes we're now running thousands of nodes. We have not had to increase the size of the team and expand and scale exponentially. >> Awesome. What's next for you guys? What's on your, your plate? >> Yeah. >> With CRO, what's some of the goals you have? >> Yeah, so growth of course as a CRO, you don't get away from that. We've got some very exciting, actually, initiatives coming up. One of the things that we are seeing a lot of demand for and is, is in the area of virtualization bringing virtual machine, virtual virtual containers, sorry I'm saying that all wrong. Bringing virtual machine, the virtual machines onto the cloud native infrastructure using Kubernetes technology. So that provides a, an excellent stepping stone for those guys who are in the virtualization world. And they can't move to containers, they can't refactor their applications and workloads fast enough. So just bring your virtual machine and put it onto the container infrastructure. So we're seeing a lot of demand for that, because it provides an excellent stepping stone. Why not use Kubernetes to orchestrate virtual the virtual world? And then we've got some really interesting cost optimization. >> So a lot of migration kind of thinking around VMs and >> Oh, tremendous. The, the VM world is just massively bigger than the container world right now. So you can't ignore that. So we are providing basically the evolution, the the journey for the customers to utilize the greatest of technologies without having to do that in a, in a in a way that just breaks the bank and they can't get there fast enough. So we provide those stepping stones for them. Yeah. >> Amelia thank you for coming on. Sharing. >> Thank you. >> The update on platform nine. Congratulations on your big accounts you have and >> thank you. >> And the world could get more complex, which Means >> indeed >> have more customers. >> Thank you, thank you John. Appreciate that. Thank you. >> I'm John Furry. You're watching Platform nine and the Cube Conversations here. Thanks for watching. (gentle music)
SUMMARY :
and the journey this Great pleasure to be here. mainly in charge of serving the customers, And this market must and we are seeing a lot and where you guys have come from, I got the Kubernetes of build out right now in the marketplace. What are you seeing? that the infrastructure is not It's interesting, the AI Working on, you know, that you guys are solving? aspects of the DevOps scale. Okay, so the bullseye is into the system to predict of the technical issues out that customers have to The app side, not so much the ops So in the DIY you get flexibility So you provide the solution of the best of both worlds. So they get choices. get ahead of the issues are some of the signs? on the infrastructure, complex problems on the engineering team and cloud native on-premises is. That's kind of the core And then large scale So let's say you have thousands of stores opportunity for you guys. from your expertise. in a box, self-provisioning. So yeah. different story, I'd love to So talk about the value of the customer. by the end of this year What's the pitch? and the value needed to your customers. What are some of the testimonials This is amazing. of the team and expand What's next for you guys? and is, is in the area of virtualization So you can't ignore Amelia thank you for coming on. big accounts you have and Thank you. and the Cube Conversations here.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amelia | PERSON | 0.99+ |
Amelia Bell | PERSON | 0.99+ |
John | PERSON | 0.99+ |
six | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
yesterday | DATE | 0.99+ |
Emilia A'Bell | PERSON | 0.99+ |
John Furry | PERSON | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Peter Frey | PERSON | 0.99+ |
12 months | QUANTITY | 0.99+ |
International Women's Day | EVENT | 0.99+ |
two engineers | QUANTITY | 0.99+ |
two options | QUANTITY | 0.99+ |
20 clusters | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
two schools | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
this year | DATE | 0.98+ |
today | DATE | 0.98+ |
20 nodes | QUANTITY | 0.97+ |
each edge | QUANTITY | 0.96+ |
Kubernetes | ORGANIZATION | 0.96+ |
thousands of stores | QUANTITY | 0.93+ |
end of this year | DATE | 0.93+ |
2200 nodes | QUANTITY | 0.93+ |
Cube | ORGANIZATION | 0.93+ |
10,000 nodes | QUANTITY | 0.93+ |
Kubernetes | TITLE | 0.92+ |
both worlds | QUANTITY | 0.91+ |
up to 10,000 nodes | QUANTITY | 0.88+ |
thousands of nodes | QUANTITY | 0.87+ |
Edge | TITLE | 0.84+ |
26,000 nodes | QUANTITY | 0.81+ |
Ed me Hora | PERSON | 0.8+ |
Platform nine | TITLE | 0.75+ |
hundred nodes | QUANTITY | 0.69+ |
DevSecOps | TITLE | 0.68+ |
Platform nine | ORGANIZATION | 0.68+ |
one thing | QUANTITY | 0.62+ |
wave | EVENT | 0.57+ |
Chief Revenue Officer | PERSON | 0.57+ |
nine | QUANTITY | 0.56+ |
CRO | PERSON | 0.54+ |
three | QUANTITY | 0.53+ |
nine | OTHER | 0.52+ |
DevOps | TITLE | 0.5+ |
next | EVENT | 0.49+ |
platform nine | OTHER | 0.49+ |
Cube | TITLE | 0.39+ |
Dave Duggal, EnterpriseWeb & Azhar Sayeed, Red Hat | MWC Barcelona 2023
>> theCUBE's live coverage is made possible by funding from Dell Technologies. Creating technologies that drive human progress. (ambient music) >> Lisa: Hey everyone, welcome back to Barcelona, Spain. It's theCUBE Live at MWC 23. Lisa Martin with Dave Vellante. This is day two of four days of cube coverage but you know that, because you've already been watching yesterday and today. We're going to have a great conversation next with EnterpriseWeb and Red Hat. We've had great conversations the last day and a half about the Telco industry, the challenges, the opportunities. We're going to unpack that from this lens. Please welcome Dave Duggal, founder and CEO of EnterpriseWeb and Azhar Sayeed is here, Senior Director Solution Architecture at Red Hat. >> Guys, it's great to have you on the program. >> Yes. >> Thank you Lisa, >> Great being here with you. >> Dave let's go ahead and start with you. Give the audience an overview of EnterpriseWeb. What kind of business is it? What's the business model? What do you guys do? >> Okay so, EnterpriseWeb is reinventing middleware, right? So the historic middleware was to build vertically integrated stacks, right? And those stacks are now such becoming the rate limiters for interoperability for so the end-to-end solutions that everybody's looking for, right? Red Hat's talking about the unified platform. You guys are talking about Supercloud, EnterpriseWeb addresses that we've built middleware based on serverless architecture, so lightweight, low latency, high performance middleware. And we're working with the world's biggest, we sell through channels and we work through partners like Red Hat Intel, Fortnet, Keysight, Tech Mahindra. So working with some of the biggest players that have recognized the value of our innovation, to deliver transformation to the Telecom industry. >> So what are you guys doing together? Is this, is this an OpenShift play? >> Is it? >> Yeah. >> Yeah, so we've got two projects right her on the floor at MWC throughout the various partners, where EnterpriseWeb is actually providing an application layer, sorry application middleware over Red Hat's, OpenShift and we're essentially generating operators so Red Hat operators, so that all our vendors, and, sorry vendors that we onboard into our catalog can be deployed easily through the OpenShift platform. And we allow those, those vendors to be flexibly composed into network services. So the real challenge for operators historically is that they, they have challenges onboarding the vendors. It takes a long time. Each one of them is a snowflake. They, you know, even though there's standards they don't all observe or follow the same standards. So we make it easier using models, right? For, in a model driven process to on boards or streamline that onboarding process, compose functions into services deploy those services seamlessly through Red Hat's OpenShift, and then manage the, the lifecycle, like the quality of service and the SLAs for those services. >> So Red Hat obviously has pretty prominent Telco business has for a while. Red Hat OpenStack actually is is pretty popular within the Telco business. People thought, "Oh, OpenStack, that's dead." Actually, no, it's actually doing quite well. We see it all over the place where for whatever reason people want to build their own cloud. And, and so, so what's happening in the industry because you have the traditional Telcos we heard in the keynotes that kind of typical narrative about, you know, we can't let the over the top vendors do this again. We're, we're going to be Apifi everything, we're going to monetize this time around, not just with connectivity but the, but the fact is they really don't have a developer community. >> Yes. >> Yet anyway. >> Then you have these disruptors over here that are saying "Yeah, we're going to enable ISVs." How do you see it? What's the landscape look like? Help us understand, you know, what the horses on the track are doing. >> Sure. I think what has happened, Dave, is that the conversation has moved a little bit from where they were just looking at IS infrastructure service with virtual machines and OpenStack, as you mentioned, to how do we move up the value chain and look at different applications. And therein comes the rub, right? You have applications with different requirements, IT network that have various different requirements that are there. So as you start to build those cloud platform, as you start to modernize those set of applications, you then start to look at microservices and how you build them. You need the ability to orchestrate them. So some of those problem statements have moved from not just refactoring those applications, but actually now to how do you reliably deploy, manage in a multicloud multi cluster way. So this conversation around Supercloud or this conversation around multicloud is very >> You could say Supercloud. That's okay >> (Dave Duggal and Azhar laughs) >> It's absolutely very real though. The reason why it's very real is, if you look at transformations around Telco, there are two things that are happening. One, Telco IT, they're looking at partnerships with hybrid cloud, I mean with public cloud players to build a hybrid environment. They're also building their own Telco Cloud environment for their network functions. Now, in both of those spaces, they end up operating two to three different environments themselves. Now how do you create a level of abstraction across those? How do you manage that particular infrastructure? And then how do you orchestrate all of those different workloads? Those are the type of problems that they're actually beginning to solve. So they've moved on from really just putting that virtualizing their application, putting it on OpenStack to now really seriously looking at "How do I build a service?" "How do I leverage the catalog that's available both in my private and public and build an overall service process?" >> And by the way what you just described as hybrid cloud and multicloud is, you know Supercloud is what multicloud should have been. And what, what it originally became is "I run on this cloud and I run on this cloud" and "I run on this cloud and I have a hybrid." And, and Supercloud is meant to create a common experience across those clouds. >> Dave Duggal: Right? >> Thanks to, you know, Supercloud middleware. >> Yeah. >> Right? And, and so that's what you guys do. >> Yeah, exactly. Exactly. Dave, I mean, even the name EnterpriseWeb, you know we started from looking from the application layer down. If you look at it, the last 10 years we've looked from the infrastructure up, right? And now everybody's looking northbound saying "You know what, actually, if I look from the infrastructure up the only thing I'll ever build is silos, right?" And those silos get in the way of the interoperability and the agility the businesses want. So we take the perspective as high level abstractions, common tools, so that if I'm a CXO, I can look down on my environments, right? When I'm really not, I honestly, if I'm an, if I'm a CEO I don't really care or CXO, I don't really care so much about my infrastructure to be honest. I care about my applications and their behavior. I care about my SLAs and my quality of service, right? Those are the things I care about. So I really want an EnterpriseWeb, right? Something that helps me connect all my distributed applications all across all of the environments. So I can have one place a consistency layer that speaks a common language. We know that there's a lot of heterogeneity down all those layers and a lot of complexity down those layers. But the business doesn't care. They don't want to care, right? They want to actually take their applications deploy them where they're the most performant where they're getting the best cost, right? The lowest and maybe sustainability concerns, all those. They want to address those problems, meet their SLAs meet their quality service. And you know what, if it's running on Amazon, great. If it's running on Google Cloud platform, great. If it, you know, we're doing one project right here that we're demonstrating here is with with Amazon Tech Mahindra and OpenShift, where we took a disaggregated 5G core, right? So this is like sort of latest telecom, you know net networking software, right? We're deploying pulling elements of that network across core, across Amazon EKS, OpenShift on Red Hat ROSA, as well as just OpenShift for cloud. And we, through a single pane of deployment and management, we deployed the elements of the 5G core across them and then connected them in an end-to-end process. That's Telco Supercloud. >> Dave Vellante: So that's an O-RAN deployment. >> Yeah that's >> So, the big advantage of that, pardon me, Dave but the big advantage of that is the customer really doesn't care where the components are being served from for them. It's a 5G capability. It happens to sit in different locations. And that's, it's, it's about how do you abstract and how do you manage all those different workloads in a cohesive way? And that's exactly what EnterpriseWeb is bringing to the table. And what we do is we abstract the underlying infrastructure which is the cloud layer. So if, because AWS operating environment is different then private cloud operating environment then Azure environment, you have the networking is set up is different in each one of them. If there is a way you can abstract all of that and present it in a common operating model it becomes a lot easier than for anybody to be able to consume. >> And what a lot of customers tell me is the way they deal with multicloud complexity is they go with mono cloud, right? And so they'll lose out on some of the best services >> Absolutely >> If best of, so that's not >> that's not ideal, but at the end of the day, agree, developers don't want to muck with all the plumbing >> Dave Duggal: Yep. >> They want to write code. >> Azhar: Correct. >> So like I come back to are the traditional Telcos leaning in on a way that they're going to enable ISVs and developers to write on top of those platforms? Or are there sort of new entrance and disruptors? And I know, I know the answer is both >> Dave Duggal: Yep. >> but I feel as though the Telcos still haven't, traditional Telcos haven't tuned in to that developer affinity, but you guys sell to them. >> What, what are you seeing? >> Yeah, so >> What we have seen is there are Telcos fall into several categories there. If you look at the most mature ones, you know they are very eager to move up the value chain. There are some smaller very nimble ones that have actually doing, they're actually doing something really interesting. For example, they've provided sandbox environments to developers to say "Go develop your applications to the sandbox environment." We'll use that to build an net service with you. I can give you some interesting examples across the globe that, where that is happening, right? In AsiaPac, particularly in Australia, ANZ region. There are a couple of providers who have who have done this, but in, in, in a very interesting way. But the challenges to them, why it's not completely open or public yet is primarily because they haven't figured out how to exactly monetize that. And, and that's the reason why. So in the absence of that, what will happen is they they have to rely on the ISV ecosystem to be able to build those capabilities which they can then bring it on as part of the catalog. But in Latin America, I was talking to one of the providers and they said, "Well look we have a public cloud, we have our own public cloud, right?" What we want do is use that to offer localized services not just bring everything in from the top >> But, but we heard from Ericson's CEO they're basically going to monetize it by what I call "gouge", the developers >> (Azhar laughs) >> access to the network telemetry as opposed to saying, "Hey, here's an open platform development on top of it and it will maybe create something like an app store and we'll take a piece of the action." >> So ours, >> to be is a better model. >> Yeah. So that's perfect. Our second project that we're showing here is with Intel, right? So Intel came to us cause they are a reputation for doing advanced automation solutions. They gave us carte blanche in their labs. So this is Intel Network Builders they said pick your partners. And we went with the Red Hat, Fort Net, Keysite this company KX doing AIML. But to address your DevX, here's Intel explicitly wants to get closer to the developers by exposing their APIs, open APIs over their infrastructure. Just like Red Hat has APIs, right? And so they can expose them northbound to developers so developers can leverage and tune their applications, right? But the challenge there is what Intel is doing at the low level network infrastructure, right? Is fundamentally complex, right? What you want is an abstraction layer where develop and this gets to, to your point Dave where you just said like "The developers just want to get their job done." or really they want to focus on the business logic and accelerate that service delivery, right? So the idea here is an EnterpriseWeb they can literally declaratively compose their services, express their intent. "I want this to run optimized for low latency. I want this to run optimized for energy consumption." Right? And that's all they say, right? That's a very high level statement. And then the run time translates it between all the elements that are participating in that service to realize the developer's intent, right? No hands, right? Zero touch, right? So that's now a movement in telecom. So you're right, it's taking a while because these are pretty fundamental shifts, right? But it's intent based networking, right? So it's almost two parts, right? One is you have to have the open APIs, right? So that the infrastructure has to expose its capabilities. Then you need abstractions over the top that make it simple for developers to take, you know, make use of them. >> See, one of the demonstrations we are doing is around AIOps. And I've had literally here on this floor, two conversations around what I call as network as a platform. Although it sounds like a cliche term, that's exactly what Dave was describing in terms of exposing APIs from the infrastructure and utilizing them. So once you get that data, then now you can do analytics and do machine learning to be able to build models and figure out how you can orchestrate better how you can monetize better, how can how you can utilize better, right? So all of those things become important. It's just not about internal optimization but it's also about how do you expose it to third party ecosystem to translate that into better delivery mechanisms or IOT capability and so on. >> But if they're going to charge me for every API call in the network I'm going to go broke (team laughs) >> And I'm going to get really pissed. I mean, I feel like, I'm just running down, Oracle. IBM tried it. Oracle, okay, they got Java, but they don't they don't have developer jobs. VMware, okay? They got Aria. EMC used to have a thing called code. IBM had to buy Red Hat to get to the developer community. (Lisa laughs) >> So I feel like the telcos don't today have those developer shops. So, so they have to partner. [Azhar] Yes. >> With guys like you and then be more open and and let a zillion flowers bloom or else they're going to get disrupted in a big way and they're going to it's going to be a repeat of the over, over the top in, in in a different model that I can't predict. >> Yeah. >> Absolutely true. I mean, look, they cannot be in the connectivity business. Telcos cannot be just in the connectivity business. It's, I think so, you know, >> Dave Vellante: You had a fry a frozen hand (Dave Daggul laughs) >> off that, you know. >> Well, you know, think about they almost have to go become over the top on themselves, right? That's what the cloud guys are doing, right? >> Yeah. >> They're riding over their backbone that by taking a creating a high level abstraction, they in turn abstract away the infrastructure underneath them, right? And that's really the end game >> Right? >> Dave Vellante: Yeah. >> Is because now, >> they're over the top it's their network, it's their infrastructure, right? They don't want to become bid pipes. >> Yep. >> Now you, they can take OpenShift, run that in any cloud. >> Yep. >> Right? >> You can run that in hybrid cloud, enterprise web can do the application layer configuration and management. And together we're running, you know, OSI layers one through seven, east to west, north to south. We're running across the the RAN, the core and the transport. And that is telco super cloud, my friend. >> Yeah. Well, >> (Dave Duggal laughs) >> I'm dominating the conversation cause I love talking super cloud. >> I knew you would. >> So speaking of super superpowers, when you're in customer or prospective customer conversations with providers and they've got, obviously they're they're in this transformative state right now. How, what do you describe as the superpower between Red Hat and EnterpriseWeb in terms of really helping these Telcos transforms. But at the end of the day, the connectivity's there the end user gets what they want, which is I want this to work wherever I am. >> Yeah, yeah. That's a great question, Lisa. So I think the way you could look at it is most software has, has been evolved to be specialized, right? So in Telcos' no different, right? We have this in the enterprise, right? All these specialized stacks, all these components that they wire together in the, in you think of Telco as a sort of a super set of enterprise problems, right? They have all those problems like magnified manyfold, right? And so you have specialized, let's say orchestrators and other tools for every Telco domain for every Telco layer. Now you have a zoo of orchestrators, right? None of them were designed to work together, right? They all speak a specific language, let's say quote unquote for doing a specific purpose. But everything that's interesting in the 21st century is across layers and across domains, right? If a siloed static application, those are dead, right? Nobody's doing those anymore. Even developers don't do those developers are doing composition today. They're not doing, nobody wants to hear about a 6 million lines of code, right? They want to hear, "How did you take these five things and bring 'em together for productive use?" >> Lisa: Right. How did you deliver faster for my enterprise? How did you save me money? How did you create business value? And that's what we're doing together. >> I mean, just to add on to Dave, I was talking to one of the providers, they have more than 30,000 nodes in their infrastructure. When I say no to your servers running, you know, Kubernetes,running open stack, running different components. If try managing that in one single entity, if you will. Not possible. You got to fragment, you got to segment in some way. Now the question is, if you are not exposing that particular infrastructure and the appropriate KPIs and appropriate things, you will not be able to efficiently utilize that across the board. So you need almost a construct that creates like a manager of managers, a hierarchical structure, which would allow you to be more intelligent in terms of how you place those, how you manage that. And so when you ask the question about what's the secret sauce between the two, well this is exactly where EnterpriseWeb brings in that capability to analyze information, be more intelligent about it. And what we do is provide an abstraction of the cloud layer so that they can, you know, then do the right job in terms of making sure that it's appropriate and it's consistent. >> Consistency is key. Guys, thank you so much. It's been a pleasure really digging through EnterpriseWeb. >> Thank you. >> What you're doing >> with Red Hat. How you're helping the organization transform and Supercloud, we can't forget Supercloud. (Dave Vellante laughs) >> Fight Supercloud. Guys, thank you so much for your time. >> Thank you so much Lisa. >> Thank you. >> Thank you guys. >> Very nice. >> Lisa: We really appreciate it. >> For our guests and for Dave Vellante, I'm Lisa Martin. You're watching theCUBE, the leader in live tech coverage coming to you live from MWC 23. We'll be back after a short break.
SUMMARY :
that drive human progress. the challenges, the opportunities. have you on the program. What's the business model? So the historic middleware So the real challenge for happening in the industry What's the landscape look like? You need the ability to orchestrate them. You could say Supercloud. And then how do you orchestrate all And by the way Thanks to, you know, And, and so that's what you guys do. even the name EnterpriseWeb, you know that's an O-RAN deployment. of that is the customer but you guys sell to them. on the ISV ecosystem to be able take a piece of the action." So that the infrastructure has and figure out how you And I'm going to get So, so they have to partner. the over, over the top in, in in the connectivity business. They don't want to become bid pipes. OpenShift, run that in any cloud. And together we're running, you know, I'm dominating the conversation the end user gets what they want, which is And so you have specialized, How did you create business value? You got to fragment, you got to segment Guys, thank you so much. and Supercloud, we Guys, thank you so much for your time. to you live from MWC 23.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Dave Duggal | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Telcos | ORGANIZATION | 0.99+ |
Telco | ORGANIZATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Fortnet | ORGANIZATION | 0.99+ |
Keysight | ORGANIZATION | 0.99+ |
EnterpriseWeb | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
21st century | DATE | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
two projects | QUANTITY | 0.99+ |
Telcos' | ORGANIZATION | 0.99+ |
Latin America | LOCATION | 0.99+ |
EMC | ORGANIZATION | 0.99+ |
Dave Daggul | PERSON | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
second project | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Fort Net | ORGANIZATION | 0.99+ |
Barcelona, Spain | LOCATION | 0.99+ |
telco | ORGANIZATION | 0.99+ |
more than 30,000 nodes | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
OpenShift | TITLE | 0.99+ |
Java | TITLE | 0.99+ |
three | QUANTITY | 0.99+ |
KX | ORGANIZATION | 0.99+ |
Azhar Sayeed | PERSON | 0.98+ |
One | QUANTITY | 0.98+ |
Tech Mahindra | ORGANIZATION | 0.98+ |
two conversations | QUANTITY | 0.98+ |
yesterday | DATE | 0.98+ |
five things | QUANTITY | 0.98+ |
telcos | ORGANIZATION | 0.97+ |
four days | QUANTITY | 0.97+ |
Azhar | PERSON | 0.97+ |
Scott Walker, Wind River & Gautam Bhagra, Dell Technologies | MWC Barcelona 2023
(light music) >> Narrator: theCUBE's live coverage is made possible by funding from Dell Technologies. Creating technologies that drive human progress. (upbeat music) >> Welcome back to Spain everyone. Lisa Martin here with theCUBE Dave Vellante, my co-host for the next four days. We're live in Barcelona, covering MWC23. This is only day one, but I'll tell you the theme of this conference this year is velocity. And I don't know about you Dave, but this day is flying by already. This is ecosystem day. We're going to have a great discussion on the ecosystem next. >> Well we're seeing the disaggregation of the hardened telco stack, and that necessitates an ecosystem open- we're going to talk about Open RAN, we've been talking about even leading up to the show. It's a critical technology enabler and it's compulsory to have an ecosystem to support that. >> Absolutely compulsory. We've got two guests here joining us, Gautam Bhagra, Vice President partnerships at Dell, and Scott Walker, Vice President of global Telco ecosystem at Wind River. Guys, welcome to the program. >> Nice to be here. >> Thanks For having us. >> Thanks for having us. >> So you've got some news, this is day one of the conference, there's some news, Gautam, and let's start with you, unpack it. >> Yeah, well there's a lot of news, as you know, on Dell World. One of the things we are very excited to announce today is the launch of the Open Telecom Ecosystems Community. I think Dave, as you mentioned, getting into an Open RAN world is a challenge. And we know some of the challenges that our customers face. To help solve for those challenges, Dell wants to work with like-minded partners and customers to build innovative solutions, and join go-to-market. So we are launching that today. Wind River is one of our flagship partners for that, and I'm excited to be here to talk about that as well. >> Can you guys talk a little bit about the partnership, maybe a little bit about Wind River so the audience gets that context? >> Sure, absolutely, and the theme of the show, Velocity, is what this partnership is all about. We create velocity for operators if they want to adopt Open RAN, right? We simplify it. Wind River as a company has been around for 40 years. We were part of Intel at one point, and now we're independent, owned by a company called Aptiv. And with that we get another round of investment to help continue our acceleration into this market. So, the Dell partnership is about, like I said, velocity, accelerating the adoption. When we talk to operators, they have told us there are many roadblocks that they face, right? Like systems integration, operating at scale. 'Cause when you buy a traditional radio access network solution from a single supplier, it's very easy. It's works, it's been tested. When you break these components apart and disaggregate 'em, as we talked about David, it creates integration points and support issues, right? And what Dell and Wind River have done together is created a cloud infrastructure solution that could host a variety of RAN workloads, and essentially create a two layer cake. What we're, overall, what we're trying to do is create a traditional RAN experience, with the innovation agility and flexibility of Open RAN. And that's really what this partnership does. >> So these work, this workload innovation is interesting to me because you've got now developers, you know, the, you know, what's the telco developer look like, you know, is to be defined, right? I mean it's like this white sheet of paper that can create all this innovation. And to do that, you've got to have, as I said earlier, an ecosystem. But you've got now, I'm interested in your Open RAN agenda and how you see that sort of maturity model taking place. 'Cause today, you got disruptors that are going to lean right in say "Hey, yeah, that's great." The traditional carriers, they have to have a, you know, they have to migrate, they have to have a hybrid world. We know that takes time. So what's that look like in the marketplace today? >> Yeah, so I mean, I can start, right? So from a Dell's perspective, what we see in the market is yes, there is a drive towards, everyone understands the benefits of being open, right? There's the agility piece, the innovation piece. That's a no-brainer. The question is how do we get there? And I think that's where partnerships become critical to get there, right? So we've been working with partners like Wind River to build solutions that make it easier for customers to start adopting some of the foundational elements of an open network. The, one of the purposes in the agenda of building this community is to bring like-minded developers, like you said like we want those guys to come and work with the customers to create new solutions, and come up with something creative, which no one's even thought about, that accelerates your option even quicker, right? So that's exactly what we want to do as well. And that's one of the reasons why we launched the community. >> Yeah, and what we find with a lot of carriers, they are used to buying, like I said, traditional RAN solutions which are provided from a single provider like Erickson or Nokia and others, right? And we break this apart, and you cloudify that network infrastructure, there's usually a skills gap we see at the operator level, right? And so from a developer standpoint, they struggle with having the expertise in order to execute on that. Wind River helps them, working with companies like Dell, simplify that bottom portion of the stack, the infrastructure stack. So, and we lifecycle manage it, we test- we're continually testing it, and integrating it, so that the operator doesn't have to do that. In addition to that, wind River also has a history and legacy of working with different RAN vendors, both disruptors like Mavenir and Parallel Wireless, as well as traditional RAN providers like Samsung, Erickson, and others soon to be announced. So what we're doing on the northbound side is making it easy by integrating that, and on the southbound side with Dell, so that again, instead of four or five solutions that you need to put together, it's simply two. >> And you think about today how we- how you consume telco services are like there's these fixed blocks of services that you can buy, that has to change. It's more like the, the app stores. It's got to be an open marketplace, and that's where the innovation's going to come in, you know, from the developers, you know, top down maybe. I don't know, how do you see that maturity model evolving? People want to know how long it's going to take. So many questions, when will Open RAN be as reliable. Does it even have to be? You know, so many interesting dynamics going on. >> Yeah, and I think that's something we at Dell are also trying to find out, right? So we have been doing a lot of good work here to help our customers move in that direction. The work with Dish is an example of that. But I think we do understand the challenges as well in terms of getting, adopting the technologies, and adopting the innovation that's being driven by Open. So one of the agendas that we have as a company this year is to work with the community to drive this a lot further, right? We want to have customers adopt the technology more broadly with the tier one, tier two telcos globally. And our sales organizations are going to be working together with Wind Rivers to figure out who's the right set of customers to have these conversations with, so we can drop, drive, start driving this agenda a lot quicker than what we've seen historically. >> And where are you having those customer conversations? Is that at the operator level, is it higher, is it both? >> Well, all operators are deploying 5G in preparation for 6G, right? And we're all looking for those killer use cases which will drive top line revenue and not just make it a TCO discussion. And that starts at a very basic level today by doing things like integrating with Juniper, for their cloud router. So instead of at the far edge cell site, having a separate device that's doing the routing function, right? We take that and we cloudify that application, run it on the same server that's hosting the RAN applications, so you eliminate a device and reduce TCO. Now with Aptiv, which is primarily known as an automotive company, we're having lots of conversations, including with Dell and Intel and others about vehicle to vehicle communication, vehicle to anything communication. And although that's a little bit futuristic, there are shorter term use cases that, like, vehicle to vehicle accident avoidance, which are going to be much nearer term than autonomous driving, for example, which will help drive traffic and new revenue streams for operators. >> So, oh, that's, wow. So many other things (Scott laughs) that's just opened up there too. But I want to come back to, sort of, the Open RAN adoption. And I think you're right, there's a lot of questions that that still have to be determined. But my question is this, based on your knowledge so far does it have to be as hardened and reliable, obviously has to be low latency as existing networks, or can flexibility, like the cloud when it first came out, wasn't better than enterprise IT, it was just more flexible and faster, and you could rent it. And, is there a similar dynamic here where it doesn't have to replicate the hardened stack, it can bring in new benefits that drive adoption, what are your thoughts on that? >> Well there's a couple of things on that, because Wind River, as you know, where our legacy and history is in embedded devices like F-15 fighter jets, right? Or the Mars Rover or the James Web telescope, all run Wind River software. So, we know about can't fail ultra reliable systems, and operators are not letting us off the hook whatsoever. It has to be as hardened and locked down, as secure as a traditional RAN environment. Otherwise they will (indistinct). >> That's table stakes. >> That's table stakes that gets us there. And when River, with our legacy and history, and having operator experience running live commercial networks with a disaggregated stack in the tens of thousands of nodes, understand what this is like because they're running live commercial traffic with live customers. So we can't fail, right? And with that, they want their cake and eat it too, right? Which is, I want ultra reliable, I want what I have today, but I want the agility and flexibility to onboard third party apps. Like for example, this JCNR, this Juniper Cloud-Native Router. You cannot do something as simple as that on a traditional RAN Appliance. In an open ecosystem you can take that workload and onboard it because it is an open ecosystem, and that's really one of the true benefits. >> So they want the mainframe, but they want (Scott laughs) the flexibility of the developer cloud, right? >> That's right. >> They want their, have their cake eat it too and not gain weight. (group laughs) >> Yeah I mean David, I come from the public cloud world. >> We all don't want to do that. >> I used to work with a public cloud company, and nine years ago, public cloud was in the same stage, where you would go to a bank, and they would be like, we don't trust the cloud. It's not secure, it's not safe. It was the digital natives that adopted it, and that that drove the industry forward, right? And that's where the enterprises that realized that they're losing business because of all these innovative new companies that came out. That's what I saw over the last nine years in the cloud space. I think in the telco space also, something similar might happen, right? So a lot of this, I mean a lot of the new age telcos are understanding the value, are looking to innovate are adopting the open technologies, but there's still some inertia and hesitancy, for the reasons as Scott mentioned, to go there so quickly. So we just have to work through and balance between both sides. >> Yeah, well with that said, if there's still some inertia, but there's a theme of velocity, how do you help organizations balance that so they trust evolving? >> Yeah, and I think this is where our solution, like infrastructure block, is a foundational pillar to make that happen, right? So if we can take away the concerns that the organizations have in terms of security, reliability from the fundamental elements that build their infrastructure, by working with partners like Wind River, but Dell takes the ownership end-to-end to make sure that service works and we have those telco grade SLAs, then the telcos can start focusing on what's next. The applications and the customer services on the top. >> Customer service customer experience. >> You know, that's an interesting point Gautam brings up, too, because support is an issue too. We all talk about when you break these things apart, it creates integration points that you need to manage, right? But there's also, so the support aspect of it. So imagine if you will, you had one vendor, you have an outage, you call that one vendor, one necktie to choke, right, for accountability for the network. Now you have four or five vendors that you have to work. You get a lot of finger pointing. So at least at the infrastructure layer, right? Dell takes first call support for both the hardware infrastructure and the Wind River cloud infrastructure for both. And we are training and spinning them up to support, but we're always behind them of course as well. >> Can you give us a favorite customer example of- that really articulates the value of the partnership and the technologies that it's delivering to customers? >> Well, Infra Block- >> (indistinct) >> Is quite new, and we do have our first customer which is LG U plus, which was announced yesterday. Out of Korea, small customer, but a very important one. Okay, and I think they saw the value of the integrated system. They don't have the (indistinct) expertise and they're leveraging Dell and Wind River in order to make that happen. But I always also say historically before this new offering was Vodafone, right? Vodafone is a leader in Europe in terms of Open RAN, been very- Yago and Paco have been very vocal about what they're doing in Open RAN, and Dell and Wind River have been there with them every step of the way. And that's what I would say, kind of, led up to where we are today. We learned from engagements like Vodafone and I think KDDI as well. And it got us where we are today and understanding what the operators need and what the impediments are. And this directly addresses that. >> Those are two very different examples. You were talking about TCO before. I mean, so the earlier example is, that's an example to me of a disruptor. They'll take some chances, you know, maybe not as focused on TCO, of course they're concerned about it. Vodafone I would think very concerned about TCO. But I'm inferring from your comments that you're trying to get the industry, you're trying to check the TCO box, get there. And then move on to higher levels of value monetization. The TCO is going to come down to how many humans it takes to run the network, is it not, is that- >> Well a lot of, okay- >> Or is it devices- >> So the big one now, particularly with Vodafone, is energy cost, right? >> Of course, greening the network. >> Two-thirds of the energy consumption in RAN is the the Radio Access Network. Okay, the OPEX, right? So any reductions, even if they're 5% or 10%, can save tens or hundreds of millions of dollars. So we do things creatively with Dell to understand if there's a lot of traffic at the cell site and if it's not, we will change the C state or P state of the server, which basically spins it down, so it's not consuming power. But that's just at the infrastructure layer. Where this gets really powerful is working with the RAN vendors like Samsung and Ericson and others, and taking data from the traffic information there, applying algorithms to that in AI to shut it down and spin it back up as needed. 'Cause the idea is you don't want that thing powered up if there's no traffic on it. >> Well there's a sustainability, ESG, benefit to that, right? >> Yes. >> And, and it's very compute intensive. >> A hundred percent. >> Which is great for Dell. But at the same time, if you're not able to manage that power consumption, the whole thing fails. I mean it's, because there's going to be so much data, and such a intense requirement. So this is a huge issue. Okay, so Scott, you're saying that in the TCO equation, a big chunk is energy consumption? >> On the OPEX piece. Now there's also the CapEx, right? And Open RAN solutions are now, what we've heard from our customers today, are they're roughly at parity. 'Cause you can do things like repurpose servers after the useful life for a lower demand application which helps the TCO, right? Then you have situations like Juniper, where you can take, now software that runs on the same device, eliminating at a whole other device at the cell site. So we're not just taking a server and software point of view, we're taking a whole cell site point of view as it relates to both CapEx and OPEX. >> And then once that infrastructure it really gets adopted, that's when the innovation occurs. The ecosystem comes in. Developers now start to think of new applications that we haven't thought of yet. >> Gautam: Exactly. >> And that's where, that's going to force the traditional carriers to respond. They're responding, but they're doing so very carefully right now, it's understandable why. >> Yeah, and I think you're already seeing some news in the, I mean Nokia's announcement yesterday with the rebranding, et cetera. That's all positive momentum in my opinion, right? >> What'd you think of the logo? >> I love the logo. >> I liked it too. (group laughs) >> It was beautiful. >> I thought it was good. You had the connectivity down below, You need pipes, right? >> Exactly. >> But you had this sort of cool letters, and then the the pink horizon or pinkish, it was like (Scott laughs) endless opportunity. It was good, I thought it was well thought out. >> Exactly. >> Well, you pick up on an interesting point there, and what we're seeing, like advanced carriers like Dish, who has one of the true Open RAN networks, publishing APIs for programmers to build in their 5G network as part of the application. But we're also seeing the network equipment providers also enable carriers do that, 'cause carriers historically have not been advanced in that way. So there is a real recognition that in order for these networks to monetize new use cases, they need to be programmable, and they need to publish standard APIs, so you can access the 5G network capabilities through software. >> Yeah, and the problem from the carriers, there's not enough APIs that the carriers have produced yet. So that's where the ecosystem comes in, is going to >> A hundred percent >> I think there's eight APIs that are published out of the traditional carriers, which is, I mean there's got to be 8,000 for a marketplace. So that's where the open ecosystem really has the advantage. >> That's right. >> That's right. >> That's right. >> Yeah. >> So it all makes sense on paper, now you just, you got a lot of work to do. >> We got to deliver. Yeah, we launched it today. We got to get some like-minded partners and customers to come together. You'll start seeing results coming out of this hopefully soon, and we'll talk more about it over time. >> Dave: Great Awesome, thanks for sharing with us. >> Excellent. Guys, thank you for sharing, stopping by, sharing what's going on with Dell and Wind River, and why the opportunity's in it for customers and the technological evolution. We appreciate it, you'll have to come back, give us an update. >> Our pleasure, thanks for having us. (Group talks over each other) >> All right, thanks guys >> Appreciate it. >> For our guests and for Dave Vellante, I'm Lisa Martin. You're watching theCUBE, Live from MWC23 in Barcelona. theCUBE is the leader in live tech coverage. (upbeat music)
SUMMARY :
that drive human progress. the theme of this conference and it's compulsory to have and Scott Walker, Vice President and let's start with you, unpack it. One of the things we are very excited and the theme of the show, Velocity, they have to have a, you know, And that's one of the reasons the operator doesn't have to do that. from the developers, you and adopting the innovation So instead of at the far edge cell site, that that still have to be determined. Or the Mars Rover or and flexibility to and not gain weight. I come from the public cloud world. and that that drove the that the organizations and the Wind River cloud of the integrated system. I mean, so the earlier example is, and taking data from the But at the same time, if that runs on the same device, Developers now start to think the traditional carriers to respond. Yeah, and I think you're I liked it too. You had the connectivity down below, and then the the pink horizon or pinkish, and they need to publish Yeah, and the problem I mean there's got to be now you just, you got a lot of work to do. and customers to come together. thanks for sharing with us. for customers and the Our pleasure, thanks for having us. Live from MWC23 in Barcelona.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Samsung | ORGANIZATION | 0.99+ |
David | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Nokia | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Erickson | ORGANIZATION | 0.99+ |
Vodafone | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
Barcelona | LOCATION | 0.99+ |
Scott Walker | PERSON | 0.99+ |
Scott | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Mavenir | ORGANIZATION | 0.99+ |
Wind River | ORGANIZATION | 0.99+ |
Parallel Wireless | ORGANIZATION | 0.99+ |
Gautam | PERSON | 0.99+ |
Korea | LOCATION | 0.99+ |
tens | QUANTITY | 0.99+ |
Gautam Bhagra | PERSON | 0.99+ |
four | QUANTITY | 0.99+ |
8,000 | QUANTITY | 0.99+ |
5% | QUANTITY | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
10% | QUANTITY | 0.99+ |
Europe | LOCATION | 0.99+ |
Wind River | ORGANIZATION | 0.99+ |
Aptiv | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
Spain | LOCATION | 0.99+ |
Ericson | ORGANIZATION | 0.99+ |
one vendor | QUANTITY | 0.99+ |
five vendors | QUANTITY | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
Wind Rivers | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
F-15 | COMMERCIAL_ITEM | 0.99+ |
both sides | QUANTITY | 0.99+ |
two guests | QUANTITY | 0.99+ |
Two-thirds | QUANTITY | 0.99+ |
wind River | ORGANIZATION | 0.98+ |
first call | QUANTITY | 0.98+ |
Mohan Rokkam & Greg Gibby | 4th Gen AMD EPYC on Dell PowerEdge: Virtualization
(cheerful music) >> Welcome to theCUBE's continuing coverage of AMD's 4th Generation EPYC launch. I'm Dave Nicholson, and I'm here in our Palo Alto studios talking to Greg Gibby, senior product manager, data center products from AMD, and Mohan Rokkam, technical marketing engineer at Dell. Welcome, gentlemen. >> Mohan: Hello, hello. >> Greg: Thank you. Glad to be here. >> Good to see each of you. Just really quickly, I want to start out. Let us know a little bit about yourselves. Mohan, let's start with you. What do you do at Dell exactly? >> So I'm a technical marketing engineer at Dell. I've been with Dell for around 15 years now and my goal is to really look at the Dell powered servers and see how do customers take advantage of some of the features we have, especially with the AMD EPYC processors that have just come out. >> Greg, and what do you do at AMD? >> Yeah, so I manage our software-defined infrastructure solutions team, and really it's a cradle to grave where we work with the ISVs in the market, so VMware, Nutanix, Microsoft, et cetera, to integrate the features that we're putting into our processors and make sure they're ready to go and enabled. And then we work with our valued partners like Dell on putting those into actual solutions that customers can buy and then we work with them to sell those solutions into the market. >> Before we get into the details on the 4th Generation EPYC launch and what that means and why people should care. Mohan, maybe you can tell us a little about the relationship between Dell and AMD, how that works, and then Greg, if you've got commentary on that afterwards, that'd be great. Yeah, Mohan. >> Absolutely. Dell and AMD have a long standing partnership, right? Especially now with EPYC series. We have had products since EPYC first generation. We have been doing solutions across the whole range of Dell ecosystem. We have integrated AMD quite thoroughly and effectively and we really love how performant these systems are. So, yeah. >> Dave: Greg, what are your thoughts? >> Yeah, I would say the other thing too is, is that we need to point out is that we both have really strong relationships across the entire ecosystem. So memory vendors, the software providers, et cetera, we have technical relationships. We're working with them to optimize solutions so that ultimately when the customer buys that, they get a great user experience right out of the box. >> So, Mohan, I know that you and your team do a lot of performance validation testing as time goes by. I suspect that you had early releases of the 4th Gen EPYC processor technology. What have you been seeing so far? What can you tell us? >> AMD has definitely knocked it out of the park. Time and again, in the past four generations, in the past five years alone, we have done some database work where in five years, we have seen five exit performance. And across the board, AMD is the leader in benchmarks. We have done virtualization where we would consolidate from five into one system. We have world records in AI, we have world records in databases, we have world records in virtualization. The AMD EPYC solutions has been absolutely performant. I'll leave you with one number here. When we went from top of Stack Milan to top of Stack Genoa, we saw a performance bump of 120%. And that number just blew my mind. >> So that prompts a question for Greg. Often we, in industry insiders, think in terms of performance gains over the last generation or the current generation. A lot of customers in the real world, however, are N - 2. They're a ways back, so I guess two points on that. First of all, the kinds of increases the average person is going to see when they move to this architecture, correct me if I'm wrong, but it's even more significant than a lot of the headline numbers because they're moving two generations, number one. Correct me if I'm wrong on that, but then the other thing is the question to you, Greg. I like very long complicated questions, as you can tell. The question is, is it okay for people to skip generations or make the case for upgrades, I guess is the problem? >> Well, yeah, so a couple thoughts on that first too. Mohan talked about that five X over the generation improvements that we've seen. The other key point with that too is that we've made significant process improvements along the way moving to seven nanocomputer to now five nanocomputer and that's really reducing the total amount of power or the performance per watt the customers can realize as well. And when we look at why would a customer want to upgrade, right? And I want to rephrase that as to why aren't you? And there is a real cost of not upgrading. And so when you look at infrastructure, the average age of a server in the data center is over five years old. And if you look at the most popular processors that were sold in that timeframe, it's 8, 10, 12 cores. So now you've got a bunch of servers that you need in order to deliver the applications and meet your SLAs to your end users, and all those servers pull power. They require maintenance. They have the opportunity to go down, et cetera. You got to pay licensing and service and support costs and all those. And when you look at all the costs that roll up, even though the hardware is paid for just to keep the lights on, and not even talking about the soft costs of unplanned downtime, and, "I'm not meeting your SLAs," et cetera, it's very expensive to keep those servers running. Now, if you refresh, and now you have processors that have 32, 64, 96 cores, now you can consolidate that infrastructure and reduce your total power bill. You can reduce your CapEx, you reduce your ongoing OpEx, you improve your performance, and you improve your security profile. So it really is more cost effective to refresh than not to refresh. >> So, Mohan, what has your experience been double clicking on this topic of consolidation? I know that we're going to talk about virtualization in some of the results that you've seen. What have you seen in that regard? Does this favor better consolidation and virtualized environments? And are you both assuring us that the ROI and TCO pencil out on these new big, bad machines? >> Greg definitely hit the nail on the head, right? We are seeing tremendous savings really, if you're consolidating from two generations old. We went from, as I said, five is to one. You're going from five full servers, probably paid off down to one single server. That itself is, if you look at licensing costs, which again, with things like VMware does get pretty expensive. If you move to a single system, yes, we are at 32, 64, 96 cores, but if you compare to the licensing costs of 10 cores, two sockets, that's still pretty significant, right? That's one huge thing. Another thing which actually really drives the thing is we are looking at security, and in today's environment, security becomes a major driving factor for upgrades. Dell has its own setups, cyber-resilient architecture, as we call it, and that really is integrated from processor all the way up into the OS. And those are some of the features which customers really can take advantage of and help protect their ecosystems. >> So what kinds of virtualized environments did you test? >> We have done virtualization across primary codes with VMware, but the Azure Stack, we have looked at Nutanix. PowerFlex is another one within Dell. We have vSAN Ready Nodes. All of these, OpenShift, we have a broad variety of solutions from Dell and AMD really fits into almost every one of them very well. >> So where does hyper-converged infrastructure fit into this puzzle? We can think of a server as something that contains not only AMD's latest architecture but also latest PCIe bus technology and all of the faster memory, faster storage cards, faster nicks, all of that comes together. But how does that play out in Dell's hyper-converged infrastructure or HCI strategy? >> Dell is a leader in hyper-converged infrastructure. We have the very popular VxRail line, we have the PowerFlex, which is now going into the AWS ecosystem as well, Nutanix, and of course, Azure Stack. With all these, when you look at AMD, we have up to 96 cores coming in. We have PCIe Gen 5 which means you can now connect dual port, 100 and 200 gig nicks and get line rate on those so you can connect to your ecosystem. And I don't know if you've seen the news, 200, 400 gig routers and switchers are selling out. That's not slowing down. The network infrastructure is booming. If you want to look at the AI/ML side of things, the VDI side of things, accelerator cards are becoming more and more powerful, more and more popular. And of course they need that higher end data path that PCIe Gen 5 brings to the table. GDDR5 is another huge improvement in terms of performance and latencies. So when we take all this together, you talk about hyper-converged, all of them add into making sure that A, with hyper-converged, you get ease of management, but B, just 'cause you have ease of management doesn't mean you need to compromise on anything. And the AMD servers effectively are a no compromise offering that we at Dell are able to offer to our customers. >> So Greg, I've got a question a little bit from left field for you. We covered Supercompute Conference 2022. We were in Dallas a couple of weeks ago, and there was a lot of discussion of the current processor manufacturer battles, and a lot of buzz around 4th Gen EPYC being launched and what's coming over the next year. Do you have any thoughts on what this architecture can deliver for us in terms of things like AI? We talk about virtualization, but if you look out over the next year, do you see this kind of architecture driving significant change in the world? >> Yeah, yeah, yeah, yeah. It has the real potential to do that from just the building blocks. So we have our chiplet architecture we call it. So you have an IO die and then you have your core complexes that go around that. And we integrate it all with our infinity fabric. That architecture allows you, if we wanted to, replace some of those CCDs with specific accelerators. And so when we look two, three, four years down the road, that architecture and that capability already built into what we're delivering and can easily be moved in. We just need to make sure that when you look at doing that, that the power that's required to do that and the software, et cetera, and those accelerators actually deliver better performance as a dedicated engine versus just using standard CPUs. The other things that I would say too is if you look at emerging workloads. So data center modernization is one of the buzzwords in cloud native, right? And these container environments, well, AMD'S architecture really just screams support for those type of environments, right? Where when you get into these larger core accounts and the consolidation that Mohan talked about. Now when I'm in a container environment, that blast radius so a lot of customers have concerns around, "Hey, having a single point of failure and having more than X number of cores concerns me." If I'm in containers, that becomes less of a concern. And so when you look at cloud native, containerized applications, data center modernization, AMD's extremely well positioned to take advantage of those use cases as well. >> Yeah, Mohan, and when we talk about virtualization, I think sometimes we have to remind everyone that yeah, we're talking about not only virtualization that has a full-blown operating system in the bucket, but also virtualization where the containers have microservices and things like that. I think you had something to add, Mohan. >> I did, and I think going back to the accelerator side of business, right? When we are looking at the current technology and looking at accelerators, AMD has done a fantastic job of adding in features like AVX-512, we have the bfloat16 and eight features. And some of what these do is they're effectively built-in accelerators for certain workloads especially in the AI and media spaces. And in some of these use cases we look at, for example, are inference. Traditionally we have used external accelerator cards, but for some of the entry level and mid-level use cases, CPU is going to work just fine especially with the newer CPUs that we are seeing this fantastic performance from. The accelerators just help get us to the point where if I'm at the edge, if I'm in certain use cases, I don't need to have an accelerator in there. I can run most of my inference workloads right on the CPU. >> Yeah, yeah. You know the game. It's an endless chase to find the bottleneck. And once we've solved the puzzle, we've created a bottleneck somewhere else. Back to the supercompute conversations we had, specifically about some of the AMD EPYC processor technology and the way that Dell is packaging it up and leveraging things like connectivity. That was one of the things that was also highlighted. This idea that increasingly connectivity is critically important, not just for supercomputing, but for high-performance computing that's finding its way out of the realms of Los Alamos and down to the enterprise level. Gentlemen, any more thoughts about the partnership or maybe a hint at what's coming in the future? I know that the original AMD announcement was announcing and previewing some things that are rolling out over the next several months. So let me just toss it to Greg. What are we going to see in 2023 in terms of rollouts that you can share with us? >> That I can share with you? Yeah, so I think look forward to see more advancements in the technology at the core level. I think we've already announced our product code name Bergamo, where we'll have up to 128 cores per socket. And then as we look in, how do we continually address this demand for data, this demand for, I need actionable insights immediately, look for us to continue to drive performance leadership in our products that are coming out and address specific workloads and accelerators where appropriate and where we see a growing market. >> Mohan, final thoughts. >> On the Dell side, of course, we have four very rich and configurable options with AMD EPYC servers. But beyond that, you'll see a lot more solutions. Some of what Greg has been talking about around the next generation of processors or the next updated processors, you'll start seeing some of those. and you'll definitely see more use cases from us and how customers can implement them and take advantage of the features that. It's just exciting stuff. >> Exciting stuff indeed. Gentlemen, we have a great year ahead of us. As we approach possibly the holiday seasons, I wish both of you well. Thank you for joining us. From here in the Palo Alto studios, again, Dave Nicholson here. Stay tuned for our continuing coverage of AMD's 4th Generation EPYC launch. Thanks for joining us. (cheerful music)
SUMMARY :
talking to Greg Gibby, Glad to be here. What do you do at Dell exactly? of some of the features in the market, so VMware, on the 4th Generation EPYC launch the whole range of Dell ecosystem. is that we need to point out is that of the 4th Gen EPYC processor technology. Time and again, in the the question to you, Greg. of servers that you need in some of the results that you've seen. really drives the thing is we have a broad variety and all of the faster We have the very popular VxRail line, over the next year, do you that the power that's required to do that in the bucket, but also but for some of the entry I know that the original AMD in the technology at the core level. and take advantage of the features that. From here in the Palo Alto studios,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Greg | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Greg Gibby | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
8 | QUANTITY | 0.99+ |
Mohan | PERSON | 0.99+ |
32 | QUANTITY | 0.99+ |
Mohan Rokkam | PERSON | 0.99+ |
100 | QUANTITY | 0.99+ |
200 | QUANTITY | 0.99+ |
10 cores | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
Dallas | LOCATION | 0.99+ |
120% | QUANTITY | 0.99+ |
two sockets | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
12 cores | QUANTITY | 0.99+ |
two generations | QUANTITY | 0.99+ |
2023 | DATE | 0.99+ |
five | QUANTITY | 0.99+ |
64 | QUANTITY | 0.99+ |
200 gig | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
five full servers | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
two points | QUANTITY | 0.99+ |
400 gig | QUANTITY | 0.99+ |
EPYC | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
one system | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
Los Alamos | LOCATION | 0.99+ |
next year | DATE | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
two generations | QUANTITY | 0.99+ |
four years | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
Azure Stack | TITLE | 0.98+ |
five nanocomputer | QUANTITY | 0.98+ |
Venkat Venkataramani, Rockset | AWS re:Invent 2022 - Global Startup Program
>>And good afternoon. Welcome back here on the Cub as to continue our coverage at aws Reinvent 22, win the Venetian here in Las Vegas, day two, it's Wednesday. Thanks. Still rolling. Quite a along. We have another segment for you as part of the Global Startup program, which is under the AWS Startup Showcase. I'm joined now by Vink at Viera, who is the CEO and co-founder of R Set. And good to see you, >>Sir. Thanks for having me here. Yeah, >>No, a real pleasure. Looking forward to it. So first off, for some of, for yours who might not be familiar with Roxette, I know you've been on the cube a little bit, so you're, you're an alum, but, but why don't you set the stage a little bit for Rock set and you know, where you're engaged with in terms of, with aws? >>Definitely. Rock Set is a realtime analytics database that is built for the cloud. You know, we make realtime applications possible in the cloud. You know, realtime applications need high concurrency, low latency query processing data needs to be fresh, your analytic needs to be fast. And, you know, we built on aws and that's why we are here. We are very, very proud partners of aws. We are in the AWS Accelerate program, and also we are in the startup program of aws. We are strategic ISV partner. And so yeah, we make real time analytics possible without all the cost and complexity barriers that are usually associated with it. And very, very happy to be part of this movement from batch to real time that is happening in the world. >>Right. Which is certainly an exciting trend. Right. I know great news for you, you made news yesterday, had an announcement involved with the intel with aws, who wants to share some of that >>With us too? Definitely. So, you know, one, one question that I always ask people is like, you know, if you go perspective that I share is like, if you go ask a hundred people, do you want fast analytics on fresh data or slow analytics on stale data? You know, a hundred out of a hundred would say fast and fresh, right? Sure. So then the question is, why hasn't this happened already? Why is this still a new trend that is emerging as opposed to something that everybody's taking for granted? It really comes down to compute efficiency, right? I think, you know, at the end of the day, real time analytics was always in using, you know, technologies that are, let's say 10 years ago using let's say processors that were available 10 years ago to, you know, three cloud, you know, days. There was a lot of complexity barriers associated with realtime analytics and also a lot of cost and, and performance barriers associated with it. >>And so Rox said from the, you know, from the very beginning, has been obsessing about building the most compute efficient realtime database in the world. And, you know, AWS on one hand, you know, allows us to make a consumption based pricing model. So you only pay for what you use. Sure. And that shatters all the cost barriers. But in terms of computer efficiency, what we announced yesterday is the Intel's third generation Zon scalable processors, it's code named Intel Ice Lake. When we port it over Rock said to that architecture, taking advantage of some of the instructions sets that Intel has, we got an 84% performance boost, 84, 84, 84. >>It's, it's incredible, right? >>It's, it's an incredible charts, it's an incredible milestone. It reduces the barrier even more in terms of cost and, you know, and, and pushes the efficiency and sets a, a really new record for how efficient realtime, you know, data processing can be in the cloud. And, and it's very, very exciting news. And so we used to benchmark ourselves against some of our other, you know, realtime, you know, did up providers and we were already faster and now we've set a, a much, much higher bar for other people to follow. >>Yep. And, and so what is, or what was it about real time that, that, you know, was such a barrier because, and now you've got the speed of, of course, obviously, and maybe that's what it was, but I think cost is probably part of that too, right? That's all part of that equation. I mean, real time, so elusive. >>Yeah. So real time has this inherent pattern that your data never stops coming. And when your data never stops coming, and you can now actually do analytics on that. Now, initially people start with saying, oh, I just want a real time dashboard. And then very quickly they realize, well, the dashboard is actually in real time. I'm not gonna be staring at the 24 7. Can you tap on my shoulder when something is off, something needs to be looked at. So in which case you're constantly also asking the question, is everything okay? Is everything all right? Do I need to, is is that something that I need to be, you know, double clicking on and, and following up on? So essentially very quickly in real time analytics, what happens is your queries never stop. The questions that you're asking on your data never stops. And it's often a program asking the question to detect anomalies and things like that. >>And your data never stops coming. And so compute is running 24 7. If you look at traditional data warehouses and data lakes, they're not really optimized for these kinds of workloads. They're optimized to store massive volumes of data and in a storage efficient format. And when an analyst comes and asks a question to generate a report, you can spin up a whole bunch of compute, generate the report and tear it all down when you're done. Well, that is not compute running 24 7 to continuously, you know, you know, keep ingesting the data or continuously keep answering questions. So the compute efficiency that is needed is, is much, much, much higher. Right? And that is why, you know, Rox was born. So from the very beginning, we're only built, you know, for these use cases, we have a, an extremely powerful SQL engine that can give you full feature SQL analytics in a very, very compute efficient way in the cloud. >>Right. So, so let's talk about the leap that you've made, say in the last two years and, and, and what's been the spur of that? What has been allowed you to, to create this, you know, obviously a, a different kind of an array for your customers from which to choose, but, but what's been the spark you think >>We touched upon this a little earlier, right? This spark is really, you know, the world going from batch to real time. So if you look at mainstream adoption of technologies like Apache, Kafka and Confluent doing a really good job at that. In, in, in growing that community and, and use cases, now businesses are now acquiring business data, really important business data in real time. Now they want to operationalize it, right? So, you know, extract based static reports and bi you know, business intelligence is getting replaced in all modern enterprises with what we call operational intelligence, right? Don't tell me what happened last quarter and how to plan this quarter better. Tell me what's happening today, what's happening right now. And it's, it's your business operations using data to make day to day decisions better that either grows your top line, compresses your bottom line, eliminates risk that are inherently creeping up in your business. >>Sure. You know, eliminate potential churn from a customer or fraud, you know, deduction and, and getting on top of, you know, that, you know, a minute into this, into, into an outage as opposed to an hour into the outage. Right? And so essentially I think businesses are now realizing that operational intelligence and operational analytics really, you know, allows them to leverage data and especially real time data to make their, you know, to grow their businesses faster and more efficiently. And especially in this kind of macro environment that is, you know, more important to have better unit economics in your business than ever before. Sure. And so that is really, I think that is the real market movement happening. And, and we are here to just serve that market. We are making it much, much easier for companies that have already adopted, you know, streaming technologies like Kafka and, and, and knows Canis MSK and all these technologies. Now businesses are acquiring these data in real time now. They can also get realtime analytics on the other end of it. Sure. >>You know, you just touched on this and, and I'd like to hear your thoughts about this, about, about the economic environment because it does drive decisions, right? And it does motivate people to look for efficiencies and maybe costs, you know, right. Cutting costs. What are you seeing right now in terms of that, that kind of looming influence, right? That the economy can have in terms of driving decisions about where investments are being made and what expectations are in terms of delivering value, more value for the buck? >>Exactly. I think we see across the board, all of our customers come back and tell us, we don't want to manage data infrastructure and we don't want to do kind of DIY open source clusters. We don't wanna manage and scale and build giant data ops and DevOps teams to manage that, because that is not really, you know, in their business. You know, we have car rental companies want to be better at car rentals, we want airlines to be a better airline, and they don't, don't want their, you know, a massive investment in DevOps and data ops, which is not really their core business. And they really want to leverage, you know, you know, fully managed and, you know, cloud offerings like Rock said, you know, built on aws, massively scalable in the cloud with zero operational overhead, very, very easy to get started and scale. >>And so that completely removes all the operational overhead. And so they can invest the resources they have, the manpower, they have, the calories that they have on actually growing their businesses because that is what really gonna allow them to have better unit economics, right? So everybody that is on my payroll is helping me grow my top line or shrink my bottom line, eliminate risk in my business and, and, and, and churn and, and fraud and other, and eliminate all those risks that are inherent in my business. So, so that is where I think a lot of the investments going. So gone are the days where, you know, you're gonna have these in like five to 10% team managing a very hard to operate, you know, open source data management clusters on EC two nodes in, in AWS and, and kind of DIYing it their way because those 10 people, you know, if all they do is just operational maintenance of infrastructure, which is a means to an end, you're way better off, you know, using a cloud, you know, a bond in the cloud built for the cloud solution like rock and eliminate all that cost and, and replace that with an operationally much, much simpler, you know, system to op, you know, to to work with such as, such as rock. >>So that is really the big trend that we are seeing why, you know, not only real time is going more and more mainstream cloud native solutions or the real future even when it comes to real time because the complexity barrier needs to be shattered and only cloud native solutions can actually, >>You get the two Cs cost and complexity, right. That you, you need to address. Exactly. Yeah, for sure. You know, what is it about building trust with your, with your clients, with your partners? Because you, you're talking about this cloud environment that, that everyone is talking about, right? Not everyone's made that commitment. There are still some foot draggers out there. How are you going about establishing confidence and establishing trust and, and, and providing them with really concrete examples of the values and the benefits that you can provide, you know, with, with these opportunities? >>So, you know, I grew up, so there's a few ways to to, to answer this question. I'll, I'll, I'll come, I'll cover all the angles. So in, in order to establish trust, you have to create value. They, you know, your customer has to see that with you. They were able to solve the problem faster, better, cheaper, and they're able to, you know, have a, the business impact they were looking for, which is why they started the project in the first place. And so establishing that and proving that, I think there's no equivalence to that. And, you know, I grew up at, at, you know, at Facebook back in the day, you know, I was managing online data infrastructure, okay. For Facebook from 2007 and 2015. And internally we always had this kind of culture of all the product teams building on top of the infrastructure that my team was responsible for. >>And so they were not ever, there was never a, a customer vendor relationship internally within Facebook that we're all like, we're all part of the same team. We're partnering here to have you, you know, to help you have a successful product launch. There's a very similar DNA that, that exists in Rock said, when our customers work with us and they come to us and we are there to make them successful, our consumption based pricing model also forces us to say they're not gonna really use Rock said and consume more. I mean, we don't make money until they consume, right? And so their success is very much integral part of our, our success. And so that I think is one really important angle on, you know, give us a shot, come and do an evaluation, and we will work with you to build the most efficient way to solve your problem. >>And then when you succeed, we succeed. So that I think is a very important aspect. The second one is AWS partnership. You know, we are an ISV partner, you know, AWS a lot of the time. That really helps us establish trust. And a lot of the time, one of the, the, the people that they look up to, when a customer comes in saying, Hey, what is, who is Rock? Said? You know, who are your friends? Yeah. Who are your friends? And then, you know, and then the AWS will go like, oh, you know, we'll tell you, you know, all these other successful case studies that R has, you know, you know, built up on, you know, the world's largest insurance provider, Europe's largest insurance provider. We have customers like, you know, JetBlue Airlines to Klarna, which is a big bator company. And so, so all these case studies help and, and, and, and platform and partners like AWS helps us, helps you amplify that, that, you know, and, and, and, and, and give more credibility. And last but not least, compliance matters. You know, being Soto type two compliant is, is a really important part of establishing trust. We are hip hop compliant now so that, you know, we can, you know, pi I phi data handling that. And so I think that will continue to be a part, a big part of our focus in improving the security, you know, functionality and, and capabilities that R set has in the cloud, and also compliance and, and the set of com, you know, you know, standards that we are gonna be compliant against. >>Well, I'm glad you hit on the AWS too, cause I did wanna bring that up. I, I appreciate that and I know they appreciate the relationship as well. Thanks for the time here. It's been a pleasure. Awesome. Learning about Rockette and what you're up to. Thank you. >>You bet. >>It's a pleasure. Thank you. Vi ka. All right. You are watching the cube coverage here at AWS Reinvent 22. And on the cube, of course, the leader, the leader in high tech coverage.
SUMMARY :
We have another segment for you as part of the Global Startup program, which is Yeah, but why don't you set the stage a little bit for Rock set and you know, where you're engaged with in terms of, And, you know, I know great news for you, you made news yesterday, you know, three cloud, you know, days. And so Rox said from the, you know, from the very beginning, has been obsessing about building benchmark ourselves against some of our other, you know, realtime, you know, did up providers That's all part of that equation. you know, double clicking on and, and following up on? And that is why, you know, to create this, you know, obviously a, a different kind of an array for your customers from which This spark is really, you know, the world going from batch you know, deduction and, and getting on top of, you know, that, you know, a minute into this, maybe costs, you know, right. And they really want to leverage, you know, you know, and, and replace that with an operationally much, much simpler, you know, system to op, that you can provide, you know, with, with these opportunities? at, you know, at Facebook back in the day, you know, I was managing online data infrastructure, you know, give us a shot, come and do an evaluation, and we will work with you to build the most efficient way and the set of com, you know, you know, standards that we are gonna be compliant against. Well, I'm glad you hit on the AWS too, cause I did wanna bring that up. And on the cube, of course, the leader, the leader in high
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
2007 | DATE | 0.99+ |
2015 | DATE | 0.99+ |
84% | QUANTITY | 0.99+ |
Venkat Venkataramani | PERSON | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
84 | QUANTITY | 0.99+ |
Wednesday | DATE | 0.99+ |
last quarter | DATE | 0.99+ |
10 years ago | DATE | 0.99+ |
Vink | PERSON | 0.99+ |
JetBlue Airlines | ORGANIZATION | 0.99+ |
10 people | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
five | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Ice Lake | COMMERCIAL_ITEM | 0.99+ |
Europe | LOCATION | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
10% | QUANTITY | 0.98+ |
Apache | ORGANIZATION | 0.98+ |
aws | ORGANIZATION | 0.98+ |
two | QUANTITY | 0.98+ |
Vi ka | PERSON | 0.98+ |
Rock | PERSON | 0.97+ |
Rox | PERSON | 0.97+ |
Confluent | ORGANIZATION | 0.97+ |
an hour | QUANTITY | 0.96+ |
R Set | ORGANIZATION | 0.96+ |
one question | QUANTITY | 0.96+ |
Klarna | ORGANIZATION | 0.95+ |
24 7 | QUANTITY | 0.95+ |
Rockset | ORGANIZATION | 0.94+ |
this quarter | DATE | 0.93+ |
day two | QUANTITY | 0.93+ |
24 7 | OTHER | 0.92+ |
SQL | TITLE | 0.91+ |
intel | ORGANIZATION | 0.91+ |
second one | QUANTITY | 0.9+ |
Rock | ORGANIZATION | 0.89+ |
DevOps | TITLE | 0.88+ |
Viera | ORGANIZATION | 0.87+ |
Startup Showcase | EVENT | 0.85+ |
zero | QUANTITY | 0.85+ |
Kafka | ORGANIZATION | 0.84+ |
a minute | QUANTITY | 0.82+ |
generation | COMMERCIAL_ITEM | 0.82+ |
EC two nodes | TITLE | 0.78+ |
a hundred people | QUANTITY | 0.77+ |
Reinvent 22 | EVENT | 0.76+ |
a hundred | QUANTITY | 0.76+ |
important angle | QUANTITY | 0.75+ |
Rockette | TITLE | 0.75+ |
re:Invent 2022 - Global Startup Program | TITLE | 0.74+ |
Kafka | TITLE | 0.73+ |
Reinvent 22 | TITLE | 0.68+ |
third | QUANTITY | 0.68+ |
Zon | COMMERCIAL_ITEM | 0.66+ |
last two years | DATE | 0.65+ |
Roxette | PERSON | 0.65+ |
rock | ORGANIZATION | 0.64+ |
Soto | ORGANIZATION | 0.53+ |
Canis MSK | ORGANIZATION | 0.52+ |
Brian Henderson, Dell Technologies & Marc Trimuschat, AWS | AWS re:Invent 2022
(techno intro music) >> Hey everyone, good afternoon from sin city. This is Lisa Martin with Dave Vellante. We are in full swing of theCUBE's four days of coverage of AWS re:invent 2022. North of 50,000 people are here. We're nearing hundreds of thousands online. Dave, this has been, this is a great event. We've had great conversations. We're going to be having more conversations. One of the things we love talking about on theCUBE is AWS and its ecosystem of partners, and we are going to do just that right now. Brian Henderson joins us, Director of Marketing at Dell Technologies. Marc Trimuschat, Director of Worldwide Storage Specialists at AWS is also here. Guys, it's great to have you. >> Great to be here. >> Great to be here, yeah. Feeling the energy of the show. >> Isn't it great? >> Mark: I know, amazing. >> It's amazing. It started out high and it has not dropped since Monday night. Brian, talk a little bit about Dell, what you're doing with customers on their Cloud journeys. Every customer, every industry is on one at different points in their journey, but what's Dell helping out with there? >> What we're here to talk about is the progression that we've seen, right, Cloud has changed a lot over the years and Dell has really put out a strategy to help people with their Cloud journey, kind of wherever they are. So a lot of people have moved full shift. A lot of people see that as another location, and what we're showing at the booth is the idea of taking these enterprise capabilities that people know and trust from Dell, courting them to the Cloud. In some cases not courting, but just delivering that software in the Cloud, as well as taking some of the Kubernetes integrations, EKS Anywhere, bringing that on-prem. So we've got some storage, data protection, and our Kubernetes integration to talk about at the show. >> Awesome, Mark, talk about the role from Amazon's point of view that third party vendors like Dell Technologies plays in AWS's expanding vision of Cloud. >> Great, well, we're really excited to be partnering with Dell. What we see that historically is, you know, AWS is focused on builders, people, and really the developer community who are building those components themselves, putting together really resilient infrastructure and applications. What we're seeing today is a shift also to the type of customers that we're seeing, more traditional enterprise customers, who are demanding really performance, the scalability, also the resiliency of what they had on-premises, and they want that on the Cloud as well. So with Dell, and we've got some great solutions that we're partnering on, including Dell PowerFlex that provides that linear scalability and some of the high performance capabilities that customers are demanding. And also, another big trend that we're seeing is customers being affected by things like unfortunately malware events, right, and data protection. So Dell provides some great solutions in both those areas that allow enterprise customers to really experience that mission critical capability and resiliency that they have on-premises in the Cloud. >> You know, Brian, we've been at this a long time. >> Brian: Oh yeah, great to see you again. >> And I've been hearing my whole career that storage is going to get commoditized. And I guess if you're talking about spinning discs or flash drives, it's probably true, but as Mark was just saying if you want resilient storage and things that are recoverable, that don't go down all the time, they're not commodities. >> Brian: Yeah. >> It's real engineering. And you built the stack up, so talk about how that connection, what value you bring to the Cloud and your customers. >> Yeah, so what we see is people are always looking out for enterprise grade capabilities. So there's going to be a set of offerings, and AWS has a fantastic foundation for building on top of with the marketplace. So what we're able to do is really bring, in some cases, decades worth of investment in software engineering and put these advanced capabilities, whether it be PowerFlex with its linear scale. We'll have a file offering very soon. These products have been built from the ground up to do a very unique purpose. Giving that to people in the Cloud is just another location for us, AWS being the market leader. We're the market leader in storage. So us working together for the benefit of customers is really where it's at. >> Can you double click on that, Brian, what Dell and AWS? Give us all those juicy details. >> Sure, sure, sure, so what we've done right before this show is we put a product called PowerFlex, if you go back to 2018 scale IO, and you're taking this really linear scaling software defined architecture, and you're putting that in the Cloud. What that allows you to do is get that really advanced linear scale performance. You can even span clusters across AWS regions, as well as zones. So it's a really unique capability that allows us to be able to check in and do that. And in the data protection space, it's a whole separate category. We've been at this actually quite a while. We've got about 14 exo bytes of data that's already being protected on the AWS Cloud. So we've been at that for quite a while. And the two levels are really, do you want to back that up? Do you want to take a traditional back up application, maybe it's a lift and shift, and I want to back it up the way I used to, and you can do that in the Cloud now. Or we're seeing cyber resiliency come up a lot more, and we were just talking right before, it's a question of when, not if, and so we have to give our customers the option to not only detect that failure event early, but also to separate that copy with a logical air gap. >> The cyber resiliency is a topic we are talking more and more about. It's absolutely critical. We've seen the threat landscape change dramatically in the last couple of years. To your point, Brian, it's no longer, when we think of ransomware, it's no longer are we going to get hit? It's when, it's how often. What's the damage going to be? I think I saw a stat recently that there's one ransomware attack every 11 seconds. The average cost of reaches is in the millions, so what you're doing together on cyber resiliency for businesses in any industry is table stakes. >> Yeah, we just saw a survey that, it was done earlier this year survey, 66% unfortunately of corporations have experienced a malware attack. And that's an 80% increase from last year. >> Lisa: Wow. >> So again, I think that's an opportunity. It's a threat, but an opportunity, and so the partnership with Dell really helps bridge that and helps our customers, our mutual customers, recover from those incidents. >> A lot of people might say, this is interesting. A storage guy from Amazon, a storage guy from Dell, two leaders. And one might think, why didn't they just throw in a dash three, right, but you guys are both customer driven, customer obsessed. In the field, what are customers saying to you in terms of how they want you to work together? >> Well I think there's a place for everything. When you say throw in to S3, so S3 today, one of the big trends when you're looking here is just the amount of data, you know, we hear that rhetoric, you know, we've been in storage for many years, and the data has all increased up and to the right. But, you know, AWSI, S3 today, we have over 280 trillion objects in our, driving a hundred million transactions per second right now, so that's scale. So there's always a place for those really, we have hundreds of thousands of customers running their data links, so that's always going to be that really, you know, highly reliable, highly durable, high available solution for data links. But customers, there's a lot of different applications out there. So where customers are asking are those enterpise. So we have EBS, for example, which is our great, you know, scalable block search, elastic block store. We introduced some new volume types, like GP2, GP2, and IO2VX, which will have that performance. But there's still single availability zone. So what customers have done historically is they maybe the application layer, they put an application layer replication or resiliency across, but customers on-prem, they've relied on storage layers to do that work for them. So, with PowerFlex, that'll stand either using instant storage or EBS, building on that really strong foundation, but provide that additional layer to make it easy for customers to get that resiliency and that scalability that Brian talked about. >> Yep, yep. >> Anything you can add to that? >> Yeah, I mean to your question, how do we work together is really, it's all customer driven. So we see customers that are shifting workloads in the Cloud for the first time. And it might make sense to take an object, like PowerFlex or another storage technology, maybe you want to compress it a little bit before you send it to the Cloud. Maybe you don't want to lift and shift everything. So we have a team of people that works very closely with AWS to be able to determine how are you going to shift that workload out there? Does this make the right sense for you? So it's a very collaborative relationship. And it's all very customer driven because our customers are saying, I've got assets in the public Cloud, and I want them to be managed in a similar fashion to how I'm doing that on-prem. >> So customer obsession is clearly on both sides there. We know that. >> It's where it starts. >> Exactly, exactly. Going back to PowerFlex for a second, Brian, and I'd love to get an example of a joint customer that really is showing the value of what Dell and AWS are doing together. The question for you on PowerFlex, talk about the value that it offers to the public Cloud. And why should customers start there if they are early in this journey? >> All right, yeah, so the two angles are basically, are you coming from PowerFlex or you're coming from Cloud. If you're Cloud native, the advantage would be things like a really, really advanced block file system that has been built from the ground up to be software defined and pretty much Cloud native. What you're getting is that really linear scale up to about 1,000 nodes. You can span that across regions, across availability zones, so it's highly resilient. So if there's a node failure in one site, you're going to rebuild really fast, depending on the size of that cluster. So it's a very advanced architecture that's been built to run, you know, we didn't have to change a single line of code to run this product in the Cloud because it was Cloud native by default, so. >> Well that's the thing. We also see, and you've seen that with some of the other solutions, but customers really want that. Enterprise customers are, they want us to make sure those mission critical applications are working and stay up. So they also want to use the same environment. So we were talking before, we also see use cases where maybe they're using PowerFlex on-premises today and they want to be able to replicate that to PowerFlex that's in the Cloud. So we're seeing those, and the familiarity with that infrastructure really is that easy path, if you will, for those more conservative mission critical customers. >> We've learned a lot over the years from AWS's entry into the marketplace. Two recent teams working backwards. We talk about customer obsession. And also the Cloud experience. It brings me to APEX. >> Oh yeah. >> Dave: How does APEX fit in here? >> Yeah, so APEX is the categorization for all the things that we're doing around a modern Cloud experience for Dell customers. So we're taking them also on a journey, kind of as a service model. There's a do-it-yourself model. And anything that we do that touches Cloud is now being kind of put under that APEX moniker. So everything that we're doing around Project Alpine, enterprise software capabilities in the Cloud. Do you want someone else to manage it for you? Do you want it in a polo? That might be the right fit for you. It's all under that APEX umbrella and journey. So we're kind of still just getting started there, but we're seeing a lot of great traction. People want to pay as they go, you know, it's a very popular model that AWS has pretty much set the foundation for. So pay as you go, utility based pricing, this is all things our customers have been asking for. >> Yeah, so APEX, you basically set a baseline. You can dial it up, dial it down, very much pay by the drink. >> Absolutely. >> And, you know, like you said, it's early days. >> Brian: Yeah. >> But that's, again, AWS has influenced the business in a lot of different ways. >> Again, with the Dell, you know, the trust customers that Dell has built over the years and having those customers come in. We obviously are getting, again, it's an accelerated option for financial services to healthcare and all these customers that have relied on Dell for years, moving to the Cloud, having that trusted name and also that infrastructure that's similar and familiar to them. And then the resilience of the foundation that we have at AWS, I think it works really well together for those customers. >> I think it underscores to the majority of both AWS and in a lot of ways Dell, right. In the early days of Cloud, it was like uh oh, and now it's like oh, actually big market. Customers are demanding this. There's new value that we can create working together. Let's do it. >> Yeah, I mean, it didn't take us that long to get to it, but I'd say we had little fits and starts over the years, and now we've recognized like, this is where the future is. It's going to be Cloud, it's going to be on-prem, it's going to be Edge, it's going to be everything. It's going to be an and world. And so just doing the right thing for customers I think is exactly where we landed. It's a great partnership. >> Do you have a favorite customer story that you think really shines the light on the value of the Dell AWS partnership in terms of the business impact they're making? >> We have several large customers that I can't always like drop the names, but one of them is a very large video game production company. And we do a lot of work together where they're rendering maybe in house, they're sending to a shared location. They're copying data over to S3. They're able to let all their editors access that. They bring it back when it's compressed down a little bit and deliver that. We're also doing a lot of work with, I think I can say this, Amazon Thursday night football games. So what they've done there, it's a partner of ours working with AWS. All the details inside of that roaming truck that they drive around, there's a lot of Dell gear within there, and then everything connects back to AWS for that exact same kind of model. We need to get to the editors on a nightly basis. They're also streaming directly form that truck while they're enabling the editors to access a shared copy of it, so it's really powerful stuff. >> Thursday night prime is pretty cool. You know, some people are complaining cause I can't just switch channels during the commercials. It's like, first of all, you can. Second of all, the stats are unbelievable, right. You can just do your own replay when you want to. There's some cool innovations there. >> Oh yeah, absolutely. >> Very cool innovations. I've got one more question for each of you before we wrap. Marc, a question for you, we're making a fun Instagram reel. So think about a sizzle reel of if you were to summarize the show so far, what is AWS's message to its massive audience this year? >> Well, that's a big question. Because we have such a wide, as we mentioned, such a wide ranging audience. I really see a couple key trends that we're trying to address. One is, again don't forget, I'm a storage guy, so it's going to come from an angle from data, right. So, I think it's just this volume of data and that customers are bringing into the Cloud, either moving in from enterprises today or organically, just growing. You know, a couple years ago, megabytes were a lot, and now, you know, we're talking about petabytes every day. Soon it's going to be exo bytes are going to become the norm. So the big, I'd say, point one is the trend that I see is just the volume of data. And so what we're doing to address that is obviously we talked a little bit about S3 and being able to manage volumes of data, but also things like DataZone that we introduced because customers are looking to make sure that the right governance and controls to be able to access that data. So I think that's one big thing that I see the theme for the show today. The second thing is around, as I said, really these enterprise customers really wanted to move in these mission critical applications into the Cloud, and having that infrastructure to be able to support that easily from what they're doing today and move in quickly. The third area is around data protection, making sure the data protection and malware recovery, that's the theme that we see is really unfortunately that's today. But being able to recover quickly, both having native services and native offerings just built in resiliency into the core platforms, like S3 with object application, et cetera. And also partnering with Dell with cyber recovery and some of the solutions with Dell. >> Excellent, and Brian, last question for you. A bumper sticker that succinctly and powerfully describes why Dell and AWS are such awesome partners for customer issues. >> Best of both worlds, right? >> Lisa: Mic drop. >> Mic drop, done. >> That's awesome. You said that a lot more succinctly. (people laughing) >> Enterprise in Cloud, Cloud comin' to enterprise. >> Yeah, leader meets leader, right? >> Yeah, right. >> Love it, leader meets leader. Guys, it's been a pleasure having you on theCUBE. We appreciate hearing the latest from AWS and Dell from a storage perspective and from a Cloud perspective and how you're helping customers manage the explosion of data that's not going to slow down. We really appreciate you coming by the set. >> Thank you. >> Great, thanks so much, appreciate it. >> My pleasure. For our guests and Dave Vellante, I'm Lisa Martin, you're watching theCUBE, the leader in live enterprise and emerging tech coverage. (techno music)
SUMMARY :
One of the things we love Feeling the energy of the show. Every customer, every industry is on one that software in the Cloud, Awesome, Mark, talk about the role and really the developer community You know, Brian, we've that don't go down all the how that connection, what value you bring Giving that to people in the Cloud Can you double click on that, Brian, putting that in the Cloud. What's the damage going to be? Yeah, we just saw a survey that, and so the partnership with customers saying to you is just the amount of data, you know, I've got assets in the public Cloud, So customer obsession is that really is showing the value that has been built from the ground up replicate that to PowerFlex And also the Cloud experience. And anything that we do that touches Cloud Yeah, so APEX, you And, you know, like has influenced the business that Dell has built over the years In the early days of and starts over the years, the editors to access Second of all, the stats the show so far, what is AWS's message and some of the solutions with Dell. A bumper sticker that succinctly You said that a lot more succinctly. Cloud comin' to enterprise. We appreciate hearing the the leader in live enterprise
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Marc Trimuschat | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Brian Henderson | PERSON | 0.99+ |
Brian | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
80% | QUANTITY | 0.99+ |
Mark | PERSON | 0.99+ |
two leaders | QUANTITY | 0.99+ |
66% | QUANTITY | 0.99+ |
Monday night | DATE | 0.99+ |
two angles | QUANTITY | 0.99+ |
Marc | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
both sides | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
second thing | QUANTITY | 0.99+ |
two levels | QUANTITY | 0.99+ |
one site | QUANTITY | 0.99+ |
S3 | TITLE | 0.99+ |
first time | QUANTITY | 0.98+ |
third area | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
millions | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
Thursday night | DATE | 0.98+ |
one more question | QUANTITY | 0.98+ |
over 280 trillion objects | QUANTITY | 0.98+ |
John Kreisa, Couchbase | AWS re:Invent 2022
(upbeat music) >> Good morning and welcome back to fabulous Las Vegas, Nevada. We're here at AWS re:Invent with wall-to-wall coverage all day long on theCUBE. My name is Savannah Peterson and I am joined this morning by the beautiful Lisa Martin. Lisa, good morning. >> Good morning. Good. >> How you feeling day three? >> Day three is we are going to be shot out of a cannon today. The amount of content coming at you from theCUBE today- >> Get ready, you all. >> Us two gals, is a lot. We're going to have some great conversations. >> And we're starting with a really great one with a Cube Alumni to the max. You've been on the show multiple times. >> John: Yeah. >> Very excited to welcome John, the CMO of Couchbase. Welcome. How you doing this morning? >> Thanks. I'm doing great. Great to be here with you. >> How do you feel about the show so far? What's your pulse? >> The show has been great. I say the energy is great. The traffic at our booth, the conversations that we're having, both with prospective customers and even just partners, right? They're all here. The ecosystem is here >> And everyone's finally back in person and it feels so good. >> John: It does. >> So, we're going to dig in a little bit but just in case the audience isn't familiar, tell us about Couchbase. >> Sure. Couchbase is a publicly traded database company. We have a cloud database platform called Capella which is hosted on AWS and GCP. It is used for building mission-critical applications. So, we have great customers, we're building apps that really matter and are using to drive their business. So, we're very excited about that. 30% of the Fortune 100 are Couchbase customers. >> Nice. Talk a little bit about the AWS relationship. >> Mm-hm. Yeah, so we have a great AWS relationship. In fact, yesterday we announced a deepening of that relationship, a strategic collaboration agreement. We're very excited. It's a multi-year agreement. It's focused on go-to market, from a sales and marketing standpoint. We're going to target, you know, various verticals and, you know, really generate joint business between the two of us. So, it's a deepening of a already strong relationship and we're really excited about that. >> Savannah: Yeah. Go ahead. >> What are some of the industry verticals that you're going to be tackling together? >> Well, gaming for one, right? Manufacturing, the workloads that Couchbase is good for are these mission-critical workloads are ones that are really suited for us to be used with AWS. So, we've done some work with them already in those areas and I'm sure we'll be digging in even deeper. >> That's exciting. Speaking of digging in deeper, tell us a little bit more about Capella. >> Capella. It's a cloud databases services I mentioned. We launched it last October and we are super excited by the uptake, the interest that we're seeing. We have a free 30 day trial, so, you know, people can come and try it and get their hands dirty just getting experience with the product and then, you know, become a customer after that. And we're seeing very strong interest from our existing customer base as well. So, we're really excited about how things are going. >> Talk about Capella and the latest release and how it's really enabling Couchbase to invest deeper into the developer experience. >> Yeah, so, at the end of October, we announced a revamp of our user interface, our user experience for Capella really focused on developers. And what we've done is make it so that it's familiar to developers, right? It's a GitHub-like experience. So, developer comes in, they're very familiar, of course, with GitHub, they are familiar with how the Couchbase Capella interface will work. And so that's something that, you know, we've really invested, in fact, we've invested in developers quite a bit. We announced a Couchbase community hub and a Couchbase ambassadors program, both focused on developers and getting out there and building our community. >> A community is a big topic that we've been talking about at all the conferences this year. We're all back in person, in community. How often are you communicating with your community to get feedback on what that experience should be like? >> Yeah, I mean, we actually have a Discord server, so we're in constant communication. (Savannah laughing) >> Savannah: Yes. (John laughing) 24/7. (laughing) >> Basically, you know, we have staff who's dedicated to making sure that the users on there are getting their answers and giving us feedback on the experience. The ambassadors are somebody who have a really strong relationship, who get early insight and give us feedback before we even release a product. So, it gives us a chance to really test-drive it with core developers and get the insight we need before we get it in the market. >> Yeah. It matters so much. You can build it, but they won't come if it's not fantastic. >> John: Exactly. >> Lisa: Right. >> Let's shift a little bit and talk about customers. How, and price, how do you guys compare? >> Customers and? >> And price, your price performance? >> Price, oh. So, customers, we also announced this week a joint customer Arthrex with AWS. Arthrex is a orthopedics medical devices company and they use our Edge capabilities along with running Couchbase on AWS. So, you think of the kinds of surgeries that orthopedic surgeons do, it's scopes and they are often inside. So, what it does is it collects the data, the video data and all of that on a medical devices and then brings it back to a centralized app for the doctors to use sort of in post when they're actually doing further medical recommendations. >> Savannah: It's so cool. >> So, it's cool, the thing about it is it can work whether it's online or offline, it's one of the reasons that Arthrex selected us because the fact that it can, you know, often sometimes there's not connectivity in the operating room, I'd say deep inside of a hospital. So, these devices work regardless and then when they get connectivity, it sinks back to that centralized service. So, it's one of the main reasons that they selected us. >> That's outstanding. You know, one of the things that John Furrier, you know, John, well, you guys go way back. >> John: Way back. >> He had a sit down with Adam Selensky, oh, about 10 days or so ago. He gets an exclusive with the CEO of AWS every pre re:Invent. And one of the things that Adam said is that the role or the title, data analyst, is going to go away, in that every role will have responsibilities of analyzing data. And I always think of that in terms of operations, marketing, finance, sales, but you just brought up physicians as data analysts in their jobs, right? Probably not, we're thinking about it in that way. >> Yeah. >> But it's so interesting how data is really being democratized. >> John: Yeah. >> And how Couchbase is an enabler of that in an operating room. >> John: Yeah, yeah, yeah. >> That's amazing. >> It's a great story. There's many others and I think, you know, we have embedded operational analytics in Couchbase Capella, and, you know, in our offerings in general. So, what that does is allows us to do real-time, highly personalized applications based on that analytics that are coming in real-time from the data from the applications. And so that's something that's actually driving a highly interactive user experience, one that's very personalized and customized. And that's one of the things that our customers really like about what we do. >> It's fascinating. I never thought about it from a medical device perspective. >> Lisa: No, no. >> John: No. >> My gosh is if doctors don't have enough cognitive burden load. >> John: I know. >> You know, right? Like, they don't need to be a data analyst. I would much rather they were just good at the surgery part. That's a piece of the puzzle I need them to do. Yeah, for sure. That's a fascinating customer example. Can you share any other joint AWS examples with us? >> Joint AW- I mean, there's many in the gaming area where, because Couchbase is memory-first architecture, we deliver very, very interactive user experiences and we're used a lot for session management, user ID management in the gaming space, specifically with AWS. It's an area we've done some joint work already and had a lot of success, you know, with small and large gaming companies. >> Yeah. It looks like you also, according to my notes here, we've got things in travel and hospitality as well. >> Yes. Also Carnival Cruises is a great example. We enable their on-ship, on-board experience, highly customized, everybody wears a device called a medallion, and as they move around the ship, it knows where they are and it's able to provide customized services. You walk up to a bar, you have your favorite drink, it can be hit the bar when you land there. >> I'll take that. >> How about that? (laugh) >> That's outstanding. >> Isn't that great? >> Can we carry that onto the AWS show floor? >> Exactly. >> Or Starbucks order? >> Yeah, yeah. Yes, please. Yes, please. Well, another thing that's so interesting these days, is that every company has to be a data company. Say they have to be a software company. They have to be a data company. You just gave some great examples. Hospitality, gaming, healthcare, where that data democratization has to happen. >> John: Yeah. >> Businesses has to transform. But one of the things that Adam also told John is that CIOs, CEOs are coming to him not wanting to talk about technology but about transformation. >> Yeah. >> Huge topic. >> And that's a journey where every customer is at different levels. >> Yeah. >> How is Couchbase helping businesses transform and where are your customer conversations these days? >> Yeah, yeah, yeah. So, I mean, the transformation of the business is a major topic of conversation. So, we completely agree with that. How Couchbase helps is, you know, in our database, one of the things we have is the SQL engine. And so as people are looking to move and modernize their infrastructure, if they're moving off of, or from like a technology that's principally based on SQL but doesn't give all the flexibility of a JSON database or document database like we do, we actually enable them to get more easily onto our platform so that they can start that transformation. And then it's a, you know, it's a journey of how they want to transform their business and it's really focused on how do they better serve their customers and clients, whether it's internal or external? >> It really matters. I mean, and that ease of use as well as the transformation journey. It takes a long time for people to adapt. So, every piece of that puzzle, every Lego being quicker or easier, more intuitive, like you said, with the user experience, we can tell you're very thoughtful. How does this improve the total cost of ownership for your customers? >> That's one of the things that we announced along with that developer changes, was a new storage engine underneath Couchbase Capella. And it's 10 X more dense storage. And what that means is fewer servers. So, fewer servers is a much better cost of ownership story. That plus just the performance of the platform itself, we find, you know, against competition, we can do things on say six nodes that take 18 nodes for others. >> Lisa: Oh wow. >> And we have a great consolidation story as well because we have, it's a multi-modal database, meaning that it has SQL engine, document database, full tech search, eventing and analytics, all these pieces on one common data layer. So, you can actually consolidate off of other technologies onto one, onto Couchbase, and that actually saves you money. So, that's a great story for us. >> There's got to be a sustainability element to that as well? >> Yeah, I mean it's, obviously, if you're using less, using fewer servers, there's a kind of power consumption aspect of it as well. Absolutely. >> Are you finding that a lot of customers and companies we talk to these days have in their RFPs, they must only work with vendors who have an actual ESG program? Are you finding more customers coming to you saying, how can you help us dial down our carbon emissions? >> John: Yeah. >> Savannah: Great question. >> We've got a sustainability program that we've got to meet, we've got commitments to our customers. >> John: Yeah. >> Is that something that's really now kind of a hard and fast requirement? >> We're hearing it, we're definitely hearing it. I wouldn't say it's, you know, massively pervasive but I would say it's a growing component of, as you said, RFPs. And it's something that we feel like we have a great story for. And so, you know, it's something that helps when we get into those conversations, we can clearly articulate how we can provide that value and how we meet some of those needs that they have. >> Yeah, that's awesome. So, we have a bit of a challenge, new to the show at re:Invent. >> John: Mm-hm. >> Where we are prompting you to give us your 30 second Instagram Reel sizzle highlight. Don't worry, I'm not actually timing you, but your thought leadership hot-take on the most important theme or takeaway from this year's show. >> From the conference here. I would say that, and I think this was talked about a little bit by AWS as well, but the convergence of analytics and operational data, you know, through the applications is one that we're certainly seeing as well. It's the reason we have analytics in our database. But as I walk around and look at it, I see that very much as a common theme as well, in terms of what other vendors are saying and just the conversations we're having. So for me, that's one of the things I think would be a takeaway from this show. >> Yeah. Embedded analytics, real-time, everybody wants to know what's going on, in context. >> Yeah. That's right. >> Right now, not last week, not what we're processing from last month. >> Exactly. >> I mean, right? (cross-talking) >> So, I can react and take advantage or take an action if I have to. >> Exactly. And then deliver that personalized experience that we all expect these days. >> Oh, yes. >> I'll take that medallion- >> It's about the medallion. I was like, okay. >> You up with that, John? >> We'll get right on it. >> Lisa: All right. (laughs) >> About this. So, what's next for Couchbase? >> John: Well- >> I know you got the partnership, you've got all this exciting momentum. >> So, we're excited heading into next year. We're going to continue to innovate on Capella, right? Continue to deliver more value, lean into our developer community that we have. We're investing heavily, not just from a product standpoint but from a company standpoint in terms of, you know, our community meetups and some of those things. We have a big community-focused event coming up in March called Connect, Couchbase Connect. So, that's something that we'll, you know, continue to drive. That'll be a major theme for us next year. Cloud and developers and, you know, continuing to enable that ecosystem. >> Lisa: Excellent. >> I just had a Microsoft moment where I saw you saying, "Cloud developers," on stage. (Lisa and Savannah laughing) >> I'm not going Steve Ballmer on you. (all laughing) >> Pardon. I was trying to get someone to sing yesterday. I was hoping you were my Ballmer dance. Oh, man. Well, this has been a really great way to start the day. John, thank you so much for being on the show with us, seriously. And it's great that you keep coming back. I'm glad we haven't scared you off. (John laughing) >> Never. >> Savannah: We will have you anytime. >> Thank you. >> And thank you all for tuning in for yet another fantastic day of all day live coverage here from AWS re:Invent. We are in Sin City, having a fabulous time with Lisa Martin. I'm Savannah Peterson. This is theCUBE and we are the leader in high-tech technology coverage. (upbeat music) (upbeat music fades)
SUMMARY :
by the beautiful Lisa Martin. Good morning. at you from theCUBE today- We're going to have some You've been on the show multiple times. How you doing this morning? Great to be here with you. I say the energy is great. and it feels so good. but just in case the So, we have great customers, the AWS relationship. We're going to target, you Manufacturing, the Speaking of digging in deeper, the product and then, you know, and the latest release And so that's something that, you know, about at all the conferences this year. Yeah, I mean, we actually Savannah: Yes. get the insight we need come if it's not fantastic. How, and price, how do you guys compare? for the doctors to use sort of in post because the fact that it can, you know, You know, one of the is that the role or the But it's so interesting how data of that in an operating room. And that's one of the things I never thought about it from My gosh is if doctors don't have enough That's a piece of the and had a lot of success, you know, and hospitality as well. it can be hit the bar when you land there. They have to be a data company. But one of the things that Adam And that's a journey one of the things we So, every piece of that puzzle, we find, you know, against competition, So, you can actually consolidate consumption aspect of it as well. program that we've got to meet, And it's something that we feel So, we have a bit of a challenge, Where we are prompting you to give us and just the conversations we're having. in context. not what we're processing from last month. So, I can react and take that we all expect these days. It's about the medallion. Lisa: All right. So, what's I know you got the partnership, So, that's something that we'll, you know, where I saw you saying, I'm not going Steve Ballmer on you. And it's great that you keep coming back. have you anytime. And thank you all for tuning in
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Adam Selensky | PERSON | 0.99+ |
Savannah | PERSON | 0.99+ |
John Kreisa | PERSON | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Sin City | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
March | DATE | 0.99+ |
30 day | QUANTITY | 0.99+ |
Arthrex | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
last month | DATE | 0.99+ |
Steve Ballmer | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
30% | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
last week | DATE | 0.99+ |
Couchbase | ORGANIZATION | 0.99+ |
30 second | QUANTITY | 0.99+ |
Ballmer | PERSON | 0.99+ |
Las Vegas, Nevada | LOCATION | 0.98+ |
Capella | ORGANIZATION | 0.98+ |
last October | DATE | 0.98+ |
today | DATE | 0.98+ |
18 nodes | QUANTITY | 0.98+ |
this week | DATE | 0.98+ |
both | QUANTITY | 0.98+ |
Microsoft | ORGANIZATION | 0.97+ |
this year | DATE | 0.97+ |
GitHub | ORGANIZATION | 0.97+ |
SQL | TITLE | 0.97+ |
Lego | ORGANIZATION | 0.96+ |
six nodes | QUANTITY | 0.96+ |
Armando Acosta, Dell Technologies and Matt Leininger, Lawrence Livermore National Laboratory
(upbeat music) >> We are back, approaching the finish line here at Supercomputing 22, our last interview of the day, our last interview of the show. And I have to say Dave Nicholson, my co-host, My name is Paul Gillin. I've been attending trade shows for 40 years Dave, I've never been to one like this. The type of people who are here, the type of problems they're solving, what they talk about, the trade shows are typically, they're so speeds and feeds. They're so financial, they're so ROI, they all sound the same after a while. This is truly a different event. Do you get that sense? >> A hundred percent. Now, I've been attending trade shows for 10 years since I was 19, in other words, so I don't have necessarily your depth. No, but seriously, Paul, totally, completely, completely different than any other conference. First of all, there's the absolute allure of looking at the latest and greatest, coolest stuff. I mean, when you have NASA lecturing on things when you have Lawrence Livermore Labs that we're going to be talking to here in a second it's a completely different story. You have all of the academics you have students who are in competition and also interviewing with organizations. It's phenomenal. I've had chills a lot this week. >> And I guess our last two guests sort of represent that cross section. Armando Acosta, director of HPC Solutions, High Performance Solutions at Dell. And Matt Leininger, who is the HPC Strategist at Lawrence Livermore National Laboratory. Now, there is perhaps, I don't know you can correct me on this, but perhaps no institution in the world that uses more computing cycles than Lawrence Livermore National Laboratory and is always on the leading edge of what's going on in Supercomputing. And so we want to talk to both of you about that. Thank you. Thank you for joining us today. >> Sure, glad to be here. >> For having us. >> Let's start with you, Armando. Well, let's talk about the juxtaposition of the two of you. I would not have thought of LLNL as being a Dell reference account in the past. Tell us about the background of your relationship and what you're providing to the laboratory. >> Yeah, so we're really excited to be working with Lawrence Livermore, working with Matt. But actually this process started about two years ago. So we started looking at essentially what was coming down the pipeline. You know, what were the customer requirements. What did we need in order to make Matt successful. And so the beauty of this project is that we've been talking about this for two years, and now it's finally coming to fruition. And now we're actually delivering systems and delivering racks of systems. But what I really appreciate is Matt coming to us, us working together for two years and really trying to understand what are the requirements, what's the schedule, what do we need to hit in order to make them successful >> At Lawrence Livermore, what drives your computing requirements I guess? You're working on some very, very big problems but a lot of very complex problems. How do you decide what you need to procure to address them? >> Well, that's a difficult challenge. I mean, our mission is a national security mission dealing with making sure that we do our part to provide the high performance computing capabilities to the US Department of Energy's National Nuclear Security Administration. We do that through the Advanced Simulation computing program. Its goal is to provide that computing power to make sure that the US nuclear rep of the stockpile is safe, secure, and effective. So how we go about doing that? There's a lot of work involved. We have multiple platform lines that we accomplish that goal with. One of them is the advanced technology systems. Those are the ones you've heard about a lot, they're pushing towards exit scale, the GPU technologies incorporated into those. We also have a second line, a platform line, called the Commodity Technology Systems. That's where right now we're partnering with Dell on the latest generation of those. Those systems are a little more conservative, they're right now CPU only driven but they're also intended to be the everyday work horses. So those are the first systems our users get on. It's very easy for them to get their applications up and running. They're the first things they use usually on a day to day basis. They run a lot of small to medium size jobs that you need to do to figure out how to most effectively use what workloads you need to move to the even larger systems to accomplish our mission goals. >> The workhorses. >> Yeah. >> What have you seen here these last few days of the show, what excites you? What are the most interesting things you've seen? >> There's all kinds of things that are interesting. Probably most interesting ones I can't talk about in public, unfortunately, 'cause of NDA agreements, of course. But it's always exciting to be here at Supercomputing. It's always exciting to see the products that we've been working with industry and co-designing with them on for, you know, several years before the public actually sees them. That's always an exciting part of the conference as well specifically with CTS-2, it's exciting. As was mentioned before, I've been working with Dell for nearly two years on this, but the systems first started being delivered this past August. And so we're just taking the initial deliveries of those. We've deployed, you know, roughly about 1600 nodes now but that'll ramp up to over 6,000 nodes over the next three or four months. >> So how does this work intersect with Sandia and Los Alamos? Explain to us the relationship there. >> Right, so those three laboratories are the laboratories under the National Nuclear Security Administration. We partner together on CTS. So the architectures, as you were asking, how do we define these things, it's the labs coming together. Those three laboratories we define what we need for that architecture. We have a joint procurement that is run out of Livermore but then the systems are deployed at all three laboratories. And then they serve the programs that I mentioned for each laboratory as well. >> I've worked in this space for a very long time you know I've worked with agencies where the closest I got to anything they were actually doing was the sort of guest suite outside the secure area. And sometimes there are challenges when you're communicating, it's like you have a partner like Dell who has all of these things to offer, all of these ideas. You have requirements, but maybe you can't share 100% of what you need to do. How do you navigate that? Who makes the decision about what can be revealed in these conversations? You talk about NDA in terms of what's been shared with you, you may be limited in terms of what you can share with vendors. Does that cause inefficiency? >> To some degree. I mean, we do a good job within the NSA of understanding what our applications need and then mapping that to technical requirements that we can talk about with vendors. We also have kind of in between that we've done this for many years. A recent example is of course with the exit scale computing program and some things it's doing creating proxy apps or mini apps that are smaller versions of some of the things that we are important to us. Some application areas are important to us, hydrodynamics, material science, things like that. And so we can collaborate with vendors on those proxy apps to co-design systems and tweak the architectures. In fact, we've done a little bit that with CTS-2, not as much in CTS as maybe in the ATS platforms but that kind of general idea of how we collaborate through these proxy applications is something we've used across platforms. >> Now is Dell one of your co-design partners? >> In CTS-2 absolutely, yep. >> And how, what aspects of CTS-2 are you working on with Dell? >> Well, the architecture itself was the first, you know thing we worked with them on, we had a procurement come out, you know they bid an architecture on that. We had worked with them, you know but previously on our requirements, understanding what our requirements are. But that architecture today is based on the fourth generation Intel Xeon that you've heard a lot about at the conference. We are one of the first customers to get those systems in. All the systems are interconnected together with the Cornell Network's Omni-Path Network that we've used before and are very excited about as well. And we build up from there. The systems get integrated in by the operations teams at the laboratory. They get integrated into our production computing environment. Dell is really responsible, you know for designing these systems and delivering to the laboratories. The laboratories then work with Dell. We have a software stack that we provide on top of that called TOSS, for Tri-Lab Operating System. It's based on Redhead Enterprise Linux. But the goal there is that it allows us, a common user environment, a common simulation environment across not only CTS-2, but maybe older systems we have and even the larger systems that we'll be deploying as well. So from a user perspective they see a common user interface, a common environment across all the different platforms that they use at Livermore and the other laboratories. >> And Armando, what does Dell get out of the co-design arrangement with the lab? >> Well, we get to make sure that they're successful. But the other big thing that we want to do, is typically when you think about Dell and HPC, a lot of people don't make that connection together. And so what we're trying to do is make sure that, you know they know that, hey, whether you're a work group customer at the smallest end or a super computer customer at the highest end, Dell wants to make sure that we have the right setup portfolio to match any needs across this. But what we were really excited about this, this is kind of our, you know big CTS-2 first thing we've done together. And so, you know, hopefully this has been successful. We've made Matt happy and we look forward to the future what we can do with bigger and bigger things. >> So will the labs be okay with Dell coming up with a marketing campaign that said something like, "We can't confirm that alien technology is being reverse engineered." >> Yeah, that would fly. >> I mean that would be right, right? And I have to ask you the question directly and the way you can answer it is by smiling like you're thinking, what a stupid question. Are you reverse engineering alien technology at the labs? >> Yeah, you'd have to suck the PR office. >> Okay, okay. (all laughing) >> Good answer. >> No, but it is fascinating because to a degree it's like you could say, yeah, we're working together but if you really want to dig into it, it's like, "Well I kind of can't tell you exactly how some of this stuff is." Do you consider anything that you do from a technology perspective, not what you're doing with it, but the actual stack, do you try to design proprietary things into the stack or do you say, "No, no, no, we're going to go with standards and then what we do with it is proprietary and secret."? >> Yeah, it's more the latter. >> Is the latter? Yeah, yeah, yeah. So you're not going to try to reverse engineer the industry? >> No, no. We want the solutions that we develop to enhance the industry to be able to apply to a broader market so that we can, you know, gain from the volume of that market, the lower cost that they would enable, right? If we go off and develop more and more customized solutions that can be extraordinarily expensive. And so we we're really looking to leverage the wider market, but do what we can to influence that, to develop key technologies that we and others need that can enable us in the high forms computing space. >> We were talking with Satish Iyer from Dell earlier about validated designs, Dell's reference designs for for pharma and for manufacturing, in HPC are you seeing that HPC, Armando, and is coming together traditionally and more of an academic research discipline beginning to come together with commercial applications? And are these two markets beginning to blend? >> Yeah, I mean so here's what's happening, is you have this convergence of HPC, AI and data analytics. And so when you have that combination of those three workloads they're applicable across many vertical markets, right? Whether it's financial services, whether it's life science, government and research. But what's interesting, and Matt won't brag about, but a lot of stuff that happens in the DoE labs trickles down to the enterprise space, trickles down to the commercial space because these guys know how to do it at scale, they know how to do it efficiently and they know how to hit the mark. And so a lot of customers say, "Hey we want what CTS-2 does," right? And so it's very interesting. The way I love it is their process the way they do the RFP process. Matt talked about the benchmarks and helping us understand, hey here's kind of the mark you have to hit. And then at the same time, you know if we make them successful then obviously it's better for all of us, right? You know, I want to secure nuclear stock pile so I hope everybody else does as well. >> The software stack you mentioned, I think Tia? >> TOSS. >> TOSS. >> Yeah. >> How did that come about? Why did you feel the need to develop your own software stack? >> It originated back, you know, even 20 years ago when we first started building Linux clusters when that was a crazy idea. Livermore and other laboratories were really the first to start doing that and then push them to larger and larger scales. And it was key to have Linux running on that at the time. And so we had the. >> So 20 years ago you knew you wanted to run on Linux? >> Was 20 years ago, yeah, yeah. And we started doing that but we needed a way to have a version of Linux that we could partner with someone on that would do, you know, the support, you know, just like you get from an EoS vendor, right? Security support and other things. But then layer on top of that, all the HPC stuff you need either to run the system, to set up the system, to support our user base. And that evolved into to TOSS which is the Tri-Lab Operating System. Now it's based on the latest version of Redhead Enterprise Linux, as I mentioned before, with all the other HPC magic, so to speak and all that HPC magic is open source things. It's not stuff, it may be things that we develop but it's nothing closed source. So all that's there we run it across all these different environments as I mentioned before. And it really originated back in the early days of, you know, Beowulf clusters, Linux clusters, as just needing something that we can use to run on multiple systems and start creating that common environment at Livermore and then eventually the other laboratories. >> How is a company like Dell, able to benefit from the open source work that's coming out of the labs? >> Well, when you look at the open source, I mean open source is good for everybody, right? Because if you make a open source tool available then people start essentially using that tool. And so if we can make that open source tool more robust and get more people using it, it gets more enterprise ready. And so with that, you know, we're all about open source we're all about standards and really about raising all boats 'cause that's what open source is all about. >> And with that, we are out of time. This is our 28th interview of SC22 and you're taking us out on a high note. Armando Acosta, director of HPC Solutions at Dell. Matt Leininger, HPC Strategist, Lawrence Livermore National Laboratories. Great discussion. Hopefully it was a good show for you. Fascinating show for us and thanks for being with us today. >> Thank you very much. >> Thank you for having us >> Dave it's been a pleasure. >> Absolutely. >> Hope we'll be back next year. >> Can't believe, went by fast. Absolutely at SC23. >> We hope you'll be back next year. This is Paul Gillin. That's a wrap, with Dave Nicholson for theCUBE. See here in next time. (soft upbear music)
SUMMARY :
And I have to say Dave You have all of the academics and is always on the leading edge about the juxtaposition of the two of you. And so the beauty of this project How do you decide what you need that you need to do but the systems first Explain to us the relationship there. So the architectures, as you were asking, 100% of what you need to do. And so we can collaborate with and the other laboratories. And so, you know, hopefully that said something like, And I have to ask you and then what we do with it reverse engineer the industry? so that we can, you know, gain And so when you have that combination running on that at the time. all the HPC stuff you need And so with that, you know, and thanks for being with us today. Absolutely at SC23. with Dave Nicholson for theCUBE.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Matt Leininger | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
National Nuclear Security Administration | ORGANIZATION | 0.99+ |
Armando Acosta | PERSON | 0.99+ |
Cornell Network | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Matt | PERSON | 0.99+ |
CTS-2 | TITLE | 0.99+ |
US Department of Energy | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
40 years | QUANTITY | 0.99+ |
two years | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
Lawrence Livermore | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
CTS | TITLE | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
Paul | PERSON | 0.99+ |
Linux | TITLE | 0.99+ |
NASA | ORGANIZATION | 0.99+ |
HPC Solutions | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
Lawrence Livermore Labs | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Los Alamos | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
Lawrence Livermore National Laboratory | ORGANIZATION | 0.99+ |
Armando | ORGANIZATION | 0.99+ |
each laboratory | QUANTITY | 0.99+ |
second line | QUANTITY | 0.99+ |
over 6,000 nodes | QUANTITY | 0.99+ |
20 years ago | DATE | 0.98+ |
three laboratories | QUANTITY | 0.98+ |
28th interview | QUANTITY | 0.98+ |
Lawrence Livermore National Laboratories | ORGANIZATION | 0.98+ |
three | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
Tri-Lab | ORGANIZATION | 0.98+ |
Sandia | ORGANIZATION | 0.97+ |
one | QUANTITY | 0.97+ |
First | QUANTITY | 0.97+ |
two markets | QUANTITY | 0.97+ |
Supercomputing | ORGANIZATION | 0.96+ |
first systems | QUANTITY | 0.96+ |
fourth generation | QUANTITY | 0.96+ |
this week | DATE | 0.96+ |
Livermore | ORGANIZATION | 0.96+ |
Omni-Path Network | ORGANIZATION | 0.95+ |
about 1600 nodes | QUANTITY | 0.95+ |
Lawrence Livermore National Laboratory | ORGANIZATION | 0.94+ |
LLNL | ORGANIZATION | 0.93+ |
NDA | ORGANIZATION | 0.93+ |
Ian Colle, AWS | SuperComputing 22
(lively music) >> Good morning. Welcome back to theCUBE's coverage at Supercomputing Conference 2022, live here in Dallas. I'm Dave Nicholson with my co-host Paul Gillin. So far so good, Paul? It's been a fascinating morning Three days in, and a fascinating guest, Ian from AWS. Welcome. >> Thanks, Dave. >> What are we going to talk about? Batch computing, HPC. >> We've got a lot, let's get started. Let's dive right in. >> Yeah, we've got a lot to talk about. I mean, first thing is we recently announced our batch support for EKS. EKS is our Kubernetes, managed Kubernetes offering at AWS. And so batch computing is still a large portion of HPC workloads. While the interactive component is growing, the vast majority of systems are just kind of fire and forget, and we want to run thousands and thousands of nodes in parallel. We want to scale out those workloads. And what's unique about our AWS batch offering, is that we can dynamically scale, based upon the queue depth. And so customers can go from seemingly nothing up to thousands of nodes, and while they're executing their work they're only paying for the instances while they're working. And then as the queue depth starts to drop and the number of jobs waiting in the queue starts to drop, then we start to dynamically scale down those resources. And so it's extremely powerful. We see lots of distributed machine learning, autonomous vehicle simulation, and traditional HPC workloads taking advantage of AWS Batch. >> So when you have a Kubernetes cluster does it have to be located in the same region as the HPC cluster that's going to be doing the batch processing, or does the nature of batch processing mean, in theory, you can move something from here to somewhere relatively far away to do the batch processing? How does that work? 'Cause look, we're walking around here and people are talking about lengths of cables in order to improve performance. So what does that look like when you peel back the cover and you look at it physically, not just logically, AWS is everywhere, but physically, what does that look like? >> Oh, physically, for us, it depends on what the customer's looking for. We have workflows that are all entirely within a single region. And so where they could have a portion of say the traditional HPC workflow, is within that region as well as the batch, and they're saving off the results, say to a shared storage file system like our Amazon FSx for Lustre, or maybe aging that back to an S3 object storage for a little lower cost storage solution. Or you can have customers that have a kind of a multi-region orchestration layer to where they say, "You know what? "I've got a portion of my workflow that occurs "over on the other side of the country "and I replicate my data between the East Coast "and the West Coast just based upon business needs. "And I want to have that available to customers over there. "And so I'll do a portion of it in the East Coast "a portion of it in the West Coast." Or you can think of that even globally. It really depends upon the customer's architecture. >> So is the intersection of Kubernetes with HPC, is this relatively new? I know you're saying you're, you're announcing it. >> It really is. I think we've seen a growing perspective. I mean, Kubernetes has been a long time kind of eating everything, right, in the enterprise space? And now a lot of CIOs in the industrial space are saying, "Why am I using one orchestration layer "to manage my HPC infrastructure and another one "to manage my enterprise infrastructure?" And so there's a growing appreciation that, you know what, why don't we just consolidate on one? And so that's where we've seen a growth of Kubernetes infrastructure and our own managed Kubernetes EKS on AWS. >> Last month you announced a general availability of Trainium, of a chip that's optimized for AI training. Talk about what's special about that chip or what is is customized to the training workloads. >> Yeah, what's unique about the Trainium, is you'll you'll see 40% price performance over any other GPU available in the AWS cloud. And so we've really geared it to be that most price performance of options for our customers. And that's what we like about the silicon team, that we're part of that Annaperna acquisition, is because it really has enabled us to have this differentiation and to not just be innovating at the software level but the entire stack. That Annaperna Labs team develops our network cards, they develop our ARM cards, they developed this Trainium chip. And so that silicon innovation has become a core part of our differentiator from other vendors. And what Trainium allows you to do is perform similar workloads, just at a lower price performance. >> And you also have a chip several years older, called Inferentia- >> Um-hmm. >> Which is for inferencing. What is the difference between, I mean, when would a customer use one versus the other? How would you move the workload? >> What we've seen is customers traditionally have looked for a certain class of machine, more of a compute type that is not as accelerated or as heavy as you would need for Trainium for their inference portion of their workload. So when they do that training they want the really beefy machines that can grind through a lot of data. But when you're doing the inference, it's a little lighter weight. And so it's a different class of machine. And so that's why we've got those two different product lines with the Inferentia being there to support those inference portions of their workflow and the Trainium to be that kind of heavy duty training work. >> And then you advise them on how to migrate their workloads from one to the other? And once the model is trained would they switch to an Inferentia-based instance? >> Definitely, definitely. We help them work through what does that design of that workflow look like? And some customers are very comfortable doing self-service and just kind of building it on their own. Other customers look for a more professional services engagement to say like, "Hey, can you come in and help me work "through how I might modify my workflow to "take full advantage of these resources?" >> The HPC world has been somewhat slower than commercial computing to migrate to the cloud because- >> You're very polite. (panelists all laughing) >> Latency issues, they want to control the workload, they want to, I mean there are even issues with moving large amounts of data back and forth. What do you say to them? I mean what's the argument for ditching the on-prem supercomputer and going all-in on AWS? >> Well, I mean, to be fair, I started at AWS five years ago. And I can tell you when I showed up at Supercomputing, even though I'd been part of this community for many years, they said, "What is AWS doing at Supercomputing?" I know you care, wait, it's Amazon Web Services. You care about the web, can you actually handle supercomputing workloads? Now the thing that very few people appreciated is that yes, we could. Even at that time in 2017, we had customers that were performing HPC workloads. Now that being said, there were some real limitations on what we could perform. And over those past five years, as we've grown as a company, we've started to really eliminate those frictions for customers to migrate their HPC workloads to the AWS cloud. When I started in 2017, we didn't have our elastic fabric adapter, our low-latency interconnect. So customers were stuck with standard TCP/IP. So for their highly demanding open MPI workloads, we just didn't have the latencies to support them. So the jobs didn't run as efficiently as they could. We didn't have Amazon FSx for Lustre, our managed lustre offering for high performant, POSIX-compliant file system, which is kind of the key to a large portion of HPC workloads is you have to have a high-performance file system. We didn't even, I mean, we had about 25 gigs of networking when I started. Now you look at, with our accelerated instances, we've got 400 gigs of networking. So we've really continued to grow across that spectrum and to eliminate a lot of those really, frictions to adoption. I mean, one of the key ones, we had a open source toolkit that was jointly developed by Intel and AWS called CFN Cluster that customers were using to even instantiate their clusters. So, and now we've migrated that all the way to a fully functional supported service at AWS called AWS Parallel Cluster. And so you've seen over those past five years we have had to develop, we've had to grow, we've had to earn the trust of these customers and say come run your workloads on us and we will demonstrate that we can meet your demanding requirements. And at the same time, there's been, I'd say, more of a cultural acceptance. People have gone away from the, again, five years ago, to what are you doing walking around the show, to say, "Okay, I'm not sure I get it. "I need to look at it. "I, okay, I, now, oh, it needs to be a part "of my architecture but the standard questions, "is it secure? "Is it price performant? "How does it compare to my on-prem?" And really culturally, a lot of it is, just getting IT administrators used to, we're not eliminating a whole field, right? We're just upskilling the people that used to rack and stack actual hardware, to now you're learning AWS services and how to operate within that environment. And it's still key to have those people that are really supporting these infrastructures. And so I'd say it's a little bit of a combination of cultural shift over the past five years, to see that cloud is a super important part of HPC workloads, and part of it's been us meeting the the market segment of where we needed to with innovating both at the hardware level and at the software level, which we're going to continue to do. >> You do have an on-prem story though. I mean, you have outposts. We don't hear a lot of talk about outposts lately, but these innovations, like Inferentia, like Trainium, like the networking innovation you're talking about, are these going to make their way into outposts as well? Will that essentially become this supercomputing solution for customers who want to stay on-prem? >> Well, we'll see what the future lies, but we believe that we've got the, as you noted, we've got the hardware, we've got the network, we've got the storage. All those put together gives you a a high-performance computer, right? And whether you want it to be redundant in your local data center or you want it to be accessible via APIs from the AWS cloud, we want to provide that service to you. >> So to be clear, that's not that's not available now, but that is something that could be made available? >> Outposts are available right now, that have this the services that you need. >> All these capabilities? >> Often a move to cloud, an impetus behind it comes from the highest levels in an organization. They're looking at the difference between OpEx versus CapEx. CapEx for a large HPC environment, can be very, very, very high. Are these HPC clusters consumed as an operational expense? Are you essentially renting time, and then a fundamental question, are these multi-tenant environments? Or when you're referring to batches being run in HPC, are these dedicated HPC environments for customers who are running batches against them? When you think about batches, you think of, there are times when batches are being run and there are times when they're not being run. So that would sort of conjure, in the imagination, multi-tenancy, what does that look like? >> Definitely, and that's been, let me start with your second part first is- >> Yeah. That's been a a core area within AWS is we do not see as, okay we're going to, we're going to carve out this super computer and then we're going to allocate that to you. We are going to dynamically allocate multi-tenant resources to you to perform the workloads you need. And especially with the batch environment, we're going to spin up containers on those, and then as the workloads complete we're going to turn those resources over to where they can be utilized by other customers. And so that's where the batch computing component really is powerful, because as you say, you're releasing resources from workloads that you're done with. I can use those for another portion of the workflow for other work. >> Okay, so it makes a huge difference, yeah. >> You mentioned, that five years ago, people couldn't quite believe that AWS was at this conference. Now you've got a booth right out in the center of the action. What kind of questions are you getting? What are people telling you? >> Well, I love being on the show floor. This is like my favorite part is talking to customers and hearing one, what do they love, what do they want more of? Two, what do they wish we were doing that we're not currently doing? And three, what are the friction points that are still exist that, like, how can I make their lives easier? And what we're hearing is, "Can you help me migrate my workloads to the cloud? "Can you give me the information that I need, "both from a price for performance, "for an operational support model, "and really help me be an internal advocate "within my environment to explain "how my resources can be operated proficiently "within the AWS cloud." And a lot of times it's, let's just take your application a subset of your applications and let's benchmark 'em. And really that, AWS, one of the key things is we are a data-driven environment. And so when you take that data and you can help a customer say like, "Let's just not look at hypothetical, "at synthetic benchmarks, let's take "actually the LS-DYNA code that you're running, perhaps. "Let's take the OpenFOAM code that you're running, "that you're running currently "in your on-premises workloads, "and let's run it on AWS cloud "and let's see how it performs." And then we can take that back to your to the decision makers and say, okay, here's the price for performance on AWS, here's what we're currently doing on-premises, how do we think about that? And then that also ties into your earlier question about CapEx versus OpEx. We have models where actual, you can capitalize a longer-term purchase at AWS. So it doesn't have to be, I mean, depending upon the accounting models you want to use, we do have a majority of customers that will stay with that OpEx model, and they like that flexibility of saying, "Okay, spend as you go." We need to have true ups, and make sure that they have insight into what they're doing. I think one of the boogeyman is that, oh, I'm going to spend all my money and I'm not going to know what's available. And so we want to provide the, the cost visibility, the cost controls, to where you feel like, as an HPC administrator you have insight into what your customers are doing and that you have control over that. And so once you kind of take away some of those fears and and give them the information that they need, what you start to see too is, you know what, we really didn't have a lot of those cost visibility and controls with our on-premises hardware. And we've had some customers tell us we had one portion of the workload where this work center was spending thousands of dollars a day. And we went back to them and said, "Hey, we started to show this, "what you were spending on-premises." They went, "Oh, I didn't realize that." And so I think that's part of a cultural thing that, at an HPC, the question was, well on-premises is free. How do you compete with free? And so we need to really change that culturally, to where people see there is no free lunch. You're paying for the resources whether it's on-premises or in the cloud. >> Data scientists don't worry about budgets. >> Wait, on-premises is free? Paul mentioned something that reminded me, you said you were here in 2017, people said AWS, web, what are you even doing here? Now in 2022, you're talking in terms of migrating to cloud. Paul mentioned outposts, let's say that a customer says, "Hey, I'd like you to put "in a thousand-node cluster in this data center "that I happen to own, but from my perspective, "I want to interact with it just like it's "in your data center." In other words, the location doesn't matter. My experience is identical to interacting with AWS in an AWS data center, in a CoLo that works with AWS, but instead it's my physical data center. When we're tracking the percentage of IT that's that is on-prem versus off-prem. What is that? Is that, what I just described, is that cloud? And in five years are you no longer going to be talking about migrating to cloud because people go, "What do you mean migrating to cloud? "What do you even talking about? "What difference does it make?" It's either something that AWS is offering or it's something that someone else is offering. Do you think we'll be at that point in five years, where in this world of virtualization and abstraction, you talked about Kubernetes, we should be there already, thinking in terms of it doesn't matter as long as it meets latency and sovereignty requirements. So that, your prediction, we're all about insights and supercomputing- >> My prediction- >> In five years, will you still be talking about migrating to cloud or will that be something from the past? >> In five years, I still think there will be a component. I think the majority of the assumption will be that things are cloud-native and you start in the cloud and that there are perhaps, an aspect of that, that will be interacting with some sort of an edge device or some sort of an on-premises device. And we hear more and more customers that are saying, "Okay, I can see the future, "I can see that I'm shrinking my footprint." And, you can see them still saying, "I'm not sure how small that beachhead will be, "but right now I want to at least say "that I'm going to operate in that hybrid environment." And so I'd say, again, the pace of this community, I'd say five years we're still going to be talking about migrations, but I'd say the vast majority will be a cloud-native, cloud-first environment. And how do you classify that? That outpost sitting in someone's data center? I'd say we'd still, at least I'll leave that up to the analysts, but I think it would probably come down as cloud spend. >> Great place to end. Ian, you and I now officially have a bet. In five years we're going to come back. My contention is, no we're not going to be talking about it anymore. >> Okay. >> And kids in college are going to be like, "What do you mean cloud, it's all IT, it's all IT." And they won't remember this whole phase of moving to cloud and back and forth. With that, join us in five years to see the result of this mega-bet between Ian and Dave. I'm Dave Nicholson with theCUBE, here at Supercomputing Conference 2022, day three of our coverage with my co-host Paul Gillin. Thanks again for joining us. Stay tuned, after this short break, we'll be back with more action. (lively music)
SUMMARY :
Welcome back to theCUBE's coverage What are we going to talk about? Let's dive right in. in the queue starts to drop, does it have to be of say the traditional HPC workflow, So is the intersection of Kubernetes And now a lot of CIOs in the to the training workloads. And what Trainium allows you What is the difference between, to be that kind of heavy to say like, "Hey, can you You're very polite. to control the workload, to what are you doing I mean, you have outposts. And whether you want it to be redundant that have this the services that you need. Often a move to cloud, to you to perform the workloads you need. Okay, so it makes a What kind of questions are you getting? the cost controls, to where you feel like, And in five years are you no And so I'd say, again, the not going to be talking of moving to cloud and back and forth.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ian | PERSON | 0.99+ |
Paul | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
400 gigs | QUANTITY | 0.99+ |
2017 | DATE | 0.99+ |
Ian Colle | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
Dallas | LOCATION | 0.99+ |
40% | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
2022 | DATE | 0.99+ |
Annaperna | ORGANIZATION | 0.99+ |
second part | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
Last month | DATE | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
five years ago | DATE | 0.98+ |
five | QUANTITY | 0.98+ |
Two | QUANTITY | 0.98+ |
Supercomputing | ORGANIZATION | 0.98+ |
Lustre | ORGANIZATION | 0.97+ |
Annaperna Labs | ORGANIZATION | 0.97+ |
Trainium | ORGANIZATION | 0.97+ |
five years | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
OpEx | TITLE | 0.96+ |
both | QUANTITY | 0.96+ |
first thing | QUANTITY | 0.96+ |
Supercomputing Conference | EVENT | 0.96+ |
first | QUANTITY | 0.96+ |
West Coast | LOCATION | 0.96+ |
thousands of dollars a day | QUANTITY | 0.96+ |
Supercomputing Conference 2022 | EVENT | 0.95+ |
CapEx | TITLE | 0.94+ |
three | QUANTITY | 0.94+ |
theCUBE | ORGANIZATION | 0.92+ |
East Coast | LOCATION | 0.91+ |
single region | QUANTITY | 0.91+ |
years | QUANTITY | 0.91+ |
thousands of nodes | QUANTITY | 0.88+ |
Parallel Cluster | TITLE | 0.87+ |
about 25 gigs | QUANTITY | 0.87+ |
Madhura Maskasky, Platform9 | Cloud Native at Scale
(uplifting music) >> Hello and welcome to The Cube, here in Palo Alto, California for a special program on cloud-native at scale, enabling next generation cloud or SuperCloud for modern application cloud-native developers. I'm John Furrier, host of The Cube. My pleasure to have here Madhura Maskasky, co-founder and VP of Product at Platform9. Thanks for coming in today for this cloud-native at scale conversation. >> Thank you for having me. >> So, cloud-native at scale, something that we're talking about because we're seeing the next level of mainstream success of containers, Kubernetes and cloud-native developers, basically DevOps in the CICD pipeline. It's changing the landscape of infrastructure as code, it's accelerating the value proposition and the SuperCloud as we call it, has been getting a lot of traction because this next generation cloud is looking a lot different, but kind of the same as the first generation. What's your view on SuperCloud as it fits to cloud-native as scales up? >> Yeah, you know, I think what's interesting, and I think the reason why SuperCloud is a really good and a really fit term for this, and I think, I know my CEO was chatting with you as well, and he was mentioning this as well, but I think there needs to be a different term than just multi-cloud or cloud. And the reason is because as cloud-native and cloud deployments have scaled, I think we've reached a point now where, instead of having the traditional data center style model where you have a few large distributors of infrastructure and workload at a few locations, I think the model is kind of flipped around, right, where you have a large number of micro sites. These micro sites could be your public cloud deployment, your private, on-prem infrastructure deployments, or it could be your edge environment, right? And every single enterprise, every single industry is moving that direction. And so you got to refer that with a terminology that indicates the scale and complexity of it. And so I think SuperCloud is an appropriate term for that. >> So, you brought a couple things I want to dig into. You mentioned edge nodes. We're seeing not only edge nodes being the next kind of area of innovation, mainly because it's just popping up everywhere. And that's just the beginning. What even know what's around the corner. You got buildings, you got IOT, OT and IT kind of coming together, but you also got this idea of regions, global infrastructure is a big part of it. I just saw some news around CloudFlare shutting down a site here. There's policies being made at scale. These new challenges there. Can you share, because you got to have edge. So, hybrid cloud is a winning formula. Everybody knows that it's a steady state. >> Madhura: Yeah. >> But across multiple clouds brings in this new un-engineered area, yet it hasn't been done yet. Spanning clouds. People say they're doing it, but you start to see the toe in the water, it's happening, it's going to happen. It's only going to get accelerated with the edge and beyond globally. So I have to ask you, what is the technical challenges in doing this? Because it's something business consequences as well, but there are technical challenges. Can you share your view on what the technical challenges are for the SuperCloud or across multiple edges and regions? >> Yeah, absolutely. So, I think, you know, in the context of this, this term of SuperCloud, I think, it's sometimes easier to visualize things in terms of two axes, right? I think on one end you can think of the scale in terms of just pure number of nodes that you have, deploy number of clusters in the Kubernetes space. And then, on the other access you would have your distribution factor, right? Which is, do you have these tens of thousands of nodes in one site or do you have them distributed across tens of thousands of sites with one node at each site? Right? And if you have just one flavor of this, there is enough complexity but potentially manageable. But when you are expanding on both these axes you really get to a point where that scale really needs some well thought out, well structured solutions to address it. Right? A combination of homegrown tooling along with your, you know, favorite distribution of Kubernetes is not a strategy that can help you in this environment. It may help you when you have one of this or when your scale is not at the level. >> Can you scope the complexity? Because I mean, I hear a lot of moving parts going on there, the technology's also getting better. We're seeing cloud-native becomes successful. There's a lot to configure, there's a lot to install. Can you scope the scale of the problem? Because about at scale, >> Madhura: Yeah. >> Challenges here. >> Yeah. Absolutely. And I think, you know, I like to call it, you know, the problem that the scale creates, you know, there's various problems, but I think one problem, one way to think about it is you know, it works on my cluster problem, right? So, you know, I come from engineering background and there's a, you know, there's a famous saying between engineers and QA and the support folks, right. Which is, it works on my laptop, which is I tested this change, everything was fantastic, it worked flawlessly on my machine, on production, it's not working. And the exact same problem now happens in these distributed environments, but at massive scale, right. Which is that, you know, developers test their applications, et cetera within the sanctity of their sandbox environments. But once you expose that change in the wild world of your production deployment, right. And the production deployment could be going at the radio cell tower at the edge location where a cluster is running there, or it could be sending, you know, these applications and having them run at my customer site where they might not have configured that cluster exactly the same way as I configured it, or they configured the cluster right. But maybe they didn't deploy the security policies or they didn't deploy the other infrastructure plugins that my app relies on. All of these various factors add their own layer of complexity. And there really isn't a simple way to solve that today. And that is just, you know, one example of an issue that happens. I think another, you know, whole new ballgame of issues come in the context of security, right? Because when you are deploying applications at scale in a distributed manner, you got to make sure someone's job is on the line to ensure that the right security policies are enforced regardless of that scale factor. So, I think that's another example of problems that occur. >> Okay. So, I have to ask about scale because there are a lot of multiple steps involved when you see the success of cloud native. You know, you see some, you know, some experimentation. They set up a cluster, say, it's containers and Kubernetes, and then you say, okay, we got this, we configure it. And then, they do it again and again, they call it day two. Some people call it day one, day two operation, whatever you call it. Once you get past the first initial thing, then you got to scale it. Then you're seeing security breaches, you're seeing configuration errors. This seems to be where the hotspot is. And when companies transition from, I got this to, oh no, it's harder than I thought at scale. Can you share your reaction to that and how you see this playing out? >> Yeah, so, you know, I think it's interesting. There's multiple problems that occur when, you know, the two factors of scale, as we talked about start expanding. I think, one of them is what I like to call the, you know, it works fine on my cluster problem, which is back in, when I was a developer, we used to call this, it works on my laptop problem, which is, you know, you have your perfectly written code that is operating just fine on your machine, your sandbox environment. But the moment it runs production, it comes back with P zeros and P ones from support teams, et cetera. And those issues can be really difficult to triage. Right. And so, in the Kubernetes environment, this problem kind of multi-folds, it goes, you know, escalates to a higher degree because you have your sandbox developer environments, they have their clusters and things work perfectly fine in those clusters because these clusters are typically handcrafted or a combination of some scripting and handcrafting. And so, as you give that change to then run at your production edge location, like say your radio cell tower site or you hand it over to a customer to run it on their cluster, they might not have configured that cluster exactly how you did, or they might not have configured some of the infrastructure plugins. And so the things don't work. And when things don't work, triaging them becomes like (indistinct) hard, right? It's just one of the examples of the problem. Another whole bucket of issues is security, which is you have these distributed clusters at scale, you got to ensure someone's job is on the line to make sure that the security policies are configured properly. >> So, this is a huge problem. I love that comment. That's not happening on my system. It's the classic, you know, debugging mentality. >> Madhura: Yeah. >> But at scale it's hard to do that with error prone. I can see that being a problem. And you guys have a solution you're launching. Can you share what Arlon is this new product? What is it all about? Talk about this new introduction. >> Yeah, absolutely. I'm very, very excited. You know, it's one of the projects that we've been working on for some time now because we are very passionate about this problem and just solving problems at scale in on-prem or at in the cloud or at edge environments. And what Arlon is, it's an open source project and it is a tool, it's a Kubernetes native tool for a complete end-to-end management of not just your clusters, but your clusters, all of the infrastructure that goes within and along the sites of those clusters, security policies, your middleware plugins, and finally your applications. So, what Arlon lets you do in a nutshell is in a declarative way, it lets you handle the configuration and management of all of these components in at scale. >> So, what's the elevator pitch simply put for what dissolves in terms of the chaos you guys are reigning in, what's the bumper sticker? >> Yeah. >> What would it do? >> There's a perfect analogy that I love to reference in this context, which is think of your assembly line, you know, in a traditional, let's say, you know, an auto manufacturing factory or et cetera, and the level of efficiency at scale that assembly line brings, right? Arlon, and if you look at the logo we've designed, it's this funny little robot, and it's because when we think of Arlon, we think of these enterprise large scale environments, you know, sprawling at scale creating chaos because there isn't necessarily a well thought through, well-structured solution that's similar to an assembly line, which is taking each component, you know, addressing them, manufacturing, processing them in a standardized way, then handing to the next stage where again, it gets, you know, processed in a standardized way. And that's what Arlon really does. That's like deliver the pitch. If you have problems of scale of managing your infrastructure, you know, that is distributed. Arlon brings the assembly line level of efficiency and consistency for those. >> So keeping it smooth, the assembly line, things are flowing, CICD, pipelining. >> Madhura: Exactly. >> So, that's what you're trying to simplify that OPS piece for the developer. I mean, it's not really OPS, it's their OPS, it's coding. >> Yeah. Not just developer, the OPS, the operations folks as well, right? Because developers, you know, there is, developers are responsible for one picture of that layer, which is my apps, and then maybe that middle layer of applications that they interface with, but then they hand it over to someone else who's then responsible to ensure that these apps are secured properly, that they are logging, logs are being collected properly, monitoring and observability is integrated. And so, it solves problems for both those teams. >> Yeah, it's DevOps. So, the DevOps is the cloud-needed developer. The option teams have to kind of set policies. Is that where the declarative piece comes in? Is that why that's important? >> Absolutely. Yeah. And, you know, Kubernetes really introduced or elevated this declarative management, right? Because you know, Kubernetes clusters are, or your, yeah, you know, specifications of components that go in Kubernetes are defined in declarative way, and Kubernetes always keeps that state consistent with your defined state. But when you go outside of that world of a single cluster, and when you actually talk about defining the clusters or defining everything that's around it, there really isn't a solution that does that today. And so Arlon addresses that problem at the heart of it, and it does that using existing open source, well-known solutions. >> And, I want get into the benefits, what's in it for me as the customer, developer, but I want to finish this out real quick and get your thoughts. You mentioned open source. Why open source? What's the current state of the product? You run the product group over there, Platform9, is it open source? And you guys have a product that's commercial. Can you explain the open-source dynamic? And first of all, why open source? >> Madhura: Yeah. >> And what is the consumption? I mean, open source is great, people want open source, they can download it, look up the code, but you know, maybe want to buy the commercial. So, I'm assuming you have that thought through, can you share? >> Madhura: Yeah. >> Open source and commercial relationship. >> Yeah. I think, you know, starting with why open source, I think, it's, you know, we as a company, we have, you know, one of the things that's absolutely critical to us is that we take mainstream open-source technologies components, and then we, you know, make them available to our customers at scale through either a SaaS model or on-prem model, right? But, so as we are a company or startup or a company that benefits, you know, in a massive way by this open-source economy, it's only right, I think in my mind that, we do our part of the duty, right? And contribute back to the community that feeds us. And so, you know, we have always held that strongly as one of our principles. And we have, you know, created and built independent products starting all the way with Fission, which was a serverless product, you know, that we had built to various other, you know, examples that I can give. But that's one of the main reasons why open source and also open source because we want the community to really firsthand engage with us on this problem, which is very difficult to achieve if your product is behind a wall, you know, behind a block box. >> Well, and that's what the developers want too. I mean, what we're seeing in reporting with SuperCloud is the new model of consumption is I want to look at the code and see what's in there. >> Madhura: That's right. >> And then also, if I want to use it, I'll do it. Great. That's open source, that's the value. But then at the end of the day, if I want to move fast, that's when people buy in. So it's a new kind of freemium, I guess, business model. I guess that's the way is, well, but that's the benefit of open source. This is why standards and open source growing so fast, you have that confluence of, you know, a way for us to try before they buy, but also actually kind of date the application, if you will. We, you know, Adrian (indistinct) uses the dating metaphor, you know, hey, you know, I want to check it out first before I get married. >> Madhura: Right. >> And that's what open source. So, this is the new, this is how people are selling. This is not just open source, this is how companies are selling. >> Absolutely. Yeah. Yeah. You know, I think in, you know, two things, I think one is just, you know, this cloud-native space is so vast that if you're building a close flow solution, sometimes there's also a risk that it may not apply to every single enterprise's use cases. And so having it open source gives them an opportunity to extend it, expand it, to make it proper to their use case if they choose to do so. Right? But at the same time, what's also critical to us is we are able to provide a supported version of it with an SLA that we, you know, that's backed by us, a Saas-hosted version of it as well, for those customers who choose to go that route, you know, once they have used the open-source version and loved it and want to take it at scale and in production and need a partner to collaborate with, who can, you know, support them for that production environment. >> I have to ask you. Now, let's get into what's in it for the customer. I'm a customer, why should I be enthused about Arlon? What's in it for me? You know. 'Cause if I'm not enthused about it, I'm not going to be confident and it's going to be hard for me to get behind this. Can you share your enthusiastic view of, you know, why I should be enthused about Arlon? I'm a customer. >> Yeah, absolutely. And so, and there's multiple, you know, enterprises that we talk to, many of them, you know, our customers, where this is a very kind of typical story that you hear, which is we have, you know, a Kubernetes distribution. It could be on premise, it could be public cloud-native Kubernetes, and then, we have our CICD pipelines that are automating the deployment of applications, et cetera. And then, there's this gray zone. And the gray zone is well before you can, your CICD pipelines can deploy the apps, somebody needs to do all of that groundwork of, you know, defining those clusters and yeah, you know, properly configuring them. And as these things start by being done hand grown. And then, as you scale, what typically enterprises would do today is they will have their homegrown DIY solutions for this. I mean, a number of folks that I talk to that have built Terraform automation, and then, you know, some of those key developers leave. So, it's a typical open source or typical, you know, DIY challenge. And the reason that they're writing it themselves is not because they want to. I mean, of course, technology is always interesting to everybody, but it's because they can't find a solution that's out there that perfectly fits the problem. And so that's that pitch. I think, (indistinct) would be delighted. The folks that we've talk, you know, spoken with, have been absolutely excited and have, you know, shared that this is a major challenge we have today because we have, you know, few hundreds of clusters on EKS Amazon, and we want to scale them to few thousands, but we don't think we are ready to do that. And this will give us the ability to, >> Yeah, I think, people are scared. I won't say scare, that's a bad word. Maybe I should say that they feel nervous because, you know, at scale, small mistakes can become large mistakes. This is something that is concerning to enterprises. And I think, this is going to come up at (indistinct) this year where enterprises are going to say, okay, I need to see SLAs. I want to see track record, I want to see other companies that have used it. >> Madhura: Yeah. >> How would you answer that question to, or challenge, you know, hey, I love this, but is there any guarantees? Is there any, what's the SLA, I'm an enterprise, I got tight, you know, I love the open source trying to free fast and loose, but I need hardened code. >> Yeah, absolutely. So, two parts to that, right? One is Arlon leverages existing open-source components, products that are extremely popular. Two specifically. One is Arlon uses ArgoCD, which is probably one of the highest rated and used CD open-source tools that's out there, right? It's created by folks that are as part of into team now, you know, really brilliant team. And it's used at scale across enterprises. That's one. Second is Arlon also makes use of cluster API (indistinct), which is a Kubernetes' sub-component, right? For life cycle management of clusters. So, there is enough of, you know, community users, et cetera, around these two products, right? Or open-source projects that will find Arlon to be right up in their alley because they're already comfortable, familiar with ArgoCD. Now, Arlon just extends the scope of what ArgoCD can do. And so, that's one. And then, the second part is going back to your point of the comfort. And that's where, you know, Platform9 has a role to play, which is when you are ready to deploy Arlon at scale, because you've been, you know, playing with it in your (indistinct) test environments, you're happy with what you get with it, then Platform9 will stand behind it and provide that SLA. >> And what's been the reaction from customers you've talked to Platform9 customers with, that are familiar with Argo and then Arlon? What's been some of the feedback? >> Yeah, I think, the feedback's been fantastic. I mean, I can give examples of customers where, you know, initially, you know, when you are telling them about your entire portfolio of solutions, it might not strike a card right away. But then we start talking about Arlon, and we talk about the fact that it uses ArgoCD they start opening up, they say, we have standardized on Argo and we have built these components, homegrown, we would be very interested. Can we co-develop? Does it support these use cases? So, we've had that kind of validation. We've had validation all the way at the beginning of Arlon before we even wrote a single line of code saying, this is something we plan on doing. And the customer said, if you had it today, I would've purchased it. So, it's been really great validation. >> All right. So, next question is, what is the solution to the customer? If I asked you, look at, I have, I'm so busy, my team's overworked. I got a skills gap, I don't need another project that's so I'm so tied up right now, and I'm just chasing my tail. How does Platform9 help me? >> Yeah, absolutely. So I think, you know, one of the core tenants of Platform9 has always been that, we try to bring that public cloud like simplicity by hosting, you know, this in a lot of such similar tools in a SaaS-hosted manner for our customers, right? So, our goal behind doing that is taking away or trying to take away all of that complexity from customer's hands and offloading it to our hands, right? And giving them that full white glove treatment as we call it. And so, from a customer's perspective, one, something like Arlon will integrate with what they have, so, they don't have to rip and replace anything. In fact, it will, even in the next versions, it may even discover your clusters that you have today, and, you know, give you an inventory. And then, >> So, customers have clusters that are growing, that's a sign, >> Correct. >> Call you guys. >> Absolutely. Either they have massive large clusters. Right. That they want to split into smaller clusters, but they're not comfortable doing that today, or they've done that already on say, public cloud or otherwise. And now, they have management challenges. >> So, especially, operationalizing the clusters, whether they want to kind of reset everything and remove things around and reconfigure >> Madhura: Yeah. >> And or scale out. >> That's right. Exactly. >> And you provide that layer of policy. >> Absolutely. Yes. >> That's the key value here. >> That's right. >> So, policy-based configuration for cluster scale up. >> Profile and policy-based, declarative configuration and life cycle management for clusters. >> If I asked you how this enables SuperCloud, what would you say to that? >> I think, this is one of the key ingredients to SuperCloud, right? If you think about a SuperCloud environment, there is at least few key ingredients that come to my mind that are really critical. Like they are, you know, life-saving ingredients at that scale. One is having a really good strategy for managing that scale. You know, in a, going back to assembly line in a very consistent, predictable way. So, that Arlon solves, then you need to compliment that with the right kind of observability and monitoring tools at scale, right? Because ultimately issues are going to happen and you're going to have to figure out, you know, how to solve them fast. And Arlon by the way, also helps in that direction, but you also need observability tools. And then, especially if you're running at on the public cloud, you need some cost management tools. In my mind, these three things are like the most necessary ingredients to make SuperCloud successful. And you know, Arlon flows in one, >> Okay, so now, the next level is, okay, that makes sense. It's under the covers kind of speak under the hood. >> Madhura: Yeah. >> How does that impact the app developers of the cloud-native modern application workflows? Because the impact to me seems the apps are going to be impacted. Are they going to be faster, stronger? I mean, what's the impact, if you do all those things as you mentioned, what's the impact of the apps? >> Yeah, the impact is that your apps are more likely to operate in production the way you expect them to, because the right checks and balances have gone through, and any discrepancies have been identified prior to those apps, prior to your customer running into them, right? Because developers run into this challenge today where there's a split responsibility, right? I'm responsible for my code, I'm responsible for some of these other plugins, but I don't own the stack end to end. I have to rely on my OPS counterpart to do their part, right? And so, this really gives them, you know, the right tooling for that. >> So, this is actually a great kind of relevant point, you know, as cloud becomes more scalable, you're starting to see this fragmentation gone of the days of the full-stack developer to the more specialized role. But this is a key point, and I have to ask you because if this Arlon solution takes place, as you say, and the apps are going to be (indistinct), they're designed to do, the question is, what does the current pain look like? Are the apps breaking? What is the signals to the customer, >> Madhura: Yeah. >> That they should be calling you guys up into implementing Arlon, Argo, and on all the other goodness to automate, what does some of the signals, is it downtime? Is it failed apps, is it latency? What are some of the things that, >> Madhura: Yeah, absolutely. >> Would be indications of things are F'ed up a little bit. >> Yeah. More frequent down times, down times that are, that take longer to triage. And so your, you know, your mean times on resolution, et cetera, are escalating or growing larger, right? Like we have environments of customers where they have a number of folks on in the field that have to take these apps and run them at customer sites. And that's one of our partners, and they're extremely interested in this because the rate of failures they're encountering for this, you know, the field when they're running these apps on site, because the field is automating their clusters that are running on sites using their own scripts. So, these are the kinds of challenges, and those are the pain points, which is, you know, if you're looking to reduce your mean time to resolution, if you're looking to reduce the number of failures that occur on your production site, that's one. And second, if you're looking to manage these at scale environments with a relatively small, focused, nimble OPS team, which has an immediate impact on your budget. So, those are the signals. >> This is the cloud-native at scale situation, the innovation going on. Final thought is your reaction to the idea that, if the world goes digital, which it is, and the confluence of physical and digital coming together, and cloud continues to do its thing, the company becomes the application, not where IT used to be supporting the business, you know, the back office and the (indistinct) terminals and some PCs and handhelds. Now, if technology's running, the business is the business. >> Yeah. >> Company is the application. >> Yeah. >> So, it can't be down. So, there's a lot of pressure on CSOs and CIOs now and boards is saying, how is technology driving the top-line revenue? That's the number one conversation. >> Yeah. >> Do you see the same thing? >> Yeah, it's interesting. I think there's multiple pressures at the CXO, CIO level, right? One is that there needs to be that visibility and clarity and guarantee almost that, you know, the technology that's, you know, that's going to drive your top line is going to drive that in a consistent, reliable, predictable manner. And then second, there is the constant pressure to do that while always lowering your costs of doing it, right? Especially, when you're talking about, let's say, retailers or those kinds of large-scale vendors, they many times make money by lowering the amount that they spend on, you know, providing those goods to their end customers. So, I think those, both those factors kind of come into play and the solution to all of them is usually in a very structured strategy around automation. >> Final question. What does cloud-native at scale look like to you? If all the things happen the way we want them to happen, the magic wand, the magic dust, what does it look like? >> What that looks like to me is a CIO sipping at his desk on coffee, production is running absolutely smooth. And he's running that at a nimble, nimble team size of at the most, a handful of folks that are just looking after things, but things are just taking care of themselves. >> John: And the CIO doesn't exist and there's no CISO, there at the beach. >> (laughs) Yeah. >> Thank you for coming on, sharing the cloud-native at scale here on The Cube. Thank you for your time. >> Fantastic. Thanks for having me. >> Okay. I'm John Furrier here, for special program presentation, special programming cloud-native at scale, enabling SuperCloud modern applications with Platform9. Thanks for watching. (gentle music)
SUMMARY :
My pleasure to have here Madhura Maskasky, and the SuperCloud as we call it, Yeah, you know, I And that's just the beginning. Can you share your view on what So, I think, you know, Can you scope the And that is just, you know, Kubernetes, and then you say, I like to call the, you know, you know, debugging mentality. And you guys have a and along the sites of those in a traditional, let's say, you know, the assembly line, piece for the developer. Because developers, you know, there is, So, the DevOps is the Because you know, Kubernetes clusters are, And you guys have a look up the code, but you know, Open source and And we have, you know, created and built the developers want too. the application, if you will. And that's what open to go that route, you know, enthusiastic view of, you know, And so, and there's multiple, you know, And I think, this is going to I'm an enterprise, I got tight, you know, And that's where, you know, of customers where, you know, and I'm just chasing my tail. clusters that you have today, And now, they have management challenges. That's right. Absolutely. So, policy-based configuration and life cycle management for clusters. at on the public cloud, you Okay, so now, the next level is, Because the impact to me seems the way you expect them to, and I have to ask you Would be indications of points, which is, you know, supporting the business, you know, That's the number one conversation. the technology that's, you know, If all the things happen the What that looks like to me John: And the CIO doesn't Thank you for your time. Thanks for having me. for special program presentation,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Madhura Maskasky | PERSON | 0.99+ |
John | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Madhura | PERSON | 0.99+ |
second part | QUANTITY | 0.99+ |
Arlon | ORGANIZATION | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
one site | QUANTITY | 0.99+ |
Two | QUANTITY | 0.99+ |
first generation | QUANTITY | 0.99+ |
two factors | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
each site | QUANTITY | 0.99+ |
each component | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
Platform9 | ORGANIZATION | 0.99+ |
one flavor | QUANTITY | 0.99+ |
Argo | ORGANIZATION | 0.98+ |
two parts | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
Second | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
SuperCloud | TITLE | 0.98+ |
Adrian | PERSON | 0.98+ |
tens of thousands of nodes | QUANTITY | 0.98+ |
one problem | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
one node | QUANTITY | 0.98+ |
two products | QUANTITY | 0.97+ |
tens of thousands of sites | QUANTITY | 0.97+ |
one picture | QUANTITY | 0.97+ |
The Cube | ORGANIZATION | 0.96+ |
one end | QUANTITY | 0.96+ |
CloudFlare | TITLE | 0.96+ |
Platform9 | TITLE | 0.95+ |
this year | DATE | 0.95+ |
CXO | ORGANIZATION | 0.95+ |
two axes | QUANTITY | 0.94+ |
three things | QUANTITY | 0.94+ |
EKS | ORGANIZATION | 0.93+ |
single line | QUANTITY | 0.92+ |
one example | QUANTITY | 0.91+ |
single cluster | QUANTITY | 0.91+ |
Platform9, Cloud Native at Scale
>>Everyone, welcome to the cube here in Palo Alto, California for a special presentation on Cloud native at scale, enabling super cloud modern applications with Platform nine. I'm John Furry, your host of The Cube. We've got a great lineup of three interviews we're streaming today. Mattor Makki, who's the co-founder and VP of Product of Platform nine. She's gonna go into detail around Arlon, the open source products, and also the value of what this means for infrastructure as code and for cloud native at scale. Bickley the chief architect of Platform nine Cube alumni. Going back to the OpenStack days. He's gonna go into why Arlon, why this infrastructure as code implication, what it means for customers and the implications in the open source community and where that value is. Really great wide ranging conversation there. And of course, Vascar, Gort, the CEO of Platform nine, is gonna talk with me about his views on Super Cloud and why Platform nine has a scalable solutions to bring cloud native at scale. So enjoy the program, see you soon. Hello and welcome to the cube here in Palo Alto, California for a special program on cloud native at scale, enabling next generation cloud or super cloud for modern application cloud native developers. I'm John Forry, host of the Cube. Pleasure to have here me Makowski, co-founder and VP of product at Platform nine. Thanks for coming in today for this Cloudnative at scale conversation. >>Thank you for having >>Me. So Cloudnative at scale, something that we're talking about because we're seeing the, the next level of mainstream success of containers Kubernetes and cloud native develop, basically DevOps in the C I C D pipeline. It's changing the landscape of infrastructure as code, it's accelerating the value proposition and the super cloud as we call it, has been getting a lot of traction because this next generation cloud is looking a lot different, but kind of the same as the first generation. What's your view on Super cloud as it fits to cloud native as scales up? >>Yeah, you know, I think what's interesting, and I think the reason why Super Cloud is a really good and a really fit term for this, and I think, I know my CEO was chatting with you as well, and he was mentioning this as well, but I think there needs to be a different term than just multi-cloud or cloud. And the reason is because as cloud native and cloud deployments have scaled, I think we've reached a point now where instead of having the traditional data center style model, where you have a few large distributors of infrastructure and workload at a few locations, I think the model is kind of flipped around, right? Where you have a large number of micro sites. These micro sites could be your public cloud deployment, your private on-prem infrastructure deployments, or it could be your edge environment, right? And every single enterprise, every single industry is moving in that direction. And so you gotta rougher that with a terminology that, that, that indicates the scale and complexity of it. And so I think super cloud is a, is an appropriate term for >>That. So you brought a couple things I want to dig into. You mentioned Edge Notes. We're seeing not only edge nodes being the next kind of area of innovation, mainly because it's just popping up everywhere. And that's just the beginning. Wouldn't even know what's around the corner. You got buildings, you got iot, o ot, and it kind of coming together, but you also got this idea of regions, global infrastructures, big part of it. I just saw some news around cloud flare shutting down a site here, there's policies being made at scale. These new challenges there. Can you share because you can have edge. So hybrid cloud is a winning formula. Everybody knows that it's a steady state. Yeah. But across multiple clouds brings in this new un engineered area, yet it hasn't been done yet. Spanning clouds. People say they're doing it, but you start to see the toe in the water, it's happening, it's gonna happen. It's only gonna get accelerated with the edge and beyond globally. So I have to ask you, what is the technical challenges in doing this? Because it's something business consequences as well, but there are technical challenge. Can you share your view on what the technical challenges are for the super cloud across multiple edges and >>Regions? Yeah, absolutely. So I think, you know, in in the context of this, the, this, this term of super cloud, I think it's sometimes easier to visualize things in terms of two access, right? I think on one end you can think of the scale in terms of just pure number of nodes that you have, deploy number of clusters in the Kubernetes space. And then on the other access you would have your distribution factor, right? Which is, do you have these tens of thousands of nodes in one site or do you have them distributed across tens of thousands of sites with one node at each site? Right? And if you have just one flavor of this, there is enough complexity, but potentially manageable. But when you are expanding on both these access, you really get to a point where that skill really needs some well thought out, well-structured solutions to address it, right? A combination of homegrown tooling along with your, you know, favorite distribution of Kubernetes is not a strategy that can help you in this environment. It may help you when you have one of this or when you, when you scale, is not at the level. >>Can you scope the complexity? Because I mean, I hear a lot of moving parts going on there, the technology's also getting better. We we're seeing cloud native become successful. There's a lot to configure, there's a lot to install. Can you scope the scale of the problem? Because we're talking about at scale Yep. Challenges here. >>Yeah, absolutely. And I think, you know, I I like to call it, you know, the, the, the problem that the scale creates, you know, there's various problems, but I think one, one problem, one way to think about it is, is, you know, it works on my cluster problem, right? So, you know, I come from engineering background and there's a, you know, there's a famous saying between engineers and QA and the support folks, right? Which is, it works on my laptop, which is I tested this change, everything was fantastic, it worked flawlessly on my machine, on production, It's not working. The exact same problem now happens and these distributed environments, but at massive scale, right? Which is that, you know, developers test their applications, et cetera within the sanctity of their sandbox environments. But once you expose that change in the wild world of your production deployment, right? >>And the production deployment could be going at the radio cell tower at the edge location where a cluster is running there, or it could be sending, you know, these applications and having them run at my customer's site where they might not have configured that cluster exactly the same way as I configured it, or they configured the cluster, right? But maybe they didn't deploy the security policies or they didn't deploy the other infrastructure plugins that my app relies on all of these various factors at their own layer of complexity. And there really isn't a simple way to solve that today. And that is just, you know, one example of an issue that happens. I think another, you know, whole new ball game of issues come in the context of security, right? Because when you are deploying applications at scale in a distributed manner, you gotta make sure someone's job is on the line to ensure that the right security policies are enforced regardless of that scale factor. So I think that's another example of problems that occur. >>Okay. So I have to ask about scale because there are a lot of multiple steps involved when you see the success cloud native, you know, you see some, you know, some experimentation. They set up a cluster, say it's containers and Kubernetes, and then you say, Okay, we got this, we can configure it. And then they do it again and again, they call it day two. Some people call it day one, day two operation, whatever you call it. Once you get past the first initial thing, then you gotta scale it. Then you're seeing security breaches, you're seeing configuration errors. This seems to be where the hotpot is. And when companies transition from, I got this to, Oh no, it's harder than I thought at scale. Can you share your reaction to that and how you see this playing out? >>Yeah, so, you know, I think it's interesting. There's multiple problems that occur when, you know, the, the two factors of scale is we talked about start expanding. I think one of them is what I like to call the, you know, it, it works fine on my cluster problem, which is back in, when I was a developer, we used to call this, it works on my laptop problem, which is, you know, you have your perfectly written code that is operating just fine on your machine, your sandbox environment. But the moment it runs production, it comes back with p zeros and POS from support teams, et cetera. And those issues can be really difficult to try us, right? And so in the Kubernetes environment, this problem kind of multi folds, it goes, you know, escalates to a higher degree because yeah, you have your sandbox developer environments, they have their clusters and things work perfectly fine in those clusters because these clusters are typically handcrafted or a combination of some scripting and handcrafting. >>And so as you give that change to then run at your production edge location, like say you radio sell tower site, or you hand it over to a customer to run it on their cluster, they might not have not have configured that cluster exactly how you did it, or they might not have configured some of the infrastructure plugins. And so the things don't work. And when things don't work, triaging them becomes like ishly hard, right? It's just one of the examples of the problem. Another whole bucket of issues is security, which is, is you have these distributed clusters at scale, you gotta ensure someone's job is on the line to make sure that these security policies are configured properly. >>So this is a huge problem. I love that comment. That's not not happening on my system. It's the classic, you know, debugging mentality. Yeah. But at scale it's hard to do that with error prone. I can see that being a problem. And you guys have a solution you're launching, Can you share what our lawn is, this new product, What is it all about? Talk about this new introduction. >>Yeah, absolutely. I'm very, very excited. You know, it's one of the projects that we've been working on for some time now because we are very passionate about this problem and just solving problems at scale in on-prem or at in the cloud or at edge environments. And what arwan is, it's an open source project and it is a tool, it's a Kubernetes native tool for complete end to end management of not just your clusters, but your clusters. All of the infrastructure that goes within and along the sites of those clusters, security policies, your middleware plugins, and finally your applications. So what alarm lets you do in a nutshell is in a declarative way, it lets you handle the configuration and management of all of these components in at scale. >>So what's the elevator pitch simply put for what this solves in, in terms of the chaos you guys are reigning in. What's the, what's the bumper sticker? Yeah, >>What would it do? There's a perfect analogy that I love to reference in this context, which is think of your assembly line, you know, in a traditional, let's say, you know, an auto manufacturing factory or et cetera, and the level of efficiency at scale that that assembly line brings, right online. And if you look at the logo we've designed, it's this funny little robot. And it's because when we think of online, we, we think of these enterprise large scale environments, you know, sprawling at scale creating chaos because there isn't necessarily a well thought through, well structured solution that's similar to an assembly line, which is taking each components, you know, addressing them, manufacturing, processing them in a standardized way, then handing to the next stage. But again, it gets, you know, processed in a standardized way. And that's what Arlon really does. That's like the I pitch. If you have problems of scale of managing your infrastructure, you know, that is distributed. Arlon brings the assembly line level of efficiency and consistency >>For those. So keeping it smooth, the assembly on things are flowing. C C I CD pipelining. Exactly. So that's what you're trying to simplify that ops piece for the developer. I mean, it's not really ops, it's their ops, it's coding. >>Yeah. Not just developer, the ops, the operations folks as well, right? Because developers, you know, there is, the developers are responsible for one picture of that layer, which is my apps, and then maybe that middleware of application that they interface with, but then they hand it over to someone else who's then responsible to ensure that these apps are secure properly, that they are logging, logs are being collected properly, monitoring and observability integrated. And so it solves problems for both those >>Teams. Yeah. It's DevOps. So the DevOps is the cloud native developer. The OP teams have to kind of set policies. Is that where the declarative piece comes in? Is that why that's important? >>Absolutely. Yeah. And, and, and, and you know, Kubernetes really in introduced or elevated this declarative management, right? Because, you know, c communities clusters are Yeah. Or your, yeah, you know, specifications of components that go in Kubernetes are defined in a declarative way. And Kubernetes always keeps that state consistent with your defined state. But when you go outside of that world of a single cluster, and when you actually talk about defining the clusters or defining everything that's around it, there really isn't a solution that does that today. And so online addresses that problem at the heart of it, and it does that using existing open source well known solutions. >>Ed, do I wanna get into the benefits? What's in it for me as the customer developer? But I want to finish this out real quick and get your thoughts. You mentioned open source. Why open source? What's the, what's the current state of the product? You run the product group over at platform nine, is it open source? And you guys have a product that's commercial? Can you explain the open source dynamic? And first of all, why open source? Yeah. And what is the consumption? I mean, open source is great, People want open source, they can download it, look up the code, but maybe wanna buy the commercial. So I'm assuming you have that thought through, can you share open source and commercial relationship? >>Yeah, I think, you know, starting with why open source? I think it's, you know, we as a company, we have, you know, one of the things that's absolutely critical to us is that we take mainstream open source technologies components and then we, you know, make them available to our customers at scale through either a SaaS model on from model, right? But, so as we are a company or startup or a company that benefits, you know, in a massive way by this open source economy, it's only right, I think in my mind that we do our part of the duty, right? And contribute back to the community that feeds us. And so, you know, we have always held that strongly as one of our principles. And we have, you know, created and built independent products starting all the way with fi, which was a serverless product, you know, that we had built to various other, you know, examples that I can give. But that's one of the main reasons why opensource and also opensource because we want the community to really firsthand engage with us on this problem, which is very difficult to achieve if your product is behind a wall, you know, behind, behind a block box. >>Well, and that's, that's what the developers want too. I mean, what we're seeing in reporting with Super Cloud is the new model of consumption is I wanna look at the code and see what's in there. That's right. And then also, if I want to use it, I, I'll do it. Great. That's open source, that's the value. But then at the end of the day, if I wanna move fast, that's when people buy in. So it's a new kind of freemium, I guess, business model. I guess that's the way that, Well, but that's, that's the benefit. Open source. This is why standards and open source is growing so fast. You have that confluence of, you know, a way for helpers to try before they buy, but also actually kind of date the application, if you will. We, you know, Adrian Karo uses the dating me metaphor, you know, Hey, you know, I wanna check it out first before I get married. Right? And that's what open source, So this is the new, this is how people are selling. This is not just open source, this is how companies are selling. >>Absolutely. Yeah. Yeah. You know, I think, and you know, two things. I think one is just, you know, this, this, this cloud native space is so vast that if you, if you're building a close flow solution, sometimes there's also a risk that it may not apply to every single enterprises use cases. And so having it open source gives them an opportunity to extend it, expand it, to make it proper to their use case if they choose to do so, right? But at the same time, what's also critical to us is we are able to provide a supported version of it with an SLA that we, you know, that's backed by us, a SAS hosted version of it as well, for those customers who choose to go that route, you know, once they have used the open source version and loved it and want to take it at scale and in production and need, need, need a partner to collaborate with, who can, you know, support them for that production >>Environment. I have to ask you now, let's get into what's in it for the customer. I'm a customer, why should I be enthused about Arlo? What's in it for me? You know? Cause if I'm not enthused about it, I'm not gonna be confident and it's gonna be hard for me to get behind this. Can you share your enthusiastic view of, you know, why I should be enthused about Arlo customer? >>Yeah, absolutely. And so, and there's multiple, you know, enterprises that we talk to, many of them, you know, our customers, where this is a very kind of typical story that you hear, which is we have, you know, a Kubernetes distribution. It could be on premise, it could be public clouds, native es, and then we have our C I CD pipelines that are automating the deployment of applications, et cetera. And then there's this gray zone. And the gray zone is well before you can you, your CS CD pipelines can deploy the apps. Somebody needs to do all of their groundwork of, you know, defining those clusters and yeah. You know, properly configuring them. And as these things, these things start by being done hand grown. And then as the, as you scale, what typically enterprises would do today is they will have their home homegrown DIY solutions for this. >>I mean, the number of folks that I talk to that have built Terra from automation, and then, you know, some of those key developers leave. So it's a typical open source or typical, you know, DIY challenge. And the reason that they're writing it themselves is not because they want to. I mean, of course technology is always interesting to everybody, but it's because they can't find a solution that's out there that perfectly fits the problem. And so that's that pitch. I think Spico would be delighted. The folks that we've talked, you know, spoken with, have been absolutely excited and have, you know, shared that this is a major challenge we have today because we have, you know, few hundreds of clusters on s Amazon and we wanna scale them to few thousands, but we don't think we are ready to do that. And this will give us >>Stability. Yeah, I think people are scared, not sc I won't say scare, that's a bad word. Maybe I should say that they feel nervous because, you know, at scale small mistakes can become large mistakes. This is something that is concerning to enterprises. And, and I think this is gonna come up at co con this year where enterprises are gonna say, Okay, I need to see SLAs. I wanna see track record, I wanna see other companies that have used it. Yeah. How would you answer that question to, or, or challenge, you know, Hey, I love this, but is there any guarantees? Is there any, what's the SLAs? I'm an enterprise, I got tight, you know, I love the open source trying to free fast and loose, but I need hardened code. >>Yeah, absolutely. So, so two parts to that, right? One is Arlan leverages existing open source components, products that are extremely popular. Two specifically. One is Lon uses Argo cd, which is probably one of the highest rated and used CD open source tools that's out there, right? It's created by folks that are as part of Intuit team now, you know, really brilliant team. And it's used at scale across enterprises. That's one. Second is arlon also makes use of cluster api capi, which is a ES sub-component, right? For lifecycle management of clusters. So there is enough of, you know, community users, et cetera, around these two products, right? Or, or, or open source projects that will find Arlan to be right up in their alley because they're already comfortable, familiar with algo cd. Now Arlan just extends the scope of what Algo CD can do. And so that's one. And then the second part is going back to a point of the comfort. And that's where, you know, Platform nine has a role to play, which is when you are ready to deploy Alon at scale, because you've been, you know, playing with it in your DEF test environments, you're happy with what you get with it, then Platform nine will stand behind it and provide that sla. >>And what's been the reaction from customers you've talked to Platform nine customers with, with, that are familiar with, with Argo and then Arlo? What's been some of the feedback? >>Yeah, I, I, I think the feedback's been fantastic. I mean, I can give you examples of customers where, you know, initially, you know, when you are, when you're telling them about your entire portfolio of solutions, it might not strike a card right away. But then we start talking about Arlan and, and we talk about the fact that it uses Argo CD and they start opening up, they say, We have standardized on Argo and we have built these components, homegrown, we would be very interested. Can we co-develop? Does it support these use cases? So we've had that kind of validation. We've had validation all the way at the beginning of our line before we even wrote a single line of code saying this is something we plan on doing. And the customer said, If you had it today, I would've purchased it. So it's been really great validation. >>All right. So next question is, what is the solution to the customer? If I asked you, Look it, I have, I'm so busy, my team's overworked. I got a skills gap. I don't need another project that's, I'm so tied up right now and I'm just chasing my tail. How does Platform nine help me? >>Yeah, absolutely. So I think, you know, one of the core tenets of Platform nine has always been that we try to bring that public cloud like simplicity by hosting, you know, this in a lot of such similar tools in a SaaS hosted manner for our customers, right? So our goal behind doing that is taking away or trying to take away all of that complexity from customer's hands and offloading it to our hands, right? And giving them that full white glove treatment as we call it. And so from a customer's perspective, one, something like arlon will integrate with what they have so they don't have to rip and replace anything. In fact, it will, even in the next versions, it may even discover your clusters that you have today and, you know, give you an inventory and that, >>So customers have clusters that are growing, that's a sign correct call you guys. >>Absolutely. Either they're, they have massive large clusters, right? That they wanna split into smaller clusters, but they're not comfortable doing that today, or they've done that already on say, public cloud or otherwise. And now they have management challenges. So >>Especially operationalizing the clusters, whether they want to kind of reset everything and remove things around and reconfigure Yeah. And or scale out. >>That's right. Exactly. >>And you provide that layer of policy. >>Absolutely. >>Yes. That's the key value >>Here. That's right. >>So policy based configuration for cluster scale up >>Profile and policy based declarative configuration and life cycle management for clusters. >>If I asked you how this enables Super club, what would you say to that? >>I think this is one of the key ingredients to super cloud, right? If you think about a super cloud environment, there's at least few key ingredients that that come to my mind that are really critical. Like they are, you know, life saving ingredients at that scale. One is having a really good strategy for managing that scale, you know, in a, going back to assembly line in a very consistent, predictable way so that our lot solves then you, you need to compliment that with the right kind of observability and monitoring tools at scale, right? Because ultimately issues are gonna happen and you're gonna have to figure out, you know, how to solve them fast. And alon by the way, also helps in that direction, but you also need observability tools. And then especially if you're running it on the public cloud, you need some cost management tools. In my mind, these three things are like the most necessary ingredients to make Super Cloud successful. And, you know, alarm flows >>In one. Okay, so now the next level is, Okay, that makes sense. There's under the covers kind of speak under the hood. Yeah. How does that impact the app developers and the cloud native modern application workflows? Because the impact to me, seems the apps are gonna be impacted. Are they gonna be faster, stronger? I mean, what's the impact if you do all those things, as you mentioned, what's the impact of the apps? >>Yeah, the impact is that your apps are more likely to operate in production the way you expect them to, because the right checks and balances have gone through, and any discrepancies have been identified prior to those apps, prior to your customer running into them, right? Because developers run into this challenge to their, where there's a split responsibility, right? I'm responsible for my code, I'm responsible for some of these other plugins, but I don't own the stack end to end. I have to rely on my ops counterpart to do their part, right? And so this really gives them, you know, the right tooling for >>That. So this is actually a great kind of relevant point, you know, as cloud becomes more scalable, you're starting to see this fragmentation gone of the days of the full stack developer to the more specialized role. But this is a key point, and I have to ask you because if this Arlo solution takes place, as you say, and the apps are gonna be stupid, there's designed to do, the question is, what did, does the current pain look like of the apps breaking? What does the signals to the customer Yeah. That they should be calling you guys up into implementing Arlo, Argo, and, and, and on all the other goodness to automate, What are some of the signals? Is it downtime? Is it, is it failed apps, Is it latency? What are some of the things that Yeah, absolutely would be in indications of things are effed up a little bit. >>Yeah. More frequent down times, down times that are, that take longer to triage. And so you are, you know, the, you know, your mean times on resolution, et cetera, are escalating or growing larger, right? Like we have environments of customers where they, they have a number of folks on in the field that have to take these apps and run them at customer sites. And that's one of our partners. And they're extremely interested in this because the, the rate of failures they're encountering for this, you know, the field when they're running these apps on site, because the field is automating their clusters that are running on sites using their own script. So these are the kinds of challenges, and those are the pain points, which is, you know, if you're looking to reduce your, your meantime to resolution, if you're looking to reduce the number of failures that occur on your production site, that's one. And second, if you are looking to manage these at scale environments with a relatively small, focused, nimble ops team, which has an immediate impact on your, So those are, those are the >>Signals. This is the cloud native at scale situation, the innovation going on. Final thought is your reaction to the idea that if the world goes digital, which it is, and the confluence of physical and digital coming together, and cloud continues to do its thing, the company becomes the application, not where it used to be supporting the business, you know, the back office and the IIA terminals and some PCs and handhelds. Now if technology's running, the business is the business. Yeah. The company's the application. Yeah. So it can't be down. So there's a lot of pressure on, on CSOs and CIOs now and see, and boards is saying, how is technology driving the top line revenue? That's the number one conversation. Yeah. Do you see that same thing? >>Yeah. It's interesting. I think there's multiple pressures at the CXO CIO level, right? One is that there needs to be that visibility and clarity and guarantee almost that, you know, that the, the technology that's, you know, that's gonna drive your top line is gonna drive that in a consistent, reliable, predictable manner. And then second, there is the constant pressure to do that while always lowering your costs of doing it, right? Especially when you're talking about, let's say retailers or those kinds of large scale vendors, they many times make money by lowering the amount that they spend on, you know, providing those goods to their end customers. So I think those, both those factors kind of come into play and the solution to all of them is usually in a very structured strategy around automation. >>Final question. What does cloudnative at scale look like to you? If all the things happen the way we want 'em to happen, The magic wand, the magic dust, what does it look like? >>What that looks like to me is a CIO sipping at his desk on coffee production is running absolutely smooth. And his, he's running that at a nimble, nimble team size of at the most, a handful of folks that are just looking after things with things. So just >>Taking care of, and the CIO doesn't exist. There's no CSO there at the beach. >>Yeah. >>Thank you for coming on, sharing the cloud native at scale here on the cube. Thank you for your time. >>Fantastic. Thanks for having >>Me. Okay. I'm John Fur here for special program presentation, special programming cloud native at scale, enabling super cloud modern applications with Platform nine. Thanks for watching. Welcome back everyone to the special presentation of cloud native at scale, the cube and platform nine special presentation going in and digging into the next generation super cloud infrastructure as code and the future of application development. We're here at Bickley, who's the chief architect and co-founder of Platform nine b. Great to see you Cube alumni. We, we met at an OpenStack event in about eight years ago, or well later, earlier when opens Stack was going. Great to see you and great to see congratulations on the success of platform nine. >>Thank you very much. >>Yeah. You guys have been at this for a while and this is really the, the, the year we're seeing the, the crossover of Kubernetes because of what happens with containers. Everyone now was realized, and you've seen what Docker's doing with the new docker, the open source Docker now just a success Exactly. Of containerization, right? And now the Kubernetes layer that we've been working on for years is coming, bearing fruit. This is huge. >>Exactly. Yes. >>And so as infrastructure's code comes in, we talked to Bacar talking about Super Cloud, I met her about, you know, the new Arlon, our R lawn you guys just launched, the infrastructure's code is going to another level. And then it's always been DevOps infrastructure is code. That's been the ethos that's been like from day one, developers just code. Then you saw the rise of serverless and you see now multi-cloud or on the horizon, connect the dots for us. What is the state of infrastructures code today? >>So I think, I think I'm, I'm glad you mentioned it, everybody or most people know about infrastructures code. But with Kubernetes, I think that project has evolved at the concept even further. And these dates, it's infrastructure as configuration, right? So, which is an evolution of infrastructure as code. So instead of telling the system, here's how I want my infrastructure by telling it, you know, do step A, B, C, and D instead with Kubernetes, you can describe your desired state declaratively using things called manifest resources. And then the system kind of magically figures it out and tries to converge the state towards the one that you specify. So I think it's, it's a even better version of infrastructures code. >>Yeah, yeah. And, and that really means it's developer just accessing resources. Okay. Not declaring, Okay, give me some compute, stand me up some, turn the lights on, turn 'em off, turn 'em on. That's kind of where we see this going. And I like the configuration piece. Some people say composability, I mean now with open source, so popular, you don't have to have to write a lot of code. It's code being developed. And so it's into integration, it's configuration. These are areas that we're starting to see computer science principles around automation, machine learning, assisting open source. Cuz you got a lot of code that's right in hearing software, supply chain issues. So infrastructure as code has to factor in these new, new dynamics. Can you share your opinion on these new dynamics of, as open source grows, the glue layers, the configurations, the integration, what are the core issues? >>I think one of the major core issues is with all that power comes complexity, right? So, you know, despite its expressive power systems like Kubernetes and declarative APIs let you express a lot of complicated and complex stacks, right? But you're dealing with hundreds if not thousands of these yamo files or resources. And so I think, you know, the emergence of systems and layers to help you manage that complexity is becoming a key challenge and opportunity in, in this space that, >>That's, I wrote a LinkedIn post today was comments about, you know, hey, enterprise is the new breed, the trend of SaaS companies moving our consumer comp consumer-like thinking into the enterprise has been happening for a long time, but now more than ever, you're seeing it the old way used to be solve complexity with more complexity and then lock the customer in. Now with open source, it's speed, simplification and integration, right? These are the new dynamic power dynamics for developers. Yeah. So as companies are starting to now deploy and look at Kubernetes, what are the things that need to be in place? Because you have some, I won't say technical debt, but maybe some shortcuts, some scripts here that make it look like infrastructure is code. People have done some things to simulate or or make infrastructure as code happen. Yes. But to do it at scale Yes. Is harder. What's your take on this? What's your >>View? It's hard because there's a per proliferation of methods, tools, technologies. So for example, today it's very common for DevOps and platform engineering tools, I mean, sorry, teams to have to deploy a large number of Kubernetes clusters, but then apply the applications and configurations on top of those clusters. And they're using a wide range of tools to do this, right? For example, maybe Ansible or Terraform or bash scripts to bring up the infrastructure and then the clusters. And then they may use a different set of tools such as Argo CD or other tools to apply configurations and applications on top of the clusters. So you have this sprawl of tools. You, you also have this sprawl of configurations and files because the more objects you're dealing with, the more resources you have to manage. And there's a risk of drift that people call that where, you know, you think you have things under control, but some people from various teams will make changes here and there and then before the end of the day systems break and you have no idea of tracking them. So I think there's real need to kind of unify, simplify, and try to solve these problems using a smaller, more unified set of tools and methodologies. And that's something that we try to do with this new project. Arlon. >>Yeah. So, so we're gonna get into Arlan in a second. I wanna get into the why Arlon. You guys announced that at our GoCon, which was put on here in Silicon Valley at the, at the by intu. They had their own little day over there at their headquarters. But before we get there, Vascar, your CEO came on and he talked about Super Cloud at our inaugural event. What's your definition of super cloud? If you had to kind of explain that to someone at a cocktail party or someone in the industry technical, how would you look at the super cloud trend that's emerging? It's become a thing. What's your, what would be your contribution to that definition or the narrative? >>Well, it's, it's, it's funny because I've actually heard of the term for the first time today, speaking to you earlier today. But I think based on what you said, I I already get kind of some of the, the gist and the, the main concepts. It seems like super cloud, the way I interpret that is, you know, clouds and infrastructure, programmable infrastructure, all of those things are becoming commodity in a way. And everyone's got their own flavor, but there's a real opportunity for people to solve real business problems by perhaps trying to abstract away, you know, all of those various implementations and then building better abstractions that are perhaps business or application specific to help companies and businesses solve real business problems. >>Yeah, I remember that's a great, great definition. I remember, not to date myself, but back in the old days, you know, IBM had a proprietary network operating system, so to deck for the mini computer vendors, deck net and SNA respectively. But T C P I P came out of the osi, the open systems interconnect and remember, ethernet beat token ring out. So not to get all nerdy for all the young kids out there, look, just look up token ring, you'll see, you've probably never heard of it. It's IBM's, you know, connection for the internet at the, the layer too is Amazon, the ethernet, right? So if T C P I P could be the Kubernetes and the container abstraction that made the industry completely change at that point in history. So at every major inflection point where there's been serious industry change and wealth creation and business value, there's been an abstraction Yes. Somewhere. Yes. What's your reaction to that? >>I think this is, I think a saying that's been heard many times in this industry and, and I forgot who originated it, but I think the saying goes like, there's no problem that can't be solved with another layer of indirection, right? And we've seen this over and over and over again where Amazon and its peers have inserted this layer that has simplified, you know, computing and, and infrastructure management. And I believe this trend is going to continue, right? The next set of problems are going to be solved with these insertions of additional abstraction layers. I think that that's really a, yeah, it's gonna continue. >>It's interesting. I just really wrote another post today on LinkedIn called the Silicon Wars AMD Stock is down arm has been on rise, we've remember pointing for many years now, that arm's gonna be hugely, it has become true. If you look at the success of the infrastructure as a service layer across the clouds, Azure, aws, Amazon's clearly way ahead of everybody. The stuff that they're doing with the silicon and the physics and the, the atoms, the pro, you know, this is where the innovation, they're going so deep and so strong at ISAs, the more that they get that gets come on, they have more performance. So if you're an app developer, wouldn't you want the best performance and you'd wanna have the best abstraction layer that gives you the most ability to do infrastructures, code or infrastructure for configuration, for provisioning, for managing services. And you're seeing that today with service MeSHs, a lot of action going on in the service mesh area in, in this community of co con, which will be a covering. So that brings up the whole what's next? You guys just announced our lawn at ar GoCon, which came out of Intuit. We've had Maria Teel at our super cloud event, She's a cto, you know, they're all in the cloud. So they contributed that project. Where did Arlon come from? What was the origination? What's the purpose? Why our lawn, why this announcement? Yeah, >>So the, the inception of the project, this was the result of us realizing that problem that we spoke about earlier, which is complexity, right? With all of this, these clouds, these infrastructure, all the variations around and you know, compute storage networks and the proliferation of tools we talked about the Ansibles and Terraforms and Kubernetes itself, you can think of that as another tool, right? We saw a need to solve that complexity problem, and especially for people and users who use Kubernetes at scale. So when you have, you know, hundreds of clusters, thousands of applications, thousands of users spread out over many, many locations, there, there needs to be a system that helps simplify that management, right? So that means fewer tools, more expressive ways of describing the state that you want and more consistency. And, and that's why, you know, we built AR lawn and we built it recognizing that many of these problems or sub problems have already been solved. So Arlon doesn't try to reinvent the wheel, it instead rests on the shoulders of several giants, right? So for example, Kubernetes is one building block, GI ops, and Argo CD is another one, which provides a very structured way of applying configuration. And then we have projects like cluster API and cross plane, which provide APIs for describing infrastructure. So arlon takes all of those building blocks and builds a thin layer, which gives users a very expressive way of defining configuration and desired state. So that's, that's kind of the inception of, And >>What's the benefit of that? What does that give the, what does that give the developer, the user, in this case, >>The developers, the, the platform engineer, team members, the DevOps engineers, they get a a ways to provision not just infrastructure and clusters, but also applications and configurations. They get a way, a system for provisioning, configuring, deploying, and doing life cycle management in a, in a much simpler way. Okay. Especially as I said, if you're dealing with a large number of applications. >>So it's like an operating fabric, if you will. Yes. For them. Okay, so let's get into what that means for up above and below the, the, this abstraction or thin layer below the infrastructure. We talked a lot about what's going on below that. Yeah. Above our workloads at the end of the day, and I talk to CXOs and IT folks that, that are now DevOps engineers. They care about the workloads and they want the infrastructure's code to work. They wanna spend their time getting in the weeds, figuring out what happened when someone made a push that that happened or something happened. They need observability and they need to, to know that it's working. That's right. And here's my workloads running effectively. So how do you guys look at the workload side of it? Cuz now you have multiple workloads on these fabric, right? >>So workloads, so Kubernetes has defined kind of a standard way to describe workloads and you can, you know, tell Kubernetes, I want to run this container this particular way, or you can use other projects that are in the Kubernetes cloud native ecosystem, like K native, where you can express your application in more at a higher level, right? But what's also happening is in addition to the workloads, DevOps and platform engineering teams, they need to very often deploy the applications with the clusters themselves. Clusters are becoming this commodity. It's, it's becoming this host for the application and it kind of comes bundled with it. In many cases it is like an appliance, right? So DevOps teams have to provision clusters at a really incredible rate and they need to tear them down. Clusters are becoming more, >>It's coming like an EC two instance, spin up a cluster. We've heard people used words like that. That's >>Right. And before arlon you kind of had to do all of that using a different set of tools as, as I explained. So with AR loan you can kind of express everything together. You can say I want a cluster with a health monitoring stack and a logging stack and this ingress controller and I want these applications and these security policies. You can describe all of that using something we call the profile. And then you can stamp out your app, your applications and your clusters and manage them in a very, So >>It's essentially standard, like creates a mechanism. Exactly. Standardized, declarative kind of configurations. And it's like a playbook, just deploy it. Now what there is between say a script like I'm, I have scripts, I can just automate scripts >>Or yes, this is where that declarative API and infrastructure as configuration comes in, right? Because scripts, yes you can automate scripts, but the order in which they run matters, right? They can break, things can break in the middle and, and sometimes you need to debug them. Whereas the declarative way is much more expressive and powerful. You just tell the system what you want and then the system kind of figures it out. And there are these things are controllers which will in the background reconcile all the state to converge towards your desire. It's a much more powerful, expressive and reliable way of getting things done. >>So infrastructure as configuration is built kind of on, it's a super set of infrastructures code because it's >>An evolution. >>You need edge's code, but then you can configure the code by just saying do it. You basically declaring saying Go, go do that. That's right. Okay, so, alright, so cloud native at scale, take me through your vision of what that means. Someone says, Hey, what does cloud native at scale mean? What's success look like? How does it roll out in the future as you, not future next couple years. I mean people are now starting to figure out, okay, it's not as easy as it sounds. Kubernetes has value. We're gonna hear this year at CubeCon a lot of this, what does cloud native at scale >>Mean? Yeah, there are different interpretations, but if you ask me, when people think of scale, they think of a large number of deployments, right? Geographies, many, you know, supporting thousands or tens or millions of, of users there, there's that aspect to scale. There's also an equally important a aspect of scale, which is also something that we try to address with Arran. And that is just complexity for the people operating this or configuring this, right? So in order to describe that desired state, and in order to perform things like maybe upgrades or updates on a very large scale, you want the humans behind that to be able to express and direct the system to do that in, in relatively simple terms, right? And so we want the tools and the abstractions and the mechanisms available to the user to be as powerful but as simple as possible. So there's, I think there's gonna be a number and there have been a number of CNCF and cloud native projects that are trying to attack that complexity problem as well. And Arlon kind of falls in in that >>Category. Okay, so I'll put you on the spot rogue, that CubeCon coming up and now this'll be shipping this segment series out before. What do you expect to see at this year? It's the big story this year. What's the, what's the most important thing happening? Is it in the open source community and also within a lot of the, the people jockeying for leadership. I know there's a lot of projects and still there's some white space in the overall systems map about the different areas get run time and there's ability in all these different areas. What's the, where's the action? Where, where's the smoke? Where's the fire? Where's the piece? Where's the tension? >>Yeah, so I think one thing that has been happening over the past couple of coupon and I expect to continue and, and that is the, the word on the street is Kubernetes is getting boring, right? Which is good, right? >>Boring means simple. >>Well, well >>Maybe, >>Yeah, >>Invisible, >>No drama, right? So, so the, the rate of change of the Kubernetes features and, and all that has slowed but in, in a, in a positive way. But there's still a general sentiment and feeling that there's just too much stuff. If you look at a stack necessary for hosting applications based on Kubernetes, there are just still too many moving parts, too many components, right? Too much complexity. I go, I keep going back to the complexity problem. So I expect Cube Con and all the vendors and the players and the startups and the people there to continue to focus on that complexity problem and introduce further simplifications to, to the stack. >>Yeah. Vic, you've had an storied career VMware over decades with them within 12 years with 14 years or something like that. Big number co-founder here a platform. I you's been around for a while at this game, man. We talked about OpenStack, that project we interviewed at one of their events. So OpenStack was the beginning of that, this new revolution. I remember the early days it was, it wasn't supposed to be an alternative to Amazon, but it was a way to do more cloud cloud native. I think we had a Cloud Aati team at that time. We would joke we, you know, about, about the dream. It's happening now, now at Platform nine. You guys have been doing this for a while. What's the, what are you most excited about as the chief architect? What did you guys double down on? What did you guys pivot from or two, did you do any pivots? Did you extend out certain areas? Cuz you guys are in a good position right now, a lot of DNA in Cloud native. What are you most excited about and what does Platform Nine bring to the table for customers and for people in the industry watching this? >>Yeah, so I think our mission really hasn't changed over the years, right? It's been always about taking complex open source software because open source software, it's powerful. It solves new problems, you know, every year and you have new things coming out all the time, right? Opens Stack was an example and then Kubernetes took the world by storm. But there's always that complexity of, you know, just configuring it, deploying it, running it, operating it. And our mission has always been that we will take all that complexity and just make it, you know, easy for users to consume regardless of the technology, right? So the successor to Kubernetes, you know, I don't have a crystal ball, but you know, you have some indications that people are coming up of new and simpler ways of running applications. There are many projects around there who knows what's coming next year or the year after that. But platform will a, platform nine will be there and we will, you know, take the innovations from the the community. We will contribute our own innovations and make all of those things very consumable to customers. >>Simpler, faster, cheaper. Exactly. Always a good business model technically to make that happen. Yes. Yeah, I think the, the reigning in the chaos is key, you know, Now we have now visibility into the scale. Final question before we depart this segment. What is at scale, how many clusters do you see that would be a watermark for an at scale conversation around an enterprise? Is it workloads we're looking at or, or clusters? How would you, Yeah, how would you describe that? When people try to squint through and evaluate what's a scale, what's the at scale kind of threshold? >>Yeah. And, and the number of clusters doesn't tell the whole story because clusters can be small in terms of the number of nodes or they can be large. But roughly speaking when we say, you know, large scale cluster deployments, we're talking about maybe hundreds, two thousands. >>Yeah. And final final question, what's the role of the hyperscalers? You got AWS continuing to do well, but they got their core ias, they got a PAs, they're not too too much putting a SaaS out there. They have some SaaS apps, but mostly it's the ecosystem. They have marketplaces doing, doing over $2 billion billions of transactions a year and, and it's just like, just sitting there. It hasn't really, they're now innovating on it, but that's gonna change ecosystems. What's the role the cloud play in the cloud need of its scale? >>The, the hyper squares? >>Yeah, yeah. A's Azure Google, >>You mean from a business perspective, they're, they have their own interests that, you know, that they're, they will keep catering to, they, they will continue to find ways to lock their users into their ecosystem of services and, and APIs. So I don't think that's gonna change, right? They're just gonna keep well, >>They got great performance. I mean, from a, from a hardware standpoint, yes. That's gonna be key, >>Right? Yes. I think the, the move from X 86 being the dominant way and platform to run workloads is changing, right? That, that, that, that, and I think the, the hyper skaters really want to be in the game in terms of, you know, the, the new risk and arm ecosystems, the platforms. >>Yeah. Not joking aside, Paul Morritz, when he was the CEO of VMware, when he took over once said, I remember our first year doing the cube. Oh the cloud is one big distributed computer. It's, it's hardware and you got software and you got middleware and he kinda over, well he's kind of tongue in cheek, but really you're talking about large compute and sets of services that is essentially a distributed computer. Yes, >>Exactly. >>It's, we're back in the same game. Thank you for coming on the segment. Appreciate your time. This is cloud native at scale special presentation with Platform nine. Really unpacking super cloud Arlon open source and how to run large scale applications on the cloud, cloud native develop for developers. And John Furrier with the cube. Thanks for Washington. We'll stay tuned for another great segment coming right up. Hey, welcome back everyone to Super Cloud 22. I'm John Fur, host of the Cuba here all day talking about the future of cloud. Where's it all going? Making it super multi-cloud is around the corner and public cloud is winning. Got the private cloud on premise and Edge. Got a great guest here, Vascar Gorde, CEO of Platform nine, just on the panel on Kubernetes. An enabler blocker. Welcome back. Great to have you on. >>Good to see you >>Again. So Kubernetes is a blocker enabler by, with a question mark I put on on there. Panel was really to discuss the role of Kubernetes. Now great conversation operations is impacted. What's just thing about what you guys are doing at Platform nine? Is your role there as CEO and the company's position, kind of like the world spun into the direction of Platform nine while you're at the helm, right? >>Absolutely. In fact, things are moving very well and since they came to us, it was an insight to call ourselves the platform company eight years ago, right? So absolutely whether you are doing it in public clouds or private clouds, you know, the application world is moving very fast in trying to become digital and cloud native. There are many options for you to run the infrastructure. The biggest blocking factor now is having a unified platform. And that's what where we come into >>Patrick, we were talking before we came on stage here about your background and we were kind of talking about the glory days in 2000, 2001 when the first ASPs application service providers came out. Kind of a SaaS vibe, but that was kind of all kind of cloud-like >>It wasn't, >>And web services started then too. So you saw that whole growth. Now, fast forward 20 years later, 22 years later, where we are now, when you look back then to here and all the different cycles, >>In fact, you know, as we were talking offline, I was in one of those ASPs in the year 2000 where it was a novel concept of saying we are providing a software and a capability as a service, right? You sign up and start using it. I think a lot has changed since then. The tooling, the tools, the technology has really skyrocketed. The app development environment has really taken off exceptionally well. There are many, many choices of infrastructure now, right? So I think things are in a way the same but also extremely different. But more importantly now for any company, regardless of size, to be a digital native, to become a digital company is extremely mission critical. It's no longer a nice to have everybody's in the journey somewhere. >>Everyone is going digital transformation here. Even on a so-called downturn recession that's upcoming inflations sea year. It's interesting. This is the first downturn, the history of the world where the hyperscale clouds have been pumping on all cylinders as an economic input. And if you look at the tech trends, GDPs down, but not tech. Nope. Cause pandemic showed everyone digital transformation is here and more spend and more growth is coming even in, in tech. So this is a unique factor which proves that that digital transformation's happening and company, every company will need a super cloud. >>Everyone, every company, regardless of size, regardless of location, has to become modernize their infrastructure. And modernizing infrastructure is not just some, you know, new servers and new application tools. It's your approach, how you're serving your customers, how you're bringing agility in your organization. I think that is becoming a necessity for every enterprise to survive. >>I wanna get your thoughts on Super Cloud because one of the things Dave Alon and I want to do with Super Cloud and calling it that was we, I, I personally, and I know Dave as well, he can, I'll speak from, he can speak for himself. We didn't like multi-cloud. I mean not because Amazon said don't call things multi-cloud, it just didn't feel right. I mean everyone has multiple clouds by default. If you're running productivity software, you have Azure and Office 365. But it wasn't truly distributed. It wasn't truly decentralized, it wasn't truly cloud enabled. It didn't, it felt like they're not ready for a market yet. Yet public clouds booming on premise. Private cloud and Edge is much more on, you know, more, More dynamic, more unreal. >>Yeah. I think the reason why we think Super cloud is a better term than multi-cloud. Multi-cloud are more than one cloud, but they're disconnected. Okay, you have a productivity cloud, you have a Salesforce cloud, you may have, everyone has an internal cloud, right? So, but they're not connected. So you can say, okay, it's more than one cloud. So it's, you know, multi-cloud. But super cloud is where you are actually trying to look at this holistically. Whether it is on-prem, whether it is public, whether it's at the edge, it's a store at the branch. You are looking at this as one unit. And that's where we see the term super cloud is more applicable because what are the qualities that you require if you're in a super cloud, right? You need choice of infrastructure, you need, but at the same time you need a single pan or a single platform for you to build your innovations on, regardless of which cloud you're doing it on, right? So I think Super Cloud is actually a more tightly integrated orchestrated management philosophy we think. >>So let's get into some of the super cloud type trends that we've been reporting on. Again, the purpose of this event is as a pilot to get the conversations flowing with, with the influencers like yourselves who are running companies and building products and the builders, Amazon and Azure are doing extremely well. Google's coming up in third Cloudworks in public cloud. We see the use cases on premises use cases. Kubernetes has been an interesting phenomenon because it's become from the developer side a little bit, but a lot of ops people love Kubernetes. It's really more of an ops thing. You mentioned OpenStack earlier. Kubernetes kind of came out of that open stack. We need an orchestration. And then containers had a good shot with, with Docker. They re pivoted the company. Now they're all in an open source. So you got containers booming and Kubernetes as a new layer there. >>What's, >>What's the take on that? What does that really mean? Is that a new defacto enabler? It >>Is here. It's for here for sure. Every enterprise somewhere in the journey is going on. And you know, most companies are, 70 plus percent of them have 1, 2, 3 container based, Kubernetes based applications now being rolled out. So it's very much here. It is in production at scale by many customers. And it, the beauty of it is yes, open source, but the biggest gating factor is the skill set. And that's where we have a phenomenal engineering team, right? So it's, it's one thing to buy a tool and >>Just be clear, you're a managed service for Kubernetes. >>We provide, provide a software platform for cloud acceleration as a service and it can run anywhere. It can run in public private. We have customers who do it in truly multi-cloud environments. It runs on the edge, it runs at this in stores about thousands of stores in a retailer. So we provide that and also for specific segments where data sovereignty and data residency are key regulatory reasons. We also un on-prem as an air gap version. Can >>You give an example on how you guys are deploying your platform to enable a super cloud experience for your customer? Right. >>So I'll give you two different examples. One is a very large networking company, public networking company. They have hundreds of products, hundreds of r and d teams that are building different, different products. And if you look at few years back, each one was doing it on a different platforms, but they really needed to bring the agility. And they worked with us now over three years where we are their build test dev pro platform where all their products are built on, right? And it has dramatically increased their agility to release new products. Number two, it actually is a light out operation. In fact, the customer says like, like the Maytag service person, cuz we provide it as a service and it barely takes one or two people to maintain it for them. >>So it's kinda like an SRE vibe. One person managing a >>Large 4,000 engineers building infrastructure >>On their tools, >>Whatever they want on their tools. They're using whatever app development tools they use, but they use our platform. What >>Benefits are they seeing? Are they seeing speed? >>Speed, definitely. Okay. Definitely they're speeding. Speed uniformity because now they're building able to build, so their customers who are using product A and product B are seeing a similar set of tools that are being used. >>So a big problem that's coming outta this super cloud event that we're, we're seeing and we heard it all here, ops and security teams. Cause they're kind of part of one thing, but option security specifically need to catch up speed wise. Are you delivering that value to ops and security? Right? >>So we, we work with ops and security teams and infrastructure teams and we layer on top of that. We have like a platform team. If you think about it, depending on where you have data centers, where you have infrastructure, you have multiple teams, okay, but you need a unified platform. Who's your buyer? Our buyer is usually, you know, the product divisions of companies that are looking at or the CTO would be a buyer for us functionally cio definitely. So it it's, it's somewhere in the DevOps to infrastructure. But the ideal one we are beginning to see now many large corporations are really looking at it as a platform and saying we have a platform group on which any app can be developed and it is run on any infrastructure. So the platform engineering teams. So >>You working two sides to that coin. You've got the dev side and then >>And then infrastructure >>Side. >>Okay. Another customer that I give an example, which I would say is kind of the edge of the store. So they have thousands of stores. Retail, retail, you know food retailer, right? They have thousands of stores that are on the globe, 50,000, 60,000. And they really want to enhance the customer experience that happens when you either order the product or go into the store and pick up your product or buy or browse or sit there. They have applications that were written in the nineties and then they have very modern AIML applications today. They want something that will not have to send an IT person to install a rack in the store or they can't move everything to the cloud because the store operations has to be local. The menu changes based on it's classic edge. It's classic edge, yeah. Right? They can't send it people to go install rack access servers then they can't sell software people to go install the software and any change you wanna put through that, you know, truck roll. So they've been working with us where all they do is they ship, depending on the size of the store, one or two or three little servers with instructions that >>You, you say little servers like how big one like a box, like a small little box, >>Right? And all the person in the store has to do like what you and I do at home and we get a, you know, a router is connect the power, connect the internet and turn the switch on. And from there we pick it up. >>Yep. >>We provide the operating system, everything and then the applications are put on it. And so that dramatically brings the velocity for them. They manage thousands of >>Them. True plug and play >>Two, plug and play thousands of stores. They manage it centrally. We do it for them, right? So, so that's another example where on the edge then we have some customers who have both a large private presence and one of the public clouds. Okay. But they want to have the same platform layer of orchestration and management that they can use regardless of the locations. >>So you guys got some success. Congratulations. Got some traction there. It's awesome. The question I want to ask you is that's come up is what is truly cloud native? Cuz there's lift and shift of the cloud >>That's not cloud native. >>Then there's cloud native. Cloud native seems to be the driver for the super cloud. How do you talk to customers? How do you explain when someone says what's cloud native, what isn't cloud native? >>Right. Look, I think first of all, the best place to look at what is the definition and what are the attributes and characteristics of what is truly a cloud native, is CNC foundation. And I think it's very well documented, very well. >>Tucan, of course Detroit's >>Coming so, so it's already there, right? So we follow that very closely, right? I think just lifting and shifting your 20 year old application onto a data center somewhere is not cloud native. Okay? You can't put to cloud, not you have to rewrite and redevelop your application in business logic using modern tools. Hopefully more open source and, and I think that's what Cloudnative is and we are seeing a lot of our customers in that journey. Now everybody wants to be cloudnative, but it's not that easy, okay? Because it's, I think it's first of all, skill set is very important. Uniformity of tools that there's so many tools there. Thousands and thousands of tools you could spend your time figuring out which tool to use. Okay? So I think the complexity is there, but the business benefits of agility and uniformity and customer experience are truly being done. >>And I'll give you an example, I don't know how clear native they are, right? And they're not a customer of ours, but you order pizzas, you do, right? If you just watch the pizza industry, how dominoes actually increase their share and mind share and wallet share was not because they were making better pizzas or not, I don't know anything about that, but the whole experience of how you order, how you watch what's happening, how it's delivered. There were a pioneer in it. To me, those are the kinds of customer experiences that cloud native can provide. >>Being agility and having that flow to the application changes what the expectations >>Are >>For the customer. Customer, >>The customer's expectations change, right? Once you get used to a better customer experience, you learn. >>That's to wrap it up. I wanna just get your perspective again. One of the benefits of chatting with you here and having you part of the Super Cloud 22 is you've seen many cycles, you have a lot of insights. I want to ask you, given your career where you've been and what you've done and now let's CEO platform nine, how would you compare what's happening now with other inflection points in the industry? And you've been, again, you've been an entrepreneur, you sold your company to Oracle, you've been seeing the big companies, you've seen the different waves. What's going on right now put into context this moment in time around Super Cloud. >>Sure. I think as you said, a lot of battles. CARSs being been in an asb, being in a real time software company, being in large enterprise software houses and a transformation. I've been on the app side, I did the infrastructure right and then tried to build our own platforms. I've gone through all of this myself with lot of lessons learned in there. I think this is an event which is happening now for companies to go through to become cloud native and digitalize. If I were to look back and look at some parallels of the tsunami that's going on is a couple of paddles come to me. One is, think of it, which was forced to honors like y2k. Everybody around the world had to have a plan, a strategy, and an execution for y2k. I would say the next big thing was e-commerce. I think e-commerce has been pervasive right across all industries. >>And disruptive. >>And disruptive, extremely disruptive. If you did not adapt and adapt and accelerate your e-commerce initiative, you were, it was an existence question. Yeah. I think we are at that pivotal moment now in companies trying to become digital and cloudnative. You know, that is what I see >>Happening there. I think that that e-commerce is interesting and I think just to riff with you on that is that it's disrupting and refactoring the business models. I think that is something that's coming out of this is that it's not just completely changing the gain, it's just changing how you operate, >>How you think and how you operate. See, if you think about the early days of e-commerce, just putting up a shopping cart that made you an e-commerce or e retailer or an e e e customer, right? Or so. I think it's the same thing now is I think this is a fundamental shift on how you're thinking about your business. How are you gonna operate? How are you gonna service your customers? I think it requires that just lift and shift is not gonna work. >>Nascar, thank you for coming on, spending the time to come in and share with our community and being part of Super Cloud 22. We really appreciate, we're gonna keep this open. We're gonna keep this conversation going even after the event, to open up and look at the structural changes happening now and continue to look at it in the open in the community. And we're gonna keep this going for, for a long, long time as we get answers to the problems that customers are looking for with cloud cloud computing. I'm Sean Fur with Super Cloud 22 in the Cube. Thanks for watching. >>Thank you. Thank you. >>Hello and welcome back. This is the end of our program, our special presentation with Platform nine on cloud native at scale, enabling the super cloud. We're continuing the theme here. You heard the interviews Super Cloud and its challenges, new opportunities around solutions around like Platform nine and others with Arlon. This is really about the edge situations on the internet and managing the edge multiple regions, avoiding vendor lock in. This is what this new super cloud is all about. The business consequences we heard and and the wide ranging conversations around what it means for open source and the complexity problem all being solved. I hope you enjoyed this program. There's a lot of moving pieces and things to configure with cloud native install, all making it easier for you here with Super Cloud and of course Platform nine contributing to that. Thank you for watching.
SUMMARY :
So enjoy the program, see you soon. a lot different, but kind of the same as the first generation. And so you gotta rougher and it kind of coming together, but you also got this idea of regions, So I think, you know, in in the context of this, the, Can you scope the scale of the problem? And I think, you know, I I like to call it, you know, And that is just, you know, one example of an issue that happens. you know, you see some, you know, some experimentation. which is, you know, you have your perfectly written code that is operating just fine on your And so as you give that change to then run at your production edge location, And you guys have a solution you're launching, Can you share what So what alarm lets you do in a in terms of the chaos you guys are reigning in. And if you look at the logo we've designed, So keeping it smooth, the assembly on things are flowing. Because developers, you know, there is, the developers are responsible for one picture of So the DevOps is the cloud native developer. And so online addresses that problem at the heart of it, and it does that using So I'm assuming you have that thought through, can you share open source and commercial relationship? products starting all the way with fi, which was a serverless product, you know, that we had built to buy, but also actually kind of date the application, if you will. I think one is just, you know, this, this, this cloud native space is so vast I have to ask you now, let's get into what's in it for the customer. And so, and there's multiple, you know, enterprises that we talk to, shared that this is a major challenge we have today because we have, you know, I'm an enterprise, I got tight, you know, I love the open source trying to It's created by folks that are as part of Intuit team now, you know, And the customer said, If you had it today, I would've purchased it. So next question is, what is the solution to the customer? So I think, you know, one of the core tenets of Platform nine has always been that And now they have management challenges. Especially operationalizing the clusters, whether they want to kind of reset everything and remove things around and reconfigure That's right. And alon by the way, also helps in that direction, but you also need I mean, what's the impact if you do all those things, as you mentioned, what's the impact of the apps? And so this really gives them, you know, the right tooling for But this is a key point, and I have to ask you because if this Arlo solution of challenges, and those are the pain points, which is, you know, if you're looking to reduce your, not where it used to be supporting the business, you know, that, you know, that the, the technology that's, you know, that's gonna drive your top line is If all the things happen the way we want 'em to happen, The magic wand, the magic dust, he's running that at a nimble, nimble team size of at the most, Taking care of, and the CIO doesn't exist. Thank you for your time. Thanks for having of Platform nine b. Great to see you Cube alumni. And now the Kubernetes layer that we've been working on for years is Exactly. you know, the new Arlon, our R lawn you guys just launched, you know, do step A, B, C, and D instead with Kubernetes, I mean now with open source, so popular, you don't have to have to write a lot of code. you know, the emergence of systems and layers to help you manage that complexity is becoming That's, I wrote a LinkedIn post today was comments about, you know, hey, enterprise is the new breed, the trend of SaaS you know, you think you have things under control, but some people from various teams will make changes here in the industry technical, how would you look at the super cloud trend that's emerging? the way I interpret that is, you know, clouds and infrastructure, It's IBM's, you know, connection for the internet at the, this layer that has simplified, you know, computing and, the physics and the, the atoms, the pro, you know, this is where the innovation, all the variations around and you know, compute storage networks the DevOps engineers, they get a a ways to So how do you guys look at the workload side of it? like K native, where you can express your application in more at a higher level, It's coming like an EC two instance, spin up a cluster. And then you can stamp out your app, your applications and your clusters and manage them And it's like a playbook, just deploy it. You just tell the system what you want and then You need edge's code, but then you can configure the code by just saying do it. And that is just complexity for the people operating this or configuring this, What do you expect to see at this year? If you look at a stack necessary for hosting We would joke we, you know, about, about the dream. So the successor to Kubernetes, you know, I don't Yeah, I think the, the reigning in the chaos is key, you know, Now we have now visibility into But roughly speaking when we say, you know, They have some SaaS apps, but mostly it's the ecosystem. you know, that they're, they will keep catering to, they, they will continue to find I mean, from a, from a hardware standpoint, yes. terms of, you know, the, the new risk and arm ecosystems, It's, it's hardware and you got software and you got middleware and he kinda over, Great to have you on. What's just thing about what you guys are doing at Platform nine? clouds, you know, the application world is moving very fast in trying to Patrick, we were talking before we came on stage here about your background and we were kind of talking about the glory days So you saw that whole growth. In fact, you know, as we were talking offline, I was in one of those And if you look at the tech trends, GDPs down, but not tech. some, you know, new servers and new application tools. you know, more, More dynamic, more unreal. So it's, you know, multi-cloud. the purpose of this event is as a pilot to get the conversations flowing with, with the influencers like yourselves And you know, most companies are, 70 plus percent of them have 1, 2, 3 container It runs on the edge, You give an example on how you guys are deploying your platform to enable a super And if you look at few years back, each one was doing So it's kinda like an SRE vibe. Whatever they want on their tools. to build, so their customers who are using product A and product B are seeing a similar set Are you delivering that value to ops and security? Our buyer is usually, you know, the product divisions of companies You've got the dev side and then enhance the customer experience that happens when you either order the product or go into And all the person in the store has to do like And so that dramatically brings the velocity for them. of the public clouds. So you guys got some success. How do you explain when someone says what's cloud native, what isn't cloud native? is the definition and what are the attributes and characteristics of what is truly a cloud native, Thousands and thousands of tools you could spend your time figuring I don't know anything about that, but the whole experience of how you order, For the customer. Once you get used to a better customer experience, One of the benefits of chatting with you here and been on the app side, I did the infrastructure right and then tried to build our If you did not adapt and adapt and accelerate I think that that e-commerce is interesting and I think just to riff with you on that is that it's disrupting How are you gonna service your Nascar, thank you for coming on, spending the time to come in and share with our community and being part of Thank you. I hope you enjoyed this program.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Vascar | PERSON | 0.99+ |
Mattor Makki | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Paul Morritz | PERSON | 0.99+ |
Sean Fur | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Patrick | PERSON | 0.99+ |
Vascar Gorde | PERSON | 0.99+ |
Adrian Karo | PERSON | 0.99+ |
John Forry | PERSON | 0.99+ |
John Furry | PERSON | 0.99+ |
John Fur | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
50,000 | QUANTITY | 0.99+ |
Dave Alon | PERSON | 0.99+ |
2000 | DATE | 0.99+ |
Maria Teel | PERSON | 0.99+ |
14 years | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
tens | QUANTITY | 0.99+ |
millions | QUANTITY | 0.99+ |
Gort | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
Nascar | PERSON | 0.99+ |
2001 | DATE | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
One | QUANTITY | 0.99+ |
4,000 engineers | QUANTITY | 0.99+ |
one site | QUANTITY | 0.99+ |
Two | QUANTITY | 0.99+ |
second part | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
two people | QUANTITY | 0.99+ |
Arlon | ORGANIZATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
Office 365 | TITLE | 0.99+ |
Makowski | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
today | DATE | 0.99+ |
Arlo | ORGANIZATION | 0.99+ |
two sides | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
two parts | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
both | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
first generation | QUANTITY | 0.99+ |
22 years later | DATE | 0.99+ |
1 | QUANTITY | 0.99+ |
first downturn | QUANTITY | 0.99+ |
Platform nine | ORGANIZATION | 0.99+ |
one unit | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
first | QUANTITY | 0.98+ |
one flavor | QUANTITY | 0.98+ |
more than one cloud | QUANTITY | 0.98+ |
two thousands | QUANTITY | 0.98+ |
One person | QUANTITY | 0.98+ |
Bickley | PERSON | 0.98+ |
Bacar | PERSON | 0.98+ |
12 years | QUANTITY | 0.98+ |
first time | QUANTITY | 0.98+ |
GoCon | EVENT | 0.98+ |
each site | QUANTITY | 0.98+ |
thousands of stores | QUANTITY | 0.98+ |
Azure | TITLE | 0.98+ |
20 years later | DATE | 0.98+ |
Platform9, Cloud Native at Scale
>>Hello, welcome to the Cube here in Palo Alto, California for a special presentation on Cloud native at scale, enabling super cloud modern applications with Platform nine. I'm John Furr, your host of The Cube. We had a great lineup of three interviews we're streaming today. Meor Ma Makowski, who's the co-founder and VP of Product of Platform nine. She's gonna go into detail around Arlon, the open source products, and also the value of what this means for infrastructure as code and for cloud native at scale. Bickley the chief architect of Platform nine Cube alumni. Going back to the OpenStack days. He's gonna go into why Arlon, why this infrastructure as code implication, what it means for customers and the implications in the open source community and where that value is. Really great wide ranging conversation there. And of course, Vascar, Gort, the CEO of Platform nine, is gonna talk with me about his views on Super Cloud and why Platform nine has a scalable solutions to bring cloudnative at scale. So enjoy the program. See you soon. Hello everyone. Welcome to the cube here in Palo Alto, California for special program on cloud native at scale, enabling next generation cloud or super cloud for modern application cloud native developers. I'm John Furry, host of the Cube. A pleasure to have here, me Makoski, co-founder and VP of product at Platform nine. Thanks for coming in today for this Cloudnative at scale conversation. Thank >>You for having me. >>So Cloudnative at scale, something that we're talking about because we're seeing the, the next level of mainstream success of containers Kubernetes and cloud native develop, basically DevOps in the C I C D pipeline. It's changing the landscape of infrastructure as code, it's accelerating the value proposition and the super cloud as we call it, has been getting a lot of traction because this next generation cloud is looking a lot different, but kind of the same as the first generation. What's your view on super cloud as it fits to cloud native as scales up? >>Yeah, you know, I think what's interesting, and I think the reason why Super Cloud is a really good, in a really fit term for this, and I think, I know my CEO was chatting with you as well, and he was mentioning this as well, but I think there needs to be a different term than just multi-cloud or cloud. And the reason is because as cloud native and cloud deployments have scaled, I think we've reached a point now where instead of having the traditional data center style model where you have a few large distributions of infrastructure and workload at a few locations, I think the model is kind of flipped around, right? Where you have a large number of microsites, these microsites could be your public cloud deployment, your private on-prem infrastructure deployments, or it could be your edge environment, right? And every single enterprise, every single industry is moving in that direction. And so you gotta rougher that with a terminology that, that, that indicates the scale and complexity of it. And so I think supercloud is a, is an appropriate term for that. >>So you brought a couple of things I want to dig into. You mentioned edge nodes. We're seeing not only edge nodes being the next kind of area of innovation, mainly because it's just popping up everywhere. And that's just the beginning. Wouldn't even know what's around the corner. You got buildings, you got iot, ot, and IT kind of coming together, but you also got this idea of regions, global infras infrastructures, big part of it. I just saw some news around CloudFlare shutting down a site here. There's policies being made at scale, These new challenges there. Can you share because you can have edge. So hybrid cloud is a winning formula. Everybody knows that it's a steady state. Yeah. But across multiple clouds brings in this new un engineered area, yet it hasn't been done yet. Spanning clouds. People say they're doing it, but you start to see the toe in the water, it's happening, it's gonna happen. It's only gonna get accelerated with the edge and beyond globally. So I have to ask you, what is the technical challenges in doing this? Because there's something business consequences as well, but there are technical challenges. Can you share your view on what the technical challenges are for the super cloud or across multiple edges and regions? >>Yeah, absolutely. So I think, you know, in in the context of this, the, this, this term of super cloud, I think it's sometimes easier to visualize things in terms of two access, right? I think on one end you can think of the scale in terms of just pure number of nodes that you have deploy a number of clusters in the Kubernetes space. And then on the other axis you would have your distribution factor, right? Which is, do you have these tens of thousands of nodes in one site or do you have them distributed across tens of thousands of sites with one node at each site? Right? And if you have just one flavor of this, there is enough complexity, but potentially manageable. But when you are expanding on both these access, you really get to a point where that scale really needs some well thought out, well structured solutions to address it, right? A combination of homegrown tooling along with your, you know, favorite distribution of Kubernetes is not a strategy that can help you in this environment. It may help you when you have one of this or when you, when you scale, is not at the level. >>Can you scope the complexity? Because I mean, I hear a lot of moving parts going on there, the technology's also getting better. We we're seeing cloud native become successful. There's a lot to configure, there's a lot to install. Can you scope the scale of the problem? Because we're talking about at scale Yep. Challenges here. Yeah, >>Absolutely. And I think, you know, I I like to call it, you know, the, the, the problem that the scale creates, you know, there's various problems, but I think one, one problem, one way to think about it is, is, you know, it works on my cluster problem, right? So I, you know, I come from engineering background and there's a, you know, there's a famous saying between engineers and QA and the support folks, right? Which is, it works on my laptop, which is I tested this chain, everything was fantastic, it worked flawlessly on my machine, on production, It's not working. The exact same problem now happens and these distributed environments, but at massive scale, right? Which is that, you know, developers test their applications, et cetera within the sanctity of their sandbox environments. But once you expose that change in the wild world of your production deployment, right? >>And the production deployment could be going at the radio cell tower at the edge location where a cluster is running there, or it could be sending, you know, these applications and having them run at my customer site where they might not have configured that cluster exactly the same way as I configured it, or they configured the cluster, right? But maybe they didn't deploy the security policies, or they didn't deploy the other infrastructure plugins that my app relies on. All of these various factors are their own layer of complexity. And there really isn't a simple way to solve that today. And that is just, you know, one example of an issue that happens. I think another, you know, whole new ball game of issues come in the context of security, right? Because when you are deploying applications at scale in a distributed manner, you gotta make sure someone's job is on the line to ensure that the right security policies are enforced regardless of that scale factor. So I think that's another example of problems that occur. >>Okay. So I have to ask about scale, because there are a lot of multiple steps involved when you see the success of cloud native. You know, you see some, you know, some experimentation. They set up a cluster, say it's containers and Kubernetes, and then you say, Okay, we got this, we can figure it. And then they do it again and again, they call it day two. Some people call it day one, day two operation, whatever you call it. Once you get past the first initial thing, then you gotta scale it. Then you're seeing security breaches, you're seeing configuration errors. This seems to be where the hotspot is in when companies transition from, I got this to, Oh no, it's harder than I thought at scale. Can you share your reaction to that and how you see this playing out? >>Yeah, so, you know, I think it's interesting. There's multiple problems that occur when, you know, the two factors of scale, as we talked about, start expanding. I think one of them is what I like to call the, you know, it, it works fine on my cluster problem, which is back in, when I was a developer, we used to call this, it works on my laptop problem, which is, you know, you have your perfectly written code that is operating just fine on your machine, your sandbox environment. But the moment it runs production, it comes back with p zeros and pos from support teams, et cetera. And those issues can be really difficult to triage us, right? And so in the Kubernetes environment, this problem kind of multi folds, it goes, you know, escalates to a higher degree because you have your sandbox developer environments, they have their clusters and things work perfectly fine in those clusters because these clusters are typically handcrafted or a combination of some scripting and handcrafting. >>And so as you give that change to then run at your production edge location, like say your radio cell tower site, or you hand it over to a customer to run it on their cluster, they might not have not have configured that cluster exactly how you did, or they might not have configured some of the infrastructure plugins. And so the things don't work. And when things don't work, triaging them becomes nightmarishly hard, right? It's just one of the examples of the problem, another whole bucket of issues is security, which is, is you have these distributed clusters at scale, you gotta ensure someone's job is on the line to make sure that these security policies are configured properly. >>So this is a huge problem. I love that comment. That's not not happening on my system. It's the classic, you know, debugging mentality. Yeah. But at scale it's hard to do that with error prone. I can see that being a problem. And you guys have a solution you're launching. Can you share what Arlon is this new product? What is it all about? Talk about this new introduction. >>Yeah, absolutely. Very, very excited. You know, it's one of the projects that we've been working on for some time now because we are very passionate about this problem and just solving problems at scale in on-prem or at in the cloud or at edge environments. And what arlon is, it's an open source project, and it is a tool, it's a Kubernetes native tool for complete end to end management of not just your clusters, but your clusters. All of the infrastructure that goes within and along the site of those clusters, security policies, your middleware, plug-ins, and finally your applications. So what our LA you do in a nutshell is in a declarative way, it lets you handle the configuration and management of all of these components in at scale. >>So what's the elevator pitch simply put for what dissolves in, in terms of the chaos you guys are reigning in, what's the, what's the bumper sticker? Yeah, what >>Would it do? There's a perfect analogy that I love to reference in this context, which is think of your assembly line, you know, in a traditional, let's say, you know, an auto manufacturing factory or et cetera, and the level of efficiency at scale that that assembly line brings, right? Our line, and if you look at the logo we've designed, it's this funny little robot. And it's because when we think of online, we think of these enterprise large scale environments, you know, sprawling at scale, creating chaos because there isn't necessarily a well thought through, well structured solution that's similar to an assembly line, which is taking each component, you know, addressing them, manufacturing, processing them in a standardized way, then handing to the next stage. But again, it gets, you know, processed in a standardized way. And that's what arlon really does. That's like the deliver pitch. If you have problems of scale of managing your infrastructure, you know, that is distributed. Arlon brings the assembly line level of efficiency and consistency for >>Those. So keeping it smooth, the assembly on things are flowing. See c i CD pipe pipelining. Exactly. So that's what you're trying to simplify that ops piece for the developer. I mean, it's not really ops, it's their ops, it's coding. >>Yeah. Not just developer, the ops, the operations folks as well, right? Because developers, you know, there is, developers are responsible for one picture of that layer, which is my apps, and then maybe that middleware of applications that they interface with, but then they hand it over to someone else who's then responsible to ensure that these apps are secure properly, that they are logging, logs are being collected properly, monitoring and observability integrated. And so it solves problems for both >>Those teams. Yeah. It's DevOps. So the DevOps is the cloud needed developer's. That's right. The option teams have to kind of set policies. Is that where the declarative piece comes in? Is that why that's important? >>Absolutely. Yeah. And, and, and, and you know, ES really in introduced or elevated this declarative management, right? Because, you know, s clusters are Yeah. Or your, yeah, you know, specifications of components that go in Kubernetes are defined a declarative way, and Kubernetes always keeps that state consistent with your defined state. But when you go outside of that world of a single cluster, and when you actually talk about defining the clusters or defining everything that's around it, there really isn't a solution that does that today. And so Arlon addresses that problem at the heart of it, and it does that using existing open source well known solutions. >>And do I want to get into the benefits? What's in it for me as the customer developer? But I want to finish this out real quick and get your thoughts. You mentioned open source. Why open source? What's the, what's the current state of the product? You run the product group over at Platform nine, is it open source? And you guys have a product that's commercial? Can you explain the open source dynamic? And first of all, why open source? Yeah. And what is the consumption? I mean, open source is great, People want open source, they can download it, look up the code, but maybe wanna buy the commercial. So I'm assuming you have that thought through, can you share open source and commercial relationship? >>Yeah, I think, you know, starting with why open source? I think it's, you know, we as a company, we have, you know, one of the things that's absolutely critical to us is that we take mainstream open source technologies components and then we, you know, make them available to our customers at scale through either a SaaS model or on-prem model, right? But, so as we are a company or startup or a company that benefits, you know, in a massive way by this open source economy, it's only right, I think in my mind that we do our part of the duty, right? And contribute back to the community that feeds us. And so, you know, we have always held that strongly as one of our principles. And we have, you know, created and built independent products starting all the way with fision, which was a serverless product, you know, that we had built to various other, you know, examples that I can give. But that's one of the main reasons why opensource and also open source, because we want the community to really firsthand engage with us on this problem, which is very difficult to achieve if your product is behind a wall, you know, behind, behind a block box. >>Well, and that's, that's what the developers want too. And what we're seeing in reporting with Super Cloud is the new model of consumption is I wanna look at the code and see what's in there. That's right. And then also, if I want to use it, I'll do it. Great. That's open source, that's the value. But then at the end of the day, if I wanna move fast, that's when people buy in. So it's a new kind of freemium, I guess, business model. I guess that's the way that long. But that's, that's the benefit. Open source. This is why standards and open source is growing so fast. You have that confluence of, you know, a way for developers to try before they buy, but also actually kind of date the application, if you will. We, you know, Adrian Karo uses the dating met metaphor, you know, Hey, you know, I wanna check it out first before I get married. Right? And that's what open source, So this is the new, this is how people are selling. This is not just open source, this is how companies are selling. >>Absolutely. Yeah. Yeah. You know, I think, and you know, two things. I think one is just, you know, this, this, this cloud native space is so vast that if you, if you're building a close flow solution, sometimes there's also a risk that it may not apply to every single enterprises use cases. And so having it open source gives them an opportunity to extend it, expand it, to make it proper to their use case if they choose to do so, right? But at the same time, what's also critical to us is we are able to provide a supported version of it with an SLA that we, you know, that's backed by us, a SAS hosted version of it as well, for those customers who choose to go that route, you know, once they have used the open source version and loved it and want to take it at scale and in production and need, need, need a partner to collaborate with, who can, you know, support them for that production >>Environment. I have to ask you now, let's get into what's in it for the customer. I'm a customer. Yep. Why should I be enthused about Arla? What's in it for me? You know? Cause if I'm not enthused about it, I'm not gonna be confident and it's gonna be hard for me to get behind this. Can you share your enthusiastic view of, you know, why I should be enthused about Arlo? I'm a >>Customer. Yeah, absolutely. And so, and there's multiple, you know, enterprises that we talk to, many of them, you know, our customers, where this is a very kind of typical story that you hear, which is we have, you know, a Kubernetes distribution. It could be on premise, it could be public clouds, native Kubernetes, and then we have our C I C D pipelines that are automating the deployment of applications, et cetera. And then there's this gray zone. And the gray zone is well before you can you, your CS c D pipelines can deploy the apps. Somebody needs to do all of that groundwork of, you know, defining those clusters and yeah. You know, properly configuring them. And as these things, these things start by being done hand grown. And then as the, as you scale, what typically enterprises would do today is they will have their home homegrown DIY solutions for this. >>I mean, the number of folks that I talk to that have built Terra from automation, and then, you know, some of those key developers leave. So it's a typical open source or typical, you know, DIY challenge. And the reason that they're writing it themselves is not because they want to. I mean, of course technology is always interesting to everybody, but it's because they can't find a solution that's out there that perfectly fits the problem. And so that's that pitch. I think Ops FICO would be delighted. The folks that we've talk, you know, spoken with, have been absolutely excited and have, you know, shared that this is a major challenge we have today because we have, you know, few hundreds of clusters on ecos Amazon, and we wanna scale them to few thousands, but we don't think we are ready to do that. And this will give us the >>Ability to, Yeah, I think people are scared. Not sc I won't say scare, that's a bad word. Maybe I should say that they feel nervous because, you know, at scale small mistakes can become large mistakes. This is something that is concerning to enterprises. And, and I think this is gonna come up at co con this year where enterprises are gonna say, Okay, I need to see SLAs. I wanna see track record, I wanna see other companies that have used it. Yeah. How would you answer that question to, or, or challenge, you know, Hey, I love this, but is there any guarantees? Is there any, what's the SLAs? I'm an enterprise, I got tight, you know, I love the open source trying to free fast and loose, but I need hardened code. >>Yeah, absolutely. So, so two parts to that, right? One is Arlan leverages existing open source components, products that are extremely popular. Two specifically. One is Arlan uses Argo cd, which is probably one of the highest and used CD open source tools that's out there. Right's created by folks that are as part of into team now, you know, really brilliant team. And it's used at scale across enterprises. That's one. Second is Alon also makes use of Cluster api cappi, which is a Kubernetes sub-component, right? For lifecycle management of clusters. So there is enough of, you know, community users, et cetera, around these two products, right? Or, or, or open source projects that will find Arlan to be right up in their alley because they're already comfortable, familiar with Argo cd. Now Arlan just extends the scope of what City can do. And so that's one. And then the second part is going back to a point of the comfort. And that's where, you know, platform line has a role to play, which is when you are ready to deploy online at scale, because you've been, you know, playing with it in your DEF test environments, you're happy with what you get with it, then Platform nine will stand behind it and provide that >>Sla. And what's been the reaction from customers you've talked to Platform nine customers with, with that are familiar with, with Argo and then rlo? What's been some of the feedback? >>Yeah, I, I think the feedback's been fantastic. I mean, I can give you examples of customers where, you know, initially, you know, when you are, when you're telling them about your entire portfolio of solutions, it might not strike a card right away. But then we start talking about Arlan and, and we talk about the fact that it uses Argo adn, they start opening up, they say, We have standardized on Argo and we have built these components, homegrown, we would be very interested. Can we co-develop? Does it support these use cases? So we've had that kind of validation. We've had validation all the way at the beginning of our land before we even wrote a single line of code saying this is something we plan on doing. And the customer said, If you had it today, I would've purchased it. So it's been really great validation. >>All right. So next question is, what is the solution to the customer? If I asked you, Look it, I have, I'm so busy, my team's overworked. I got a skills gap. I don't need another project that's, I'm so tied up right now and I'm just chasing my tail. How does Platform nine help me? >>Yeah, absolutely. So I think, you know, one of the core tenets of Platform nine has always been been that we try to bring that public cloud like simplicity by hosting, you know, this in a lot of such similar tools in a SaaS hosted manner for our customers, right? So our goal behind doing that is taking away or trying to take away all of that complexity from customers' hands and offloading it to our hands, right? And giving them that full white glove treatment, as we call it. And so from a customer's perspective, one, something like arlon will integrate with what they have so they don't have to rip and replace anything. In fact, it will, even in the next versions, it may even discover your clusters that you have today and you know, give you an inventory. And that will, >>So if customers have clusters that are growing, that's a sign correct call you guys. >>Absolutely. Either they're, they have massive large clusters, right? That they wanna split into smaller clusters, but they're not comfortable doing that today, or they've done that already on say, public cloud or otherwise. And now they have management challenges. So >>Especially operationalizing the clusters, whether they want to kind of reset everything and remove things around and reconfigure Yep. And or scale out. >>That's right. Exactly. And >>You provide that layer of policy. >>Absolutely. >>Yes. That's the key value here. >>That's right. >>So policy based configuration for cluster scale up, >>Well profile and policy based declarative configuration and lifecycle management for clusters. >>If I asked you how this enables supercloud, what would you say to that? >>I think this is one of the key ingredients to super cloud, right? If you think about a super cloud environment, there's at least few key ingredients that that come to my mind that are really critical. Like they are, you know, life saving ingredients at that scale. One is having a really good strategy for managing that scale, you know, in a, going back to assembly line in a very consistent, predictable way so that our lot solves then you, you need to compliment that with the right kind of observability and monitoring tools at scale, right? Because ultimately issues are gonna happen and you're gonna have to figure out, you know, how to solve them fast. And arlon by the way, also helps in that direction, but you also need observability tools. And then especially if you're running it on the public cloud, you need some cost management tools. In my mind, these three things are like the most necessary ingredients to make Super Cloud successful. And you know, our alarm fills in >>One. Okay. So now the next level is, Okay, that makes sense. Is under the covers kind of speak under the hood. Yeah. How does that impact the app developers and the cloud native modern application workflows? Because the impact to me, seems the apps are gonna be impacted. Are they gonna be faster, stronger? I mean, what's the impact if you do all those things, as you mentioned, what's the impact of the apps? >>Yeah, the impact is that your apps are more likely to operate in production the way you expect them to, because the right checks and balances have gone through, and any discrepancies have been identified prior to those apps, prior to your customer running into them, right? Because developers run into this challenge to their, where there's a split responsibility, right? I'm responsible for my code, I'm responsible for some of these other plugins, but I don't own the stack end to end. I have to rely on my ops counterpart to do their part, right? And so this really gives them, you know, the right tooling for that. >>So this is actually a great kind of relevant point, you know, as cloud becomes more scalable, you're starting to see this fragmentation gone of the days of the full stack developer to the more specialized role. But this is a key point, and I have to ask you because if this RLO solution takes place, as you say, and the apps are gonna be stupid, they're designed to do, the question is, what did does the current pain look like of the apps breaking? What does the signals to the customer Yeah. That they should be calling you guys up into implementing Arlo, Argo and, and all the other goodness to automate? What are some of the signals? Is it downtime? Is it, is it failed apps, Is it latency? What are some of the things that Yeah, absolutely would be indications of things are effed up a little bit. Yeah. >>More frequent down times, down times that are, that take longer to triage. And so you are, you know, the, you know, your mean times on resolution, et cetera, are escalating or growing larger, right? Like we have environments of customers where they're, they have a number of folks on in the field that have to take these apps and run them at customer sites. And that's one of our partners. And they're extremely interested in this because they're the, the rate of failures they're encountering for this, you know, the field when they're running these apps on site, because the field is automating their clusters that are running on sites using their own script. So these are the kinds of challenges, and those are the pain points, which is, you know, if you're looking to reduce your meantime to resolution, if you're looking to reduce the number of failures that occur on your production site, that's one. And second, if you are looking to manage these at scale environments with a relatively small, focused, nimble ops team, which has an immediate impact on your budget. So those are, those are the signals. >>This is the cloud native at scale situation, the innovation going on. Final thought is your reaction to the idea that if the world goes digital, which it is, and the confluence of physical and digital coming together, and cloud continues to do its thing, the company becomes the application, not where it used to be supporting the business, you know, the back office and the maybe terminals and some PCs and handhelds. Now if technology's running, the business is the business. Yeah. Company's the application. Yeah. So it can't be down. So there's a lot of pressure on, on CSOs and CIOs now and boards is saying, How is technology driving the top line revenue? That's the number one conversation. Yep. Do you see that same thing? >>Yeah. It's interesting. I think there's multiple pressures at the CXO CIO level, right? One is that there needs to be that visibility and clarity and guarantee almost that, you know, that the, the technology that's, you know, that's gonna drive your top line is gonna drive that in a consistent, reliable, predictable manner. And then second, there is the constant pressure to do that while always lowering your costs of doing it, right? Especially when you're talking about, let's say retailers or those kinds of large scale vendors, they many times make money by lowering the amount that they spend on, you know, providing those goods to their end customers. So I think those, both those factors kind of come into play and the solution to all of them is usually in a very structured strategy around automation. >>Final question. What does cloudnative at scale look like to you? If all the things happen the way we want 'em to happen, The magic wand, the magic dust, what does it look like? >>What that looks like to me is a CIO sipping at his desk on coffee production is running absolutely smooth. And his, he's running that at a nimble, nimble team size of at the most, a handful of folks that are just looking after things, but things are >>Just taking care of the CIO doesn't exist. There's no ciso, they're at the beach. >>Yep. >>Thank you for coming on, sharing the cloud native at scale here on the cube. Thank you for your time. >>Fantastic. Thanks for >>Having me. Okay. I'm John Fur here for special program presentation, special programming cloud native at scale, enabling super cloud modern applications with Platform nine. Thanks for watching. Welcome back everyone to the special presentation of cloud native at scale, the cube and platform nine special presentation going in and digging into the next generation super cloud infrastructure as code and the future of application development. We're here with Bickley, who's the chief architect and co-founder of Platform nine Pick. Great to see you Cube alumni. We, we met at an OpenStack event in about eight years ago, or later, earlier when OpenStack was going. Great to see you and great to see congratulations on the success of platform nine. >>Thank you very much. >>Yeah. You guys have been at this for a while and this is really the, the, the year we're seeing the, the crossover of Kubernetes because of what happens with containers. Everyone now has realized, and you've seen what Docker's doing with the new docker, the open source Docker now just the success Exactly. Of containerization, right? And now the Kubernetes layer that we've been working on for years is coming, bearing fruit. This is huge. >>Exactly. Yes. >>And so as infrastructures code comes in, we talked to Bacar talking about Super Cloud, I met her about, you know, the new Arlon, our, our lawn, and you guys just launched the infrastructures code is going to another level, and then it's always been DevOps infrastructures code. That's been the ethos that's been like from day one, developers just code. Then you saw the rise of serverless and you see now multi-cloud or on the horizon, connect the dots for us. What is the state of infrastructure as code today? >>So I think, I think I'm, I'm glad you mentioned it, everybody or most people know about infrastructures code. But with Kubernetes, I think that project has evolved at the concept even further. And these dates, it's infrastructure is configuration, right? So, which is an evolution of infrastructure as code. So instead of telling the system, here's how I want my infrastructure by telling it, you know, do step A, B, C, and D instead with Kubernetes, you can describe your desired state declaratively using things called manifest resources. And then the system kind of magically figures it out and tries to converge the state towards the one that you specified. So I think it's, it's a even better version of infrastructures code. >>Yeah. And that really means it's developer just accessing resources. Okay. That declare, Okay, give me some compute, stand me up some, turn the lights on, turn 'em off, turn 'em on. That's kind of where we see this going. And I like the configuration piece. Some people say composability, I mean now with open source so popular, you don't have to have to write a lot of code, this code being developed. And so it's into integration, it's configuration. These are areas that we're starting to see computer science principles around automation, machine learning, assisting open source. Cuz you got a lot of code that's right in hearing software, supply chain issues. So infrastructure as code has to factor in these new dynamics. Can you share your opinion on these new dynamics of, as open source grows, the glue layers, the configurations, the integration, what are the core issues? >>I think one of the major core issues is with all that power comes complexity, right? So, you know, despite its expressive power systems like Kubernetes and declarative APIs let you express a lot of complicated and complex stacks, right? But you're dealing with hundreds if not thousands of these yamo files or resources. And so I think, you know, the emergence of systems and layers to help you manage that complexity is becoming a key challenge and opportunity in, in this space. >>That's, I wrote a LinkedIn post today was comments about, you know, hey, enterprise is a new breed. The trend of SaaS companies moving our consumer comp consumer-like thinking into the enterprise has been happening for a long time, but now more than ever, you're seeing it the old way used to be solve complexity with more complexity and then lock the customer in. Now with open source, it's speed, simplification and integration, right? These are the new dynamic power dynamics for developers. Yeah. So as companies are starting to now deploy and look at Kubernetes, what are the things that need to be in place? Because you have some, I won't say technical debt, but maybe some shortcuts, some scripts here that make it look like infrastructure is code. People have done some things to simulate or or make infrastructure as code happen. Yes. But to do it at scale Yes. Is harder. What's your take on this? What's your view? >>It's hard because there's a per proliferation of methods, tools, technologies. So for example, today it's very common for DevOps and platform engineering tools, I mean, sorry, teams to have to deploy a large number of Kubernetes clusters, but then apply the applications and configurations on top of those clusters. And they're using a wide range of tools to do this, right? For example, maybe Ansible or Terraform or bash scripts to bring up the infrastructure and then the clusters. And then they may use a different set of tools such as Argo CD or other tools to apply configurations and applications on top of the clusters. So you have this sprawl of tools. You, you also have this sprawl of configurations and files because the more objects you're dealing with, the more resources you have to manage. And there's a risk of drift that people call that where, you know, you think you have things under control, but some people from various teams will make changes here and there and then before the end of the day systems break and you have no idea of tracking them. So I think there's real need to kind of unify, simplify, and try to solve these problems using a smaller, more unified set of tools and methodologies. And that's something that we try to do with this new project. Arlon. >>Yeah. So, so we're gonna get into Arlan in a second. I wanna get into the why Arlon. You guys announced that at AR GoCon, which was put on here in Silicon Valley at the, at the community meeting by in two, they had their own little day over there at their headquarters. But before we get there, vascar, your CEO came on and he talked about Super Cloud at our in AAL event. What's your definition of super cloud? If you had to kind of explain that to someone at a cocktail party or someone in the industry technical, how would you look at the super cloud trend that's emerging? It's become a thing. What's your, what would be your contribution to that definition or the narrative? >>Well, it's, it's, it's funny because I've actually heard of the term for the first time today, speaking to you earlier today. But I think based on what you said, I I already get kind of some of the, the gist and the, the main concepts. It seems like super cloud, the way I interpret that is, you know, clouds and infrastructure, programmable infrastructure, all of those things are becoming commodity in a way. And everyone's got their own flavor, but there's a real opportunity for people to solve real business problems by perhaps trying to abstract away, you know, all of those various implementations and then building better abstractions that are perhaps business or applications specific to help companies and businesses solve real business problems. >>Yeah, I remember that's a great, great definition. I remember, not to date myself, but back in the old days, you know, IBM had a proprietary network operating system, so of deck for the mini computer vendors, deck net and SNA respectively. But T C P I P came out of the osi, the open systems interconnect and remember, ethernet beat token ring out. So not to get all nerdy for all the young kids out there, look, just look up token ring, you'll see, you've probably never heard of it. It's IBM's, you know, connection for the internet at the, the layer two is Amazon, the ethernet, right? So if T C P I P could be the Kubernetes and the container abstraction that made the industry completely change at that point in history. So at every major inflection point where there's been serious industry change and wealth creation and business value, there's been an abstraction Yes. Somewhere. Yes. What's your reaction to that? >>I think this is, I think a saying that's been heard many times in this industry and, and I forgot who originated it, but I think that the saying goes like, there's no problem that can't be solved with another layer of indirection, right? And we've seen this over and over and over again where Amazon and its peers have inserted this layer that has simplified, you know, computing and, and infrastructure management. And I believe this trend is going to continue, right? The next set of problems are going to be solved with these insertions of additional abstraction layers. I think that that's really a, yeah, it's gonna >>Continue. It's interesting. I just, when I wrote another post today on LinkedIn called the Silicon Wars AMD stock is down arm has been on a rise. We remember pointing for many years now that arm's gonna be hugely, it has become true. If you look at the success of the infrastructure as a service layer across the clouds, Azure, aws, Amazon's clearly way ahead of everybody. The stuff that they're doing with the silicon and the physics and the, the atoms, the pro, you know, this is where the innovation, they're going so deep and so strong at ISAs, the more that they get that gets come on, they have more performance. So if you're an app developer, wouldn't you want the best performance and you'd wanna have the best abstraction layer that gives you the most ability to do infrastructures, code or infrastructure for configuration, for provisioning, for managing services. And you're seeing that today with service MeSHs, a lot of action going on in the service mesh area in in this community of, of co con, which will be a covering. So that brings up the whole what's next? You guys just announced our lawn at Argo Con, which came out of Intuit. We've had Mariana Tessel at our super cloud event. She's the cto, you know, they're all in the cloud. So they contributed that project. Where did Arlon come from? What was the origination? What's the purpose? Why our lawn, why this announcement? >>Yeah, so the, the inception of the project, this was the result of us realizing that problem that we spoke about earlier, which is complexity, right? With all of this, these clouds, these infrastructure, all the variations around and, you know, compute storage networks and the proliferation of tools we talked about the Ansibles and Terraforms and Kubernetes itself. You can, you can think of that as another tool, right? We saw a need to solve that complexity problem, and especially for people and users who use Kubernetes at scale. So when you have, you know, hundreds of clusters, thousands of applications, thousands of users spread out over many, many locations, there, there needs to be a system that helps simplify that management, right? So that means fewer tools, more expressive ways of describing the state that you want and more consistency. And, and that's why, you know, we built our lawn and we built it recognizing that many of these problems or sub problems have already been solved. So Arlon doesn't try to reinvent the wheel, it instead rests on the shoulders of several giants, right? So for example, Kubernetes is one building block, GI ops, and Argo CD is another one, which provides a very structured way of applying configuration. And then we have projects like cluster API and cross plane, which provide APIs for describing infrastructure. So arlon takes all of those building blocks and builds a thin layer, which gives users a very expressive way of defining configuration and desired state. So that's, that's kind of the inception of, And >>What's the benefit of that? What does that give the, what does that give the developer, the user, in this case, >>The developers, the, the platform engineer, team members, the DevOps engineers, they get a a ways to provision not just infrastructure and clusters, but also applications and configurations. They get a way, a system for provisioning, configuring, deploying, and doing life cycle management in a, in a much simpler way. Okay. Especially as I said, if you're dealing with a large number of applications. >>So it's like an operating fabric, if you will. Yes. For them. Okay, so let's get into what that means for up above and below the the, this abstraction or thin layer below as the infrastructure. We talked a lot about what's going on below that. Yeah. Above our workloads. At the end of the day, you know, I talk to CXOs and IT folks that are now DevOps engineers. They care about the workloads and they want the infrastructures code to work. They wanna spend their time getting in the weeds, figuring out what happened when someone made a push that that happened or something happened. They need observability and they need to, to know that it's working. That's right. And is my workloads running effectively? So how do you guys look at the workload side of it? Cuz now you have multiple workloads on these fabric, >>Right? So workloads, so Kubernetes has defined kind of a standard way to describe workloads and you can, you know, tell Kubernetes, I want to run this container this particular way, or you can use other projects that are in the Kubernetes cloud native ecosystem like K native, where you can express your application in more at a higher level, right? But what's also happening is in addition to the workloads, DevOps and platform engineering teams, they need to very often deploy the applications with the clusters themselves. Clusters are becoming this commodity. It's, it's becoming this host for the application and it kind of comes bundled with it. In many cases it is like an appliance, right? So DevOps teams have to provision clusters at a really incredible rate and they need to tear them down. Clusters are becoming more, >>It's kinda like an EC two instance, spin up a cluster. We very, people used words like that. That's >>Right. And before arlon you kind of had to do all of that using a different set of tools as, as I explained. So with Armon you can kind of express everything together. You can say I want a cluster with a health monitoring stack and a logging stack and this ingress controller and I want these applications and these security policies. You can describe all of that using something we call a profile. And then you can stamp out your app, your applications and your clusters and manage them in a very, so >>Essentially standard creates a mechanism. Exactly. Standardized, declarative kind of configurations. And it's like a playbook. You deploy it. Now what's there is between say a script like I'm, I have scripts, I could just automate scripts >>Or yes, this is where that declarative API and infrastructures configuration comes in, right? Because scripts, yes you can automate scripts, but the order in which they run matters, right? They can break, things can break in the middle and, and sometimes you need to debug them. Whereas the declarative way is much more expressive and powerful. You just tell the system what you want and then the system kind of figures it out. And there are these things about controllers which will in the background reconcile all the state to converge towards your desire. It's a much more powerful, expressive and reliable way of getting things done. >>So infrastructure has configuration is built kind of on, it's as super set of infrastructures code because it's >>An evolution. >>You need edge's code, but then you can configure the code by just saying do it. You basically declaring and saying Go, go do that. That's right. Okay, so, alright, so cloud native at scale, take me through your vision of what that means. Someone says, Hey, what does cloud native at scale mean? What's success look like? How does it roll out in the future as you, not future next couple years? I mean people are now starting to figure out, okay, it's not as easy as it sounds. Could be nice, it has value. We're gonna hear this year coan a lot of this. What does cloud native at scale >>Mean? Yeah, there are different interpretations, but if you ask me, when people think of scale, they think of a large number of deployments, right? Geographies, many, you know, supporting thousands or tens or millions of, of users there, there's that aspect to scale. There's also an equally important a aspect of scale, which is also something that we try to address with Arran. And that is just complexity for the people operating this or configuring this, right? So in order to describe that desired state and in order to perform things like maybe upgrades or updates on a very large scale, you want the humans behind that to be able to express and direct the system to do that in, in relatively simple terms, right? And so we want the tools and the abstractions and the mechanisms available to the user to be as powerful but as simple as possible. So there's, I think there's gonna be a number and there have been a number of CNCF and cloud native projects that are trying to attack that complexity problem as well. And Arlon kind of falls in in that >>Category. Okay, so I'll put you on the spot road that CubeCon coming up and obviously this will be shipping this segment series out before. What do you expect to see at Coan this year? What's the big story this year? What's the, what's the most important thing happening? Is it in the open source community and also within a lot of the, the people jogging for leadership. I know there's a lot of projects and still there's some white space in the overall systems map about the different areas get run time and there's ability in all these different areas. What's the, where's the action? Where, where's the smoke? Where's the fire? Where's the piece? Where's the tension? >>Yeah, so I think one thing that has been happening over the past couple of cons and I expect to continue and, and that is the, the word on the street is Kubernetes is getting boring, right? Which is good, right? >>Boring means simple. >>Well, well >>Maybe, >>Yeah, >>Invisible, >>No drama, right? So, so the, the rate of change of the Kubernetes features and, and all that has slowed but in, in a, in a positive way. But there's still a general sentiment and feeling that there's just too much stuff. If you look at a stack necessary for hosting applications based on Kubernetes, there are just still too many moving parts, too many components, right? Too much complexity. I go, I keep going back to the complexity problem. So I expect Cube Con and all the vendors and the players and the startups and the people there to continue to focus on that complexity problem and introduce further simplifications to, to the stack. >>Yeah. Vic, you've had an storied career, VMware over decades with them obviously in 12 years with 14 years or something like that. Big number co-founder here at Platform. Now you guys have been around for a while at this game. We, man, we talked about OpenStack, that project you, we interviewed at one of their events. So OpenStack was the beginning of that, this new revolution. And I remember the early days it was, it wasn't supposed to be an alternative to Amazon, but it was a way to do more cloud cloud native. I think we had a cloud ERO team at that time. We would to joke we, you know, about, about the dream. It's happening now, now at Platform nine. You guys have been doing this for a while. What's the, what are you most excited about as the chief architect? What did you guys double down on? What did you guys tr pivot from or two, did you do any pivots? Did you extend out certain areas? Cuz you guys are in a good position right now, a lot of DNA in Cloud native. What are you most excited about and what does Platform nine bring to the table for customers and for people in the industry watching this? >>Yeah, so I think our mission really hasn't changed over the years, right? It's been always about taking complex open source software because open source software, it's powerful. It solves new problems, you know, every year and you have new things coming out all the time, right? OpenStack was an example when the Kubernetes took the world by storm. But there's always that complexity of, you know, just configuring it, deploying it, running it, operating it. And our mission has always been that we will take all that complexity and just make it, you know, easy for users to consume regardless of the technology, right? So the successor to Kubernetes, you know, I don't have a crystal ball, but you know, you have some indications that people are coming up of new and simpler ways of running applications. There are many projects around there who knows what's coming next year or the year after that. But platform will a, platform nine will be there and we will, you know, take the innovations from the the community. We will contribute our own innovations and make all of those things very consumable to customers. >>Simpler, faster, cheaper. Exactly. Always a good business model technically to make that happen. Yes. Yeah, I think the, the reigning in the chaos is key, you know, Now we have now visibility into the scale. Final question before we depart this segment. What is at scale, how many clusters do you see that would be a watermark for an at scale conversation around an enterprise? Is it workloads we're looking at or, or clusters? How would you, Yeah, how would you describe that? When people try to squint through and evaluate what's a scale, what's the at scale kind of threshold? >>Yeah. And, and the number of clusters doesn't tell the whole story because clusters can be small in terms of the number of nodes or they can be large. But roughly speaking when we say, you know, large scale cluster deployments, we're talking about maybe hundreds, two thousands. >>Yeah. And final final question, what's the role of the hyperscalers? You got AWS continuing to do well, but they got their core ias, they got a PAs, they're not too too much putting a SaaS out there. They have some SaaS apps, but mostly it's the ecosystem. They have marketplaces doing over $2 billion billions of transactions a year and, and it's just like, just sitting there. It hasn't really, they're now innovating on it, but that's gonna change ecosystems. What's the role the cloud play in the cloud native of its scale? >>The, the hyperscalers, >>Yeahs Azure, Google. >>You mean from a business perspective? Yeah, they're, they have their own interests that, you know, that they're, they will keep catering to, they, they will continue to find ways to lock their users into their ecosystem of services and, and APIs. So I don't think that's gonna change, right? They're just gonna keep, >>Well they got great I performance, I mean from a, from a hardware standpoint, yes, that's gonna be key, right? >>Yes. I think the, the move from X 86 being the dominant way and platform to run workloads is changing, right? That, that, that, that, and I think the, the hyperscalers really want to be in the game in terms of, you know, the the new risk and arm ecosystems and the platforms. >>Yeah, not joking aside, Paul Morritz, when he was the CEO of VMware, when he took over once said, I remember our first year doing the cube. Oh the cloud is one big distributed computer, it's, it's hardware and he got software and you got middleware and he kind over, well he's kind of tongue in cheek, but really you're talking about large compute and sets of services that is essentially a distributed computer. >>Yes, >>Exactly. It's, we're back on the same game. Vic, thank you for coming on the segment. Appreciate your time. This is cloud native at scale special presentation with Platform nine. Really unpacking super cloud Arlon open source and how to run large scale applications on the cloud Cloud Native Phil for developers and John Furrier with the cube. Thanks for Washington. We'll stay tuned for another great segment coming right up. Hey, welcome back everyone to Super Cloud 22. I'm John Fur, host of the Cuba here all day talking about the future of cloud. Where's it all going? Making it super multi-cloud clouds around the corner and public cloud is winning. Got the private cloud on premise and edge. Got a great guest here, Vascar Gorde, CEO of Platform nine, just on the panel on Kubernetes. An enabler blocker. Welcome back. Great to have you on. >>Good to see you >>Again. So Kubernetes is a blocker enabler by, with a question mark. I put on on that panel was really to discuss the role of Kubernetes. Now great conversation operations is impacted. What's interest thing about what you guys are doing at Platform nine? Is your role there as CEO and the company's position, kind of like the world spun into the direction of Platform nine while you're at the helm? Yeah, right. >>Absolutely. In fact, things are moving very well and since they came to us, it was an insight to call ourselves the platform company eight years ago, right? So absolutely whether you are doing it in public clouds or private clouds, you know, the application world is moving very fast in trying to become digital and cloud native. There are many options for you do on the infrastructure. The biggest blocking factor now is having a unified platform. And that's what we, we come into, >>Patrick, we were talking before we came on stage here about your background and we were gonna talk about the glory days in 2000, 2001, when the first as piece application service providers came out, kind of a SaaS vibe, but that was kind of all kind of cloudlike. >>It wasn't, >>And and web services started then too. So you saw that whole growth. Now, fast forward 20 years later, 22 years later, where we are now, when you look back then to here and all the different cycles, >>I, in fact you, you know, as we were talking offline, I was in one of those ASPs in the year 2000 where it was a novel concept of saying we are providing a software and a capability as a service, right? You sign up and start using it. I think a lot has changed since then. The tooling, the tools, the technology has really skyrocketed. The app development environment has really taken off exceptionally well. There are many, many choices of infrastructure now, right? So I think things are in a way the same but also extremely different. But more importantly now for any company, regardless of size, to be a digital native, to become a digital company is extremely mission critical. It's no longer a nice to have everybody's in the journey somewhere. >>Everyone is going digital transformation here. Even on a so-called downturn recession that's upcoming inflation's here. It's interesting. This is the first downturn in the history of the world where the hyperscale clouds have been pumping on all cylinders as an economic input. And if you look at the tech trends, GDPs down, but not tech. >>Nope. >>Cuz the pandemic showed everyone digital transformation is here and more spend and more growth is coming even in, in tech. So this is a unique factor which proves that that digital transformation's happening and company, every company will need a super cloud. >>Everyone, every company, regardless of size, regardless of location, has to become modernize their infrastructure. And modernizing Infras infrastructure is not just some new servers and new application tools, It's your approach, how you're serving your customers, how you're bringing agility in your organization. I think that is becoming a necessity for every enterprise to survive. >>I wanna get your thoughts on Super Cloud because one of the things Dave Ante and I want to do with Super Cloud and calling it that was we, I, I personally, and I know Dave as well, he can, I'll speak from, he can speak for himself. We didn't like multi-cloud. I mean not because Amazon said don't call things multi-cloud, it just didn't feel right. I mean everyone has multiple clouds by default. If you're running productivity software, you have Azure and Office 365. But it wasn't truly distributed. It wasn't truly decentralized, it wasn't truly cloud enabled. It didn't, it felt like they're not ready for a market yet. Yet public clouds booming on premise. Private cloud and Edge is much more on, you know, more, more dynamic, more real. >>Yeah. I think the reason why we think super cloud is a better term than multi-cloud. Multi-cloud are more than one cloud, but they're disconnected. Okay, you have a productivity cloud, you have a Salesforce cloud, you may have, everyone has an internal cloud, right? So, but they're not connected. So you can say okay, it's more than one cloud. So it's you know, multi-cloud. But super cloud is where you are actually trying to look at this holistically. Whether it is on-prem, whether it is public, whether it's at the edge, it's a store at the branch. You are looking at this as one unit. And that's where we see the term super cloud is more applicable because what are the qualities that you require if you're in a super cloud, right? You need choice of infrastructure, you need, but at the same time you need a single pain, a single platform for you to build your innovations on regardless of which cloud you're doing it on, right? So I think Super Cloud is actually a more tightly integrated orchestrated management philosophy we think. >>So let's get into some of the super cloud type trends that we've been reporting on. Again, the purpose of this event is to, as a pilots, to get the conversations flowing with with the influencers like yourselves who are running companies and building products and the builders, Amazon and Azure are doing extremely well. Google's coming up in third cloudworks in public cloud. We see the use cases on premises use cases. Kubernetes has been an interesting phenomenon because it's become from the developer side a little bit, but a lot of ops people love Kubernetes. It's really more of an ops thing. You mentioned OpenStack earlier. Kubernetes kind of came out of that open stack. We need an orchestration and then containers had a good shot with, with Docker. They re pivoted the company. Now they're all in an open source. So you got containers booming and Kubernetes as a new layer there. What's the, what's the take on that? What does that really mean? Is that a new defacto enabler? It >>Is here. It's for here for sure. Every enterprise somewhere else in the journey is going on. And you know, most companies are, 70 plus percent of them have won two, three container based, Kubernetes based applications now being rolled out. So it's very much here, it is in production at scale by many customers. And the beauty of it is, yes, open source, but the biggest gating factor is the skill set. And that's where we have a phenomenal engineering team, right? So it's, it's one thing to buy a tool >>And just be clear, you're a managed service for Kubernetes. >>We provide, provide a software platform for cloud acceleration as a service and it can run anywhere. It can run in public private. We have customers who do it in truly multi-cloud environments. It runs on the edge, it runs at this in stores are thousands of stores in a retailer. So we provide that and also for specific segments where data sovereignty and data residency are key regulatory reasons. We also un OnPrem as an air gap version. >>Can you give an example on how you guys are deploying your platform to enable a super cloud experience for your >>Customer? Right. So I'll give you two different examples. One is a very large networking company, public networking company. They have, I dunno, hundreds of products, hundreds of r and d teams that are building different, different products. And if you look at few years back, each one was doing it on a different platforms but they really needed to bring the agility and they worked with us now over three years where we are their build test dev pro platform where all their products are built on, right? And it has dramatically increased their agility to release new products. Number two, it actually is a light out operation. In fact the customer says like, like the Maytag service person cuz we provide it as a service and it barely takes one or two people to maintain it for them. >>So it's kinda like an SRE vibe. One person managing a >>Large 4,000 engineers building infrastructure >>On their tools, >>Whatever they want on their tools. They're using whatever app development tools they use, but they use our platform. >>What benefits are they seeing? Are they seeing speed? >>Speed, definitely. Okay. Definitely they're speeding. Speed uniformity because now they're building able to build, so their customers who are using product A and product B are seeing a similar set of tools that are being used. >>So a big problem that's coming outta this super cloud event that we're, we're seeing and we've heard it all here, ops and security teams cuz they're kind of too part of one theme, but ops and security specifically need to catch up speed wise. Are you delivering that value to ops and security? Right. >>So we, we work with ops and security teams and infrastructure teams and we layer on top of that. We have like a platform team. If you think about it, depending on where you have data centers, where you have infrastructure, you have multiple teams, okay, but you need a unified platform. Who's your buyer? Our buyer is usually, you know, the product divisions of companies that are looking at or the CTO would be a buyer for us functionally cio definitely. So it it's, it's somewhere in the DevOps to infrastructure. But the ideal one we are beginning to see now many large corporations are really looking at it as a platform and saying we have a platform group on which any app can be developed and it is run on any infrastructure. So the platform engineering teams, >>You working two sides of that coin. You've got the dev side and then >>And then infrastructure >>Side side, okay. >>Another customer like give you an example, which I would say is kind of the edge of the store. So they have thousands of stores. Retail, retail, you know food retailer, right? They have thousands of stores that are on the globe, 50,000, 60,000. And they really want to enhance the customer experience that happens when you either order the product or go into the store and pick up your product or buy or browse or sit there. They have applications that were written in the nineties and then they have very modern AIML applications today. They want something that will not have to send an IT person to install a rack in the store or they can't move everything to the cloud because the store operations has to be local. The menu changes based on, It's a classic edge. It's classic edge. Yeah. Right. They can't send it people to go install rack access servers then they can't sell software people to go install the software and any change you wanna put through that, you know, truck roll. So they've been working with us where all they do is they ship, depending on the size of the store, one or two or three little servers with instructions that >>You, you say little servers like how big one like a net box box, like a small little >>Box and all the person in the store has to do like what you and I do at home and we get a, you know, a router is connect the power, connect the internet and turn the switch on. And from there we pick it up. >>Yep. >>We provide the operating system, everything and then the applications are put on it. And so that dramatically brings the velocity for them. They manage >>Thousands of them. True plug and play >>Two, plug and play thousands of stores. They manage it centrally. We do it for them, right? So, so that's another example where on the edge then we have some customers who have both a large private presence and one of the public clouds. Okay. But they want to have the same platform layer of orchestration and management that they can use regardless of the location. So >>You guys got some success. Congratulations. Got some traction there. It's awesome. The question I want to ask you is that's come up is what is truly cloud native? Cuz there's lift and shift of the cloud >>That's not cloud native. >>Then there's cloud native. Cloud native seems to be the driver for the super cloud. How do you talk to customers? How do you explain when someone says what's cloud native, what isn't cloud native? >>Right. Look, I think first of all, the best place to look at what is the definition and what are the attributes and characteristics of what is truly a cloud native, is CNC foundation. And I think it's very well documented where you, well >>Con of course Detroit's >>Coming here, so, so it's already there, right? So, so we follow that very closely, right? I think just lifting and shifting your 20 year old application onto a data center somewhere is not cloud native. Okay? You can't put to cloud native, you have to rewrite and redevelop your application and business logic using modern tools. Hopefully more open source and, and I think that's what Cloudnative is and we are seeing a lot of our customers in that journey. Now everybody wants to be cloudnative, but it's not that easy, okay? Because it's, I think it's first of all, skill set is very important. Uniformity of tools that there's so many tools there. Thousands and thousands of tools you could spend your time figuring out which tool to use. Okay? So I think the complexities there, but the business benefits of agility and uniformity and customer experience are truly them. >>And I'll give you an example. I don't know how clear native they are, right? And they're not a customer of ours, but you order pizzas, you do, right? If you just watch the pizza industry, how dominoes actually increase their share and mind share and wallet share was not because they were making better pizzas or not, I don't know anything about that, but the whole experience of how you order, how you watch what's happening, how it's delivered. There were a pioneer in it. To me, those are the kinds of customer experiences that cloud native can provide. >>Being agility and having that flow to the application changes what the expectations of the, for the customer. >>Customer, the customer's expectations change, right? Once you get used to a better customer experience, you learn >>Best car. To wrap it up, I wanna just get your perspective again. One of the benefits of chatting with you here and having you part of the Super Cloud 22 is you've seen many cycles, you have a lot of insights. I want to ask you, given your career where you've been and what you've done and now the CEO platform nine, how would you compare what's happening now with other inflection points in the industry? And you've been, again, you've been an entrepreneur, you sold your company to Oracle, you've been seeing the big companies, you've seen the different waves. What's going on right now put into context this moment in time around Super >>Cloud. Sure. I think as you said, a lot of battles. Cars being been, been in an asp, been in a realtime software company, being in large enterprise software houses and a transformation. I've been on the app side, I did the infrastructure right and then tried to build our own platforms. I've gone through all of this myself with a lot of lessons learned in there. I think this is an event which is happening now for companies to go through to become cloud native and digitalize. If I were to look back and look at some parallels of the tsunami that's going on is a couple of paddles come to me. One is, think of it, which was forced to honors like y2k. Everybody around the world had to have a plan, a strategy, and an execution for y2k. I would say the next big thing was e-commerce. I think e-commerce has been pervasive right across all industries. >>And disruptive. >>And disruptive, extremely disruptive. If you did not adapt and adapt and accelerate your e-commerce initiative, you were, it was an existence question. Yeah. I think we are at that pivotal moment now in companies trying to become digital and cloudnative that know that is what I see >>Happening there. I think that that e-commerce was interesting and I think just to riff with you on that is that it's disrupting and refactoring the business models. I think that is something that's coming out of this is that it's not just completely changing the game, it's just changing how you operate, >>How you think, and how you operate. See, if you think about the early days of eCommerce, just putting up a shopping cart didn't made you an eCommerce or an E retailer or an e e customer, right? Or so. I think it's the same thing now is I think this is a fundamental shift on how you're thinking about your business. How are you gonna operate? How are you gonna service your customers? I think it requires that just lift and shift is not gonna work. >>Mascar, thank you for coming on, spending the time to come in and share with our community and being part of Super Cloud 22. We really appreciate, we're gonna keep this open. We're gonna keep this conversation going even after the event, to open up and look at the structural changes happening now and continue to look at it in the open in the community. And we're gonna keep this going for, for a long, long time as we get answers to the problems that customers are looking for with cloud cloud computing. I'm Sean Feer with Super Cloud 22 in the Cube. Thanks for watching. >>Thank you. Thank you, John. >>Hello. Welcome back. This is the end of our program, our special presentation with Platform nine on cloud native at scale, enabling the super cloud. We're continuing the theme here. You heard the interviews Super Cloud and its challenges, new opportunities around the solutions around like Platform nine and others with Arlon. This is really about the edge situations on the internet and managing the edge multiple regions, avoiding vendor lock in. This is what this new super cloud is all about. The business consequences we heard and and the wide ranging conversations around what it means for open source and the complexity problem all being solved. I hope you enjoyed this program. There's a lot of moving pieces and things to configure with cloud native install, all making it easier for you here with Super Cloud and of course Platform nine contributing to that. Thank you for watching.
SUMMARY :
See you soon. but kind of the same as the first generation. And so you gotta rougher and IT kind of coming together, but you also got this idea of regions, So I think, you know, in in the context of this, the, this, Can you scope the scale of the problem? the problem that the scale creates, you know, there's various problems, but I think one, And that is just, you know, one example of an issue that happens. Can you share your reaction to that and how you see this playing out? which is, you know, you have your perfectly written code that is operating just fine on your And so as you give that change to then run at your production edge location, And you guys have a solution you're launching. So what our LA you do in a But again, it gets, you know, processed in a standardized way. So keeping it smooth, the assembly on things are flowing. Because developers, you know, there is, developers are responsible for one picture of So the DevOps is the cloud needed developer's. And so Arlon addresses that problem at the heart of it, and it does that using existing So I'm assuming you have that thought through, can you share open source and commercial relationship? products starting all the way with fision, which was a serverless product, you know, that we had built to buy, but also actually kind of date the application, if you will. I think one is just, you know, this, this, this cloud native space is so vast I have to ask you now, let's get into what's in it for the customer. And so, and there's multiple, you know, enterprises that we talk to, shared that this is a major challenge we have today because we have, you know, I'm an enterprise, I got tight, you know, I love the open source trying And that's where, you know, platform line has a role to play, which is when been some of the feedback? And the customer said, If you had it today, I would've purchased it. So next question is, what is the solution to the customer? So I think, you know, one of the core tenets of Platform nine has always been been that And now they have management challenges. Especially operationalizing the clusters, whether they want to kind of reset everything and remove things around and And And arlon by the way, also helps in that direction, but you also need I mean, what's the impact if you do all those things, as you mentioned, what's the impact of the apps? And so this really gives them, you know, the right tooling for that. So this is actually a great kind of relevant point, you know, as cloud becomes more scalable, So these are the kinds of challenges, and those are the pain points, which is, you know, if you're looking to to be supporting the business, you know, the back office and the maybe terminals and that, you know, that the, the technology that's, you know, that's gonna drive your top line is If all the things happen the way we want 'em to happen, The magic wand, the magic dust, he's running that at a nimble, nimble team size of at the most, Just taking care of the CIO doesn't exist. Thank you for your time. Thanks for Great to see you and great to see congratulations on the success And now the Kubernetes layer that we've been working on for years is Exactly. you know, the new Arlon, our, our lawn, and you guys just launched the So I think, I think I'm, I'm glad you mentioned it, everybody or most people know about infrastructures I mean now with open source so popular, you don't have to have to write a lot of code, you know, the emergence of systems and layers to help you manage that complexity is becoming That's, I wrote a LinkedIn post today was comments about, you know, hey, enterprise is a new breed. you know, you think you have things under control, but some people from various teams will make changes here in the industry technical, how would you look at the super cloud trend that's emerging? the way I interpret that is, you know, clouds and infrastructure, It's IBM's, you know, connection for the internet at the, this layer that has simplified, you know, computing and, the physics and the, the atoms, the pro, you know, this is where the innovation, the state that you want and more consistency. the DevOps engineers, they get a a ways to So how do you guys look at the workload native ecosystem like K native, where you can express your application in more at It's kinda like an EC two instance, spin up a cluster. And then you can stamp out your app, your applications and your clusters and manage them And it's like a playbook. You just tell the system what you want and then You need edge's code, but then you can configure the code by just saying do it. And that is just complexity for the people operating this or configuring this, What do you expect to see at Coan this year? If you look at a stack necessary for hosting We would to joke we, you know, about, about the dream. So the successor to Kubernetes, you know, I don't Yeah, I think the, the reigning in the chaos is key, you know, Now we have now visibility into But roughly speaking when we say, you know, They have some SaaS apps, but mostly it's the ecosystem. you know, that they're, they will keep catering to, they, they will continue to find terms of, you know, the the new risk and arm ecosystems it's, it's hardware and he got software and you got middleware and he kind over, Great to have you on. What's interest thing about what you guys are doing at Platform nine? clouds, you know, the application world is moving very fast in trying to Patrick, we were talking before we came on stage here about your background and we were gonna talk about the glory days in So you saw that whole growth. So I think things are in And if you look at the tech trends, GDPs down, but not tech. Cuz the pandemic showed everyone digital transformation is here and more And modernizing Infras infrastructure is not you know, more, more dynamic, more real. So it's you know, multi-cloud. So you got containers And you know, most companies are, 70 plus percent of them have won two, It runs on the edge, And if you look at few years back, each one was doing So it's kinda like an SRE vibe. Whatever they want on their tools. to build, so their customers who are using product A and product B are seeing a similar set Are you delivering that value to ops and security? Our buyer is usually, you know, the product divisions of companies You've got the dev side and then that happens when you either order the product or go into the store and pick up your product or like what you and I do at home and we get a, you know, a router is And so that dramatically brings the velocity for them. Thousands of them. of the public clouds. The question I want to ask you is that's How do you explain when someone says what's cloud native, what isn't cloud native? is the definition and what are the attributes and characteristics of what is truly a cloud native, Thousands and thousands of tools you could spend your time figuring out which I don't know anything about that, but the whole experience of how you order, Being agility and having that flow to the application changes what the expectations of One of the benefits of chatting with you here and been on the app side, I did the infrastructure right and then tried to build our own If you did not adapt and adapt and accelerate I think that that e-commerce was interesting and I think just to riff with you on that is that it's disrupting How are you gonna service your Mascar, thank you for coming on, spending the time to come in and share with our community and being part of Thank you, John. I hope you enjoyed this program.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Patrick | PERSON | 0.99+ |
Paul Morritz | PERSON | 0.99+ |
Vascar | PERSON | 0.99+ |
Adrian Karo | PERSON | 0.99+ |
Sean Feer | PERSON | 0.99+ |
2000 | DATE | 0.99+ |
John Furry | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
50,000 | QUANTITY | 0.99+ |
John | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
John Furr | PERSON | 0.99+ |
Vascar Gorde | PERSON | 0.99+ |
John Fur | PERSON | 0.99+ |
Meor Ma Makowski | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Makoski | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
14 years | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
12 years | QUANTITY | 0.99+ |
2001 | DATE | 0.99+ |
Gort | PERSON | 0.99+ |
Mascar | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Mariana Tessel | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
hundreds | QUANTITY | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Two | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
millions | QUANTITY | 0.99+ |
two parts | QUANTITY | 0.99+ |
tens | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
next year | DATE | 0.99+ |
Arlon | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Kubernetes | TITLE | 0.99+ |
eight years ago | DATE | 0.99+ |
one site | QUANTITY | 0.99+ |
Thousands | QUANTITY | 0.99+ |
second part | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
each component | QUANTITY | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Office 365 | TITLE | 0.99+ |
one unit | QUANTITY | 0.99+ |
one flavor | QUANTITY | 0.99+ |
4,000 engineers | QUANTITY | 0.99+ |
first generation | QUANTITY | 0.99+ |
Super Cloud | TITLE | 0.99+ |
Dave Ante | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
Vic | PERSON | 0.99+ |
two sides | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
two thousands | QUANTITY | 0.99+ |
Bickley | PERSON | 0.98+ |
tens of thousands of nodes | QUANTITY | 0.98+ |
Azure | TITLE | 0.98+ |
two people | QUANTITY | 0.98+ |
each site | QUANTITY | 0.98+ |
Kubernetes | PERSON | 0.98+ |
super cloud | TITLE | 0.98+ |
One person | QUANTITY | 0.98+ |
two factors | QUANTITY | 0.98+ |
Arlan | ORGANIZATION | 0.98+ |
Cloud native at scale: A Supercloud conversation with Madhura Maskasky, Platform9
(upbeat music) >> Hello, and welcome to theCUBE here in Palo Alto, California, for a special program on Cloud Native at Scale, Enabling Next Generation Cloud or Supercloud for Modern Application Cloud Native Developers. I'm John Furrier, host of theCUBE. My pleasure to have here, me Madhura Maskasky, Co-founder and VP of Product at Platform9. Thanks for coming in today for this cloud native at scale conversation. >> Thank you for having me. >> So cloud native at scale, something that we're talking about because we're seeing the next level of mainstream success of containers, Kubernetes and cloud native develop, basically DevOps in the CI/CD pipeline. It's changing the landscape of infrastructure as code. It's accelerating the value proposition. And the Supercloud as we call it, has been getting a lot of traction because this next generation cloud is looking a lot different, but kind of the same as the first generation. What's your view on Supercloud as it fits to cloud native, it scales up. >> Yeah, you know, I think what's interesting. And I think the reason why Supercloud is a really good and a really fit term for this. And I think I know my CEO was chatting with you as well, and he was mentioning this as well, but I think there needs to be a different term than just multicloud or cloud. And the reason is because as cloud native and cloud deployments have scaled, I think we've reached a point now where instead of having the traditional data center style model, where you have a few large distributions of infrastructure and workload at a few locations, I think the model's kind of flipped around, right? Where you have a large number of micro-sites. These micro-sites could be your public cloud deployment, your private OnPrem infrastructure deployment, or it could be your Edge environment, right? And every single enterprise, every single industry is moving in that direction. And so you got to refer that with a terminology that indicates the scale and complexity of it. And so I think Supercloud is an appropriate term for that. >> So you brought a couple things I want to dig into. You mentioned Edge nodes. We're seeing not only Edge nodes being the next kind of area of innovation, mainly because it's just popping up everywhere. And that's just the beginning, wouldn't even know what's around the corner. You got buildings, you got IoT, OT and IT kind of coming together, but you also got this idea of regions. Global infrastructure is a big part of it. I just saw some news around CloudFlare shutting down a site here. There's policies being made at scale, these new challenges there. Can you share, because you got to have Edge. So hybrid cloud is a winning formula. Everybody knows that, it's a steady state. But across multiple clouds brings in this new un-engineered area yet, It hasn't been done yet, Spanning Clouds. People say they're doing it, but you start to see the toe in the water. It's happening, it's going to happen. It's only going to get accelerated with the Edge and beyond globally. So I have to ask you, what is the technical challenges in doing this? Because there's something, business consequences as well, but there are technical challenges. Can you share your view on what the technical challenges are for the Supercloud across multiple edges and regions? >> Yeah, absolutely. So I think, you know, in the context of this term of Supercloud, I think it's sometimes easier to visualize things in terms of two axis, right? I think on one end you can think of the scale in terms of just pure number of nodes that you have deployed, a number of clusters in the Kubernetes space. And then on the other axis, you would have your distribution factor, right? Which is, do you have these tens of thousands of nodes in one site, or do you have them distributed across tens of thousands of sites, with one node at each site, right? And if you have just one flare of this, there is enough complexity, but potentially manageable. But when you are expanding on both these axis, you really get to a point where that scale really needs some well thought out, well structured solutions to address it, right? A combination of homegrown tooling, along with your, you know, favorite distribution of Kubernetes is not a strategy that can help you in this environment. It may help you when you have one of this, or when your scale is not at the level. >> Can you scope the complexity? Because, I mean, I hear a lot of moving parts going on there. The technology is also getting better. We're seeing cloud native become successful. There's a lot to configure. There's lot to install. Can you scope the scale of the problem because we're about at scale challenges here. >> Yeah absolutely, and I think I like to call it, you know, the problem that the scale creates, there's various problems. But I think one problem, one way to think about it is it works on my cluster problem, right? So, you know, I come from engineering background and there's a famous saying between engineers and QA, and the support folks, right. Which is, it works on my laptop, which is I tested this change, everything was fantastic. It worked flawlessly on my machine. On production, it's not working. The exact same problem now happens in these distributed environments, but at massive scale, right. Which is that, you know, developers test their applications, et cetera within these sanctity of their sandbox environments. But once you expose that change in the wild world of your production deployment, right. And the production deployment could be going at the radio cell tower at the Edge location where a cluster is running there. Or it could be sending, you know, these applications and having them run at my customer site, where they might not have configured that cluster exactly the same way as I configured it. Or they configured the cluster right. But maybe they didn't deploy the security policies, or they didn't deploy the other infrastructure plugins that my app relies on. All of these various factors add their own layer of complexity. And there really isn't a simple way to solve that today. And that is just, you know, one example of an issue that happens. I think another, you know, whole new ballgame of issues come in the context of security, right? Because when you are deploying applications at scale, in a distributed manner, you got to make sure someone's job is on the line to ensure that the right security policies are enforced regardless of that scale factor. So I think that's another example of problems that occur. >> Okay, so I have to ask about scale, because there are a lot of multiple steps involved when you see the success of cloud native, you know, you see some experimentation, they set up a cluster, say it's containers and Kubernetes. And then you say, okay, we got this. We configure it. And then they do it again, and again, they call it day two. Some people call it day one, day two operation, whatever you call it. Once you get past the first initial thing, then you got to scale it. Then you're seeing security breaches. You're seeing configuration errors. This seems to be where the hotspot is, in when companies transition from, I got this, to oh no, it's harder than I thought at scale. Can you share your reaction to that and how you see this playing out? >> Yeah, so, you know, I think it's interesting. There's multiple problems that occur when the two factors of scale, as we talked about, start expanding. I think one of them is what I like to call the, it works fine on my cluster problem, which is back in, when I was a developer, we used to call this, it works on my laptop problem. Which is, you know, you have your perfectly written code that is operating just fine on your machine, your sandbox environment. But the moment it runs production, it comes back with P 0s and POS from support teams, et cetera. And those issues can be really difficult to try us, right. And so in the Kubernetes environment, this problem kind of multi-folds. It goes, you know, escalates to a higher degree because you have your sandbox developer environments, they have their clusters, and things work perfectly fine in those clusters, because these clusters are typically handcrafted or a combination of some scripting and handcrafting. And so as you give that change to then run at your production Edge location, like say your radial cell power site, or you hand it over to a customer to run it on their cluster, they might not have configured that cluster exactly how you did, or they might not have configured some of the infrastructure plugins. And so things don't work. And when things don't work, triaging them becomes nightmarishly hard, right? It's just one of the examples of the problem. Another whole bucket of issues is security, which is, as you have these distributed clusters at scale. You got to ensure someone's job is on the line to make sure that the security policies are configured properly. >> So this is a huge problem. I love that comment. That's not happening on my system. It's the classic, you know, debugging mentality. But at scale, it's hard to do that with error prone. I can see that being a problem. And you guys have a solution you're launching, can you share what Arlon is? This new product? What is it all about? Talk about this new introduction. >> Yeah absolutely, I'm very, very excited. You know, it's one of the projects that we've been working on for some time now. Because we are very passionate about this problem and just solving problems at scale in OnPrem or in the cloud or at Edge environments. And what Arlon is, it's an open source project, and it is a tool, a Kubernetes native tool for complete end-to-end management of not just your clusters, but your clusters, all of the infrastructure that goes within and along the sites of those clusters, security policies, your middleware plugins, and finally your applications. So what Arlon lets you do in a nutshell is in a declarative way, it lets you handle the configuration and management of all of these components in at scale. >> So what's the elevator pitch simply put for what this solves in terms of the chaos you guys are reigning in, what's the bumper sticker. What did it do? >> There's a perfect analogy that I love to reference in this context, which is, think of your assembly line, you know, in a traditional, let's say an auto manufacturing factory, or et cetera, and the level of efficiency at scale that that assembly line brings, right. Arlon, and if you look at the logo we've designed, it's this funny little robot. And it's because when we think of Arlon, we think of these enterprise large scale environments, you know, sprawling at scale, creating chaos, because there isn't necessarily a well thought through, well-structured solution that's similar to an assembly line, which is taking each component, you know, addressing them, manufacturing, processing them in a standardized way, then handing to the next stage where again, it gets processed in a standardized way. And that's what Arlon really does. That's like the elevator pitch. If you have problems of scale, of managing your infrastructure, you know, that is distributed, Arlon brings the assembly line level of efficiency and consistency for those problems. >> So keeping it smooth, the assembly line, things are flowing, see CI/CD pipe-lining. So that's what you're trying to simplify that OPS piece for the developer. I mean, it's not really OPS, it's their OPS, it's coding. >> Yeah, not just developer the OPS, the operations folks as well, right. Because developers, you know, developers are responsible for one picture of that layer, which is my apps. And then maybe that middleware of applications that they interface with. But then they hand it over to someone else who's then responsible to ensure that these apps are secured properly, that they are logging, logs are being collected properly. Monitoring and observability is integrated. And so it solves problems for both those teams. >> Yeah, it's DevOps. So the DevOps is the cloud native developer. The OPS team have to kind of set policies. Is that where the declarative piece comes in? Is that why that's important? >> Absolutely, yeah. And you know, Kubernetes really introduced or elevated this declarative management, right. Because you know, Kubernetes clusters are you know your specifications of components that go in Kubernetes are defined in a declarative way. And Kubernetes always keeps that state consistent with your defined state. But when you go outside of that world of a single cluster, and when you actually talk about defining the clusters or defining everything that's around it, there really isn't a solution that does that today. And so Arlon addresses that problem at the heart of it. And it does that using existing open source, well known solutions. >> And do I want to get into the benefits, what's in it for me as the customer, developer, but I want to finish this out real quick and get your thoughts. You mentioned open source. Why open source? What's the current state of the product? You run the product group over there at Platform9. Is it open source, and you guys have a product that's commercial? Can you explain the open source dynamic? And first of all, why open source? And what is the consumption? I mean open source is great. People want opensource, they can download and look up the code, but maybe want to buy the commercial. So I'm assuming you have that thought through. Can you share open source and commercial relationship? >> Yeah, I think, you know, starting with why opensource? I think it's, you know, we, as a company, we have one of the things that's absolutely critical to us is that we take mainstream open source technologies, components, and then we make them available to our customers at scale through either a SaaS model or OnPrem model, right. But so as we are a company or startup, or a company that benefits, you know, in a massive way by this open source economy, it's only right I think in my mind that we do are part of the duty, right. And contribute back to the community that feeds us. And so, you know, we have always held that strongly as one of our principles. And we have, you know, created and built independent products, starting all the way with Fission, which was a serverless product that we had built, to various other examples that I can give. But that's one of the main reasons why open source. And also open source because we want the community to really first-hand engage with us on this problem, which is very difficult to achieve if your product is behind a wall, you know, behind a black box. >> Well, and that's what the developers want too. What we're seeing in reporting with Supercloud is the new model of consumption is I want to look at the code and see what's in there. >> That's right. >> And then also if I want to use it, I'll do it, great. That's open source, that's the value. But then at the end of the day, if I want to move fast, that's when people buy in. So it's a new kind of freemium, I guess, business model. I guess that's the way it is, but that's the benefit of open source. This is why standards and open source is growing so fast. You have that confluence of, you know, a way for developers to try before they buy, but also actually kind of date the application, if you will. We, you know, Adrian Kakroff uses the dating metaphor, you know, hey, you know, I want to check it out first before I get married. And that's what open source is. So this is the new, this is how people are selling. This is not just open source. This is how companies are selling. >> Absolutely, yeah, yeah. You know, I think two things, I think one is just, you know, this cloud native space is so vast that if you're building a cluster solution, sometimes there's also a risk that it may not apply to every single enterprises use cases. And so having it open source gives them an opportunity to extend it, expand it, to make it proper to their use case, if they choose to do so, right. But at the same time, what's also critical to us, is we are able to provide a supported version of it, with an SLA that's backed by us, a SaaS-hosted version of it as well for those customers who choose to go that route. You know, once they have used the open source version and loved it and want to take it at scale and in production and need a partner to collaborate with who can support them for that production environment. >> I have to ask you. Now let's get into what's in it for the customer? I'm a customer. Why should I be enthused about Arlon? What's in it for me? You know, 'cause if I'm not enthused about it, I'm not going to be confident, and it's going to be hard for me to get behind this. Can you share your enthusiastic view of, you know, why I should be enthused about Arlon, if I'm a customer. >> Yeah, absolutely. And so, and there's multiple, you know, enterprises that we talk to, many of them, are customers where this is a very kind of typical story that you will hear, which is we have a Kubernetes distribution. It could be On-Premise. It could be public cloud native Kubernetes. And then we have our CI/CD pipelines that are automating the deployment of applications, et cetera. And then there's this gray zone. And the gray zone is, well before you can, your CI/CD pipelines can deploy the apps, somebody needs to do all of their groundwork of, you know, defining those clusters, and yeah properly configuring them. And as these things start by being done hand-grown. And then as you scale, what typically enterprises would do today is they will have their homegrown DIY solutions for this. I mean, the number of folks that I talk to that have built Terraform automation, and then, you know, some of those key developers leave. So it's a typical open source, or typical, you know, DIY challenge. And the reason that they're writing it themselves is not because they want to. I mean, of course technology is always interesting to everybody, but it's because they can't find a solution that's out there that perfectly fits their problem. And so that's that pitch. I think OPS people would be delighted. The folks that we've talked, you know, spoken with have been absolutely excited and have shared that this is a major challenge we have today, because we have few hundreds of clusters on EKS, Amazon, and we want to scale them to few thousands, but we don't think we are ready to do that. And this will give us the ability to do that. >> Yeah, I think people are scared. I won't say scared, that's a bad word. Maybe I should say that they feel nervous because you know, at scale, small mistakes can become large mistakes. This is something that is concerning to enterprises. And I think this is going to come up at KubeCon this year where enterprises are going to say, okay, I need to see SLAs. I want to see track record. I want to see other companies that have used it. How would you answer that question to, or challenge, you know, hey I love this, but is there any guarantees? Is there any, what's the SLAs? I'm an enterprise, I got tight. You know, I love the open source trying to free, fast and loose, but I need hardened code. >> Yeah, absolutely. So two parts to that, right? One is Arlon leverages, existing opensource components, products that are extremely popular. Two specifically, one is Arlon uses Argo CD, which is probably one of the highest rated and used CD opensource tools that's out there, right. Created by folks that are as part of Intuit team now, you know, really brilliant team, and it's used at scale across enterprises. That's one. Second is Arlon also makes use of cluster API, CAPI, which is a Kubernetes sub-component, right for lifecycle management of clusters. So there is enough of, you know, community users, et cetera, around these two products or open source projects that will find Arlon to be right up in their alley, because they're already comfortable, familiar with Argo CD. Now Arlon just extends the scope of what Argo CD can do. And so that's one. And then the second part is going back to your point of the comfort. And that's where, you know, Platform9 has a role to play, which is when you are ready to deploy Arlon at scale, because you've been, you know playing with it in your DEV test environments, you're happy with what you get with it. Then Platform9 will stand behind it and provide that SLA. >> And what's been the reaction from customers you've talked to, Platform9 customers that are familiar with Argo, and then Arlo? What's been some of the feedback? >> Yeah, I think the feedback's been fantastic. I mean, I can give you examples of customers where you know, initially, when you're telling them about your entire portfolio of solutions, it might not strike a chord right away. But then we start talking about Arlon, and we talk about the fact that it uses Argo CD. They start opening up, they say, we have standardized on Argo, and we have built these components homegrown. We would be very interested. Can we co-develop? Does it support these use cases? So we've had that kind of validation. We've had validation all the way at the beginning of Arlon, before we even wrote a single line of code, saying this is something we plan on doing. And the customer said, if you had it today, I would've purchased it. So it's been really great validation. >> All right, so next question is what is the solution to the customer? If I asked you, look, I'm so busy. My team's overworked, I got a skills gap. I don't need another project. I'm so tied up right now, and I'm just chasing my tail. How does Platform9 help me? >> Yeah, absolutely. So I think, you know, one of the core tenants of Platform9 has always been, that we try to bring that public cloud like simplicity by hosting, you know, this and a lot of such similar tools in a SaaS hosted manner for our customers, right. So our goal behind doing that is taking away, or trying to take away all of that complexity from customer's hands and offloading it to our hands, right. And giving them that full white glove treatment as we call it. And so from a customer's perspective, one, something like Arlon will integrate with what they have, so they don't have to rip and replace anything. In fact, it will even in the next versions, it may even discover your clusters that you have today, and give you an inventory. >> So customers have clusters that are growing. That's a sign, call you guys. >> Absolutely, either they have massive, large clusters, right, that they want to split into smaller clusters, but they're not comfortable doing that today. Or they've done that already on say public cloud or otherwise. And now they have management challenges. >> So, especially operationalizing the clusters, whether they want to kind of reset everything and move things around, and reconfigure, and or scale out. >> That's right, exactly. >> And you provide that layer of policy. >> Absolutely, yes. >> That's the key value here. >> That's right. >> So policy based configuration for cluster scale up. >> Profile and policy based declarative configuration and life cycle management for clusters. >> If I asked you how this enables Supercloud, what would you say to that? >> I think this is one of the key ingredients to Supercloud, right? If you think about a Supercloud environment, there is at least few key ingredients that come to my mind that are really critical. Like they are, you know, life saving ingredients at that scale. One is having a really good strategy for managing that scale, you know, in a going back to assembly line, in a very consistent, predictable way. So that, Arlon solves. Then you need to compliment that with the right kind of observability and monitoring tools at scale, right? Because ultimately issues are going to happen, and you're going to have to figure out, you know, how to solve them fast. And Arlon, by the way also helps in that direction. But you also need observability tools. And then especially if you're running it on the public cloud, you need some cost management tools. In my mind, these three things are like the most necessary ingredients to make Supercloud successful. And you know, Arlon is one of them. >> Okay so now the next level is, okay, that makes sense is under the covers, kind of speak under the hood. How does that impact the app developers of the cloud native modern application workflows? Because the impact to me seems, the apps are going to be impacted. Are they going to be faster, stronger? I mean, what's the impact if you do all those things, as you mentioned, what's the impact of the apps? >> Yeah, the impact is that your apps are more likely to operate in production the way you expect them to, because the right checks and balances have gone through. And any discrepancies have been identified prior to those apps, prior to your customer running into them, right? Because developers run into this challenge today where there's a split responsibility, right. I'm responsible for my code. I'm responsible for some of these other plugins, but I don't own these stack end to end. I have to rely on my OPS counterpart to do their part, right. And so this really gives them the right tooling for that. >> This is actually a great kind of relevant point. You know, as cloud becomes more scalable, you're starting to see this fragmentation, gone are the days of the full stack developer, to the more specialized role. But this is a key point. And I have to ask you, because if this Arlo solution takes place, as you say, and the apps are going to do what they're designed to do, the question is what does the current pain look like? Are the apps breaking? What is the signals to the customer that they should be calling you guys up and implementing Arlo, Argo, and all the other goodness to automate, what are some of the signals? Is it downtime? Is it failed apps? Is it latency? What are some of the things that would be indications of things are effed up a little bit. >> Yeah, more frequent down times, down times that take longer to triage. And so your, you know, your mean times on resolution, et cetera, are escalating or growing larger, right? Like we have environments of customers where they have a number of folks in the field that have to take these apps, and run them at customer sites. And that's one of our partners. And they're extremely interested in this, because the rate of failures they're encountering for this, you know, the field when they're running these apps on site, because the field is automating their clusters that are running on sites using their own script. So these are the kinds of challenges. So those are the pain points, which is, you know, if you're looking to reduce your meantime to resolution. If you're looking to reduce the number of failures that occur on your production site, that's one. And second, if you're looking to manage these at scale environments with a relatively small focused nimble OPS team, which has an immediate impact on your budget. So those are the signals. >> This is the cloud native at scale situation. The innovation going on. Final thought is your reaction to the idea that if the world goes digital, which it is, and the confluence of physical and digital coming together, and cloud continues to do its thing, the company becomes the application. Not where IT used to be supporting the business, you know, the back office, and the immediate terminals and some PCs and handhelds. Now, if technology's running the business, is the business, company's the application. So it can't be down. So there's a lot of pressure on CSOs and CIOs now, and boards are saying, how is technology driving the top line revenue? That's the number one conversation. Do you see the same thing? >> Yeah, it's interesting. I think there's multiple pressures at the CSO, CIO level, right? One, is that there needs to be that visibility and clarity and guarantee almost that, you know, the technology that's going to drive your top line is going to drive that in a consistent, reliable, predictable manner. And then second, there is the constant pressure to do that while always lowering your costs of doing it, right. Especially when you're talking about, let's say retailers, or those kinds of large scale vendors, they many times make money by lowering the amount that they spend providing those goods to their end customers. So I think both those factors kind of come into play and the solution to all of them is usually in a very structured strategy around automation. >> Final question. What does cloud native at scale look like to you? If all the things happen the way we want 'em to happen, the magic wand, the magic dust, what does it look like? >> What that looks like to me is a CIO sipping at his desk on coffee. Production is running absolutely smooth. And he's running that at a nimble, nimble team size of, at the most, a handful of folks that are just looking after things, but things are just taking care of themselves. >> And the CIO doesn't exist. There's no CISO, they're at the beach. >> (laughing) Yeah. >> Madhura, thank you for coming on, sharing the cloud native at scale here on theCUBE. Thank you for your time. >> Fantastic, thanks for having me. >> Okay, I'm John Furrier here for special program presentation, special programming Cloud Native at Scale, Enabling Supercloud Modern Applications with Platform9. Thanks for watching. (upbeat music)
SUMMARY :
Co-founder and VP of Product at Platform9. And the Supercloud as we call it, And so you got to refer And that's just the beginning, So I think, you know, in the context Can you scope the complexity? And that is just, you know, And then you say, okay, we got this. And so as you give that change to then run It's the classic, you So what Arlon lets you do in a nutshell you guys are reigning in, Arlon, and if you look at that OPS piece for the developer. Because developers, you know, So the DevOps is the And you know, Kubernetes really introduced So I'm assuming you have or a company that benefits, you know, is the new model of consumption You have that confluence of, you know, I think one is just, you Can you share your enthusiastic view I mean, the number of folks that I talk to And I think this is going to And that's where, you know, where you know, initially, is what is the solution to the customer? clusters that you have today, That's a sign, call you guys. that they want to split operationalizing the clusters, So policy based configuration and life cycle management for clusters. for managing that scale, you know, Because the impact to me seems, the way you expect them to, and the apps are going to do for this, you know, the field that if the world goes and the solution to all of them If all the things happen the What that looks like to me And the CIO doesn't exist. Thank you for your time. for special program presentation,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Madhura Maskasky | PERSON | 0.99+ |
Adrian Kakroff | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Madhura | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
second part | QUANTITY | 0.99+ |
Arlon | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
tens of thousands of sites | QUANTITY | 0.99+ |
one site | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
two parts | QUANTITY | 0.99+ |
two factors | QUANTITY | 0.99+ |
one node | QUANTITY | 0.99+ |
Two | QUANTITY | 0.99+ |
first generation | QUANTITY | 0.99+ |
two products | QUANTITY | 0.98+ |
two things | QUANTITY | 0.98+ |
each site | QUANTITY | 0.98+ |
one problem | QUANTITY | 0.98+ |
each component | QUANTITY | 0.98+ |
Supercloud | ORGANIZATION | 0.98+ |
Second | QUANTITY | 0.98+ |
tens of thousands of nodes | QUANTITY | 0.98+ |
Arlo | ORGANIZATION | 0.97+ |
KubeCon | EVENT | 0.97+ |
Platform9 | ORGANIZATION | 0.97+ |
single line | QUANTITY | 0.97+ |
one end | QUANTITY | 0.96+ |
CloudFlare | TITLE | 0.96+ |
one way | QUANTITY | 0.96+ |
Argo | ORGANIZATION | 0.96+ |
three things | QUANTITY | 0.96+ |
One | QUANTITY | 0.95+ |
Kubernetes | TITLE | 0.94+ |
one flare | QUANTITY | 0.94+ |
Fission | ORGANIZATION | 0.93+ |
single cluster | QUANTITY | 0.93+ |
one picture | QUANTITY | 0.93+ |
DevOps | TITLE | 0.92+ |
EKS | ORGANIZATION | 0.91+ |
this year | DATE | 0.91+ |
one example | QUANTITY | 0.91+ |
Cloud | TITLE | 0.9+ |
Digging into HeatWave ML Performance
(upbeat music) >> Hello everyone. This is Dave Vellante. We're diving into the deep end with AMD and Oracle on the topic of mySQL HeatWave performance. And we want to explore the important issues around machine learning. As applications become more data intensive and machine intelligence continues to evolve, workloads increasingly are seeing a major shift where data and AI are being infused into applications. And having a database that simplifies the convergence of transaction and analytics data without the need to context, switch and move data out of and into different data stores. And eliminating the need to perform extensive ETL operations is becoming an industry trend that customers are demanding. At the same time, workloads are becoming more automated and intelligent. And to explore these issues further, we're happy to have back in theCUBE Nipun Agarwal, who's the Senior Vice President of mySQL HeatWave and Kumaran Siva, who's the Corporate Vice President Strategic Business Development at AMD. Gents, hello again. Welcome back. >> Hello. Hi Dave. >> Thank you, Dave. >> Okay. Nipun, obviously machine learning has become a must have for analytics offerings. It's integrated into mySQL HeatWave. Why did you take this approach and not the specialized database approach as many competitors do right tool for the right job? >> Right? So, there are a lot of customers of mySQL who have the need to run machine learning on the data which is store in mySQL database. So in the past, customers would need to extract the data out of mySQL and they would take it to a specialized service for running machine learning. Now, the reason we decided to incorporate machine learning inside the database, there are multiple reasons. One, customers don't need to move the data. And if they don't need to move the data, it is more secure because it's protected by the same access controlled mechanisms as rest of the data There is no need for customers to manage multiple services. But in addition to that, when we run the machine learning inside the database customers are able to leverage the same service the same hardware, which has been provisioned for OTP analytics and use machine learning capabilities at no additional charge. So from a customer's perspective, they get the benefits that it is a single database. They don't need to manage multiple services. And it is offered at no additional charge. And then as another aspect, which is kind of hard to learn which is based on the IP, the work we have done it is also significantly faster than what customers would get by having a separate service. >> Just to follow up on that. How are you seeing customers use HeatWaves machine learning capabilities today? How is that evolving? >> Right. So one of the things which, you know customers very often want to do is to train their models based on the data. Now, one of the things is that data in a database or in a transaction database changes quite rapidly. So we have introduced support for auto machine learning as a part of HeatWave ML. And what it does is that it fully automates the process of training. And this is something which is very important to database users, very important to mySQL users that they don't really want to hire or data scientists or specialists for doing training. So that's the first part that training in HeatWave ML is fully automated. Doesn't require the user to provide any like specific parameters, just the source data and the task which they want to train. The second aspect is the training is really fast. So the training is really fast. The benefit is that customers can retrain quite often. They can make sure that the model is up to date with any changes which have been made to their transaction database. And as a result of the models being up to date, the accuracy of the prediction is high. Right? So that's the first aspect, which is training. The second aspect is inference, which customers run once they have the models trained. And the third thing, which is perhaps been the most sought after request from the mySQL customers is the ability to provide explanations. So, HeatWave ML provides explanations for any model which has been generated or trained by HeatWave ML. So these are the three capabilities- training, inference and explanations. And this whole process is completely automated, doesn't require a specialist or a data scientist. >> Yeah, that's nice. I mean, training obviously very popular today. I've said inference I think is going to explode in the coming decade. And then of course, AI explainable AI is a very important issue. Kumaran, what are the relevant capabilities of the AMD chips that are used in OCI to support HeatWave ML? Are they different from say the specs for HeatWave in general? >> So, actually they aren't. And this is one of the key features of this architecture or this implementation that is really exciting. Um, there with HeatWave ML, you're using the same CPU. And by the way, it's not a GPU, it's a CPU for both for all three of the functions that Nipun just talked about- inference, training and explanation all done on CPU. You know, bigger picture with the capabilities we bring here we're really providing a balance, you know between the CPU cores, memory and the networking. And what that allows you to do here is be able to feed the CPU cores appropriately. And within the cores, we have these AVX instruc... extensions in with the Zen 2 and Zen 3 cores. We had AVX 2, and then with the Zen 4 core coming out we're going to have AVX 512. But we were able to with that balance of being able to bring in the data and utilize the high memory bandwidth and then use the computation to its maximum we're able to provide, you know, build pride enough AI processing that we are able to get the job done. And then we're built to build a fit into that larger pipeline that that we build out here with the HeatWave. >> Got it. Nipun you know, you and I every time we have a conversation we've got to talk benchmarks. So you've done machine learning benchmarks with HeatWave. You might even be the first in the industry to publish you know, transparent, open ML benchmarks on GitHub. I mean, I, I wouldn't know for sure but I've not seen that as common. Can you describe the benchmarks and the data sets that you used here? >> Sure. So what we did was we took a bunch of open data sets for two categories of tasks- classification and regression. So we took about a dozen data sets for classification and about six for regression. So to give an example, the kind of data sets we used for classifications like the airlines data set, hex sensors bank, right? So these are open data sets. And what we did was for on these data sets we did a comparison of what would it take to train using HeatWave ML? And then the other service we compared with is that RedShift ML. So, there were two observations. One is that with HeatWave ML, the user does not need to provide any tuning parameters, right? The HeatWave ML using RML fully generates a train model, figures out what are the right algorithms? What are the right features? What are the right hyper parameters and sets, right? So no need for any manual intervention not so the case with Redshift ML. The second thing is the performance, right? So the performance of HeatWave ML aggregate on these 12 data sets for classification and the six data sets on regression. On an average, it is 25 times faster than Redshift ML. And note that Redshift ML in turn involves SageMaker, right? So on an average, HeatWave ML provides 25 times better performance for training. And the other point to note is that there is no need for any human intervention. That's fully automated. But in the case of Redshift ML, many of these data sets did not even complete in the set duration. If you look at price performance, one of the things again I want to highlight is because of the fact that AMD does pretty well in all kinds of workloads. We are able to use the same cluster users and use the same cluster for analytics, for OTP or for machine learning. So there is no additional cost for customers to run HeatWave ML if they have provision HeatWave. But assuming a user is provisioning a HeatWave cluster only to run HeatWave ML, right? That's the case, even in that case the price performance advantage of HeatWave ML over Redshift ML is 97 times, right? So 25 times faster at 1% of the cost compared to Redshift ML And all these scripts and all this information is available on GitHub for customers to try to modify and like, see, like what are the advantages they would get on their workloads? >> Every time I hear these numbers, I shake my head. I mean, they're just so overwhelming. Um, and so we'll see how the competition responds when, and if they respond. So, but thank you for sharing those results. Kumaran, can you elaborate on how the specs that you talked about earlier contribute to HeatWave ML's you know, benchmark results. I'm particularly interested in scalability, you know Typically things degrade as you push the system harder. What are you seeing? >> No, I think, I think it's good. Look, yeah. That's by those numbers, just blow me, blow my head too. That's crazy good performance. So look from, from an AMD perspective, we have really built an architecture. Like if you think about the chiplet architecture to begin with, it is fundamentally, you know, it's kind of scaling by design, right? And, and one of the things that we've done here is been able to work with, with the HeatWave team and heat well ML team, and then been able to, to within within the CPU package itself, be able to scale up to take very efficient use of all of the course. And then of course, work with them on how you go between nodes. So you can have these very large systems that can run ML very, very efficiently. So it's really, you know, building on the building blocks of the chiplet architecture and how scaling happens there. >> Yeah. So it's you're saying it's near linear scaling or essentially. >> So, let Nipun comment on that. >> Yeah. >> Is it... So, how about as cluster sizes grow, Nipun? >> Right. >> What happens there? >> So one of the design points for HeatWave is scale out architecture, right? So as you said, that as we add more data set or increase the size of the data, or we add the number of nodes to the cluster, we want the performance to scale. So we show that we have near linear scale factor, or nearly near scale scalability for SQL workloads in the case of HeatWave ML, as well. As users add more nodes to the cluster so the size of the cluster the performance of HeatWave ML improves. So I was giving you this example that HeatWave ML is 25 times faster compared to Redshift ML. Well, that was on a cluster size of two. If you increase the cluster size of HeatWave ML to a larger number. But I think the number is 16. The performance advantage over Redshift ML increases from 25 times faster to 45 times faster. So what that means is that on a cluster size of 16 nodes HeatWave ML is 45 times faster for training these again, dozen data sets. So this shows that HeatWave ML skills better than the computation. >> So you're saying adding nodes offsets any management complexity that you would think of as getting in the way. Is that right? >> Right. So one is the management complexity and which is why by features like last customers can scale up or scale down, you know, very easily. The second aspect is, okay What gives us this advantage, right, of scalability? Or how are we able to scale? Now, the techniques which we use for HeatWave ML scalability are a bit different from what we use for SQL processing. So in the case of HeatWave ML, they really like, you know, three, two trade offs which we have to be careful about. One is the accuracy. Because we want to provide better performance for machine learning without compromising on the accuracy. So accuracy would require like more synchronization if you have multiple threads. But if you have too much of synchronization that can slow down the degree of patterns that we get. Right? So we have to strike a fine balance. So what we do is that in HeatWave ML, there are different phases of training, like algorithm selection, feature selection, hyper probability training. Each of these phases is analyzed. And for instance, one of the ways techniques we use is that if you're trying to figure out what's the optimal hyper parameter to be used? We start up with the search space. And then each of the VMs gets a part of the search space. And then we synchronize only when needed, right? So these are some of the techniques which we have developed over the years. And there are actually paper's filed, research publications filed on this. And this is what we do to achieve good scalability. And what that results to the customer is that if they have some amount of training time and they want to make it better they can just provision a larger cluster and they will get better performance. >> Got it. Thank you. Kumaran, when I think of machine learning, machine intelligence, AI, I think GPU but you're not using GPU. So how are you able to get this type of performance or price performance without using GPU's? >> Yeah, definitely. So yeah, that's a good point. And you think about what is going on here and you consider the whole pipeline that Nipun has just described in terms of how you get you know, your training, your algorithms And using the mySQL pieces of it to get to the point where the AI can be effective. In that process what happens is you have to have a lot of memory to transactions. A lot of memory bandwidth comes into play. And then bringing all that data together, feeding the actual complex that does the AI calculations that in itself could be the bottleneck, right? And you can have multiple bottlenecks along the way. And I think what you see in the AMD architecture for epic for this use case is the balance. And the fact that you are able to do the pre-processing, the AI, and then the post-processing all kind of seamlessly together, that has a huge value. And that goes back to what Nipun was saying about using the same infrastructure, gets you the better TCO but it also gets you gets you better performance. And that's because of the fact that you're bringing the data to the computation. So the computation in this case is not strictly the bottleneck. It's really about how you pull together what you need and to do the AI computation. And that is, that's probably a more, you know, it's a common case. And so, you know, you're going to start I think the least start to see this especially for inference applications. But in this case we're doing both inference explanation and training. All using the the CPU in the same OCI infrastructure. >> Interesting. Now Nipun, is the secret sauce for HeatWave ML performance different than what we've discussed before you and I with with HeatWave generally? Is there some, you know, additive engine additive that you're putting in? >> Right? Yes. The secret sauce is indeed different, right? Just the way I was saying that for SQL processing. The reason we get very good performance and price performance is because we have come up with new algorithms which help the SQL process can scale out. Similarly for HeatWave ML, we have come up with new IP, new like algorithms. One example is that we use meta-learn proxy models, right? That's the technique we use for automating the training process, right? So think of this meta-learn proxy models to be like, you know using machine learning for machine learning training. And this is an IP which we developed. And again, we have published the results and the techniques. But having such kind of like techniques is what gives us a better performance. Similarly, another thing which we use is adaptive sampling that you can have a large data set. But we intelligently sample to figure out that how can we train on a small subset without compromising on the accuracy? So, yes, there are many techniques that you have developed specifically for machine learning which is what gives us the better performance, better price performance, and also better scalability. >> What about mySQL autopilot? Is there anything that differs from HeatWave ML that is relevant? >> Okay. Interesting you should ask. So mySQL Autopilot is think of it to be an application using machine learning. So mySQL Autopilot uses machine learning to automate various aspects of the database service. So for instance, if you want to figure out that what's the right partitioning scheme to partition the data in memory? We use machine learning techniques to figure out that what's the right, the best column based on the user's workload to partition the data in memory Or given a workload, if you want to figure out what is the right cluster size to provision? That's something we use mySQL autopilot for. And I want to highlight that we don't aware of any other database service which provides this level of machine learning based automation which customers get with mySQL Autopilot. >> Hmm. Interesting. Okay. Last question for both of you. What are you guys working on next? What can customers expect from this collaboration specifically in this space? Maybe Nipun, you can start and then Kamaran can bring us home. >> Sure. So there are two things we are working on. One is based on the feedback we have gotten from customers, we are going to keep making the machine learning capabilities richer in HeatWave ML. That's one dimension. And the second thing is which Kamaran was alluding to earlier, We are looking at the next generation of like processes coming from AMD. And we will be seeing as to how we can more benefit from these processes whether it's the size of the L3 cache, the memory bandwidth, the network bandwidth, and such or the newer effects. And make sure that we leverage the all the greatness which the new generation of processes will offer. >> It's like an engineering playground. Kumaran, let's give you the final word. >> No, that's great. Now look with the Zen 4 CPU cores, we're also bringing in AVX 512 instruction capability. Now our implementation is a little different. It was in, in Rome and Milan, too where we use a double pump implementation. What that means is, you know, we take two cycles to do these instructions. But the key thing there is we don't lower our speed of the CPU. So there's no noisy neighbor effects. And it's something that OCI and the HeatWave has taken full advantage of. And so like, as we go out in time and we see the Zen 4 core, we can... we see up to 96 CPUs that that's going to work really well. So we're collaborating closely with, with OCI and with the HeatWave team here to make sure that we can take advantage of that. And we're also going to upgrade the memory subsystem to get to 12 channels of DDR 5. So it should be, you know there should be a fairly significant boost in absolute performance. But more important or just as importantly in TCO value for the customers, the end customers who are going to adopt this great service. >> I love their relentless innovation guys. Thanks so much for your time. We're going to have to leave it there. Appreciate it. >> Thank you, David. >> Thank you, David. >> Okay. Thank you for watching this special presentation on theCUBE. Your leader in enterprise and emerging tech coverage.
SUMMARY :
And eliminating the need and not the specialized database approach So in the past, customers How are you seeing customers use So one of the things of the AMD chips that are used in OCI And by the way, it's not and the data sets that you used here? And the other point to note elaborate on how the specs And, and one of the things or essentially. So, how about as So one of the design complexity that you would So in the case of HeatWave ML, So how are you able to get And the fact that you are Nipun, is the secret sauce That's the technique we use for automating of the database service. What are you guys working on next? And the second thing is which Kamaran Kumaran, let's give you the final word. OCI and the HeatWave We're going to have to leave it there. and emerging tech coverage.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Rome | LOCATION | 0.99+ |
Dave | PERSON | 0.99+ |
David | PERSON | 0.99+ |
OCI | ORGANIZATION | 0.99+ |
Nipun Agarwal | PERSON | 0.99+ |
Milan | LOCATION | 0.99+ |
45 times | QUANTITY | 0.99+ |
25 times | QUANTITY | 0.99+ |
12 channels | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Zen 4 | COMMERCIAL_ITEM | 0.99+ |
Kumaran | PERSON | 0.99+ |
HeatWave | ORGANIZATION | 0.99+ |
Zen 3 | COMMERCIAL_ITEM | 0.99+ |
second aspect | QUANTITY | 0.99+ |
Kumaran Siva | PERSON | 0.99+ |
12 data sets | QUANTITY | 0.99+ |
first aspect | QUANTITY | 0.99+ |
97 times | QUANTITY | 0.99+ |
Zen 2 | COMMERCIAL_ITEM | 0.99+ |
both | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Each | QUANTITY | 0.99+ |
1% | QUANTITY | 0.99+ |
two cycles | QUANTITY | 0.99+ |
three capabilities | QUANTITY | 0.99+ |
third thing | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
AVX 2 | COMMERCIAL_ITEM | 0.99+ |
AVX 512 | COMMERCIAL_ITEM | 0.99+ |
second thing | QUANTITY | 0.99+ |
Redshift ML | TITLE | 0.99+ |
six data sets | QUANTITY | 0.98+ |
HeatWave | TITLE | 0.98+ |
mySQL Autopilot | TITLE | 0.98+ |
two | QUANTITY | 0.98+ |
Nipun | PERSON | 0.98+ |
two categories | QUANTITY | 0.98+ |
mySQL | TITLE | 0.98+ |
two observations | QUANTITY | 0.98+ |
first part | QUANTITY | 0.98+ |
mySQL autopilot | TITLE | 0.98+ |
three | QUANTITY | 0.97+ |
SQL | TITLE | 0.97+ |
One example | QUANTITY | 0.97+ |
single database | QUANTITY | 0.95+ |
16 | QUANTITY | 0.95+ |
today | DATE | 0.95+ |
about six | QUANTITY | 0.95+ |
HeatWaves | ORGANIZATION | 0.94+ |
about a dozen data sets | QUANTITY | 0.94+ |
16 nodes | QUANTITY | 0.93+ |
mySQL HeatWave | TITLE | 0.93+ |
AMD Oracle Partnership Elevates MySQLHeatwave
(upbeat music) >> For those of you who've been following the cloud database space, you know that MySQL HeatWave has been on a technology tear over the last 24 months with Oracle claiming record breaking benchmarks relative to other database platforms. So far, those benchmarks remain industry leading as competitors have chosen not to respond, perhaps because they don't feel the need to, or maybe they don't feel that doing so would serve their interest. Regardless, the HeatWave team at Oracle has been very aggressive about its performance claims, making lots of noise, challenging the competition to respond, publishing their scripts to GitHub. But so far, there are no takers, but customers seem to be picking up on these moves by Oracle and it's likely the performance numbers resonate with them. Now, the other area we want to explore, which we haven't thus far, is the engine behind HeatWave and that is AMD. AMD's epic processors have been the powerhouse on OCI, running MySQL HeatWave since day one. And today we're going to explore how these two technology companies are working together to deliver these performance gains and some compelling TCO metrics. In fact, a recent Wikibon analysis from senior analyst Marc Staimer made some TCO comparisons in OLAP workloads relative to AWS, Snowflake, GCP, and Azure databases, you can find that research on wikibon.com. And with that, let me introduce today's guest, Nipun Agarwal senior vice president of MySQL HeatWave and Kumaran Siva, who's the corporate vice president for strategic business development at AMD. Welcome to theCUBE gentlemen. >> Welcome. Thank you. >> Thank you, Dave. >> Hey Nipun, you and I have talked a lot about this. You've been on theCUBE a number of times talking about MySQL HeatWave. But for viewers who may not have seen those episodes maybe you could give us an overview of HeatWave and how it's different from competitive cloud database offerings. >> Sure. So MySQL HeatWave is a fully managed MySQL database service offering from Oracle. It's a single database, which can be used to run transactional processing, analytics and machine learning workloads. So, in the past, MySQL has been designed and optimized for transaction processing. So customers of MySQL when they had to run, analytics machine learning, would need to extract the data out of MySQL, into some other database or service, to run analytics or machine learning. MySQL HeatWave offers a single database for running all kinds of workloads so customers don't need to extract data into some of the database. In addition to having a single database, MySQL HeatWave is also very performant compared to one up databases and also it is very price competitive. So the advantages are; single database, very performant, and very good price performance. >> Yes. And you've published some pretty impressive price performance numbers against competitors. Maybe you could describe those benchmarks and highlight some of the results, please. >> Sure. So one thing to notice that the performance of any database is going to like vary, the performance advantage is going to vary based on, the size of the data and the specific workloads, so the mileage varies, that's the first thing to know. So what we have done is, we have published multiple benchmarks. So we have benchmarks on PPCH or PPCDS and we have benchmarks on different data sizes because based on the customer's workload, the mileage is going to vary, so we want to give customers a broad range of comparisons so that they can decide for themselves. So in a specific case, where we are running on a 30 terabyte PPCH workload, HeatWave is about 18 times better price performance compared to Redshift. 18 times better compared to Redshift, about 33 times better price performance, compared to Snowflake, and 42 times better price performance compared to Google BigQuery. So, this is on 30 Terabyte PPCH. Now, if the data size is different, or the workload is different, the characteristics may vary slightly but this is just to give a flavor of the kind of performance advantage MySQL HeatWave offers. >> And then my last question before we bring in Kumaran. We've talked about the secret sauce being the tight integration between hardware and software, but would you add anything to that? What is that secret sauce in HeatWave that enables you to achieve these performance results and what does it mean for customers? >> So there are three parts to this. One is HeatWave has been designed with a scale out architecture in mind. So we have invented and implemented new algorithms for skill out query processing for analytics. The second aspect is that HeatWave has been really optimized for cloud, commodity cloud, and that's where AMD comes in. So for instance, many of the partitioning schemes we have for processing HeatWave, we optimize them for the L3 cache of the AMD processor. The thing which is very important to our customers is not just the sheer performance but the price performance, and that's where we have had a very good partnership with AMD because not only does AMD help us provide very good performance, but the price performance, right? And that all these numbers which I was showing, big part of it is because we are running on AMD which provides very good price performance. So that's the second aspect. And the third aspect is, MySQL autopilot, which provides machine learning based automation. So it's really these three things, a combination of new algorithms, design for scale out query processing, optimized for commodity cloud hardware, specifically AMD processors, and third, MySQL auto pilot which gives us this performance advantage. >> Great, thank you. So that's a good segue for AMD and Kumaran. So Kumaran, what is AMD bringing to the table? What are the, like, for instance, relevance specs of the chips that are used in Oracle cloud infrastructure and what makes them unique? >> Yeah, thanks Dave. That's a good question. So, OCI is a great customer of ours. They use what we call the top of stack devices meaning that they have the highest core count and they also are very, very fast cores. So these are currently Zen 3 cores. I think the HeatWave product is right now deployed on Zen 2 but will shortly be also on the Zen 3 core as well. But we provide in the case of OCI 64 cores. So that's the largest devices that we build. What actually happens is, because these large number of CPUs in a single package and therefore increasing the density of the node, you end up with this fantastic TCO equation and the cost per performance, the cost per for deployed services like HeatWave actually ends up being extraordinarily competitive and that's a big part of the contribution that we're bringing in here. >> So Zen 3 is the AMD micro architecture which you introduced, I think in 2017, and it's the basis for EPIC, which is sort of the enterprise grade that you really attacked the enterprise with. Maybe you could elaborate a little bit, double click on how your chips contribute specifically to HeatWave's, price performance results. >> Yeah, absolutely. So in the case of HeatWave, so as Nipun alluded to, we have very large L3 caches, right? So in our very, very top end parts just like the Milan X devices, we can go all the way up to like 768 megabytes of L3 cache. And that gives you just enormous performance and performance gains. And that's part of what we're seeing with HeatWave today and that not that they're currently on the second generation ROM based product, 'cause it's a 7,002 based product line running with the 64 cores. But as time goes on, they'll be adopting the next generation Milan as well. And the other part of it too is, as our chip led architecture has evolved, we know, so from the first generation Naples way back in 2017, we went from having multiple memory domains and a sort of NUMA architecture at the time, today we've really optimized that architecture. We use a common I/O Die that has all of the memory channels attached to it. And what that means is that, these scale out applications like HeatWave, are able to really scale very efficiently as they go from a small domain of CPUs to, for example the entire chip, all 64 cores that scaling, is been a key focus for AMD and being able to design and build architectures that can take advantage of that and then have applications like HeatWave that scale so well on it, has been, a key aim of ours. >> And Gen 3 moving up the Italian countryside. Nipun, you've taken the somewhat unusual step of posting the benchmark parameters, making them public on GitHub. Now, HeatWave is relatively new. So people felt that when Oracle gained ownership of MySQL it would let it wilt on the vine in favor of Oracle database, so you lost some ground and now, you're getting very aggressive with HeatWave. What's the reason for publishing those benchmark parameters on GitHub? >> So, the main reason for us to publish price performance numbers for HeatWave is to communicate to our customers a sense of what are the benefits they're going to get when they use HeatWave. But we want to be very transparent because as I said the performance advantages for the customers may vary, based on the data size, based on the specific workloads. So one of the reasons for us to publish, all these scripts on GitHub is for transparency. So we want customers to take a look at the scripts, know what we have done, and be confident that we stand by the numbers which we are publishing, and they're very welcome, to try these numbers themselves. In fact, we have had customers who have downloaded the scripts from GitHub and run them on our service to kind of validate. The second aspect is in some cases, they may be some deviations from what we are publishing versus what the customer would like to run in the production deployments so it provides an easy way, for customers to take the scripts, modify them in some ways which may suit their real world scenario and run to see what the performance advantages are. So that's the main reason, first, is transparency, so the customers can see what we are doing, because of the comparison, and B, if they want to modify it to suit their needs, and then see what is the performance of HeatWave, they're very welcome to do so. >> So have customers done that? Have they taken the benchmarks? And I mean, if I were a competitor, honestly, I wouldn't get into that food fight because of the impressive performance, but unless I had to, I mean, have customers picked up on that, Nipun? >> Absolutely. In fact, we have had many customers who have benchmarked the performance of MySQL HeatWave, with other services. And the fact that the scripts are available, gives them a very good starting point, and then they've also tweaked those queries in some cases, to see what the Delta would be. And in some cases, customers got back to us saying, hey the performance advantage of HeatWave is actually slightly higher than what was published and what is the reason. And the reason was, when the customers were trying, they were trying on the latest version of the service, and our benchmark results were posted let's say, two months back. So the service had improved in those two to three months and customers actually saw better performance. So yes, absolutely. We have seen customers download the scripts, try them and also modify them to some extent and then do the comparison of HeatWave with other services. >> Interesting. Maybe a question for both of you how is the competition responding to this? They haven't said, "Hey, we're going to come up "with our own benchmarks." Which is very common, you oftentimes see that. Although, for instance, Snowflake hasn't responded to data bricks, so that's not their game, but if the customers are actually, putting a lot of faith in the benchmarks and actually using that for buying decisions, then it's inevitable. But how have you seen the competition respond to the MySQL HeatWave and AMD combo? >> So maybe I can take the first track from the database service standpoint. When customers have more choice, it is invariably advantages for the customer because then the competition is going to react, right? So the way we have seen the reaction is that we do believe, that the other database services are going to take a closer eye to the price performance, right? Because if you're offering such good price performance, the vendors are already looking at it. And, you know, instances where they have offered let's say discount to the customers, to kind of at least like close the gap to some extent. And the second thing would be in terms of the capability. So like one of the things which I should have mentioned even early on, is that not only does MySQL HeatWave on AMD, provide very good price performance, say on like a small cluster, but it's all the way up to a cluster size of 64 nodes, which has about 1000 cores. So the point is, that HeatWave performs very well, both on a small system, as well as a huge scale out. And this is again, one of those things which is a differentiation compared to other services so we expect that even other database services will have to improve their offerings to provide the same good scale factor, which customers are now starting to expectancy, with MySQL HeatWave. >> Kumaran, anything you'd add to that? I mean, you guys are an arms dealer, you love all your OEMs, but at the same time, you've got chip competitors, Silicon competitors. How do you see the competitive-- >> I'd say the broader answer and the big picture for AMD, we're very maniacally focused on our customers, right? And OCI and Oracle are huge and important customers for us, and this particular use cases is extremely interesting both in that it takes advantage, very well of our architecture and it pulls out some of the value that AMD bring. I think from a big picture standpoint, our aim is to execute, to build to bring out generations of CPUs, kind of, you know, do what we say and say, sorry, say what we do and do what we say. And from that point of view, we're hitting, the schedules that we say, and being able to bring out the latest technology and bring it in a TCO value proposition that generationally keeps OCI and HeatWave ahead. That's the crux of our partnership here. >> Yeah, the execution's been obvious for the last several years. Kumaran, staying with you, how would you characterize the collaboration between, the AMD engineers and the HeatWave engineering team? How do you guys work together? >> No, I'd say we're in a very, very deep collaboration. So, there's a few aspects where, we've actually been working together very closely on the code and being able to optimize for both the large L3 cache that AMD has, and so to be able to take advantage of that. And then also, to be able to take advantage of the scaling. So going between, you know, our architecture is chip like based, so we have these, the CPU cores on, we call 'em CCDs and the inter CCD communication, there's opportunities to optimize an application level and that's something we've been engaged with. In the broader engagement, we are going back now for multiple generations with OCI, and there's a lot of input that now, kind of resonates in the product line itself. And so we value this very close collaboration with HeatWave and OCI. >> Yeah, and the cadence, Nip, and you and I have talked about this quite a bit. The cadence has been quite rapid. It's like this constant cycle every couple of months I turn around, is something new on HeatWave. But for question again, for both of you, what new things do you think that organizations, customers, are going to be able to do with MySQL HeatWave if you could look out next 12 to 18 months, is there anything you can share at this time about future collaborations? >> Right, look, 12 to 18 months is a long time. There's going to be a lot of innovation, a lot of new capabilities coming out on in MySQL HeatWave. But even based on what we are currently offering, and the trend we are seeing is that customers are bringing, more classes of workloads. So we started off with OLTP for MySQL, then it went to analytics. Then we increased it to mixed workloads, and now we offer like machine learning as alike. So one is we are seeing, more and more classes of workloads come to MySQL HeatWave. And the second is a scale, that kind of data volumes people are using HeatWave for, to process these mixed workloads, analytics machine learning OLTP, that's increasing. Now, along the way we are making it simpler to use, we are making it more cost effective use. So for instance, last time, when we talked about, we had introduced this real time elasticity and that's something which is a very, very popular feature because customers want the ability to be able to scale out, or scale down very efficiently. That's something we provided. We provided support for compression. So all of these capabilities are making it more efficient for customers to run a larger part of their workloads on MySQL HeatWave, and we will continue to make it richer in the next 12 to 18 months. >> Thank you. Kumaran, anything you'd add to that, we'll give you the last word as we got to wrap it. >> No, absolutely. So, you know, next 12 to 18 months we will have our Zen 4 CPUs out. So this could potentially go into the next generation of the OCI infrastructure. This would be with the Genoa and then Bergamo CPUs taking us to 96 and 128 cores with 12 channels at DDR five. This capability, you know, when applied to an application like HeatWave, you can see that it'll open up another order of magnitude potentially of use cases, right? And we're excited to see what customers can do do with that. It certainly will make, kind of the, this service, and the cloud in general, that this cloud migration, I think even more attractive. So we're pretty excited to see how things evolve in this period of time. >> Yeah, the innovations are coming together. Guys, thanks so much, we got to leave it there really appreciate your time. >> Thank you. >> All right, and thank you for watching this special Cube conversation, this is Dave Vellante, and we'll see you next time. (soft calm music)
SUMMARY :
and it's likely the performance Thank you. and how it's different from So the advantages are; single and highlight some of the results, please. the first thing to know. We've talked about the secret sauce So for instance, many of the relevance specs of the chips that are used and that's a big part of the contribution and it's the basis for EPIC, So in the case of HeatWave, of posting the benchmark parameters, So one of the reasons for us to publish, So the service had improved how is the competition responding to this? So the way we have seen the but at the same time, and the big picture for AMD, for the last several years. and so to be able to Yeah, and the cadence, and the trend we are seeing is we'll give you the last and the cloud in general, Yeah, the innovations we'll see you next time.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Marc Staimer | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Nipun | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
2017 | DATE | 0.99+ |
Dave | PERSON | 0.99+ |
OCI | ORGANIZATION | 0.99+ |
Zen 3 | COMMERCIAL_ITEM | 0.99+ |
7,002 | QUANTITY | 0.99+ |
Kumaran | PERSON | 0.99+ |
second aspect | QUANTITY | 0.99+ |
Nipun Agarwal | PERSON | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
12 | QUANTITY | 0.99+ |
64 cores | QUANTITY | 0.99+ |
768 megabytes | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
MySQL | TITLE | 0.99+ |
third aspect | QUANTITY | 0.99+ |
12 channels | QUANTITY | 0.99+ |
Kumaran Siva | PERSON | 0.99+ |
HeatWave | ORGANIZATION | 0.99+ |
96 | QUANTITY | 0.99+ |
18 times | QUANTITY | 0.99+ |
Bergamo | ORGANIZATION | 0.99+ |
three parts | QUANTITY | 0.99+ |
Delta | ORGANIZATION | 0.99+ |
three months | QUANTITY | 0.99+ |
MySQL HeatWave | TITLE | 0.99+ |
42 times | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
18 months | QUANTITY | 0.99+ |
Zen 2 | COMMERCIAL_ITEM | 0.99+ |
one | QUANTITY | 0.99+ |
GitHub | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.98+ |
second generation | QUANTITY | 0.98+ |
single database | QUANTITY | 0.98+ |
128 cores | QUANTITY | 0.98+ |
18 months | QUANTITY | 0.98+ |
three things | QUANTITY | 0.98+ |
Said Ouissal, Zededa | VMware Explore 2022
>>Hey, everyone. Welcome back to San Francisco. Lisa Martin and John furrier live on the floor at VMware Explorer, 2022. This is our third day of wall to wall coverage on the cube. But you know that cuz you've been here the whole time. We're pleased to welcome up. First timer to the cubes we saw is here. The CEO and founder of ZDA. Saed welcome to the program. >>Thank you for having me >>Talk to me a little bit about what ZDA does in edge. >>Sure. So ZDA is a company purely focused in edge computing. I started a company about five years ago, go after edge. So what we do is we help customers with orchestrating their edge, helping them to deploy secure monitor application services and devices at the edge. >>What's the business model for you guys. We get that out there. So the targeting the edge, which is everything from telco to whatever. Yeah. What's the business model. Yeah. >>Maybe before we go there, let's talk about edge itself. Cuz edge is complex. There's a lot of companies. I call 'em lens company nowadays, if you're not a cloud company, you're probably an edge company at this point. So we are focusing something called the distributed edge. So distributed edge. When you start putting tiny servers in environments like factory floors, solar farms, wind farms, even inside machines or well sites, et cetera. And a question that people always ask me, like why, why would you want to put, you know, servers there on servers supposed to be in a data center in the cloud? And the answer to the question actually is data gravity. So traditionally wherever the data gets created is where your applications live. But as we're connecting more and more devices to the edge of the network, we basically customers now are required to push the applications to the edge cause they can't go all the data to the cloud. So basically that's where we focus on people call it the far edge as well. You know, that's the term we've heard in the past as well. And what we do in our business model is provide customers a, a software as a service solution where they can basically deploy and monitor these applications at these highly distributed environments. >>Data, gravity comes up a lot and I want you to take a minute to explain the definition as it is today. And people have used that term, you know, with big data, going back to 2010 leads when we covering the Hadoop wave, which ended up becoming, you know, data, data, bricks, and snowflake now, but, but a lots changed, but what does it mean to be data gravity? It means that staying local, it's just what specifically describe and, and define what data gravity is. >>Yeah. So for me, data gravity is where you need to process the data, right? It's where the data usually gets created. So if you think about a web app, where does the data get created? Where people click on buttons, they, they interface with it. They, they upload content to it, et cetera. So that's where the data gravity therefore is therefore that's where you do your analytics. That's where you do your visualization processing, machine learning and all of those pieces. So it's really where that data gets created is where the data gravity in my view says, >>What are some of the challenges that data and opportunities that data gravity presents to customers? >>Well, obviously I think every enterprise in this day is trying to take data and make it a competitive advantage, right? Like faster decisions, better decisions, outcompete your competition by, you know, being first with a product or being first with a product with the future, et cetera. So, so I think, you know, if you're not a data driven enterprise by now, then I think the future may be a little bit bleak. >>Okay. So you're targeting the market distributed edge business model, SAS technology, secret sauce. What's that piece. >>Yeah. So that's, that's what the interesting part comes in. I think, you know, if you kind of look at the data center in the cloud, we've had these virtualization and orchestration stacks create, I mean, we're here in VMware Explorer. And as an example, what we basically, what we saw is that the edge is so unique and so different than what we've seen in the data center, in the cloud that we needed to build a complete brand new purpose-built illustration and virtualization solution. So that's really what we, we set off to do. So there's two components that we do. One end is we built a purpose-built edge operating system for the edge and we actually open sourced it. And the reason we opensource it, we said, Hey, you know, edge is so diverse. You know, depending on the environment you're running in a machine or in a vehicle or in a well site, you have different hardware, different networks, different applications you need to enable. >>And we will never be able to support all of them ourselves. As a matter of fact, we actually think there's a need for standardization at the edge. We need to kind of cut through all these silos that have been created traditionally from the embedded way of thinking. So we created basically an open source project in the Linux foundation in LFS, which is a sister organization through the CNCF it's called project Eve. And the idea is to create the Android of the edge, basically what Android became for mobile computing, an a common operating system. So you build one app. You can run in any phone in the world that runs Android, build an architecture. You build one app. You can run in any Eve powered node in the world, >>So distributed edge and you get the tech here, get the secret sauce. We'll get more into that in a second, but I wanna just tie one kick quick point and get your clarification on edge is becoming much more about the physical side too. I mean, absolutely. So when you talk about Android, you're making the reference of a phone. I get that's metaphor to what you're doing at the edge, wind farms, factories, alarms, light bulbs, buildings. I mean, that's what you're talking about, right? Yes. We're getting down to that very, >>Very physical, dark distributed locations. >>We're gonna come back to the CISO CSO. We're gonna come back to the CISO versus CSO question because is the CISO or CIO or who runs that anyway? So that's true. What's the important thing that's happening because that sounds like old OT world, like yes. Operating technology, not it information technology, is it a complete reset of those worlds or is it a collision? >>It's a great question. So what we're seeing is first of all, there is already compute in these environments, industrial PCs of existed well beyond, you know, an industrial automation has been done for many, many decades. The point is that that stuff has been done. Collect data has been collected, but never connected, right? So with edge computing, we're connecting now this data from an industrial machine and industrial process to the cloud, right? And one of the problems is it's data that comes of that industrial process too much to upload to the cloud. So I gotta analyze, analyze it locally. So one of the, the things we saw early on in edge is there's a lot of brownfield. Most of our customers today actually have applications running on windows and they would love to make in Linux and containers and Kubernetes, but it took them 20, 30 years to build those apps. And they basically are the money makers of the enterprise. So they are in a, in a transitionary phase and they need something that can take them from the brown to the Greenfield. So to your point, you gotta support all of these types of unique brownfield applications. >>So you're, you're saying I don't really care if this is a customer, how you get the data, you wanna start new start fresh. That's cool. But if you wanna take your old data, you'll >>Take that. Yeah. You don't wanna rebuild the whole machine. You're >>Just, they can life cycle it out on their own timetable. Yeah. >>So we had to learn, first of all, how do we take and lift and shift windows based industrial application and make it run at the edge on, on our architecture. Right? And then the second step is how do we then Sen off that data that this application is generating and do we fuse it with cloud native capability? Like, >>So your cloud, so your staff is your open source that you're giving to the Linux foundation as part of that Eve project that's available to everybody. So they can, they can look at the code, which is great by the way. Yeah. So people wanna do that. Yeah. Your self source, I'm assuming, is your hardened version with support? >>Well, we took what we took, what the open source companies did, opensource companies traditionally have sold, you know, basically a support model around the open source. We actually saw another problem. Customers has like, okay, now I have this node running and I can, you know, do this data analytics, but what if I have 15 or 20,000 of these node? And they're all around the world in remote locations on satellite links or wireless connectivity, how do I orchestrate them? So we actually build an orchestration service for these nodes running this open source >>Software. So that's a key secret sauce right there. >>That is the business model that taking open store and a lot. >>And you're taking your own code that you have. Okay. Got it. Cool. And then the customer's customer piece is, is key. So that's the final piece, I guess who's using it. >>Yeah. Well, and, >>And, and one of the business outcomes that they're achieving. Oh >>Yeah. Well, so maybe start with that first. I mean, we are deployed in customers in all and gas, for instance, helping them with the transition to renewable energy, right? So basically we, we have customers for instance, that deploy us in the, how they drill Wells is one use case and doing that better, faster, and cheaper and, and less environmental impacting. But we also have customers that use us in wind farms. We have, and solar farms, like we, one of the leading solar energy companies in the world is using us to bring down the cost of power by predicting failures ahead of time, for >>Instance. And when you're working with customers to create the optimal solution at the distributed edge, who are you working with in, within an organization? Yeah. >>It's usually a mix of OT and it people. Okay. So the OT people typically they're >>Arm wrestling, well, or they're getting along, actually, >>I think they're getting along very well. Okay, good. But they also agree that they have to have swim lanes. The it folks, obviously their job is to make sure, you know, everything is secure. Everything is according to the compliance it's, it's, you know, the, the best TCO on the infrastructure, those type of things, the OT guy, they, they, or girl, they care about the application. They care about the services. They care about the support new business. So how can you create a model that too can coexist? And if you do that, they get along really well. >>You know, we had an event called Supercloud and@theurlsupercloud.world, if you're watching check it out, it's our version of what we think multicloud will merge into including edge cuz edge is just another node in the, in the, in the network. As far as we're concerned, hybrid is the steady state. That's distributed computing on premise, private cloud, public cloud. We know what that looks like. People love that things are happening. Edge is like a whole nother new area. That's blossoming and with disruption, yeah. There's a lot of existing market and incumbents that need to be disrupted. And there's also a new capabilities that are coming that we don't yet see. So we're seeing it with the super cloud idea that these new kinds of clouds are emerging. Like there could be an edge cloud. Yeah. Why isn't there a security cloud, whereas the financial services cloud, whereas the insurance cloud, whereas the, so these become super clouds where the CapEx could be done by the Amazon, whatnot you've been following them is edge cloud. Can you make that a cloud? Is that what you guys are trying to do? And if so, what does that look like? Cause we we're adding a new track to our super cloud site. I mentioned on edge specifically, we're trying to figure out you and if you share your opinion, it'd be great. Can the E can edge clouds exist and be run by companies? Yeah. Or is that what you guys are trying to do? >>I, I, I mean, I think first of all, there is no edge without cloud, right? So when I meet any customer who says, Hey, we're gonna do edge without cloud. Then I'm like, you're probably not gonna do edge computing. Right. And, and the way we built the company and the way we think about it, it's about extending the cloud experience all the way into these embedded distributed environments. That's really, I think what customers are looking for, cuz customers love the simplicity of the cloud. They love the ease of use agility, all of that greatness. And they're like, Hey, I want that. But not in a, you know, in an Amazon or Azure data center. I want that in my factories. I want that in my wealth sites, in my vehicles. And that's really what I think the future >>Is gonna. And how long have you guys been around? What's the, what's the history of the company because you might actually be that cloud. Yeah. And are you on AWS or Azure? You're building your own. What's the, >>Yeah. Yeah. So >>Take it through the, the architecture because yeah, yeah, sure. You're a modern startup. I mean you gotta, and the edges you're going after you gotta be geared up. Yeah. To win that. Yeah. >>So, so the company's about five years old. So we, when we started focusing on edge, people didn't necessarily talk as much about edge. We kind of identified the it's like, you know, how do you find a black hole in, in the universe? Cuz you can't see it, but you sort of look around that's why you in it. And so we were like looking at it, like there's something gonna happen here at the edge of the network, because everybody's saying we're connecting these vice upload the data to the cloud's never gonna work. My background is networking. I worked at companies like Juniper and Ericsson ran several products there. So I know how the internet networks have built. And it was very Evan to me. It's not gonna be possible. My co-founders come from open source companies like pivotal and Cloudera. My auto co-founder was a, an engineer at sun Microsystems built the first network stack in the solar is operating system. So a lot of experience that kind of came together to build this. >>Yeah. Cloudera is a big day. That's where the cube started by the way. Yeah. >>Yeah. So, so we, we, we have, I think a good view on the stack, the cloud stack and therefore a good view of what the ed stack needs to look like. And then I think, you know, to answer your other question, our orchestration service runs in the cloud. We have, we actually are multi-cloud company. So we offer customers choice where they want to orchestrate the node from the nodes themself, never sit in a data center. They always highly embedded. We have customers are putting machines or inside these factory lines, et cetera. Are >>You running your SAS on Amazon web services or which >>Cloud we're running it on several clouds, including Amazon, all of, pretty much the cloud. So some customers say, Hey, I'd prefer to be on the Amazon set. And others customers say, I wanna be on Azure set. >>And you leverage their CapEx on that side. Yes. On behalf of yeah. >>Yeah. We, yes. Yes. But the majority of the customer data and, and all the data that the nodes process, the customer send it to their clouds. They don't send it to us. We don't get a copy of the camera feed analytics or the machine data. We actually decouple those though. So basically the, the team production data go straight to the customer's cloud and that's why they love us. >>And they choose that they can control their own desktop. >>Yeah. So we separate the management plane from the data plane at the edge. Yeah. >>That's a good call >>Actually. Yeah. That was another very important part of the architecture early on. Cause customers don't want us to see their, you know, highly confidential production data and we don't wanna have it either. So >>We had a great chat with Chris Wolf who works with kit culvert about control plane, data, plane. So that seems to be the trend data, plane customers want full yeah. Management of that. Yeah. Control plane. Maybe give multiple >>Versions. Yeah. Yeah. So our cloud consumption what the data we stories about the apps, their behavior, the networking, the security, all of that. That's what we store in our cloud. And then customers can access that and monitor. But the actual machine that I go somewhere else >>Here we are at VMware. Explore. Talk a little bit about the VMware relationship. You just had some big news the other day. >>Yeah. So two days ago we actually made a big announcement with VMware. So we signed an OEM agreement with VMware. So we're part now of VMware's edge compute stack. So VMware customers, as they start using the recently announced edge compute stack 2.0, that was announced here. Basically it's powered by Edda technology. So it's a really exciting partnership as part of this, we actually building integrations with the VMware organization products. So that's basically now extending to more, you know, other groups inside VMware. >>So what's the value in it for VMware customers. >>Yeah. So I think the, the, the benefit of, of VMware customers, I think cus VMware customers want that multi-cloud multi edge orchestration experience. So they wanna be able to deploy workloads in the cloud. They wanna deploy the workloads in the data center. And of course also at the edge. So by us integrating in that vision customers now can have that unified experience from cloud to edge and anywhere in between. >>What's the big vision that you see happening at the edge. I mean, a lot of the VMware customers here, they're classic it that have evolved into ops now, dev ops. Now you've got second data ops coming. The edge is gonna right around the corner for them. They're dealing with it now, probably just kicking the tires, towing the water kind of thing. Where do you see the vision going? Cuz now, no matter what happens with VMware, the Broadcom, this wave is still here. You got AWS, got Azure, got Google cloud, you got Oracle, Alibaba internationally. And the cloud native surges here. How do you see that disrupting the existing edge? Because let's face it the O some of those OT players, a little bit old and antiquated, a little bit outdated. I mean, I was talking to a telco person. They, they puked the word open source. I mean, these people are so dogmatic on, on their architecture. Yeah. They're gonna get disrupted. It's a matter of time. Yeah. Where's the new guard come in. How do you see the configuration changing in the landscape? Because some people will cross over to the right side of the street here. Yeah. Some won't yeah. Open circle. Dominate cloud native will be key. Yeah. >>Well, I mean, I think, again, let's, let's take an example of a vertical that's heavily disrupted now as the automotive market, right? The, so look at Tesla and look at all these companies, they built, they built software first cars, right? Software, first delivery of capabilities and everything else. And the, and the incumbents. They have only two options, right? Either they try to respond by adopting open source cloud, native technologies. Like the, these new entrants have done and really, you know, compete with them at that level, or they can become commodity. Right. So, and I think that's the customers we're seeing the smart customers go like, we need to compete with these guys. We need to figure out how to take this technology in. And they need partners like us and partners like VMware for them. >>Do you see customers becoming cloud super cloud players? If they continue to keep leveraging the CapEx of the clouds and focus all their operational capital on top line revenue, generating activities. >>Yeah. I, so I think the CapEx model of the cloud is a great benefit of the cloud, but I think that is not, what's the longer term future of the cloud. I think the op the cloud operating model is the future. Like the agility, the ability imagine embedded software that, you know, you do an over the year update to fix a bug, but it's very hard to make a, an embedded device smarter over time. And then imagine if you can run cloud native software, you can roll out every two weeks new features and make that thing smarter, intelligent, and continue to help you in your business. That I think is what cloud did ultimately. And I think that is what really these customers are gonna need at their edge. >>Well, we talked about the value within it for customers with the VMware partnership, but what are some of your expectations? Obviously, this is a pretty powerful partnership for you guys. Yeah. What are some of the things that you're expecting that this is gonna drive? Yeah, >>So we, we, we have always operated at the more OT layer, distributed organizations in retail, energy, industrial automotive. Those are the verticals we, so we've developed. I think a lot of experience there, what, what we're seeing as we talk to those customers is they obviously have it organizations and the it organizations, Hey, that's great. You're looking at its computing, but how do we tie this into the existing investments we made with VMware? And how do we kind of take that also to this new environment? And I think that's the expectation I have is that I think we will be able to, to talk to the it folks and say, Hey, you can actually talk to the OT person. And both of you will speak the same language. You probably will both standardize on the same architecture and you'll be together deploying and enabling this new agility at the edge. >>What are some of the next things coming up for ZDA and the team? >>Well, so we've had a really amazing few quarters. We just close a series B round. So we've raised the companies raised over 55 million so far, we're growing very rapidly. We opened up no new international offices. I would say the, the early customers that we started deploying, wait a while back, they're now going into mass scale deployment. So we have now deployments underway in, you know, the 10 to hundred thousands of nodes at certain customers and in amazing environments. And so, so for us, it's continuing to prove the product in more and more verticals. Our, our product is really built for the largest of the largest. So, you know, for the size of the company, we are, we have a high concentration of fortune 500 global 500 customers, and some of them even invested in our rounds recently. So we we've been really, you know, honored with that support. Well, congratulations. Good stuff, edges popping. All right. Thank you. >>Thank you so much for joining us, talking about what you're doing in distributed edge. What's in it for customers, the VMware partnership, and by the way, congratulations on >>That too. Thank you. Thank you so much. Nice to meet you. Thank >>You. All right. Nice to meet you as well for our guest and John furrier. I'm Lisa Martin. You're watching the cube live from VMware Explorer, 22, John and I will be right back with our next guest.
SUMMARY :
But you know that cuz you've been here the whole time. So what we do is we help customers with orchestrating What's the business model for you guys. And the answer to the question actually And people have used that term, you know, with big data, going back to 2010 leads when we covering the Hadoop So that's where the data gravity therefore is therefore that's where you do your analytics. so I think, you know, if you're not a data driven enterprise by now, then I think the future may be a little bit bleak. What's that piece. And the reason we opensource it, And the idea is to create the Android of the edge, basically what Android became for mobile computing, So when you talk about Android, you're making the reference of a phone. So that's true. So one of the, the things we saw early But if you wanna take your old data, you'll You're Just, they can life cycle it out on their own timetable. So we had to learn, first of all, how do we take and lift and shift windows based industrial application So they can, they can look at the code, which is great by the way. So we actually build an orchestration service for these nodes running this open source So that's a key secret sauce right there. So that's the final piece, I guess who's using it. And, and one of the business outcomes that they're achieving. I mean, we are deployed in customers in all and gas, edge, who are you working with in, within an organization? So the OT people typically they're So how can you create a model that too can coexist? Or is that what you guys are trying to do? And, and the way we built the company and And are you on AWS or Azure? I mean you gotta, and the edges you're going after you gotta be We kind of identified the it's like, you know, how do you find a black hole in, That's where the cube started by the way. And then I think, you know, to answer your other question, So some customers say, And you leverage their CapEx on that side. the team production data go straight to the customer's cloud and that's why they love us. you know, highly confidential production data and we don't wanna have it either. So that seems to be the trend data, plane customers want full yeah. But the actual machine that I go somewhere else You just had some big news the other day. So that's basically now extending to more, you know, other groups inside VMware. And of course also at the edge. What's the big vision that you see happening at the edge. Like the, these new entrants have done and really, you know, compete with them at that level, Do you see customers becoming cloud super cloud players? that thing smarter, intelligent, and continue to help you in your business. What are some of the things that you're expecting that this is gonna drive? And I think that's the expectation I have is that I think we will be able to, to talk to the it folks and say, So we we've been really, you know, honored with that support. Thank you so much for joining us, talking about what you're doing in distributed edge. Thank you so much. Nice to meet you as well for our guest and John furrier.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Ericsson | ORGANIZATION | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Juniper | ORGANIZATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Chris Wolf | PERSON | 0.99+ |
Tesla | ORGANIZATION | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
2010 | DATE | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
15 | QUANTITY | 0.99+ |
Android | TITLE | 0.99+ |
20 | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Zededa | PERSON | 0.99+ |
John | PERSON | 0.99+ |
both | QUANTITY | 0.99+ |
two components | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
second step | QUANTITY | 0.99+ |
third day | QUANTITY | 0.99+ |
sun Microsystems | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
CNCF | ORGANIZATION | 0.99+ |
20,000 | QUANTITY | 0.99+ |
Linux | TITLE | 0.99+ |
CapEx | ORGANIZATION | 0.99+ |
windows | TITLE | 0.99+ |
Cloudera | ORGANIZATION | 0.99+ |
John furrier | PERSON | 0.99+ |
two days ago | DATE | 0.98+ |
telco | ORGANIZATION | 0.98+ |
over 55 million | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
two options | QUANTITY | 0.98+ |
one app | QUANTITY | 0.98+ |
500 customers | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
One end | QUANTITY | 0.98+ |
Hadoop wave | EVENT | 0.98+ |
Broadcom | ORGANIZATION | 0.97+ |
Kubernetes | TITLE | 0.97+ |
first network | QUANTITY | 0.96+ |
LFS | ORGANIZATION | 0.96+ |
multicloud | ORGANIZATION | 0.95+ |
VMware Explorer | TITLE | 0.95+ |
first cars | QUANTITY | 0.93+ |
one use case | QUANTITY | 0.91+ |
Ouissal | PERSON | 0.9+ |
about five years old | QUANTITY | 0.9+ |
2022 | DATE | 0.89+ |
ZDA | ORGANIZATION | 0.88+ |
pivotal | ORGANIZATION | 0.87+ |
about five years ago | DATE | 0.87+ |
series B round | OTHER | 0.86+ |
hundred thousands | QUANTITY | 0.85+ |
30 years | QUANTITY | 0.81+ |
Jason Collier, AMD | VMware Explore 2022
(upbeat music) >> Welcome back to San Francisco, "theCUBE" is live, our day two coverage of VMware Explore 2022 continues. Lisa Martin with Dave Nicholson. Dave and I are pleased to welcome Jason Collier, principal member of technical staff at AMD to the program. Jason, it's great to have you. >> Thank you, it's great to be here. >> So what's going on at AMD? I hear you have some juicy stuff to talk about. >> Oh, we've got a ton of juicy stuff to talk about. Clearly the Project Monterey announcement was big for us, so we've got that to talk about. Another thing that I really wanted to talk about was a tool that we created and we call it, it's the VMware Architecture Migration Tool, call it VAMT for short. It's a tool that we created and we worked together with VMware and some of their professional services crew to actually develop this tool. And it is also an open source based tool. And really the primary purpose is to easily enable you to move from one CPU architecture to another CPU architecture, and do that in a cold migration fashion. >> So we're probably not talking about CPUs from Tandy, Radio Shack systems, likely this would be what we might refer to as other X86 systems. >> Other X86 systems is a good way to refer to it. >> So it's interesting timing for the development and the release of a tool like this, because in this sort of X86 universe, there are players who have been delayed in terms of delivering their next gen stuff. My understanding is AMD has been public with the idea that they're on track for by the end of the year, Genoa, next gen architecture. So can you imagine a situation where someone has an existing set of infrastructure and they're like, hey, you know what I want to get on board, the AMD train, is this something they can use from the VMware environment? >> Absolutely, and when you think about- >> Tell us exactly what that would look like, walk us through 100 servers, VMware, 1000 VMs, just to make the math easy. What do you do? How does it work? >> So one, there's several things that the tool can do, we actually went through, the design process was quite extensive on this. And we went through all of the planning phases that you need to go through to do these VM migrations. Now this has to be a cold migration, it's not a live migration. You can't do that between the CPU architectures. But what we do is you create a list of all of the virtual machines that you want to migrate. So we take this CSV file, we import this CSV file, and we ask for things like, okay, what's the name? Where do you want to migrate it to? So from one cluster to another, what do you want to migrate it to? What are the networks that you want to move it to? And then the storage platform. So we can move storage, it could either be shared storage, or we could move say from VSAN to VSAN, however you want to set it up. So it will do those storage migrations as well. And then what happens is it's actually going to go through, it's going to shut down the VM, it's going to take a snapshot, it is going to then basically move the compute and/or storage resources over. And once it does that, it's going to power 'em back up. And it's going to check, we've got some validation tools, where it's going to make sure VM Tools comes back up where everything is copacetic, it didn't blue screen or anything like that. And once it comes back up, then everything's good, it moves onto the next one. Now a couple of things that we've got feature wise, we built into it. You can parallelize these tasks. So you can say, how many of these machines do you want to do at any given time? So it could be, say 10 machines, 50 machines, 100 machines at a time, that you want to go through and do this move. Now, if it did blue screen, it will actually roll it back to that snapshot on the origin cluster. So that there is some protection on that. A couple other things that are actually in there are things like audit tracking. So we do full audit logging on this stuff, we take a snapshot, there's basically kind of an audit trail of what happens. There's also full logging, SYS logging, and then also we'll do email reporting. So you can say, run this and then shoot me a report when this is over. Now, one other cool thing is you can also actually define a change window. So I don't want to do this in the middle of the afternoon on a Tuesday. So I want to do this later at night, over the weekend, you can actually just queue this up, set it, schedule it, it'll run. You can also define how long you want that change window to be. And what it'll do, it'll do as many as it can, then it'll effectively stop, finish up, clean up the tasks and then send you a report on what all was successfully moved. >> Okay, I'm going to go down the rabbit hole a little bit on this, 'cause I think it's important. And if I say something incorrect, you correct me. >> No problem. >> In terms of my technical understanding. >> I got you. >> So you've got a VM, essentially a virtual machine typically will consist of an entire operating system within that virtual machine. So there's a construct that containerizes, if you will, the operating system, what is the difference, where is the difference in the instruction set? Where does it lie? Is it in the OS' interaction with the CPU or is it between the construct that is the sort of wrapper around the VM that is the difference? >> It's really primarily the OS, right? And we've not really had too many issues doing this and most of the time, what is going to happen, that OS is going to boot up, it's going to recognize the architecture that it's on, it's going to see the underlying architecture, and boot up. All the major operating systems that we test worked fine. I mean, typically they're going to work on all the X86 platforms. But there might be instruction sets that are kind of enabled in one architecture that may not be in another architecture. >> And you're looking for that during this process. >> Well usually the OS itself is going to kind of detect that. So if it pops up, the one thing that is kind of a caution that you need to look for. If you've got an application that's explicitly using an instruction set that's on one CPU vendor and not the other CPU vendor. That's the one thing where you're probably going to see some application differences. That said, it'll probably be compatible, but you may not get that instruction set advantage in it. >> But this tool remediates against that. >> Yeah, and what we do, we're actually using VM Tools itself to go through and validate a lot of those components. So we'll look and make sure VM Tools is enabled in the first place, on the source system. And then when it gets to the destination system, we also look at VM Tools to see what is and what is not enabled. >> Okay, I'm going to put you on the spot here. What's the zinger, where doesn't it work? You already said cold, we understand, you can schedule for cold migrations, that's not a zinger. What's the zinger, where doesn't it work? >> It doesn't work like, live migrations just don't work. >> No live, okay, okay, no live. What about something else? What's the oh, you've got that version, you've got that version of X86 architecture, it-won't work, anything? >> A majority of those cases work, where it would fail, where it's going to kick back and say, hey, VM Tools is not installed. So where you would see this is if you're running a virtual appliance from some vendor, like insert vendor here that say, got a firewall, or got something like that, and they don't have VM Tools enabled. It's going to fail it out of the gate, and say, hey, VM Tools is not on this, you might want to manually do it. >> But you can figure out how to fix that? >> You can figure out how to do that. You can also, and there's a flag in there, so in kind of the options that you give it, you say, ignore VM Tools, don't care, move it anyway. So if you've got less, some VMs that are in there, but they're not a priority VM, then it's going to migrate just fine. >> Got It. >> Can you elaborate a little bit on the joint development work that AMD and VMware are doing together and the value in it for customers? >> Yeah, so it's one of those things we worked with VMware to basically produce this open source tool. So we did a lot of the core component and design and we actually engaged VMware Professional Services. And a big shout out to Austin Browder. He helped us a ton in this project specifically. And we basically worked, we created this, kind of co-designed, what it was going to look like. And then jointly worked together on the coding, of pulling this thing together. And then after that, and this is actually posted up on VMware's public repos now in GitHub. So you can go to GitHub, you can go to the VMware samples code, and you can download this thing that we've created. And it's really built to help ease migrations from one architecture to another. So if you're looking for a big data center move and you got a bunch of VMs to move. I mean, even if it's same architecture to same architecture, it's definitely going to ease the pain of going through and doing a migration of, it's one thing when you're doing 10 machines, but when you're doing 10,000 virtual machines, that's a different story. It gets to be quite operationally inefficient. >> I lose track after three. >> Yeah. >> So I'm good for three, not four. >> I was going to ask you what your target market segment is here. Expand on that a little bit and talk to me about who you're working with and those organizations. >> So really this is targeted toward organizations that have large deployments in enterprise, but also I think this is a big play with channel partners as well. So folks out there in the channel that are doing these migrations and they do a lot of these, when you're thinking about the small and mid-size organizations, it's a great fit for that. Especially if they're kind of doing that upgrade, the lift and shift upgrade, from here's where you've been five to seven years on an architecture and you want to move to a new architecture. This is really going to help. And this is not a point and click GUI kind of thing. It's command line driven, it's using PowerShell, we're using PowerCLI to do the majority of this work. And for channel partners, this is an excellent opportunity to put the value and the value add and VAR, And there's a lot of opportunity for, I think, channel partners to really go and take this. And once again, being open source. We expect this to be extensible, we want the community to contribute and put back into this to basically help grow it and make it a more useful tool for doing these cold migrations between CPU architectures. >> Have you seen any in the last couple of years of dynamics, obviously across the world, any industries in particular that are really leading edge for what you guys are doing? >> Yeah, that's really, really interesting. I mean, we've seen it, it's honestly been a very horizontal problem, pretty much across all vertical markets. I mean, we've seen it in financial services, we've seen it in, honestly, pretty much across the board. Manufacturing, financial services, healthcare, we have seen kind of a strong interest in that. And then also we we've actually taken this and presented this to some of our channel partners as well. And there's been a lot of interest in it. I think we presented it to about 30 different channel partners, a couple of weeks back about this. And I got contact from 30 different channel partners that said they're interested in basically helping us work on it. >> Tagging on to Lisa's question, do you have visibility into the AMD thought process around the timing of your next gen release versus others that are competitors in the marketplace? How you might leverage that in terms of programs where partners are going out and saying, hey, perfect time, you need a refresh, perfect time to look at AMD, if you haven't looked at them recently. Do you have any insight into that in what's going on? I know you're focused on this area. But what are your thoughts on, well, what's the buzz? What's the buzz inside AMD on that? >> Well, when you look overall, if you look at the Gartner Hype Cycle, when VMware was being broadly adopted, when VMware was being broadly adopted, I'm going to be blunt, and I'm going to be honest right here, AMD didn't have a horse in the race. And the majority of those VMware deployments we see are not running on AMD. Now that said, there's an extreme interest in the fact that we've got these very cored in systems that are now coming up on, now you're at that five to seven year refresh window of pulling in new hardware. And we have extremely attractive hardware when it comes to running virtualized workloads. The test cluster that I'm running at home, I've got that five to seven year old gear, and I've got some of the, even just the Milan systems that we've got. And I've got three nodes of another architecture going onto AMD. And when I got these three nodes completely maxed to the number of VMs that I can run on 'em, I'm at a quarter of the capacity of what I'm putting on the new stuff. So what you get is, I mean, we worked the numbers, and it's definitely, it's like a 30% decrease in the amount of resources that you need. >> That's a compelling number. >> It's a compelling number. >> 5%, 10%, nobody's going to do anything for that. You talk 30%. >> 30%. It's meaningful, it's meaningful. Now you you're out of Austin, right? >> Yes. >> So first thing I thought of when you talk about running clusters in your home is the cost of electricity, but you're okay. >> I'm okay. >> You don't live here, you don't live here, you don't need to worry about that. >> I'm okay. >> Do you have a favorite customer example that you think really articulates the value of AMD when you're in customer conversations and they go, why AMD and you hit back with this? >> Yeah. Actually it's funny because I had a conversation like that last night, kind of random person I met later on in the evening. We were going through this discussion and they were facing exactly this problem. They had that five to seven year infrastructure. It's funny, because the guy was a gamer too, and he's like, man, I've always been a big AMD fan, I love the CPUs all the way since back in basically the Opterons and Athlons right. He's like, I've always loved the AMD systems, loved the graphics cards. And now with what we're doing with Ryzen and all that stuff. He's always been a big AMD fan. He's like, and I'm going through doing my infrastructure refresh. And I told him, I'm just like, well, hey, talk to your VAR and have 'em plug some AMD SKUs in there from the Dells, HPs and Lenovos. And then we've got this tool to basically help make that migration easier on you. And so once we had that discussion and it was great, then he swung by the booth today and I was able to just go over, hey, this is the tool, this is how you use it, here's all the info. Call me if you need any help. >> Yeah, when we were talking earlier, we learned that you were at Scale. So what are you liking about AMD? How does that relate? >> The funny thing is this is actually the first time in my career that I've actually had a job where I didn't work for myself. I've been doing venture backed startups the last 25 years and we've raised couple hundred million dollars worth of investment over the years. And so one, I figured, here I am going to AMD, a larger corporation. I'm just like, am I going to be able to make it a year? And I have been here longer than a year and I absolutely love it. The culture at AMD is amazing. We still have that really, I mean, almost it's like that underdog mentality within the organization. And the team that I'm working with is a phenomenal team. And it's actually, our EVP and our Corp VP, were actually my executive sponsors, we were at a prior company. They were one of my executive sponsors when I was at Scale. And so my now VP boss calls me up and says, hey, I'm putting a band together, are you interested? And I was kind of enjoying a semi-retirement lifestyle. And then I'm just like, man, because it's you, yes, I am interested. And the group that we're in, the work that we're doing, the way that we're really focusing on forward looking things that are affecting the data center, what's going to be the data center like three to five years from now. It's exciting, and I am having a blast, I'm having the time of my life. I absolutely love it. >> Well, that relationship and the trust that you will have with each other, that bleeds into the customer conversations, the partner conversations, the employee conversations, it's all inextricably linked. >> Yes it is. >> And we want to know, you said three to five years out, like what? Like what? Just general futurist stuff, where do you think this is going. >> Well, it's interesting. >> So moon collides with the earth in 2025, we already know that. >> So we dialed this back to the Pensando acquisition. When you look at the Pensando acquisition and you look at basically where data centers are today, but then you look at where basically the big hyperscalers are. You look at an AWS, you look at their architecture, you specifically wrap Nitro around that, that's a very different architecture than what's being run in the data center. And when you look at what Pensando does, that's a lot of starting to bring what these real clouds out there, what these big hyperscalers are running into the grasps of the data center. And so I think you're going to see a fundamental shift. The next 10 years are going to be exciting because the way you look at a data center now, when you think of what CPUs do, what shared storage, how the networking is all set up, it ain't going to look the same. >> Okay, so the competing vision with that, to play devil's advocate, would be DPUs are kind of expensive. Why don't we just use NICs, give 'em some more bandwidth, and use the cheapest stuff. That's the competing vision. >> That could be. >> Or the alternative vision, and I imagine everything else we've experienced in our careers, they will run in parallel paths, fit for function. >> Well, parallel paths always exist, right? Otherwise, 'cause you know how many times you've heard mainframe's dead, tape's dead, spinning disk is dead. None of 'em dead, right? The reality is you get to a point within an industry where it basically goes from instead of a growth curve like that, it goes to a growth curve of like that, it's pretty flat. So from a revenue growth perspective, I don't think you're going to see the revenue growth there. I think you're going to see the revenue growth in DPUs. And when you actually take, they may be expensive now, but you look at what Monterey's doing and you look at the way that those DPUs are getting integrated in at the OEM level. It's going to be a part of it. You're going to order your VxRail and VSAN style boxes, they're going to come with them. It's going to be an integrated component. Because when you start to offload things off the CPU, you've driven your overall utilization up. When you don't have to process NSX on basically the X86, you've just freed up cores and a considerable amount of them. And you've also moved that to where there's a more intelligent place for that pack to be processed right, out here on this edge. 'Cause you know what, that might not need to go into the host bus at all. So you have just alleviated any transfers over a PCI bus, over the PCI lanes, into DRAM, all of these components, when you're like, but all to come with, oh, that bit needs to be on this other machine. So now it's coming in and it's making that decision there. And then you take and integrate that into things like the Aruba Smart Switch, that's running the Pensando technology. So now you got top of rack that is already making those intelligent routing decisions on where packets really need to go. >> Jason, thank you so much for joining us. I know you guys could keep talking. >> No, I was going to say, you're going to have to come back. You're going to have to come back. >> We've just started to peel the layers of the onion, but we really appreciate you coming by the show, talking about what AMD and VMware are doing, what you're enabling customers to achieve. Sounds like there's a lot of tailwind behind you. That's awesome. >> Yeah. >> Great stuff, thank you. >> It's a great time to be at AMD, I can tell you that. >> Oh, that's good to hear, we like it. Well, thank you again for joining us, we appreciate it. For our guest and Dave Nicholson, I'm Lisa Martin. You're watching "theCUBE Live" from San Francisco, VMware Explore 2022. We'll be back with our next guest in just a minute. (upbeat music)
SUMMARY :
Jason, it's great to have you. I hear you have some to easily enable you to move So we're probably good way to refer to it. and the release of a tool like this, 1000 VMs, just to make the math easy. And it's going to check, we've Okay, I'm going to In terms of my that is the sort of wrapper and most of the time, that during this process. that you need to look for. in the first place, on the source system. What's the zinger, where doesn't it work? It doesn't work like, live What's the oh, you've got that version, So where you would see options that you give it, And a big shout out to Austin Browder. I was going to ask you what and the value add and VAR, and presented this to some of competitors in the marketplace? in the amount of resources that you need. nobody's going to do anything for that. Now you you're out of Austin, right? is the cost of electricity, you don't live here, you don't They had that five to So what are you liking about AMD? that are affecting the data center, Well, that relationship and the trust where do you think this is going. we already know that. because the way you look Okay, so the competing Or the alternative vision, And when you actually take, I know you guys could keep talking. You're going to have to come back. peel the layers of the onion, to be at AMD, I can tell you that. Oh, that's good to hear, we like it.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Jason Collier | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
50 machines | QUANTITY | 0.99+ |
10 machines | QUANTITY | 0.99+ |
Jason | PERSON | 0.99+ |
10 machines | QUANTITY | 0.99+ |
100 machines | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Austin | LOCATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
five | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
100 servers | QUANTITY | 0.99+ |
seven year | QUANTITY | 0.99+ |
theCUBE Live | TITLE | 0.99+ |
10,000 virtual machines | QUANTITY | 0.99+ |
Lenovos | ORGANIZATION | 0.99+ |
30% | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
four | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
10% | QUANTITY | 0.99+ |
30 different channel partners | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
earth | LOCATION | 0.99+ |
5% | QUANTITY | 0.99+ |
1000 VMs | QUANTITY | 0.99+ |
Dells | ORGANIZATION | 0.99+ |
GitHub | ORGANIZATION | 0.99+ |
seven years | QUANTITY | 0.98+ |
Austin Browder | PERSON | 0.98+ |
a year | QUANTITY | 0.98+ |
Tandy | ORGANIZATION | 0.98+ |
Radio Shack | ORGANIZATION | 0.98+ |
VMware | ORGANIZATION | 0.98+ |
Monterey | ORGANIZATION | 0.98+ |
today | DATE | 0.97+ |
HPs | ORGANIZATION | 0.97+ |
first time | QUANTITY | 0.97+ |
Tuesday | DATE | 0.97+ |
Scale | ORGANIZATION | 0.97+ |
VM Tools | TITLE | 0.97+ |
one thing | QUANTITY | 0.96+ |
last night | DATE | 0.96+ |
about 30 different channel partners | QUANTITY | 0.95+ |
first | QUANTITY | 0.95+ |
Athlons | COMMERCIAL_ITEM | 0.95+ |
VxRail | COMMERCIAL_ITEM | 0.95+ |
X86 | TITLE | 0.94+ |
Pensando | ORGANIZATION | 0.94+ |
VMware Explore 2022 | TITLE | 0.94+ |
Ryzen | COMMERCIAL_ITEM | 0.94+ |
five years | QUANTITY | 0.93+ |
Ash McCarty, Dell Technologies & Josh Prewitt, Rackspace Technology | VMware Explore 2022
(modern music) >> Welcome back, everyone to theCUBE's live coverage here in San Francisco for VMware Explore, formerly VMworld. theCUBE's been here 12 years today, we've been watching the evolution of the user conference. It's been quite a journey to see and, you know, virtualization just explode. We got two great guests here, we're going to break it all down. Ash McCarty, director of Multicloud Product Management Dell Technologies, no stranger to the VMworld, now VMware Explore, and Josh Prewitt, Chief Product Officer at Rackspace Technology. Great to see you guys, thanks for coming on. >> Absolutely. >> Yeah, thanks so much, thanks for having us. >> So, you know, the theme this year is multicloud, but it's really all about vSphere 8's out, you got VxRail, you got containers, you got the magic going on around cloud native, which it really points to the future state of where this is going, which is agile enterprises, infrastructure as code, high performance under the hood, I mean, all the things that you guys have been doing for many, many years and decades and business, but now with VMware putting it all together, it feels like, this year, it's like you got visibility into the value proposition, people have clear line of sight into where the performances are from the hardware software and now Cloud, it's kind of coming together, feels like it's coming together. Let's talk about that and the relationship between you guys, Rackspace and Dell and VMware. >> Perfect. That sounds great. Well, thanks so much for having us. You know, I'll sort of kick that off. We've got a huge lifelong partnership and relationship with Dell and VMware and the technologies that these guys create that we're able to put in front of our customers are really what allows us to go drive those business outcomes. So, yeah, happy to dive into it. >> Yeah, and I think to add to that, we understand that customers have a tremendously complex challenge ahead of them on managing their infrastructure. That's why with VxRail, we have intelligent infrastructure. We want it to simplify the outcomes for customers no matter if they're managing VMware or if they're managing the actual hardware infrastructure underneath it. >> Yeah, one of the things that we always talk about, you know, you read about it on the blogs and the news and the startup world, is "Oh, product-market fit," and, well, it kind of applies here, if you think about what's going on on the product side with the Edge emerging, hybrid cloud on pace with private cloud, and obviously, cloud native is great too if you have native applications in there, but now, putting it all together, you're hearing things like the telco cloud, I hear buzzwords like that, I hear supercloud, which we promoting, which you see in companies becoming cloud themselves, with the CapEx being handled by either public cloud or optimized on premise or hosted hardware. I mean, this is now, this is not all about everything's going to the cloud, this is now cloud operations on premise and in hosting hardware, so I'd love to get your perspective on that because you guys are huge hosting, you've got huge experience there, modernizing all the time. What does the modern era look like for the customer? >> Yeah, yeah, so, I mean, I think it's very clear to everybody that it's a multicloud world, right? I think the main question is, are you multicloud as a strategy, or are you multicloud as a situation? Because everybody's multicloud. That ship has sailed, right? >> Yeah, exactly. >> And so, when I look at the capabilities that we have with the partnership with Dell and the VxRail technologies, you know, life-cycle management that you have to go and perform across your fleet can be extremely difficult, and whenever you take something like the VxRail and you add, you know, you have the hardware and you have the software all fully integrated there, it makes it much easier to do life-cycle management, so for a company like Rackspace, where we have tens of thousands of nodes that we're managing for customers across 29 global data centers, and we're all over the place, the ability to have that strength with Dell's hardware, the VMware platform improve life-cycle management makes it so much easier for us to manage our fleet and be able to deliver those outcomes even faster for customers. >> So assuming that VxRail isn't a virtual railroad that delivers data to Rackspace data centers, if it's not that, what is it, Ash? Give us a little premier on what VxRail is. >> Well, VxRail is the first and only jointly engineered HCI system with VMware, so everything we do with VMware is better. >> So hyperconverged infrastructure. >> Hyperconverged infrastructure. >> What we used to call a server because all the bits are in the box, right? >> All the storage is computed in there. >> Everything's in there. Right. >> Simplifies management. And we built in with the VxRail HCI system software, which is really our secret sauce, we built in to actually add those automation capabilities with VMware, so it allows you to scale out very quickly, scale up very quickly. And one of our big capabilities is our life-cycle management, which is full stack, meaning it life-cycles the entire vSphere stack as well as the hardware infrastructure underneath as one continuously validated state, meaning that customers can focus more on their business outcomes and driving their business forward versus spending time managing their infrastructure. >> And when you talk about customers, it's also the value proposition that's flowing through Rackspace because Rackspace, when you install these systems, how long does it take to spin up to have a VM available for use when you install one of these systems? >> Oh, so you can have the system up and running very quickly. So we automate all the day one deployment, so you can have the system up and running in your labs, in your data centers in 45 minutes, and you can have VMs up in provision very shortly after that. >> So what do you do with that kind of agility? >> Oh my gosh, so we've actually taken that, and we've taken the VxRail platform and we've created what we call Rackspace Services for VMware Cloud, and this is our platform that is based on VxRail, it's based on vCloud Director from VMware, and by having the VxRail is already RackStacked, ready to go for our customers, we're able to sign a customer up today, and then, within a matter of minutes, give them access to a vCloud Director portal where they can go in and spin up a new VM anytime they want, but then, it also integrates into all of those cloud management platforms and tools, right? It integrates into your Terraform, so you've got, you know, your full CI/CD pipeline, and so you have that full end-to-end capability. If you want to go click around on a portal, you can using vCloud Director and using vSphere and all that great stuff. If you want to automate it, you can do that too. And we do it all in the backs of that VxRail hyperconverged infrastructure. >> Talk about the DPU dynamic. We're hearing a lot about DPUs. VxRail, you guys have some HCI-like vibe there with DPUs. How is that impacting performance, can you guys see? 'Cause we're hearing a lot of buzz around the VxRail and the VMware DPUs really making things much faster. >> I mean, it's the thing we talk about most with customers now is their challenges with scaling their infrastructure, and VxRail is going to be the first and only jointly engineered system that will have vSphere 8 with DPUs functionality and will have the full life-cycle management, and what this really empowers customers to do is, as they're growing their environments that they're scaling out their workloads in the data center, they need a way to scale to that next generation of networking and network security, and that's what DPUs allow you to do. They give you that offload and that high performance capability. >> Talk about the... I'd love to get your guys' perspective, while we're just riffing on this real quick sidebar for a second, if VxRail has these capabilities which you guys are promoting it does and some of the things go on in the modern era, the next gen apps are going to look a lot different. We're kind of calling it supercloud, if you will, for lack of a better description. Yeah, multicloud is a state, I agree. It's a situation and a state, but supercloud is really the functionality of what cloud does. So what do you guys see as, maybe it's tea leaves reading now or dots connecting, what are some of those next gen apps? I mean the Edge is there with, "Oh, the Edge is going to explode," and I can see the Edge having new kinds of apps that we've never seen before, whether it's on premise building lights and however they work or IoT changing. What do you guys see as the next gen app/apps coming out that's not looking the same as now, or how are apps today changing for next gen? 'Cause you get more performance at the Edge, you get more action, you get more co-locations in GEOS, so it's clear multicloud multi-presence is happening too, right? So what are you guys seeing? What's this... >> Yeah, I would say two areas that resonate most with customers is customers transitioning to their cloud native journey, so beginning it and using things like Tanzu for Kubernetes Operations, which we fully support and have a white paper out there list for customers, another area is really in the AIML space, so we've been partnering with both VMware and Nvidia to simplify how customers deploy new AIML infrastructure. I mean, it's challenging, complex, a lot of customers are wanting to dive in because it really enables them to better operate and operate on insights and analytics they get from running their business. >> Josh? >> And, you know, I think it really comes down to, whether you want to call it Edge or IoT or, you know, smart things, whatever, right? It all comes down to how we are expected, now, to capture all of the data to create a better user experience, and that's what we're seeing the modern applications being built around, right, is how do you leverage all of the data that's now at your fingertips, whether it's from wearables, machine vision, whatever it may be, and drive that improved user experience. And so that's the apps that we're seeing now, right? You know, of course, you still have all your business apps, all your ERP capabilities that need to exist and all of that great stuff, but at the same time, I also expect that, whenever, you know, now, whenever I'm walking into a store and their machine vision picks me up and they're pinging my phone and pushing me push notifications, I expect to have a better user experience. >> And do a database search on you too, by the way. >> Yeah, exactly, right? >> No search warrants out for 'em, you know, you're good. >> That's exactly it, so, you know, you kind of expect that better user experience and that's where I'm seeing a lot of the new app development. >> Yeah, it's fun, as these cases are intoxicating to think about all the weird coolness around it. The thing that I want to get your thoughts on is, we were just talking on the analyst session earlier in theCUBE, if DevOps is here and won, which we believe it has and infrastructure as code is happening, the cloud native discussion, shifting left CI/CD pipeline, that's DevOps in my mind, that's like cloud native developers, that's like traditional IT in my mind, so that's all part of the coding. DataOps and Security Ops seem to be the most robust areas of conversations where that's the new Ops, right? So, I mean, I made the term up, but new Ops, in terms of the focus, what are you making more efficient? What are you optimizing for? What's your guys reaction to that? Because all the conversations that we talk about is data, security, and then the rest seems to be cool, all good on the developer's side. Yeah, shift left events happening up there, Kubernetes containers, but all the action on the Ops side seems to be data and security. >> Yeah. >> What's your reaction to that? Is that right? >> So personally, I do think that it's right. I think that, you know with great power comes great responsibility, right? And so the clouds have brought that to us, all of your infrastructure as code has brought that to us. We have that great power now, right? But then you start to see, kind of, the pipeline attacks that are starting to become more and more popular. And so how you secure something that is as complex as, you know, a cloud native development pipeline is really hard, it's really challenging, so I do think that it warrants the attention. Then on the data side, I think that that matters because when I talked about those examples of a better user experience, I don't want my better user experience tomorrow, I don't want it 20 minutes from now. I want that real time capability, and so with that comes massive requirements from a compute and hardware perspective, massive requirements from a software perspective, and from, you know, what folks are now calling DataOps perspective >> Data addressability, having the data available to be delivered in real time. >> You know, there there's been a lot of talk, here at the conference, about the disaggregation of, you know, the brainularism, if we're going to make up words, you know, the horsepower that's involved, CPU, DPU, GPU. I'll make up another word. We're familiar with the thermometers used during COVID to measure temperature. Pretend that I've invented a device called a Care-o-meter and I'm pointing at various people's foreheads, who needs to care about DPUs and GPUs and CPUs? You know, John was referencing the idea of security at the Edge, data. Well, wow, we've got GPUs that can do things. Who needs to care about that? Obviously, we care about it. You care about it. You care about it. You're building this stuff, you're deploying this stuff, but at what level in the customer stack do they need to care about it? Are you going in, is RackSpace engaging customers and saying, "Look, here's the value proposition: we understand your mission to be this. We believe we can achieve your mission." How far down in the organization do you go before you get to someone where you have to have the DPU conversation? 'Cause we didn't even define DPU yet here, which is always offensive to me. >> I think I defined it actually. >> Did you define DPU? Good. Thank you John. >> Yeah, yeah. >> But so who should care? Who should really care about that? >> Oh, that's such a complex question, right? Because everybody, Rackspace included >> But a good one. But a good question. >> Oh, it's a great question. >> Thank you. >> Great question. (laughing) >> Everybody, Rackspace included, is talking about selling business outcomes, right? And ultimately, that is what matters. It is what matters, is selling those business outcomes to the customer. And so of course we're dealing with our business buyers who are just looking for, "Hey, improve my KPIs, make this run faster, better, stronger, all of that great stuff," but ultimately you get down to an IT staff, and to the IT staff, these things matter because the IT staff, they all have budgets that they have to hit. The realities start to hit them and they can't just go and spend whatever they want, you know, trying to hit the KPIs of the marketing department or the finance department, right? And so you have your business buyers that do care significantly about buying their outcomes, and so we're having, you know, the business outcomes conversations with them and then, oftentimes, they will come back to us and say, "Okay, but now we need you to talk to this person over in our IT organization. We need you to talk with our CIO, with our VP of infrastructure," whatever that may be, where we really get down to the nuts and bolts and we talk about how, you know, we can stretch the hardware coming from Dell, we can stretch the software coming from VMware, and we can deliver a higher caliber experience, a lower TCO, by taking advantage of some of the new technologies coming out. >> Yeah, so there's a reason why I ask that awesome question, and it's because I can imagine a scenario where, and this speaks to RackSpace's position in the market today and moving forward and what your history has been, people want to know, "Well, why should I work with Rackspace instead of some mega-hyper-monster-cloud?" If part of the answer is: well, it's because, for very specific application environments, like healthcare we talked about earlier, that might be a conversation where you're actually bringing in Dell to have a conversation about how you are specifically optimizing hardware and software to achieve things that otherwise can't be achieved with t-shirt sizes of servers in a hyperscale cloud. I mean, is that part of the Rackspace value proposition moving forward, that you can do things like that with partners like Dell that the other folks aren't going to focus on? >> Absolutely, it is, right? And a lot of the power of Rackspace is that, you know, we're the best-in-class pure play cloud solutions provider, and we can talk to you about your AWS, your Azure, your GCP, all of that great stuff, but we can also talk to you about private cloud solutions that are built on the backs of Dell Technologies, and in this multicloud world, you don't have that one size fits all for every single application. There are some things that run great in a hyperscale provider, and we can help you get there, but just exactly like you said, there are these verticals where you have applications that don't necessarily run all that well or they're not modernized, they haven't been refactored to be able to take advantage of cloud native services. And if all you're going to do is run that on bare metal in VMs, a hosted private cloud is, by far, the best way to do that, right? And Rackspace provides that hosted private cloud on the backs of Dell technology, on the backs of VMware technology, and we can go deliver those custom bespoke solutions to customers. >> So the infrastructure and the hardware still matters, Ash, yes? >> Absolutely, and I think he just highlighted, while what he does with his customers and what's important to his internal organization is being to deliver faster outcomes, better outcomes, give those customers, to meet those KPIs of those customers consuming their infrastructure at Rackspace, so I think, really, what the DPU and the underlying infrastructure enables is all that full stack integration to allow them to quickly scale to the demands of those customers and what they need in their infrastructure. >> Guys, while we got you here, what do you think about this year's VMware Explore, a lot of anticipation around how many people are going to show up and, you know, all kinds of things around the new name and Broadcom. Big attendance here, I mean, I was very surprised about the size of the attendance and the show floor, the ecosystem, this train is not stopping. I mean, this is VMware's third act, no matter what the contextual situation is. What's your observation of the show? Do you agree, or is there anything that you could want to share about for folks who didn't make it, what they missed? >> Yeah, I mean it really highlights, I mean, you've seen the breadth of the show, I know people that aren't here that aren't able to see it are really missing the excitement. So there's a lot of great announcements around multicloud, around all the announcements, around the vSphere 8 with the DPUs, the vSAN Express Storage architecture, ton of new exciting technologies that are really empowering how customers, you know, the future of how customers are going to consume their workloads in their data centers. >> Josh, they're not short on products and stuff. A lot of moving parts. vSphere 8, a bunch of new stuff. And the cloud native stuff's looking pretty good too, off the tee. >> You know, it does feel like a focus on the core, though, in a way. So I don't think there's been a lot of peripheral noise at the show. Sometimes it's, you know, "And we got this, and this, and this, and this." It's vSphere 8, vSAN 8, cloud software, you know, really hammering it home and refining it. >> But you don't think of it as a little bit of a circus act. I mean the general keynote was theatrical, I thought, I mean, I thought they did a good job on that. I think vSphere 8 was buried a little bit, I thought they could have... They checked the box at the beginning. >> That's true, that's true. >> I mean, they mentioned it, but we didn't see the demos. You know? Demos are usually great. But that's my only criticism. >> Well, that's why we supplemented it with the VxRail announcements, right? With our big announcements around vSphere 8 and with the DPUs as well as the vSAN Express Storage architecture being integrated into VxRail, so I think, you know, it's always that ongoing partnership and, you know, doing what's best for our customers, showing them the next generation and how they consume that technology. >> Yeah, you guys got good props on VxRail. We had a great chat about it yesterday. Rackspace, you guys doing good? Quick update on what's happening with you guys. Give a quick plug. What's going on at Rackspace? What's hot? What's going on? Give a quick plug for what the services are and the products you got going on there. >> Yeah, absolutely. So we are that end-to-end cloud provider, right? And so we've got really exciting offers in market, helping customers take advantage of all the hyperscale providers, and then giving them that private cloud experience. We've got everything from single-tenant running in our data centers on the backs of vSphere, vCloud Director, and VxRails, all the way through to, like, multi-tenant burstable capability that runs within our own data centers as well. It's a really exciting time for technology, a really exciting time for Rackspace. >> Congratulations, we've been following your journey for a long time. Dell, you guys do continue to do a great job and end-to-end phenomenal work. The telco thing's a huge opportunity, we didn't even go there. But Ash, thanks. Josh, thanks for coming on. Appreciate it. >> Yeah, thanks so much. Thanks for having us. >> Thank you very much. >> Okay, thanks for watching theCUBE. We're live, day two of three days of wall-to-wall coverage. Two sets here in Moscone West on the ground level, in the lobby, checking out all the action. Stay with us for more coverage after this short break. (modern music)
SUMMARY :
to see and, you know, Yeah, thanks so much, Let's talk about that and the and the technologies Yeah, and I think to add to that, and the startup world, or are you multicloud as a situation? and you have the software that delivers data to Well, VxRail is the first and only infrastructure. All the storage Everything's in there. so it allows you to and you can have VMs up in provision and so you have that full and the VMware DPUs really and that's what DPUs allow you to do. and some of the things another area is really in the AIML space, And so that's the apps that on you too, by the way. 'em, you know, you're good. a lot of the new app development. the rest seems to be cool, And so the clouds have brought that to us, having the data available to How far down in the organization do you go Thank you John. But a good question. Great question. and we talk about how, you know, I mean, is that part of the and we can talk to you about and the underlying infrastructure enables to show up and, you know, around the vSphere 8 with the DPUs, And the cloud native stuff's like a focus on the core, I mean the general keynote but we didn't see the demos. VxRail, so I think, you know, and the products you got going on there. centers on the backs of Dell, you guys do Yeah, thanks so much. West on the ground level,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Josh | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Josh Prewitt | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Ash McCarty | PERSON | 0.99+ |
Ash | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Rackspace | ORGANIZATION | 0.99+ |
29 global data centers | QUANTITY | 0.99+ |
three days | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
12 years | QUANTITY | 0.99+ |
vCloud Director | TITLE | 0.99+ |
yesterday | DATE | 0.99+ |
45 minutes | QUANTITY | 0.99+ |
Rackspace Technology | ORGANIZATION | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
vCloud Director | TITLE | 0.99+ |
third act | QUANTITY | 0.99+ |
vSphere | TITLE | 0.99+ |
VxRail | TITLE | 0.99+ |
RackSpace | ORGANIZATION | 0.98+ |
vSphere 8 | TITLE | 0.98+ |
tomorrow | DATE | 0.98+ |
today | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
vSphere 8 | TITLE | 0.98+ |
telco cloud | ORGANIZATION | 0.97+ |
Moscone West | LOCATION | 0.97+ |
two areas | QUANTITY | 0.97+ |
both | QUANTITY | 0.96+ |
Tanzu | ORGANIZATION | 0.96+ |
vSAN 8 | TITLE | 0.96+ |
this year | DATE | 0.96+ |
telco | ORGANIZATION | 0.96+ |
single | QUANTITY | 0.96+ |
two great guests | QUANTITY | 0.96+ |
Multicloud Product Management | ORGANIZATION | 0.96+ |
theCUBE | ORGANIZATION | 0.95+ |
tens of thousands | QUANTITY | 0.95+ |
Kubernetes Operations | ORGANIZATION | 0.95+ |
Broadcom | ORGANIZATION | 0.95+ |
multicloud | ORGANIZATION | 0.93+ |
20 minutes | QUANTITY | 0.93+ |
RackStacked | TITLE | 0.93+ |
DevOps | TITLE | 0.91+ |
Hillary Ashton, Teradata | Amazon re:MARS
(upbeat music) >> And welcome back. I'm John Furrier, host of theCUBE. We're excited to welcome Teradata back to theCUBE and today with us at the ARIA is re:MARS conference coverage. It's great to hear with Hillary Ashton, Chief Product Officer of Teradata. Great to have you on. Thanks for coming on. >> John, thanks so much for having me. I'm super excited to be joining you today. >> So re:MARS, what a great event. It brings together the confluence of machine learning, which is data, automation, robotics, and space. Which is to me, is a whole new genre of conversations, around technology and business value. It is going to be a big kind of area. And it's just, again just getting started any one, as they say, and super excited. Tell us about what you guys are doing there and yourself. >> About two and a half years ago I head up the products organization. That means I have responsibility for our roadmap and our and our strategy overall on the product side. Prior to coming Teradata, gosh, I have spent the last 20 years, if I can say that, in the data and analytics space. I grew up in marketing application space, spent 11 years at SaaS, really cut my teeth on hardcore AI, ML and analytics at SaaS, and most recently was at PTC, where I was in charge of, I was a general manager of augmented reality, the business unit at PTC, focused on IOT data and how IOT data and augmented reality can really bring machines to life. >> It's interesting. You talked about SaaS and kind of your background, you know everything SaaSified with the cloud now. So you think about platform as a service, SaaS models emerging, software is an open source game now. So it's an integration cloud-scale data conversation we're seeing. What's your reaction to that? What's your reaction to that kind of idea that, okay, everything's open to source, software value integrating in with data. What's your reaction to that? >> Yeah, I mean, I think open source absolutely has some awesome things going on there. I think there's great opportunities for commercial, reliable, governed software and open source capabilities to come together in an open ecosystem that allow our customers to choose the best way to deliver the analytic outcomes that they're focused on. >> So you guys have been in the news lately around connecting multicloud data analytics platforms and transforming businesses around there, obviously, the background with Teradata is well documented. What's this news about? What's really going on there? You got Vantage platform. What's happening? Take us through that story. What's the key point? >> Yeah, we've worked super hard to deliver a true, multicloud, hybrid, data platform. So, if you think customers, many of our enterprise customers started with on-premises data systems and are moving violently to the cloud, right? So they're super excited about moving to the cloud but being able to deploy on multiple clouds, I think is important and then importantly, sort of this hybrid notion of being able to leverage data that's on-premises and combine it with data in the cloud on AWS, for example. And so being able to do those hybrid use cases you may have data that's like older and kind of archaic, needs to stay on-premises. There's not a lot of value in moving it to the cloud but you want to combine it with some of the innovative, analytic capabilities that perhaps you're doing on AWS. And so Teradata allows you to live in that hybrid multicloud environment and deliver analytic outcomes wherever your data is. >> Hillary, one of the top conversations is data cloud. You got to have a data cloud. I want to deal with this, move this around, but there's a lot of now integration opportunities to bring data from different sources together whether you're in healthcare, all the verticals have the same use case, multiple access to different databases, bringing them all together, ETL, all that old-school stuff is coming back in and being kind of refactored with machine learning, with cloud scale, with platforms like AWS, there's now this new commitment to bringing this to the next level for enterprises. And you mentioned some of those partnerships. What specifically is going on in the cloud that's notable, that's realistically that customers are executing on now? Not the hype, the reality. >> The reality. Yeah, absolutely. So I mean, I think today with Teradata our customers are leveraging something that we call a query fabric. And so this is the idea, as you said, John, that data might be in a lot of different places and you want to be able to get value out of that data without the difficulty of moving it around unnecessarily. Sometimes you want to move it around but unnecessary data movement is both expensive and an inefficient use of precious time. And so I think that there's an opportunity for this query fabric to be able to do remote push-down queries, wherever that data is and return back the results that you are looking for, analytic results, AI and ML results, combining different data that's in different locations to deliver that analytic outcome quickly without having to move the data around. So I would say query fabric is one of the areas that we are super invested in and, today, is delivering real value for our customers. >> It's really interesting. Data being addressable and available, low latency. I mean, we're talking about space, automation, robotics, real-time, so you have different data types stored in different data vehicles or mechanisms that need to be real-time and available. Because machine learning only works as good as the data they has available to it. So again, this is a key, kind of new way that folks are re-architecting. And again, we're here at, at re:MARS, right? I mean to machine learning automation, robotics and space, kind of the real world, physical, digital, trust, scale, huge concepts here. What's the partnership? How's it working with AWS? Take us through that strong partnership that you guys are developing. >> Yeah. I mean, we have a fantastic relationship with AWS. We're really excited that we signed a strategic collaboration agreement at the end of last year that really puts us in an elite category of AWS partners. We're really committed to co-investing and co-engineering with Amazon and our product development organization and also in go-to market and marketing and other parts of our business. As the Chief Product Officer, I'm really excited about three key areas. First is we've optimized Teradata Vantage to run in the AWS cloud at great scale, with unparalleled scale at the highest level for our customers. And so we've partnered with them to be able to handle some of the complex analytic workloads. And we think of analytic models are one part of a workload. There may be other ELT that you talked about, right? Workloads that you may need to run, all of that running at tremendous scale with AWS in the cloud. The second area is deep integration. So Teradata used to think that we were the ecosystem. We built everything soup connects end-to-end. Today, we live in a really exciting data and analytics space and we partner closely with CSPs like AWS, where we are deeply integrated. We have dozens of AWS native integrations in our AWS offer today. And that lets customers take advantage of AWS X3 for Cloud Lake, for example Amazon Kinesis for data ingestion and streaming and on and on. So we're really focused on the integration area there. And then finally, we've developed, co-developed with AWS, a fast and low risk migration approach to move from on-premises to the cloud for our enterprise customers. >> You know, what's interesting is as we kind of weave together, I hear you talking about those three areas. I mentioned earlier at the top of the interview, how integration is now the competitive advantage. Software is almost going commodity with open source because you mentioned that. All good, right? All good stuff. But when you think about kind of the big trends in this new computing world, it's hybrid cloud, it's edge, and IOT, okay? Again, cloud-scale and these new connected points, trust, access, all these things have to be integrated. So integration, you guys have been in the middle, Teradata has been around for a long time, leader in data warehousing, but now with cloud and in the data types, this is a game changer. I mean, this is notable. Can you share more about how you see this evolving with customers because at the end of the day the integration becomes super critical. >> Yeah, absolutely. And I'm super passionate about the opportunities of IOT streaming data. And that's one of the key areas of partnership with Amazon is taking that streaming data, leveraging the analytic opportunities with Amazon. We'll talk about that in just a second, but I think some of the examples that I could share with you, everyone loves to hear, I love to hear, about what actual customers are doing. So Brinker International, they're one of the world's largest casual dining restaurants. If you've ever been to a Chili's Grill or Maggiano's Little Italy those are the guys, Brinker International owns those brands. So we leveraged Amazon SageMaker and Teradata Vantage together to apply advanced analytic and predictive modeling to be able to understand things like demand. And you're in the middle of COVID and trying to understand how many people should you have on staff today? What is the demand going to look like? What should sales look like? What's foot traffic look like? So that demand forecasting capability across their 1,600 different store fronts or restaurant fronts is one of the examples that I could share with you. The other one is Hertz. So one of the world's largest vehicle rental companies. They are using Vantage and AWS together to track and analyze transaction data across all of its global locations and manage again that complex inventory. And some of that is streaming data, some of that is data that we're getting from the cars themselves, and then create a new value-added program to their loyalty members which is sort of the name of the game. Is customer acquisition and extension of brand across those customers. So those are two examples I can share with you. There's many, many others but I know you probably had some other questions. >> Yeah. I want to come back to the SageMaker thing. I think that's important partnership there because it's been one of the fastest growing services. It's always at the top or in the top two or three whenever I talk to Andy Jassy and the team over there. But I want to talk about scalability and I want to ask you, if you can scope for me the scalability of what's going on with this data challenging, 'cause where are we on that scale? Can you share how you would scope the scale? >> Absolutely. And I love talking about scale because it is a home run for Teradata. I think many customers start looking at the cloud and they start with kind of a little tiny baby footprints but we are an enterprise solution, an enterprise platform. And so I think that we're looking at tens of thousands of users and thousands of business critical applications. That's what our customers are doing and have done for decades with Teradata and bringing all of that scale to the cloud. And with AWS in particular, we recently did 1,000 node testing. I'm going to walk through this a little bit slowly, which is hard for me, as you can tell, but it was a single system of more than 1,000 nodes which is just to give you a sense, that's double our largest on-premises system. So it's huge. It was the single largest system. >> John: Double is your largest customer deployment? >> Double our largest customer deployment on-premises. Yeah, that's right. So it was 1,000 nodes with more than 1,000 different users submitting thousands of concurrent queries. So huge enterprise scale. And this was a real-world use case. We took not a traditional benchmark but a real world customer set of mixed workloads. So lots of long running strategic queries and lots of fast running queries that needed really tight SLAs. All of that running simultaneously. We saw no system down times, we were able to roll out and roll back new capabilities seamlessly in a true software as a service fashion. So that was an awesome test all run on AWS. And I think that their team was just as excited as we were about it. >> Well, I love the scale. I love that test you guys ran. I see you're sponsoring re:MARS which is great, congratulations. We love covering since the beginning, we believe of kind of a whole new genre of programming brings together the confluence of exciting technologies that just a decade ago weren't always working together. They were bespoke. >> That's right. Yeah. >> So now it's all integrated in at cloud scale, you got the test, got thousands of concurrents queries. What else are you showcasing? You mentioned the SageMaker because that's really where Amazon's connecting all these tools. How are you integrating in? It sounds like you're bringing all that Amazon goodness in with Teradata and vice versa. >> Absolutely. We're delivering sort of the best in class to our customers jointly. So here re:MARS today, we're really excited to be talking about SageMaker and our relationship with AWS to be able to deliver that seamless integration between our solutions for machine learning services and Teradata Vantage. So I'm sure it won't come as any surprise to you as we just talked about, but we're finding that massive investments in AI and ML and other advanced analytic capabilities are out there, and many organizations are really only experimenting. They're just starting to explore some of these opportunities. We think that there's tremendous value in this scale that we just talked about, that we can offer, combined with best in class AI and ML capabilities like SageMaker. And so we are excited to talk about it. If you want to see it, we've got a booth set up, you can come and take a look at what we're doing there but I think there's huge opportunities for customers to get to the analytic value with Teradata Vantage and AWS SageMaker. >> Yeah, it's great to see Teradata seeing that headroom opportunity to extend the value proposition to kind of new territory with your customers. I can definitely see it. Love the connection here. Where can they learn more about the Teradata partnership with AWS and Amazon? Is there a site? Is there a program coming? Is there any more content that they can be expecting to see? Take a little plug time to plug the company. >> If you insist, I will, John. Thank you. I think, if you're at the event right now, you can swing by Teradata's booth. We're at booth 111. You can get a demo of our SageMaker integration and learn more about both our enterprise scale and the advanced outcomes that we're able to provide to our customers. If you're not at re:MARS and we really think you should be, we would encourage you to sign up for one of our upcoming SageMaker webinars that we're doing with AWS this year. And if you'd like to, you can also just email us at aws@teradata.com. Again, that's aws@teradata.com and we'll set up a private demo for you. >> Well, Hillary Ashton, great to have you on. Chief Product Officer, Teradata, you must be feeling good. You got a lot to work with. You've got an install base. You have new territory to take down. As the Chief Product Officer, you got the keys to the kingdom. Give us a quick bumper sticker of where you guys are going with the product. >> We are fast and furious. My team will tell you, we are so excited to be here with AWS and Teradata is on an epic trajectory forward in our cloud first approach, so we are so excited about our roadmap. If you'd like to learn more, please swing by teradata.com. >> Lot of innovation happening. Thanks for coming on theCUBE. Okay, this is theCUBE coverage of Amazon re:MARS machine learning, automation, robotics, and space. It cuts the confluence of digital, virtual data and real-world and space. You can't get any more than this. That's a big edge out there in space. Talk about edge computing and space. Of course, theCUBE's here covering it. I'm John Furrier, your host. Stay with us for more coverage here at Amazon re:MARS. (upbeat music)
SUMMARY :
Great to have you on. I'm super excited to be joining you today. It is going to be a big kind of area. I have spent the last 20 So you think about platform as a service, to choose the best way to obviously, the background with of being able to leverage and being kind of refactored for this query fabric to be able to do or mechanisms that need to and we partner closely with CSPs like AWS, and in the data types, What is the demand going to look like? and the team over there. that scale to the cloud. All of that running simultaneously. love that test you guys ran. That's right. You mentioned the SageMaker as any surprise to you to extend the value proposition that we're doing with AWS this year. great to have you on. so excited to be here with AWS It cuts the confluence
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Hillary Ashton | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
Teradata | ORGANIZATION | 0.99+ |
Brinker International | ORGANIZATION | 0.99+ |
11 years | QUANTITY | 0.99+ |
Hillary | PERSON | 0.99+ |
aws@teradata.com | OTHER | 0.99+ |
today | DATE | 0.99+ |
First | QUANTITY | 0.99+ |
two examples | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
Hertz | ORGANIZATION | 0.99+ |
teradata.com | OTHER | 0.99+ |
Double | QUANTITY | 0.99+ |
more than 1,000 different users | QUANTITY | 0.99+ |
Chili's Grill | ORGANIZATION | 0.99+ |
Today | DATE | 0.99+ |
one | QUANTITY | 0.99+ |
second area | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
thousands | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
single | QUANTITY | 0.97+ |
About two and a half years ago | DATE | 0.97+ |
single system | QUANTITY | 0.96+ |
PTC | ORGANIZATION | 0.96+ |
Vantage | ORGANIZATION | 0.96+ |
double | QUANTITY | 0.95+ |
thousands of business | QUANTITY | 0.95+ |
SageMaker | TITLE | 0.94+ |
three areas | QUANTITY | 0.93+ |
a decade ago | DATE | 0.93+ |
first approach | QUANTITY | 0.93+ |
Teradata Vantage | ORGANIZATION | 0.92+ |
1,000 nodes | QUANTITY | 0.92+ |
Regina Manfredi, Teradata | Amazon re:MARS 2022
(light techno music) >> Okay, welcome back, everyone from theCUBE's coverage of AWS re:Mars here in Las Vegas. Back in person, I'm John Furrier, host of theCUBE. Re:MARS stands or Machine learning, Automation, Robotics, and Space. And we're covering all the action two days, day two. And we're here with Regina Manfredi, who's the VP of global CSPs, Cloud Service Providers Alliances with Teradata. Great to see you. Cloud service providers or- >> Cloud services providers, the hyperscalers. >> Hyperscalers, the big guys. All the CapEx, Amazon. >> Yes. >> The big guys. >> Indeed, thanks for having me. >> Yeah, Thanks for coming on. So tell about your role. So alliances, you're here with AWS. What's the role with AWS and Teradata? >> So AWS and Teradata have recently entered into a strategic collaboration agreement where we're really focused on building solutions together, leveraging AWS services, as well as Teradata's outstanding architecture, as it relates to the data analytics platform that we provide for our customers in the cloud today. And we're really trying to drive better outcomes for data scientists, business analysts, etc. >> You know, just recently, did a CUBE conversation with Teradata, and I was really surprised to find, not shocked, but kind of surprised, the scale of the computation that's going on in some of the cloud things you're doing. And you have the legacy on-premises data warehouse traditional business as well. >> Regina: We do. >> And there's a huge shift going on. A lot of the kind of upstarts, "Oh, data warehouse, old school. Data warehouse, it's antiquated, old," but that's not true. You guys have a lot of cloud action. >> We do, we have substantial cloud action that's occurring with our customers today. We actually just released earlier this year an announcement around 1,000 node tests in the cloud together with AWS, and had success, no downtime, no failures at all. And so we're pretty proud about that, and excited about what that's going to hold for our customers who need that level of scale. >> Well, Regina, I got to tell you, I have a little bit of a confession here. I'm a cloud data nerd by my training. And, you know, I've always watched all the different kind of levels of transformation with the industry, and you know, this is going to change that, that's going to kill that. Everything's going to be killed and then it never dies, but it just changes. Even today, SQL is still like the prominent language, it's never going to, in fact it's amplified further because that's what people like. So that just proves that things don't always get replaced. And so I wanted to ask you this because as we're here at this event at re:MARS, you have space, you have all these ambitious positive goals, and they just need to do some machine learning. They need some cloud, they need some, they need to have the solutions. >> Regina: Yes. They're not going to like get in the weed and say, "Oh, this is a better Hadoop cluster than this Kubernetes cluster. So it's not about sometimes the tech, it's about the solution. >> It is, and one of the things that was interesting for us in our session earlier this week was the fact that we had so many customers approach us after that session and say, "I just need help preparing my data. Running my models, training my models, and making sure that they run and can be deployed. And I don't want to move all this data all the time and have all this failure rate that I'm experiencing." And so it was very basic requirements and needs as people begin into their journey on AI/ML for their business. And so it was reaffirming that we're on the right track and driving the right tools for them. I want to get your perspective on what you're thinking about the show, but first, I want to ask this since you brought that up. Swami was on stage and he said, "You can spend your entire time and your career just trying to figure out what's going on, machine learning." >> Regina: Yup. >> "Which open source framework's going to be better than the other one." I mean, it's just a lot of work to even figure it out. We just had the Fiddler's AI CEO on who worked out all the hyperscalers, say Facebook tend to, you know, real, you know, super alpha geek, if you will. And he was saying, and we were talking about open source, free software, integrations are a big part of where cloud scale, and the value is being captured for companies and people who are doing projects. Integrating some managed services, so this is where I see you, guys, going right now with Teradata, having all these cloud services built on the install base. >> Right. Which is not, doesn't hurt that at all. It just only helps it as they would migrate to cloud, its integrations, so you take a little bit of Amazon here, a little bit of Teradata there. >> Regina: Absolutely. >> What's your perspective, what's your reaction to that? >> So, I agree. And we think that's part of our secret sauce. You know, what we want to have is a data analytics platform in the cloud that allows data scientists, and architects, etc., to bring their own tools. So whatever they're utilizing today, we want them to be able to utilize it in vantage, and make sure that, A, can drive some efficiencies, and also, some better, smarter economics, as it relates to their particular projects. And so I agree with you 100% , and would tell you that we view that as somewhat our competitive advantage. It's not about being all proprietary. We want those integrations, and we've got dozens of them with AWS, and- >> Can you give example, can you give a couple examples of some integrations that highlight that? >> Sure, so right now we've got an integration with SageMaker today that allows our customers or data scientists to come in, prepare the data, and actually leverage SageMaker to build and train the models, and then deploy very quickly and easily without having to do all the data movement within their architecture. >> It's just so fascinating. I can't wait to have more conversation with you guys about this because I just think the world's spinning in a direction where, with low code, no code, >> Regina: Yup. >> you can see code, companion whisperer, that they have CodeWhisperer they launched today, they're writing subroutines for machine learning. And so it's not autocomplete, it's subroutine. So you're seeing all these advances on the technology. So it comes back to the building blocks, the integration. It just seems like going to be like a plug and play. That's old, were all, are old words. Mix and match, plug and play, interoperability, were old words, like, in the old days. Now they're becoming more relevant. What's your take on all that? >> Yeah, I would agree. I don't think that we should be competing against the algorithms, and neither do we. We want to just actually build out the toolsets that drive the enablement based on what a customer's requirements and needs are, and based on what the investments that they've already made within their own enterprises. >> You know, what's interesting about this event, I love to get your reaction to what re:MARS means to you because it's machine learning, automation, robotics, and space. Not your typical tech conference. >> Regina: No. >> Okay, little bit of a mixed bag there, so to speak. I love it. I think it's like super alpha geek, very nerdy, super nerds are here. And the topics kind of reflect the future. For the people that are watching that aren't here, what's your vibe on the show? What's your takeaway? How would you explain what's going on here from a market perspective, from a vibe perspective, what's happening? >> This is my first re:MARS actually, and I would have to tell you that I feel like it just, general observation, a few things, one, the conversations are more meaningful and we're getting into the meat of what a data scientist truly needs in order to be successful in their role and help drive their enterprise. That's number one. So I think, to your point, we're all kind of geeking out together here. The other thing that I think is pretty exciting is the amount of use cases, and ways in which we are driving impact. AWS and Teradata driving impact for the business analysts in the enterprise environment, but also for the people, their customers. That's pretty exciting to see. >> You know, it's interesting. When I first, was kind of like thinking about the show and what I was going to expect, it kind of overexceeded my expectations in the sense of what I was thinking about IOT, industrial, and digital innovation. 'Cause that's going to scale. I think now we're at a tipping point with machine learning that the industrial, IOT markets is going to explode 'cause machine learning's ready. But there was a whole positive, save the earth angle >> Regina: Yes. >> that caught my attention. >> Regina: Yes. You know, the discoveries from space are going to potentially have impact for the good, not just a cliche some sustainability messaging. It was actually real. >> Right, I think that that's exciting in an area in which we're excited to explore. We're doing a lot of work behind the scenes around sustainability and ESG initiatives for our customers, but also for the greater good. It's about driving outcomes for the greater good and being responsible with how we approach that. You know, the other thing I noticed too from a robotics standpoint, given I live in California, is a huge robotics culture there, you know. It's like bigger than football and baseball, and some sports. They provide A and B team and people get cut from the B team. There's so much demand to be on the robotics team. It's not a club, it's a team. >> Regina: Right. And so, you look at what's going on robotics, it's so exciting in the sense that if you're young and you're into tech, this is like- >> Regina: This is the place to be. >> I mean, why wouldn't you be hanging out here? >> Yeah, well, and I visited the booth over at University of Michigan, and how they're driving robotics to help support the human body to go further distances, and to drive better performance and health for individuals, and was really impressed with the work that they're doing, and even saw a use case and a need where I thought, you know, I have a quadriplegic sister-in-law, who I thought, "Wow, someday, maybe she'll be upright and walking again." >> John: Yeah. >> And those were exciting conversations to have while I was here. >> The advances on the material management robots I think is fascinating to see that growth. Well, let's get back to Teradata real quick to kind of close out future of what's next. Obviously, a lot of migration to the cloud happening. What's the outlook on the landscape and where do you see it evolving? Because you're seeing what the hyperscalers are doing, the cloud service providers, they're providing the CapEx. In fact, we coined the term supercloud, last re:Invent, that's become a thing. And Charles Fitzgerald would think it's not a thing, he debates us online all the time on Twitter. But it's, you can build on top of a CapEx. >> Regina: Yup. >> They did all the heavy lifting. You know, Snowflake, Databricks, the list goes on and on. So building on top of that to build proprietary advantages or even just sustainable advantages is now easier to do. So superclouds are kind of in play. So that means whoever's got the playbook can win. So you guys seem to be executing that playbook of having the installed base, and then working with AWS >> Regina: Yes. >> to ride that wave. Tell us about the migration strategies you're seeing, and what are your customers doing specifically, and take us through a customer that's leaning into the cloud and driving. >> So when I think about specific customers that are leaning in, you know, the first and most important thing that we're hearing is, you've got to be able to scale. I've got 1,000 nodes or 100 nodes, or whatnot. And so we're addressing that because we think that there's a place for hybrid cloud. We think everyone's moving and rushing towards the cloud, but even one of our competitors last week announced that there's a place for on-prem, and we would agree. >> John: Yeah. >> So that is something that we're really focused on, and you take, for example, the automotive industry. We're seeing a lot of work being done together with our joint customers, AWS and Teradata, and some of these auto manufacturers who are experiencing supply chain issues and challenges today, and also need to drive better quality control measures within their own lines, in the manufacturing lines. And so we're working together with them to look at what type of machine learning and AI can we be leveraging together as part of the overall solution to drive those analytics, and make sure that they have better quality control >> You know, that's really good insight about the on-premise thing. And I think that supports what we're seeing around hybrid. We see hybrid as a steady state going forward, period. >> Regina: Yeah. >> And that will evolve into multi thing. Multi-cloud, you want to call it, or superclouds, and more things. Basically, distributed computing. So if you look at the edge here, the edge is just on-premise. What is the premise? It's an edge or big device, small device, data center is a large edge. >> Regina: Right. >> And so if you're using cloud hybrid, the distinction kind of goes away. And I think this is where we'll going to see the winners emerge in data. Because remember, you go back to 2010, Hadoop was the big thing, big data. And that kind of crashed and burned. And then now you're seeing Databricks picking up a lot of that. Snowflake, you guys are there. And so it's still going on, this transformation in data. >> Regina: It is. And I think hybrid's a huge deal. What are customers saying around that? Because I think they're just trying to figure out cloud scale. >> I think they're trying to figure out cloud scale, I think they're also trying to figure out security. And so, you know, when we're talking to our customers, that absolutely is critical. And I would also suggest that the customer base is really looking for, "Hey, don't just help me migrate, I really need to modernize." And so driving the right use cases for the customer is important. >> You know, another thing that you, guys, have a lot of core expertise in is governance. And we've seen how that has played in all the compliance, and all these conversations are kind of converging. Do you have closed, do you have open? Machine learning needs more data, dow do you protect it? So that set a hot area that I see as well. And that's something that's emerging, 'cause cyber's also involved too, like, you have cyber security threats on code, so I'm curious to see how that turns out. What's your perspective on, what's Teradata's perspective on the security, open, closed perspective? Any- >> It's a priority for, security is a priority for us. And I don't think that we've officially made that determination yet, right? We're still exploring, and we're going to do whatever our customers require of us. In terms of an open, closed perspective, I think we want to be flexible. Again, like I said before, it's about being open and supportive of whatever the customer requirement is especially across the different industries. >> Well, Regina, great to have you on theCUBE. Thanks for coming. I really appreciate it. Great insight, great to catch up on Teradata, cloud play. Very strong move. I think it's a good one. Final question I want to ask you though, is a little bit more about the personnel in the industry, like, obviously, if you're young, you're seeing all this space here, machine learning's not obvious. I know schools now are training it, but you start to see new personas come into the workforce. Where are the gaps? I mean, obviously, we have a lot of new opportunities, like, cybersecurity has a lot of job openings. Is there any observations that you have around or advice to younger folks coming in, from a career standpoint? Because a lot of job openings are skills that weren't even taught in school. >> Regina: Right, that's- >> You know. >> And then you got the women in check, and you have all kinds of opportunities now that aren't just engineering, right? >> Regina: Yes. >> It's not just engineering. It's computer science, so there's a whole in-migration of new talent coming in the industry. >> Yes, I think maintaining a curious mind is really critical, and taking time to invest in learning. You know, there are so many resources available to us at our disposal that that don't cost us a dime. And so my advice to anybody who is curious, remain curious, dig in, and get some experience, and don't be afraid to stick your neck out, and try it. >> Well, in this conference we have robots welcome, you know, in this out there. >> Yeah. (laughs) >> Regina, thanks for coming out here. Really appreciate it >> John, thank you, it's a pleasure. >> CUBE coverage here in Las Vegas for Amazon re:MARS. I'm John Furrier, your host. Stay with more live coverage after this short break. (upbeat bright music)
SUMMARY :
And we're here with Regina Manfredi, providers, the hyperscalers. Hyperscalers, the big guys. What's the role with AWS and Teradata? customers in the cloud today. in some of the cloud things you're doing. A lot of the kind of upstarts, in the cloud together with AWS, and they just need to do So it's not about sometimes the tech, and driving the right tools for them. and the value is being captured so you take a little bit of Amazon here, And so I agree with you 100% , prepare the data, with you guys about this advances on the technology. that drive the enablement to what re:MARS means to you And the topics kind of reflect the future. but also for the people, their customers. in the sense of what I You know, the discoveries from space You know, the other thing I noticed too it's so exciting in the and to drive better performance And those I think is fascinating to see that growth. of having the installed base, that's leaning into the cloud and driving. and we would agree. and also need to drive better And I think that supports what What is the premise? And I think this is where And I think hybrid's a huge deal. And so driving the right use cases in all the compliance, And I don't think that to have you on theCUBE. coming in the industry. and don't be afraid to we have robots welcome, you Really appreciate it I'm John Furrier, your host.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Regina Manfredi | PERSON | 0.99+ |
California | LOCATION | 0.99+ |
Regina | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
100% | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
Charles Fitzgerald | PERSON | 0.99+ |
Teradata | ORGANIZATION | 0.99+ |
last week | DATE | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Las Vegas | LOCATION | 0.99+ |
first | QUANTITY | 0.99+ |
2010 | DATE | 0.99+ |
today | DATE | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
Swami | PERSON | 0.98+ |
1,000 nodes | QUANTITY | 0.98+ |
two days | QUANTITY | 0.98+ |
CapEx | ORGANIZATION | 0.98+ |
earlier this year | DATE | 0.98+ |
ORGANIZATION | 0.97+ | |
100 nodes | QUANTITY | 0.97+ |
earlier this week | DATE | 0.96+ |
SageMaker | TITLE | 0.91+ |
day two | QUANTITY | 0.88+ |
Fiddler | ORGANIZATION | 0.87+ |
around 1,000 node tests | QUANTITY | 0.86+ |
dozens | QUANTITY | 0.84+ |
SQL | TITLE | 0.8+ |
MARS | TITLE | 0.78+ |
earth | LOCATION | 0.77+ |
ESG | ORGANIZATION | 0.74+ |
Michigan | LOCATION | 0.69+ |
Ian Massingham, MongoDB and Robbie Belson, Verizon | MongoDB World 2022
>>Welcome back to NYC the Cube's coverage of Mongo DB 2022, a few thousand people here at least bigger than many people, perhaps expected, and a lot of buzz going on and we're gonna talk devs. I'm really excited to welcome back. Robbie Bellson who's the developer relations lead at Verizon and Ian Massingham. Who's the vice president of developer relations at Mongo DB Jens. Good to see you. Great >>To be here. >>Thanks having you. So Robbie, we just met a few weeks ago at the, the red hat summit in Boston and was blown away by what Verizon is doing in, in developer land. And of course, Ian, you know, Mongo it's rayon Detra is, is developers start there? Why is Mongo so developer friendly from your perspective? >>Well, it's been the ethos of MongoDB since day one. You know, back when we launched the first version of MongoDB back in 2009, we've always been about making developers lives easier. And then in 2016, we announced and released MongoDB Atlas, which is our cloud managed service for MongoDB, you know, starting with a small number of regions built on top of AWS and about 2,500 adoption events per week for MongoDB Atlas. After the first year today, MongoDB Atlas provides a managed service for MongoDB developers around the world. We're present in almost a hundred cloud regions across S DCP and Azure. And that adoption number is now running at about 25,000 developers a week. So, you know, the proof are in proof is really in the metrics. MongoDB is an incredibly popular platform for developers that wanna build data-centric applications. You just can't argue with the metrics really, >>You know, Ravi, sometimes there's an analyst who come up with these theories and one of the theories I've been spouting for a long time is that developers are gonna win the edge. And now to, to see you at Verizon building out this developer community was really exciting to me. So explain how you got this started with this journey. >>Absolutely. As you think about Verizon 5g edge or mobile edge computing portfolio, we knew from the start that developers would play a central role and not only consuming the service, but shaping the roadmap for what it means to build a 5g future. And so we started this journey back in late 20, 19 and fast forward to about a year ago with Mongo, we realized, well, wait a minute, you look at the core service offerings available at the edge. We didn't know really what to do with data. We wanted to figure it out. We wanted the vote of confidence from developers. So there I was in an apartment in Colorado racing, your open source Mongo against that in the region edge versus region, what would you see? And we saw tremendous performance improvements. It was so much faster. It's more than 40% faster for thousands and thousands of rights. And we said, well, wait a minute. There's something here. So what often starts is an organic developer, led intuition or hypothesis can really expand to a much broader go to market motion that really brings in the enterprise. And that's been our strategy from day one. Well, >>It's interesting. You talk about the performance. I, I just got off of a session talking about benchmarks in the financial services industry, you know, amazing numbers. And that's one of the hallmarks of, of Mongo is it can play in a lot of different places. So you guys both have developer relations in your title. Is that how you met some formal developer relations? >>We were a >>Program. >>Yeah, I would say that Verizon is one of the few customers that we also collaborate with on a developer relations effort. You know, it's in our mutual best interest to try to drive MongoDB consumption amongst developers using Verizon's 5g edge network and their platform. So of course we work together to help, to increase awareness of MongoDB amongst mobile developers that want to use that kind of technology. >>But so what's your story on this? >>I mean, as I, as I mentioned, everything starts with an organic developer discovery. It all started. I just cold messaged a developer advocate on Twitter and here we are at MongoDB world. It's amazing how things turn out. But one of the things that's really resonated with me as I was speaking with one of, one of your leads within your organization, they were mentioning that as Mongo DVIA developed over the years, the mantra really became, we wanna make software development easy. Yep. And that really stuck with me because from a network perspective, we wanna make networking easy. Developers are not gonna care about the internals of 5g network. In fact, they want us to abstract away those complexities so that they can focus on building their apps. So what better co-innovation opportunity than taking MongoDB, making software easy, and we make the network easy. >>So how do you think about the edge? How does you know variety? I mean, to me, you know, there's a lot of edge use cases, you know, think about the home Depot or lows. Okay, great. I can put like a little mini data center in there. That's cool. That's that's edge. Like, but when I think of Verizon, I mean, you got cell towers, you've got the far edge. How do you think about edge Robbie? >>Well, the edge is a, I believe a very ambiguous term by design. The edge is the device, the mobile device, an IOT device, right? It could be the radio towers that you mentioned. It could be in the Metro edge. The CDN, no one edge is better than the other. They're all just serving different use cases. So when we talk about the edge, we're focused on the mobile edge, which we believe is most conducive to B2B applications, a fleet of IOT devices that you can control a manufacturing plant, a fleet of ground and aerial robotics. And in doing so you can create a powerful compute mesh where you could have a private network and private mobile edge computing by way of say an AWS outpost and then public mobile edge computing by way of AWS wavelength. And why keep them separate. You could have a single compute mesh even with MongoDB. And this is something that we've been exploring. You can extend Atlas, take a cluster, leave it in the region and then use realm the mobile portfolio and spread it all across the edge. So you're creating that unified compute and data mesh together. >>So you're describing what we've been expecting is a new architecture emerging, and that's gonna probably bring new economics of new use cases, right? Where are we today in that first of all, is that a reasonable premise that this is a sort of a new architecture that's being built out and where are we in that build out? How, how do you think about the, the future of >>That? Absolutely. It's definitely early days. I think we're still trying to figure it out, but the architecture is definitely changing the idea to rip out a mobile device that was initially built and envisioned for the device and only for the device and say, well, wait a minute. Why can't it live at the edge? And ultimately become multi-tenant if that's the data volume that may be produced to each of those edge zones with hypothesis that was validated by developers that we continue to build out, but we recognize that we can't, we can't get that static. We gotta keep evolving. So one of our newest ideas as we think about, well, wait a minute, how can Mongo play in the 5g future? We started to get really clever with our 5g network APIs. And I, I think we talked about this briefly last time, 5g, programmability and network APIs have been talked about for a while, but developers haven't had a chance to really use them and our edge discovery service answering the question in this case of which database is the closest database, doesn't have to be invoked by the device anymore. You can take a thin client model and invoke it from the cloud using Atlas functions. So we're constantly permuting across the entire portfolio edge or otherwise for what it means to build at the edge. We've seen such tremendous results. >>So how does Mongo think about the edge and, and, and playing, you know, we've been wondering, okay, which database is actually gonna be positioned best for the edge? >>Well, I think if you've got an ultra low latency access network using data technology, that adds latency is probably not a great idea. So MongoDB since the very formative years of the company and product has been built with performance and scalability in mind, including things like in memory storage for the storage engine that we run as well. So really trying to match the performance characteristics of the data infrastructure with the evolution in the mobile network, I think is really fundamentally important. And that first principles build of MongoDB with performance and scalability in mind is actually really important here. >>So was that a lighter weight instance of, of Mongo or not >>Necessarily? No, not necessarily. No, no, not necessarily. We do have edge cashing with realm, the mobile databases Robbie's already mentioned, but the core database is designed from day one with those performance and scalability characteristics in mind, >>I've been playing around with this. This is kind of a, I get a lot of heat for this term, but super cloud. So super cloud, you might have data on Preem. You might have data in various clouds. You're gonna have data out at the edge. And, and you've got an abstraction that allows a developer to, to, to tap services without necessarily if, if he or she wants to go deep into the S great, but then there's a higher level of services that they can actually build for their customers. So is that a technical reality from a developer standpoint, in your view, >>We support that with the Mongo DB multi-cloud deployment model. So you can place Mongo DB, Atlas nodes in any one of the three hyperscalers that we mentioned, AWS, GCP or Azure, and you can distribute your data across nodes within a cluster that is spread across different cloud providers. So that kinds of an kind of answers the question about how you do data placement inside the MongoDB clustered environment that you run across the different providers. And then for the abstraction layer. When you say that I hear, you know, drivers ODMs the other intermediary software components that we provide to make developers more productive in manipulating data in MongoDB. This is one of the most interesting things about the technology. We're not forcing developers to learn a different dialect or language in order to interact with MongoDB. We meet them where they are by providing idiomatic interfaces to MongoDB in JavaScript in C sharp, in Python, in rust, in that in fact in 12 different pro programming languages that we support as a first party plus additional community contributed programming languages that the community have created drivers for ODMs for. So there's really that model that you've described in hypothesis exist in reality, using >>Those different Compli. It's not just a series of siloed instances in, >>In different it's the, it's the fabric essentially. Yeah. >>What, what does the Verizon developer look like? Where does that individual come from? We talked about this a little bit a few weeks ago, but I wonder if you could describe it. >>Absolutely. My view is that the Verizon or just mobile edge ecosystem in general for developers are present at this very conference. They're everywhere. They're building apps. And as Ian mentioned, those idiomatic interfaces, we need to take our network APIs, take the infrastructure that's being exposed and make sure that it's leveraging languages, frameworks, automation, tools, the likes of Terraform and beyond. We wanna meet developers where they are and build tools that are easy for them to use. And so you had talked about the super cloud. I often call it the cloud continuum. So we, we took it P abstraction by abstraction. We started with, will it work in one edge? Will it work in multiple edges, public and private? Will it work in all of the edges for a given region, public or private, will it work in multiple regions? Could it work in multi clouds? We've taken it piece by piece by piece and in doing so abstracting way, the complexity of the network, meaning developers, where they are providing those idiomatic interfaces to interact with our API. So think the edge discovery, but not in a silo within Atlas functions. So the way that we're able to converge portfolios, using tools that dev developers already use know and love just makes it that much easier. Do, >>Do you feel like I like the cloud continuum cause that's really what it is. The super cloud does the security model, how does the security model evolve with that? >>At least in the context of the mobile edge, the attack surface is a lot smaller because it's only for mobile traffic not to say that there couldn't be various configuration and human error that could be entertained by a given application experience, but it is a much more secure and also reliable environment from a failure domain perspective, there's more edge zones. So it's less conducive to a regionwide failure because there's so many more availability zones. And that goes hand in hand with security. Mm. >>Thoughts on security from your perspective, I mean, you added, you've made some announcements this week, the, the, the encryption component that you guys announced. >>Yeah. We, we issued a press release this morning about a capability called queryable encryption, which actually as we record this Mark Porter, our CTO is talking about in his keynote, and this is really the next generation of security for data stored within databases. So the trade off within field level encryption within databases has always been very hard, very, very rigid. Either you have keys stored within your database, which means that your memory, so your data is decrypted while it's resident in memory on your database engine. This allow, of course, allows you to perform query operations on that data. Or you have keys that are managed and stored in the client, which means the data is permanently OBS from the engine. And therefore you can't offload query capabilities to your data platform. You've gotta do everything in the client. So if you want 10 records, but you've got a million encrypted records, you have to pull a million encrypted records to the client, decrypt them all and see performance hit in there. Big performance hit what we've got with queryable encryption, which we announced today is the ability to keep data encrypted in memory in the engine, in the database, in the data platform, issue queries from the client, but use a technology called structural encryption to allow the database engine, to make decisions, operate queries, and find data without ever being able to see it without it ever being decrypted in the memory of the engine. So it's groundbreaking technology based on research in the field of structured encryption with a first commercial database provided to bring this to market. >>So how does the mobile edge developer think about that? I mean, you hear a lot about shifting left and not bolting on security. I mean, is this, is this an example of that? >>It certainly could be, but I think the mobile edge developer still stuck with how does this stuff even work? And I think we need to, we need to be mindful of that as we build out learning journeys. So one of my favorite moments with Mongo was an immersion day. We had hosted earlier last year where we, our, from an enterprise perspective, we're focused on BW BS, but there's nothing stopping us. You're building a B2C app based on the theme of the winner Olympics. At the time, you could take a picture of Sean White or of Nathan Chen and see that it was in fact that athlete and then overlaid on that web app was the number of medals they accrued with the little trumpeteer congratulating you for selecting that athlete. So I think it's important to build trust and drive education with developers with a more simple experience and then rapidly evolve overlaying the features that Ian just mentioned over time. >>I think one of the keys with cryptography is back to the familiar messaging for the cloud offloading heavy lifting. You actually need to make it difficult to impossible for developers to get this wrong, and you wanna make it as easy as possible for developers to deal with cryptography. And that of course is what we're trying to do with our driver technology combined with structure encryption, with query encryption. >>But Robbie, your point is lots of opportunity for education. I mean, I have to say the developers that I work with, it's, I'm, I'm in awe of how they solve problems and I, and the way they solve problems, if they don't know the answer, they figure out how to go get it. So how, how are your two communities and other communities, you know, how are they coming together to, to solve such problems and share whether it's best practices or how do I do this? >>Well, I'm not gonna lie in person. Events are a bunch of fun. And one of the easiest domain knowledge exchange opportunities, when you're all in person, you can ideate, you can whiteboard, you can brainstorm. And often those conversations are what leads to that infrastructure module that an immersion day features. And it's just amazing what in person events can do, but community groups of interest, whether it's a Twitch stream, whether it's a particular code sample, we rely heavily on digital means today to upscale the developer community, but also build on by, by means of a simple port request, introduce new features that maybe you weren't even thinking of before. >>Yeah. You know, that's a really important point because when you meet people face to face, you build a connection. And so if you ask a question, you're more likely perhaps to get an answer, or if one doesn't exist in a, in a search, you know, you, oh, Hey, we met at the, at the conference and let's collaborate on this guys. Congratulations on, on this brave new world. You're in a really interesting spot. You know, developers, developers, developers, as Steve bomber says screamed. And I was glad to see Dave was not screaming and jumping up and down on the stage like that, but, but the message still resonates. So thank you, definitely appreciate. All right, keep it right there. This is Dave ante for the cubes coverage of Mago DB world 2022 from New York city. We'll be right back.
SUMMARY :
Who's the vice president of developer relations at Mongo DB Jens. And of course, Ian, you know, Mongo it's rayon Detra is, is developers start Well, it's been the ethos of MongoDB since day one. So explain how you versus region, what would you see? So you guys both have developer relations in your So of course we But one of the things that's really resonated with me as I was speaking with one So how do you think about the edge? It could be the radio towers that you mentioned. the idea to rip out a mobile device that was initially built and envisioned for the of the company and product has been built with performance and scalability in mind, including things like the mobile databases Robbie's already mentioned, but the core database is designed from day one So super cloud, you might have data on Preem. So that kinds of an kind of answers the question about how It's not just a series of siloed instances in, In different it's the, it's the fabric essentially. but I wonder if you could describe it. So the way that we're able to model, how does the security model evolve with that? And that goes hand in hand with security. week, the, the, the encryption component that you guys announced. So it's groundbreaking technology based on research in the field of structured So how does the mobile edge developer think about that? At the time, you could take a picture of Sean White or of Nathan Chen And that of course is what we're trying to do with our driver technology combined with structure encryption, with query encryption. and other communities, you know, how are they coming together to, to solve such problems And one of the easiest domain knowledge exchange And so if you ask a question, you're more likely perhaps to get an answer, or if one doesn't exist
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Steve | PERSON | 0.99+ |
Verizon | ORGANIZATION | 0.99+ |
Robbie Bellson | PERSON | 0.99+ |
Ian Massingham | PERSON | 0.99+ |
Ian | PERSON | 0.99+ |
10 records | QUANTITY | 0.99+ |
Robbie | PERSON | 0.99+ |
Robbie Belson | PERSON | 0.99+ |
Colorado | LOCATION | 0.99+ |
2009 | DATE | 0.99+ |
Dave | PERSON | 0.99+ |
2016 | DATE | 0.99+ |
Mark Porter | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
MongoDB | ORGANIZATION | 0.99+ |
Sean White | PERSON | 0.99+ |
Nathan Chen | PERSON | 0.99+ |
Olympics | EVENT | 0.99+ |
Python | TITLE | 0.99+ |
MongoDB | TITLE | 0.99+ |
today | DATE | 0.99+ |
NYC | LOCATION | 0.99+ |
late 20 | DATE | 0.99+ |
more than 40% | QUANTITY | 0.99+ |
two communities | QUANTITY | 0.99+ |
Ravi | PERSON | 0.98+ |
MongoDB Atlas | TITLE | 0.98+ |
Mongo DB | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
JavaScript | TITLE | 0.98+ |
this morning | DATE | 0.98+ |
one edge | QUANTITY | 0.97+ |
12 different pro programming languages | QUANTITY | 0.97+ |
New York city | LOCATION | 0.97+ |
first version | QUANTITY | 0.97+ |
this week | DATE | 0.97+ |
both | QUANTITY | 0.97+ |
Azure | TITLE | 0.96+ |
ORGANIZATION | 0.95+ | |
Atlas | TITLE | 0.95+ |
C sharp | TITLE | 0.95+ |
a million encrypted records | QUANTITY | 0.95+ |
about 25,000 developers a week | QUANTITY | 0.93+ |
Twitch | ORGANIZATION | 0.93+ |
first year | QUANTITY | 0.93+ |
19 | DATE | 0.89+ |
Christopher Voss, Microsoft | Kubecon + Cloudnativecon Europe 2022
>> theCUBE presents KubeCon and CloudNativeCon, Europe, 2022. Brought to you by Red Hat, the cloud-native computing foundation and its ecosystem partners. >> Welcome to Valencia, Spain in KubeCon, CloudNativeCon, Europe, 2022. I'm Keith Townsend with my cohosts, Enrico Signoretti, Senior IT Analyst at GigaOm. >> Exactly. >> 7,500 people I'm told, Enrico. What's the flavor of the show so far? >> It's a fantastic mood, I mean, I found a lot of people wanting to track, talk about what they're doing with Kubernetes, sharing their you know, stories, some war stories that bit tough. And you know, this is where you learn actually. Because we had a lot of Zoom calls, webinar and stuff. But it is when you talk a video, "Oh, I did it this way, and it didn't work out very well." So, and, you start a conversation like this that is really different from learning from Zoom, when, you know, everybody talks about things that work it well, they did it right. No, it's here that you learn from other experiences. >> So we're talking to amazing people the whole week, talking about those experiences here on theCUBE. Fresh on the theCUBE for the first time, Chris Voss, senior software engineer at Microsoft Xbox. Chris, welcome to the theCUBE. >> Thank you so much for having me. >> So first off, give us a high level picture of the environment that you're running at Microsoft. >> Yeah. So, you know, we've got 20 well probably close to 30 clusters at this point around the globe, you know 700 to 1,000 pods per cluster, roughly. So about 22,000 pods total. So yeah, it's pretty, pretty sizable footprint and yeah. So we've been running on Kubernetes since 2018 and well actually might be 2017, but anyways, so yeah, that's kind of our footprint. Yeah. >> So all of that, let's talk about the basics which is security across multiple I'm assuming containers, microservices, etcetera. Why did you and the team settle on Linkerd? >> Yeah, so previously we had our own kind of solution for managing TLS certs and things like that. And we found it to be pretty painful, pretty quickly. And so we knew, you know we wanted something that was a little bit more abstracted away from the developers and things like that, that allowed us to move quickly. And so we began investigating, you know, solutions to that. And a few of our colleagues went to Kubecon in San Diego in 2019, Cloudnativecon as well. And basically they just, you know, sponged it all up. And actually funny enough, my old manager was one of the people who was there and he went to the Linkerd booth and they had a thing going that was like, "Hey, get set up with MTLS in five minutes." And he was like, "This is something we want to do, why not check this out?" And he was able to do it. And so that put it on our radar. And so yeah, we investigated several others and Linkerd just perfectly fit exactly what we needed. >> So, in general we are talking about, you know, security at scale. So how you manage security scale and also flexibility. Right? So, but you know, what is the... You told us about the five minutes to start using there but you know, again, we are talking about war stories. We're talking about, you know, all these. So what kind of challenges you found at the beginning when you started adopting this technology? >> So the biggest ones were around getting up and running with like a new service, especially in the beginning, right, we were, you know, adding a new service almost every day. It felt like. And so, you know, basically it took someone going through a whole bunch of different repos, getting approvals from everyone to get the certs minted, all that fun stuff getting them put into the right environments and in the right clusters, to make sure that, you know, everybody is talking appropriately. And just the amount of work that that took alone was just a huge headache and a huge barrier to entry for us to, quickly move up the number of services we have. >> So, I'm trying to wrap my head around the scale of the challenge. When I think about certification or certificate management, I have to do it on a small scale. And every now and again, when a certificate expires it is just a troubleshooting pain. >> Yes. >> So as I think about that, it costs it's not just certificates across 22,000 pods, or it's certificates across 22,000 pods in multiple applications. How were you doing that before Linkerd? Like, what was the... And what were the pain points? Like what happens when a certificate either fails? Or expired up? Not updated? >> So, I mean, to be completely honest, the biggest thing is we're just unable to make the calls, you know, out or in, based on yeah, what is failing basically. But, you know, we saw essentially an uptick in failures around a certain service and pretty quickly, pretty quickly, we got used to the fact that it was like, oh, it's probably a cert expiration issue. And so we tried, you know, a few things in order to make that a little bit more automated and things like that. But we never came to a solution that like didn't require every engineer on the team to know essentially quite a bit about this, just to get into it, which was a huge issue. >> So talk about day two, after you've deployed Linkerd, how did this alleviate software engineers? And what was like the benefits of now having this automated way of managing certs? >> So the biggest thing is like, there is no touch from developers, everyone on our team... Well, I mean, there are a lot of people who are familiar with security and certs and all of that stuff. But no one has to know it. Like it's not a requirement. Like for instance, I knew nothing about it when I joined the team. And even when I was setting up our newer clusters, I knew very little about it. And I was still able to really quickly set up Linkerd, which was really nice. And it's been, you know, essentially we've been able to just kind of set it, and not think about it too much. Obviously, you know, there're parts of it that you have to think about, we monitor it and all that fun stuff, but yeah, it's been pretty painless almost day one. It took a long time to trust it for developers. You know, anytime there was a failure, it's like, "Oh, could this be Linkerd?" you know. But after a while, like now we don't have that immediate assumption because people have built up that trust, but. >> Also you have this massive infrastructure I mean, 30 clusters. So, I guess, that it's quite different to manage a single cluster in 30. So what are the, you know, consideration that you have to do to install this software on, you know, 30 different cluster, manage different, you know versions probably, et cetera, et cetera, et cetera. >> So, I mean, you know, as far as like... I guess, just to clarify, are you asking specifically with Linkerd? Or are you just asking in more in general? >> Well, I mean, you can take that the question in two ways. >> Okay. >> Sure, yeah, so Linkerd in particular but the 30 cluster also quite interesting. >> Yeah. So, I mean, you know, more generally, you know how we manage our clusters and things like that. We have, you know, a CLI tool that we use in order to like change context very quickly, and switch and communicate with whatever cluster we're trying to connect to and you know, are we debugging or getting logs, whatever. And then, you know, with Linkerd it's nice because again, you know, we aren't having to worry about like, oh, how is this cert being inserted in the right node? Or not the right node, but in the right cluster or things like that. Whereas with Linkerd, we don't really have that concern. When we spin up our clusters, essentially we get the route certificate and everything like that packaged up, passed along to Linkerd on installation. And then essentially, there's not much we have to do after that. >> So talk to me about your upcoming section here at Kubecon. what's the high level talking points? Like what attendees learn? >> Yeah. So it's a journey. Those are the sorts of talks that I find useful. Having not been, you know, I'm not a deep Kubernetes expert from, you know decades or whatever of experience, but-- >> I think nobody is. >> (indistinct). >> True, yes. >> That's also true. >> That's another story >> That's a job posting decades of requirements for-- >> Of course, yeah. But so, you know, it's a journey. It's really just like, hey, what made us decide on a service mesh in the first place? What made us choose Linkerd? And then what are the ways in which, you know, we use Linkerd? So what are those, you know we use some of the extra plugins and things like that. And then finally, a little bit about more what we're going to do in the future. >> Let's talk about not just necessarily the future as in two or three days from now, or two or three years from now. Well, the future after you immediately solve the low level problems with Linkerd, what were some of the surprises? Because Linkerd in service mesh and in general have side benefits. Do you experience any of those side benefits as well? >> Yeah, it's funny, you know, writing the blog post, you know, I hadn't really looked at a lot of the data in years on, you know when we did our investigations and things like that. And we had seen that we like had very low latency and low CPU utilization and things like that. And looking at some of that, I found that we were actually saving time off of requests. And I couldn't really think of why that was and I was talking with someone else and the biggest, unfortunately all that data's gone now, like the source data. So I can't go back and verify this but it makes sense, you know, there's the availability zone routing that Linkerd supports. And so I think that's actually doing it where, you know essentially, if a node is closer to another node, it's essentially, you know, routing to those ones. So when one service is talking to another service and maybe they're on the same node, you know, it short circuits that and allows us to gain some time there. It's not huge, but it adds up after, you know, 10, 20 calls down the line. >> Right. In general, so you are saying that it's smooth operations at this very, you know, simplifying your life. >> And again, we didn't have to really do anything for that. It handled that for us. >> It was there? >> Yep. Yeah, exactly. >> So we know one thing when I do it on my laptop it works fine. When I do it with across 22,000 pods, that's a different experience. What were some of the lessons learned coming out of Kubecon 2018 in San Diego? I was there. I wish I would've ran into the Microsoft folks, but what were some of the hard lessons learned scaling Linkerd across the 22,000 nodes? >> So, you know, the first one and this seems pretty obvious, but was just not something I knew about was the high availability mode of Linkerd. So obviously makes sense. You would want that in, you know a large scale environment. So like, that's one of the big lessons that like, we didn't ride away. No. Like one of the mistakes we made in one of our pre-production clusters was not turning that on. And we were kind of surprised. We were like, whoa, like all of these pods are spinning up but they're having issues, like actually getting injected and things like that. And we found, oh, okay. Yeah, you need to actually give it some more resources. But it's still very lightweight considering, you know, they have high availability mode but it's just a few instances still. >> So from, even from, you know, binary perspective and running Linkerd how much overhead is it? >> That is a great question. So I don't remember off the top of my head, the numbers but it's very lightweight. We evaluated a few different service missions and it was the lightest weight that we encountered at that point. >> And then from a resource perspective, is it a team of Linkerd people? Is it a couple of people? Like how? >> To be completely honest for a long time, it was one person Abraham, who actually is the person who proposed this talk. He couldn't make it to Valencia, but he essentially did probably 95% of the work to get into production. And then this was before, we even had a team dedicated to our infrastructure. And so we have, now we have a team dedicated, we're all kind of Linkerd folks, if not Linkerd experts, we at least can troubleshoot basically. And things like that. So it's, I think a group of six people on our team and then, you know various people who've had experience with it on other teams. >> But others, dedicated just to that. >> No one is dedicated just to it. No, it's pretty like pretty light touch once it's up and running. It took a very long time for us to really understand it and to, you know, get like not getting started, but like getting to where we really felt comfortable letting it go in production. But once it was there, like, it is very, very light touch. >> Well, I really appreciate you stopping by Chris. It's been an amazing conversation to hear how Microsoft is using a open source project. >> Exactly. >> At scale, it's just a few years ago when you would've heard the concept of Microsoft and open source together and like OS, just, you know-- >> They have changed a lot in the last few years. Now, there are huge contributors. And, you know, if you go to Azure, it's full of open source stuff, everywhere so. >> Yeah. >> Wow. The Kubecon 2022, how the world has changed in so many ways. From Valencia Spain, I'm Keith Townsend, along with Enrico Signoretti. You're watching theCUBE, the leader in high tech coverage. (upbeat music)
SUMMARY :
Brought to you by Red Hat, Welcome to Valencia, Spain What's the flavor of the show so far? And you know, this is Fresh on the theCUBE for the first time, of the environment that at this point around the globe, you know Why did you and the And so we knew, you know So, but you know, what is the... right, we were, you know, I have to do it on a small scale. How were you doing that before Linkerd? And so we tried, you know, And it's been, you know, So what are the, you know, So, I mean, you know, as far as like... Well, I mean, you can take that but the 30 cluster also quite interesting. And then, you know, with Linkerd So talk to me about Having not been, you know, But so, you know, you immediately solve but it makes sense, you know, you know, simplifying your life. And again, we didn't have So we know one thing So, you know, the first one and it was the lightest and then, you know dedicated just to that. and to, you know, get you stopping by Chris. And, you know, if you go to Azure, how the world has changed in so many ways.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Enrico | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Enrico Signoretti | PERSON | 0.99+ |
Christopher Voss | PERSON | 0.99+ |
Chris Voss | PERSON | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
95% | QUANTITY | 0.99+ |
700 | QUANTITY | 0.99+ |
2017 | DATE | 0.99+ |
Linkerd | ORGANIZATION | 0.99+ |
San Diego | LOCATION | 0.99+ |
30 clusters | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Abraham | PERSON | 0.99+ |
10 | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
20 | QUANTITY | 0.99+ |
Valencia | LOCATION | 0.99+ |
six people | QUANTITY | 0.99+ |
22,000 pods | QUANTITY | 0.99+ |
30 | QUANTITY | 0.99+ |
Valencia, Spain | LOCATION | 0.99+ |
Valencia Spain | LOCATION | 0.99+ |
KubeCon | EVENT | 0.99+ |
7,500 people | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
1,000 pods | QUANTITY | 0.99+ |
two ways | QUANTITY | 0.99+ |
five minutes | QUANTITY | 0.99+ |
Europe | LOCATION | 0.99+ |
CloudNativeCon | EVENT | 0.98+ |
Enrico Signore | PERSON | 0.98+ |
three days | QUANTITY | 0.98+ |
GigaOm | ORGANIZATION | 0.98+ |
two | QUANTITY | 0.98+ |
first time | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
Cloudnativecon | ORGANIZATION | 0.97+ |
one service | QUANTITY | 0.97+ |
Kubecon | ORGANIZATION | 0.97+ |
three years | QUANTITY | 0.97+ |
30 different cluster | QUANTITY | 0.96+ |
first one | QUANTITY | 0.96+ |
22,000 nodes | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
30 cluster | QUANTITY | 0.95+ |
one thing | QUANTITY | 0.94+ |
Xbox | COMMERCIAL_ITEM | 0.93+ |
about 22,000 pods | QUANTITY | 0.92+ |
single cluster | QUANTITY | 0.92+ |
20 calls | QUANTITY | 0.91+ |
day two | QUANTITY | 0.91+ |
one person | QUANTITY | 0.89+ |
few years ago | DATE | 0.88+ |
decades | QUANTITY | 0.87+ |
2022 | DATE | 0.85+ |
Azure | TITLE | 0.79+ |
Kubernetes | TITLE | 0.77+ |
Christopher Voss, Microsoft | Kubecon + Cloudnativecon Europe 2022
>>The cube presents, Coon and cloud native con Europe 22, brought to you by the cloud native computing foundation. >>Welcome to Valencia Spain in co con cloud native con Europe, 2022. I'm Keith Townsend with my cohos on Rico senior. Etti senior it analyst at gig home. Exactly 7,500 people I'm told en Rico. What's the flavor of the show so far, >>It's a fantastic mood. I mean, I found a lot of people wanting to track talk about what they're doing with Kubernetes, sharing their, you know, stories, some word stories that meet tough. And you know, this is where you learn actually, because we had a lot of zoom calls, webinar and stuff, but it is when you talk a video, oh, I did it this way and it didn't work out very well. So, and, and you start a conversation like this that is really different from learning from zoom. When, you know, everybody talks about things that working well, they did it, right. No, it's here that you learn from other experiences. >>So we're talking to amazing people the whole week, talking about those experiences here on the queue, fresh on the queue for the first time, Chris Vos, senior software engineer at Microsoft Xbox, Chris, welcome to the queue. >>Thank you so much for having >>Me. So first off, give us a high level picture of the environment that you're running at Microsoft. >>Yeah. So, you know, we've got 20, well probably close to 30 clusters at this point around the globe, you know, 700 to a thousand pods per cluster, roughly. So about 22,000 pods total. So yeah, it's pretty pretty sizable footprint and yeah. So we've been running on Kubernetes since 2018 and well actually might be 2017, but anyways, so yeah, that, that's kind of our, our footprint. >>Yeah. So all of that, let's talk about the basics, which is security across multiple I'm assuming containers, work, microservices, et cetera. Why did you and the team settle on link or do >>Yeah, so previously we had our own kind of solution for managing TLS certs and things like that. And we found it to be pretty painful pretty quickly. And so we knew, you know, we wanted something that was a little bit more abstracted away from the developers and, and things like that that allowed us to move quickly. And so we began investigating, you know, solutions to that. And a few of our colleagues went to Cuban in San Diego in 2019 cloud native con as well. And basically they just, you know, sped it all up. And actually funny enough, my, my old manager was one of the people who was there and he went to the link D booth and they had a thing going that was like, Hey, get set up with MTLS in five minutes. And he was like, this is something we want to do, why not check this out? And he was able to do it. And so that, that put it on our radar. And so yeah, we investigated several others and Leer D just perfectly fit exactly what we needed. >>So, so in general, we are talking about, you know, security at scale. So how you manage security to scale and also flexibility, right. But you know, what is the you, this there, you told us about the five minutes to start using there, but you know, again, we are talking about word stories. We talk about, you know, all these. So what, what, what kind of challenges you found at the beginning when you start adopting this technology? >>So the biggest ones were around getting up and running with like a new service, especially in the beginning, right. We were, you know, adding a new service almost every day. It felt like. And so, you know, basically it took someone going through a whole bunch of different repos, getting approvals from everyone to get the SEARCHs minted, all that fun stuff, getting them put into the right environments and in the right clusters to make sure that, you know, everybody is talking appropriately. And just the amount of work that, that took alone was just a huge headache and a huge barrier to entry for us to, you know, quickly move up the number of services we have. So, >>So I'm, I'm trying to wrap my head around the scale of the challenge. When I think about certification or certificate management, I have to do it on a small scale and the, the, every now and again, when a certificate expires, it is just a troubleshooting pain. Yes. So as I think about that, it costs, it's not just certificates across 22,000 pods or it's certificates across 22,000 pods in multiple applications. How were you doing that before link D like, what was the, what and what were the pain points? Like? What happens when a certificate either fails or expired up not, not updated? >>So, I mean, to be completely honest, the biggest thing is we're just unable to make the calls, you know, out or, or in, based on yeah. What is failing basically. But, you know, we saw essentially an uptick in failures around a certain service and pretty quickly, I pretty quickly, we got used to the fact that it was like, oh, it's probably a cert expiration issue. And so we tried, you know, a few things in order to make that a little bit more automated and things like that, but we never came to a solution that like didn't require every engineer on the team to know essentially quite a bit about this, just to get into it, which was a huge issue. >>So talk about day two after you've deployed link D how did this alleviate software engineers and what was like the, the benefits of now having this automated way of managing >>Certs? So the biggest thing is like, there is no touch from developers, everyone on our team. Well, I mean, there are a lot of people who are familiar with security and certs and all of that stuff, but no one has to know it. Like it's not a requirement. Like for instance, I knew nothing about it when I joined the team. And even when I was setting up our newer clusters, I knew very little about it. And I was still able to really quickly set up blinker D, which was really nice. And, and it's been, you know, essentially we've been able to just kind of set it and not think about it too much. Obviously, you know, there are parts of it that you have to think about. We monitor it and all that fun stuff, but, but yeah, it's been pretty painless almost day one. It took a lot, a long time to trust it for developers. You know, anytime there was a failure, it's like, oh, could this be link or D you know, but after a while, like now we don't have that immediate assumption because people have built up that trust, but >>Also you have this massive infrastructure, I mean, 30 cluster. So I guess that it's quite different to manage a single cluster and 30. So what are the, you know, consideration that you have to do to install this software on, you know, 30 different cluster manage different, you know, versions probably etcetera, etcetera, et cetera. >>So, I mean, you know, the, the, as far as like, I guess, just to clarify, are you asking specifically with Linky or are you just asking in more in general? Well, >>I mean, you, you can take the, the question in the, in two ways, so, okay. Yeah. Yes. Link in particular, but the 30 cluster also quite interesting. >>Yeah. So, I mean, you know, more generally, you know, how we manage our clusters and things like that. We have, you know, a CLI tool that we use in order to like, change context very quickly and switch and communicate with whatever cluster we're trying to connect to and, you know, are we debugging or getting logs, whatever. And then, you know, with link D it's nice because again, you know, we, we, aren't having to worry about like, oh, how is this cert being inserted in the right node or, or not the right node, but in the right cluster or things like that. Whereas with link D we don't, we don't really have that concern when we spin up our, our clusters, essentially we get the root certificate and, and everything like that packaged up, passed along to link D on installation. And then essentially there's not much we have to do after that. >>So talk to me about your upcoming coming section here at Q con what's the, what's the high level talking points? Like what, what will attendees learn? >>Yeah. So it's, it's a journey. Those are the sorts of talks that I find useful. Having not been, you know, I, I'm not a deep Kubernetes expert from, you know, decades or whatever of experience, but I think >>Nobody is >>Also true. That's another story. That's a, that's, that's a job posting decades of requirements for >>Of course. Yeah. But so, you know, it, it's a journey it's really just like, Hey, what made us decide on a service mesh in the first place? What made us choose link D and then what are the ways in which, you know, we, we use link D so what are those, you know, we use some of the extra plugins and things like that. And then finally, a little bit about more, what we're gonna do in the future. >>Let's talk about not just necessarily the future as in two or three days from now, or two or three years from now. Well, the future after you immediately solve the, the low level problems with link D what were some of the, the surprises, because link D in service me in general has have side benefits. Do you experience any of those side benefits as well? >>Yeah, it's funny, you know, writing the, the blog post, you know, I hadn't really looked at a lot of the data in years on, you know, when we did our investigations and things like that. And we had seen that we like had very low latency and low CPU utilization and things like that. And looking at some of that, I found that we were actually saving time off of requests. And I couldn't really think of why that was, and I was talking with someone else and the biggest, unfortunately, all that data's gone now, like the source data. So I can't go back and verify this, but it, it makes sense, you know, there's the availability zone routing that linker D supports. And so I think that's actually doing it where, you know, essentially if a node is closer to another node, it's essentially, you know, routing to those ones. So when one service is talking to another service and maybe on they're on the same node, you know, it, it short circuits that, and allows us to gain some, some time there. It's not huge, but it adds up after, you know, 10, 20 calls down the line. Right. >>In general. So you are saying that it's smooth operations in, in ATS, very, you know, simplifying your life. >>And again, we didn't have to really do anything for that. It, it, it handled that for it was there. Yeah. Yep. Yeah, exactly. >>So we know one thing when I do it on my laptop, it works fine when I do it with across 22,000 pods, that's a different experience. What were some of the lessons learned coming out of KU con 2018 in San Diego was there? I wish I would've ran to the microphone folks, but what were some of the hard lessons learned scaling link D across the 22,000 nodes? >>So, you know, the, the first one, and this seems pretty obvious, but was just not something I knew about was the high availability mode of link D so obviously makes sense. You would want that in a, you know, a large scale environment. So like, that's one of the big lessons that like, we didn't ride away. No. Like one of the mistakes we made in, in one of our pre-production clusters was not turning that on. And we were kind of surprised. We were like, whoa, like all of these pods are spinning up, but they're having issues like actually getting injected and things like that. And we found, oh, okay. Yeah, you need to actually give it some, some more resources, but it's still very lightweight considering, you know, they have high availability mode, but it's just a few instances still. >>So from, even from a, you know, binary perspective and running link D how much overhead is it? >>That is a great question. So I don't remember off the top of my head, the numbers, but it's very lightweight. We, we evaluated a few different service missions and it was the lightest weight that we encountered at that point. >>And then from a resource perspective, is it a team of link D people? Is it a couple of people, like how >>To be completely honest for a long time, it was one person, Abraham who actually is the person who proposed this talk. He couldn't make it to Valencia, but he essentially did probably 95% of the work to get a into production. And then this was before we even had a team dedicated to our infrastructure. And so we have, now we have a team dedicated, we're all kind of Linky folks, if not Linky experts, we at least can troubleshoot basically. And things like that. So it's, I think a group of six people on our team, and then, you know, various people who've had experience with it >>On other teams, but I'm not dedicated just to that. >>I mean, >>No one is dedicated just to it. No, it's pretty like pretty light touch once it's, once it's up and running, it took a very long time for us to really understand it and, and to, you know, get like, not getting started, but like getting to where we really felt comfortable letting it go in production. But once it was there, like, it is very, very light touch. >>Well, I really appreciate you stopping by Chris. It's been an amazing conversation to hear how Microsoft is using a open source project. Exactly. At scale. It's just a few years ago, when you would've heard the concept of Microsoft and open source together and like, oh, that's just, you know, but >>They have changed a lot in the last few years now, there are huge contributors. And, you know, if you go to Azure, it's full of open source stuff, every >>So, yeah. Wow. The Cuban 2022, how the world has changed in so many ways from Licia Spain, I'm Keith Townsend, along with a Rico senior, you're watching the, the leader in high tech coverage.
SUMMARY :
brought to you by the cloud native computing foundation. What's the flavor of the show so far, And you know, on the queue, fresh on the queue for the first time, Chris Vos, Me. So first off, give us a high level picture of the environment that you're at this point around the globe, you know, 700 to a thousand pods per you and the team settle on link or do And so we began investigating, you know, solutions to that. So, so in general, we are talking about, you know, security at scale. And so, you know, basically it took someone going through a whole How were you doing that before link D like, what was the, what and what were the pain points? we tried, you know, a few things in order to make that a little bit more automated and things like that, You know, anytime there was a failure, it's like, oh, could this be link or D you know, but after a while, you know, consideration that you have to do to install this software on, Link in particular, but the 30 cluster also quite interesting. And then, you know, with link D it's nice Having not been, you know, I, I'm not a deep Kubernetes expert from, Also true. What made us choose link D and then what are the ways in which, you know, we, we use link D so what Well, the future after you immediately solve I hadn't really looked at a lot of the data in years on, you know, when we did our investigations and very, you know, simplifying your life. And again, we didn't have to really do anything for that. So we know one thing when I do it on my laptop, it works fine when I do it with across 22,000 So, you know, the, the first one, and this seems pretty obvious, but was just not something I knew about was So I don't remember our team, and then, you know, various people who've had experience with it you know, get like, not getting started, but like getting to where together and like, oh, that's just, you know, but you know, if you go to Azure, it's full of open source stuff, every how the world has changed in so many ways from Licia Spain,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Keith Townsend | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Christopher Voss | PERSON | 0.99+ |
2017 | DATE | 0.99+ |
Chris Vos | PERSON | 0.99+ |
Abraham | PERSON | 0.99+ |
20 | QUANTITY | 0.99+ |
95% | QUANTITY | 0.99+ |
700 | QUANTITY | 0.99+ |
San Diego | LOCATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
10 | QUANTITY | 0.99+ |
30 | QUANTITY | 0.99+ |
five minutes | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
22,000 pods | QUANTITY | 0.99+ |
six people | QUANTITY | 0.99+ |
Valencia | LOCATION | 0.99+ |
two | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
two ways | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
20 calls | QUANTITY | 0.99+ |
7,500 people | QUANTITY | 0.99+ |
22,000 pods | QUANTITY | 0.99+ |
first time | QUANTITY | 0.98+ |
Cuban | LOCATION | 0.98+ |
first | QUANTITY | 0.98+ |
one service | QUANTITY | 0.98+ |
Valencia Spain | LOCATION | 0.98+ |
Europe | LOCATION | 0.98+ |
Linky | ORGANIZATION | 0.97+ |
three days | QUANTITY | 0.97+ |
2022 | DATE | 0.97+ |
one person | QUANTITY | 0.97+ |
first one | QUANTITY | 0.97+ |
link D | ORGANIZATION | 0.96+ |
Kubecon | ORGANIZATION | 0.96+ |
30 cluster | QUANTITY | 0.96+ |
22,000 nodes | QUANTITY | 0.96+ |
KU con 2018 | EVENT | 0.95+ |
Coon | ORGANIZATION | 0.94+ |
Licia Spain | PERSON | 0.94+ |
30 clusters | QUANTITY | 0.94+ |
day two | QUANTITY | 0.92+ |
link D | OTHER | 0.92+ |
Xbox | COMMERCIAL_ITEM | 0.91+ |
Rico | ORGANIZATION | 0.91+ |
Q con | ORGANIZATION | 0.91+ |
about 22,000 pods | QUANTITY | 0.91+ |
Kubernetes | PERSON | 0.9+ |
few years ago | DATE | 0.9+ |
three years | QUANTITY | 0.89+ |
link | ORGANIZATION | 0.86+ |
single cluster | QUANTITY | 0.85+ |
one thing | QUANTITY | 0.82+ |
Leer D | ORGANIZATION | 0.79+ |
a thousand pods | QUANTITY | 0.77+ |
Cloudnativecon | ORGANIZATION | 0.75+ |
last | DATE | 0.74+ |
cluster | QUANTITY | 0.74+ |
MTLS | ORGANIZATION | 0.72+ |
Etti | ORGANIZATION | 0.72+ |
Azure | TITLE | 0.71+ |
Rico | LOCATION | 0.69+ |
ATS | ORGANIZATION | 0.68+ |
years | DATE | 0.64+ |
cloud native con | ORGANIZATION | 0.61+ |
Cuban | PERSON | 0.6+ |
day one | QUANTITY | 0.59+ |
decades | QUANTITY | 0.56+ |
link | OTHER | 0.56+ |
Kubernetes | ORGANIZATION | 0.53+ |
link | TITLE | 0.52+ |
22 | EVENT | 0.5+ |
Tushar Katarki & Justin Boitano | Red Hat Summit 2022
(upbeat music) >> We're back. You're watching theCUBE's coverage of Red Hat Summit 2022 here in the Seaport in Boston. I'm Dave Vellante with my co-host, Paul Gillin. Justin Boitano is here. He's the Vice President of Enterprise and Edge Computing at NVIDIA. Maybe you've heard of him. And Tushar Katarki who's the Director of Product Management at Red Hat. Gentlemen, welcome to theCUBE, good to see you. >> Thank you. >> Great to be here, thanks >> Justin, you are a keynote this morning. You got interviewed and shared your thoughts on AI. You encourage people to got to think bigger on AI. I know it's kind of self-serving but why? Why should we think bigger? >> When you think of AI, I mean, it's a monumental change. It's going to affect every industry. And so when we think of AI, you step back, you're challenging companies to build intelligence and AI factories, and factories that can produce intelligence. And so it, you know, forces you to rethink how you build data centers, how you build applications. It's a very data centric process where you're bringing in, you know, an exponential amount of data. You have to label that data. You got to train a model. You got to test the model to make sure that it's accurate and delivers business value. Then you push it into production, it's going to generate more data, and you kind of work through that cycle over and over and over. So, you know, just as Red Hat talks about, you know, CI/CD of applications, we're talking about CI/CD of the AI model itself, right? So it becomes a continuous improvement of AI models in production which is a big, big business transformation. >> Yeah, Chris Wright was talking about basically take your typical application development, you know, pipeline, and life cycle, and apply that type of thinking to AI. I was saying those two worlds have to come together. Actually, you know, the application stack and the data stack including AI need to come together. What's the role of Red Hat? What's your sort of posture on AI? Where do you fit with OpenShift? >> Yeah, so we're really excited about AI. I mean, a lot of our customers obviously are looking to take that data and make meaning out of it using AI is definitely a big important tool. And OpenShift, and our approach to Open Hybrid Cloud really forms a successful platform to base all your AI journey on with the partners such as NVIDIA whom we are working very closely with. And so the idea really is as Justin was saying, you know, the end to end, when you think about life of a model, you've got data, you mine that data, you create models, you deploy it into production. That whole thing, what we call CI/CD, as he was saying DevOps, DevSecOps, and the hybrid cloud that Red Hat has been talking about, although with OpenShift as the center forms a good basis for that. >> So somebody said the other day, I'm going to ask you, is INVIDIA a hardware company or a software company? >> We are a company that people know for our hardware but, you know, predominantly now we're a software company. And that's what we were on stage talking about. I mean, ultimately, a lot of these customers know that they've got to embark on this journey to apply AI, to transform their business with it. It's such a big competitive advantage going into, you know, the next decade. And so the faster they get ahead of it, the more they're going to win, right? But some of them, they're just not really sure how to get going. And so a lot of this is we want to lower the barrier to entry. We built this program, we call it Launchpad to basically make it so they get instant access to the servers, the AI servers, with OpenShift, with the MLOps tooling, with example applications. And then we walk them through examples like how do you build a chatbot? How do you build a vision system for quality control? How do you build a price recommendation model? And they can do hands on labs and walk out of, you know, Launchpad with all the software they need, I'll say the blueprint for building their application. They've got a way to have the software and containers supported in production, and they know the blueprint for the infrastructure and operating that a scale with OpenShift. So more and more, you know, to come back to your question is we're focused on the software layers and making that easy to help, you know, either enterprises build their apps or work with our ecosystem and developers to buy, you know, solutions off the shelf. >> On the harbor side though, I mean, clearly NVIDIA has prospered on the backs of GPUs, as the engines of AI development. Is that how it's going to be for the foreseeable future? Will GPUs continue to be core to building and training AI models or do you see something more specific to AI workloads? >> Yeah, I mean, it's a good question. So I think for the next decade, well, plus, I mean not forever, we're going to always monetize hardware. It's a big, you know, market opportunity. I mean, Jensen talks about a $100 billion, you know, market opportunity for NVIDIA just on hardware. It's probably another a $100 billion opportunity on the software. So the reality is we're getting going on the software side, so it's still kind of early days, but that's, you know, a big area of growth for us in the future and we're making big investments in that area. On the hardware side, and in the data center, you know, the reality is since Moore's law has ended, acceleration is really the thing that's going to advance all data centers. So I think in the future, every server will have GPUs, every server will have DPUs, and we can talk a bit about what DPUs are. And so there's really kind of three primary processors that have to be there to form the foundation of the enterprise data center in the future. >> Did you bring up an interesting point about DPUs and MPUs, and sort of the variations of GPUs that are coming about? Do you see those different PU types continuing to proliferate? >> Oh, absolutely. I mean, we've done a bunch of work with Red Hat, and we've got a, I'll say a beta of OpenShift 4.10 that now supports DPUs as the, I'll call it the control plane like software defined networking offload in the data center. So it takes all the software defined networking off of CPUs. When everybody talks about, I'll call it software defined, you know, networking and core data centers, you can think of that as just a CPU tax up to this point. So what's nice is it's all moving over to DPU to, you know, offload and isolate it from the x86 cores. It increases security of data center. It improves the throughput of your data center. And so, yeah, DPUs, we see everybody copying that model. And, you know to give credit where credit is due, I think, you know, companies like AWS, you know, they bought Annapurna, they turned it into Nitro which is the foundation of their data centers. And everybody wants the, I'll call it democratized version of that to run their data centers. And so every financial institution and bank around the world sees the value of this technology, but running in their data centers. >> Hey, everybody needs a Nitro. I've written about it. It's Annapurna acquisition, 350 million. I mean, peanuts in the grand scheme of things. It's interesting, you said Moore's law is dead. You know, we have that conversation all the time. Pat Gelsinger promised that Moore's law is alive and well. But the interesting thing is when you look at the numbers, that's, you know, Moore's law, we all know it, doubling of the transistor densities every 18 to 24 months. Let's say that, that promise that he made is true. What I think the industry maybe doesn't appreciate, I'm sure you do, being in NVIDIA, when you combine what you were just saying, the CPU, the GPU, Paul, the MPU, accelerators, all the XPUs, you're talking about, I mean, look at Apple with the M1, I mean 6X in 15 months versus doubling every 18 to 24. The A15 is probably averaging over the last five years, a 110% performance improvement each year versus the historical Moore's law which is 40%. It's probably down to the low 30s now. So it's a completely different world that we're entering now. And the new applications are going to be developed on these capabilities. It's just not your general purpose market anymore. From an application development standpoint, what does that mean to the world? >> Yeah, I mean, yeah, it is a great point. I mean, from an application, I mean first of all, I mean, just talk about AI. I mean, they are all very compute intensive. They're data intensive. And I mean to move data focus so much in to compute and crunch those numbers. I mean, I'd say you need all the PUs that you mentioned in the world. And also there are other concerns that will augment that, right? Like we want to, you know, security is so important so we want to secure everything. Cryptography is going to take off to new levels, you know, that we are talking about, for example, in the case of DPUs, we are talking about, you know, can that be used to offload your encryption and firewalling, and so on and so forth. So I think there are a lot of opportunities even from an application point of view to take of this capacity. So I'd say we've never run out of the need for PUs if you will. >> So is OpenShift the layer that's going to simplify all that for the developer. >> That's right. You know, so one of the things that we worked with NVIDIA, and in fact was we developed this concept of an operator for GPUs, but you can use that pattern for any of the PUs. And so the idea really is that, how do you, yeah-- (all giggle) >> That's a new term. >> Yeah, it's a new term. (all giggle) >> XPUs. >> XPUs, yeah. And so that pattern becomes very easy for GPUs or any other such accelerators to be easily added as a capacity. And for the Kubernetes scaler to understand that there is that capacity so that an application which says that I want to run on a GPU then it becomes very easy for it to run on that GPU. And so that's the abstraction to your point about how we are making that happen. >> And to add to this. So the operator model, it's this, you know, open source model that does the orchestration. So Kubernetes will say, oh, there's a GPU in that node, let me run the operator, and it installs our entire run time. And our run time now, you know, it's got a MIG configuration utility. It's got the driver. It's got, you know, telemetry and metering of the actual GPU and the workload, you know, along with a bunch of other components, right? They get installed in that Kubernetes cluster. So instead of somebody trying to chase down all the little pieces and parts, it just happens automatically in seconds. We've extended the operator model to DPUs and networking cards as well, and we have all of those in the operator hub. So for somebody that's running OpenShift in their data centers, it's really simple to, you know, turn on Node Feature Discovery, you point to the operators. And when you see new accelerated nodes, the entire run time is automatically installed for you. So it really makes, you know, GPUs and our networking, our advanced networking capabilities really first class citizens in the data center. >> So you can kind of connect the dots and see how NVIDIA and the Red Hat partnership are sort of aiming at the enterprise. I mean, NVIDIA, obviously, they got the AI piece. I always thought maybe 25% of the compute cycles in the data center were wasted doing storage offloads or networking offload, security. I think Jensen says it's 30%, probably a better number than I have. But so now you're seeing a lot of new innovation in new hardware devices that are attacking that with alternative processors. And then my question is, what about the edge? Is that a blue field out at the edge? What does that look like to NVIDIA and where does OpenShift play? >> Yeah, so when we talk about the edge, we always going to start talking about like which edge are we talking about 'cause it's everything outside the core data center. I mean, some of the trends that we see with regard to the edges is, you know, when you get to the far edge, it's single nodes. You don't have the guards, gates, and guns protection of the data center. So you start having to worry about physical security of the hardware. So you can imagine there's really stringent requirements on protecting the intellectual property of the AI model itself. You spend millions of dollars to build it. If I push that out to an edge data center, how do I make sure that that's fully protected? And that's the area that we just announced a new processor that we call Hopper H100. It supports confidential computing so that you can basically ensure that model is always encrypted in system memory across the bus, of the PCI bus to the GPU, and it's run in a confidential way on the GPU. So you're protecting your data which is your model plus the data flowing through it, you know, in transit, wallet stored, and then in use. So that really adds to that edge security model. >> I wanted to ask you about the cloud, correct me if I'm wrong. But it seems to me that that AI workloads have been slower than most to make their way to the cloud. There are a lot of concerns about data transfer capacity and even cost. Do you see that? First of all, do you agree with that? And secondly, is that going to change in the short-term? >> Yeah, so I think there's different classes of problems. So we'll take, there's some companies where their data's generated in the cloud and we see a ton of, I'll say, adoption of AI by cloud service providers, right? Recommendation engines, translation engines, conversational AI services, that all the clouds are building. That's all, you know, our processors. There's also problems that enterprises have where now I'm trying to take some of these automation capabilities but I'm trying to create an intelligent factory where I want to, you know, merge kind of AI with the physical world. And that really has to run at the edge 'cause there's too much data being generated by cameras to bring that all the way back into the cloud. So, you know, I think we're seeing mass adoption in the cloud today. I think at the edge a lot of businesses are trying to understand how do I deploy that reliably and securely and scale it. So I do think, you know, there's different problems that are going to run in different places, and ultimately we want to help anybody apply AI where the business is generating the data. >> So obviously very memory intensive applications as well. We've seen you, NVIDIA, architecturally kind of move away from the traditional, you know, x86 approach, take better advantage of memories where obviously you have relationships with Arm. So you've got a very diverse set of capabilities. And then all these other components that come into use, to just be a kind of x86 centric world. And now it's all these other supporting components to support these new applications and it's... How should we think about the future? >> Yeah, I mean, it's very exciting for sure, right? Like, you know, the future, the data is out there at the edge, the data can be in the data center. And so we are trying to weave a hybrid cloud footprint that spans that. I mean, you heard Paul come here, talk about it. But, you know, we've talked about it for some time now. And so the paradigm really that is, that be it an application, and when I say application, it could be even an AI model as a service. It can think about that as an application. How does an application span that entire paradigm from the core to the edge and beyond is where the future is. And, of course, there's a lot of technical challenges, you know, for us to get there. And I think partnerships like this are going to help us and our customers to get there. So the world is very exciting. You know, I'm very bullish on how this will play out, right? >> Justin, we'll give you the last word, closing thoughts. >> Well, you know, I think a lot of this is like I said, it's how do we reduce the complexity for enterprises to get started which is why Launchpad is so fundamental. It gives, you know, access to the entire stack instantly with like hands on curated labs for both IT and data scientists. So they can, again, walk out with the blueprints they need to set this up and, you know, start on a successful AI journey. >> Just a position, is Launchpad more of a Sandbox, more of a school, or more of an actual development environment. >> Yeah, think of it as it's, again, it's really for trial, like hands on labs to help people learn all the foundational skills they need to like build an AI practice and get it into production. And again, it's like, you don't need to go champion to your executive team that you need access to expensive infrastructure and, you know, and bring in Red Hat to set up OpenShift. Everything's there for you so you can instantly get started. Do kind of a pilot project and then use that to explain to your executive team everything that you need to then go do to get this into production and drive business value for the company. >> All right, great stuff, guys. Thanks so much for coming to theCUBE. >> Yeah, thanks. >> Thank you for having us. >> All right, thank you for watching. Keep it right there, Dave Vellante and Paul Gillin. We'll be back right after this short break at the Red Hat Summit 2022. (upbeat music)
SUMMARY :
here in the Seaport in Boston. Justin, you are a keynote this morning. And so it, you know, forces you to rethink Actually, you know, the application And so the idea really to buy, you know, solutions off the shelf. Is that how it's going to be the data center, you know, of that to run their data centers. I mean, peanuts in the of the need for PUs if you will. all that for the developer. And so the idea really is Yeah, it's a new term. And so that's the So it really makes, you know, Is that a blue field out at the edge? across the bus, of the PCI bus to the GPU, First of all, do you agree with that? And that really has to run at the edge you know, x86 approach, from the core to the edge and beyond Justin, we'll give you the Well, you know, I think a lot of this is Launchpad more of a that you need access to Thanks so much for coming to theCUBE. at the Red Hat Summit 2022.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Tushar Katarki | PERSON | 0.99+ |
Justin | PERSON | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Justin Boitano | PERSON | 0.99+ |
Chris Wright | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Paul | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Pat Gelsinger | PERSON | 0.99+ |
110% | QUANTITY | 0.99+ |
25% | QUANTITY | 0.99+ |
30% | QUANTITY | 0.99+ |
40% | QUANTITY | 0.99+ |
$100 billion | QUANTITY | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
INVIDIA | ORGANIZATION | 0.99+ |
Annapurna | ORGANIZATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Seaport | LOCATION | 0.99+ |
350 million | QUANTITY | 0.99+ |
15 months | QUANTITY | 0.99+ |
24 | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
24 months | QUANTITY | 0.99+ |
next decade | DATE | 0.99+ |
Red Hat Summit 2022 | EVENT | 0.98+ |
18 | QUANTITY | 0.98+ |
Boston | LOCATION | 0.98+ |
OpenShift | TITLE | 0.98+ |
30s | QUANTITY | 0.97+ |
each year | QUANTITY | 0.97+ |
A15 | COMMERCIAL_ITEM | 0.97+ |
secondly | QUANTITY | 0.97+ |
First | QUANTITY | 0.97+ |
today | DATE | 0.96+ |
6X | QUANTITY | 0.96+ |
next decade | DATE | 0.96+ |
both | QUANTITY | 0.96+ |
Open Hybrid Cloud | TITLE | 0.95+ |
Kubernetes | TITLE | 0.95+ |
theCUBE | ORGANIZATION | 0.94+ |
Launchpad | TITLE | 0.94+ |
two worlds | QUANTITY | 0.93+ |
millions of dollars | QUANTITY | 0.92+ |
M1 | COMMERCIAL_ITEM | 0.92+ |
Nitro | ORGANIZATION | 0.91+ |
Vice President | PERSON | 0.91+ |
OpenShift 4.10 | TITLE | 0.89+ |
single nodes | QUANTITY | 0.88+ |
DevSecOps | TITLE | 0.86+ |
Jensen | ORGANIZATION | 0.83+ |
one | QUANTITY | 0.82+ |
three primary processors | QUANTITY | 0.82+ |
DevOps | TITLE | 0.81+ |
first | QUANTITY | 0.8+ |
last five years | DATE | 0.79+ |
this morning | DATE | 0.79+ |
Moore | PERSON | 0.77+ |
x86 cores | QUANTITY | 0.71+ |
Michael Dell, Dell Technologies | Dell Technologies World 2022
>>The cube presents, Dell technologies world brought to you by Dell. >>Hello. Welcome to the cube here at Dell tech world. I'm John furry host of the cube with Dave Alon here with Michael Dell, the CEO of Dell technologies cube alumni comes on every year. We have the cube here. It's been two years. Michael, welcome to the cube. Get to see you. >>Hey, John, Dave, great to be with you guys. Thanks for being here. Wonderful to be back here in Vegas with >>You. Well, great to be in person two years ago, we had the cue with the pandemic a lot's happened. We were talking end to end solutions here at Dell tech world in person two years ago, pandemic hits. Thank God you had all that supply for the, for the people having the remote remote end to work now back in person. What's it look like now with, with Dell tech end to end, the edge is important. What's the story, >>You know, edge is, is the physical world. And if you, if you step back from clouds and, you know, multi-cloud, you sort of think about what is the purpose of a cloud or a data center? Well, it's to take data out of the physical world and move it to this place, to somehow enhance it or do something with it and create business value and hopefully create better outcomes. Well, it turns out that, you know, increasingly a lot of that data is gonna stay in the physical world and all of those nodes are gonna be connected. They're gonna be intelligent and we're seeing it in manufacturing and retail and healthcare, transportation, logistics. We're seeing this rapidly intelligent edge being formed. And then of course, with the new networks, the 5g we're seeing, you know, all, all this develop. And so here on the show floor, we're showing a lot of those solutions, but our customers are, are highly engaged. And certainly we think that's a, a big, a big growth factor for the next decade. >>And it's been ING to watch the transformation of the it world and cloudification and the as service, uh, consumption model, which you guys are putting out there has been very successful, but cloud operations is more prominent now on premises and edge and cloud. So the combination of cloud on-premise and edge hardware matters more now than ever before Silicon advances, um, abstraction layers from modern cloud native applications are what people are focused on. What's the story that you cite to the CIOs saying, we're here to help you with that new architecture cloud multi-cloud on premise and edge. What's the main story for you guys with the customers? >>Well, you know, customers want to go faster, right? And they want to accelerate their transformation. And so they wanna shift more resources over to developers, to applications, to access their data, to create competitive advantage. And so we talk a lot about the value line and what are those things below the value line, where we can provide that as a service on a consumption based model and accelerate their transformation, kind of, you know, do for them what we've done inside our own business. And, you know, it's absolutely resonating. We're seeing great growth there. People continue to, to need the solutions, but as we can automate the management and deployment of infrastructure and make it super easy, it gives them a lot of cycles back. >>You know, Michael, my, the favorite part, my favorite part of your book was you were in, I think you were in his, in his home court, in his dining room at Carl Icahn's house. And you said, well, why don't you just buy the company? And then you'll do what you're doing. I I'll buy it back for cheaper. Now, thankfully, you didn't have to do that. Cuz you had an environment of low interest rates and you obviously took it into the other direction, added tremendous value, 101 billion in revenue last year, 17% revenue growth, which was out astounding. When you think about that, um, now we're entering a new chapter with VMware untethered of course you're the chairman of both companies. So how should we think about the new Dell what's next? >>Well, so look, we, we have some unbelievable core businesses, right? We have our client system business and we've all learned during these last two years, how incredibly important it is to enable and empower your workforce with the right tools in the remote and high hybrid work. And we're showing off all kinds of new innovations here. That's a huge business force continues to grow, continues to be super important. Then we have our ISG, the cloud data center, the network of the future, the edge, you know, the, the sort of epicenter of where we're embracing, consumption based business models. That's absolutely huge. Then we have these new, new businesses that we're building with telco with edge, put it all together. It's a 1.3 trillion Tam that we operate in, as you said, more than a hundred billion dollars last year. So there's plenty of room for us to continue to grow and, and expand. And you know, as we make this shift to outcomes, it's obviously more valuable for customers and that, you know, increases our opportunity, increases the, the value we can create for all our stakeholders. >>And number one, number one, share in PCs, by the way, congratulations, again, hit that milestone. All of our gamer, uh, fans in our discord want to know what's the hottest chips coming. What's the fastest machines. What, how's the monitors coming? They want faster, cheaper. What's the coolest, uh, monitors out there right now and, and machines. >>Well, uh, you know, what what's, what's amazing is the, the pace of innovation continues to improve. So whether it's in the GPU, the CPU, the, the resolution, I I'm pretty partial to our 41, uh, display 11 million pixels of fun. And look, I mean, we, we it's, it's, it's clear that people are more productive when they have large screens and all the performance is enabling photo realistic, uh, you know, uh, gaming and photo realistic, everything. And these are immersive experiences. And, you know, again, uh, what companies have figured out to bring it back to, to, to a little bit of business here, John, is that when you, uh, give people the right tools, they're more productive, they're more engaged and look, people are smart. They know what tools are available. And, you know, uh, the thing that actually is most representative of how a person thinks about the tools they have at their organization is actually the thing that's right in front of 'em. And so, you know, this ability for us to provide a pool set of solutions for organizations to keep their workforce productive, to run their applications and infrastructure securely anywhere they want. That's, that's a winning proposition. >>Michael trust was a big theme of your keynote yesterday. And when you acquired EMC and got VMware, it really changed the dynamic with regard to your ability to, into new parts of organizations. You became a much more strategic supplier. I, I would argue. And now with VMware as a separate company, do you feel like you have built up over the, you know, five or whatever years that muscle memory you kinda earn that trust. So how do you see the customer relationship with that regard to that integration that they, they loved the eco. So system competitors might not have loved it so much, but the customers really did love. In fact, the, the U S a, a gentleman yesterday kind of mentioned that, how do you see it? >>You know, customers, uh, are not as interested in the balance sheet and what you know, where different holdings are, what they, they want things to work together, right? And they want partnerships in ecosystems. And certainly, you know, with VMware, even before the combination, we had a powerful partnership. It obviously solidified in a super special way. And now we have this first and best relationship and I've remained the chairman of VMware and super excited about their future. But our ecosystem is incredibly broad. And you see that here in this show floor, and again, making things work together better and more effectively building these engineered solutions that allow people to very quickly deploy the kind of capabilities they want, whether it's, you know, snowflake now working with the on premise and the edge data and more of these, you know, multi-cloud, uh, eco of systems that are being built. It's not gonna be just one company >>You called the edge a couple years ago. You're really prominent in your, in your speeches. And your keynotes data also is a big theme. You mentioned data now, data engineering seems to be the hottest track of, of, of students graduating with data engineering skills, not data science, data engineering, large scale data as code concepts. So what's your vision now with data, how's that fitting into the solutions and the role of data, obviously data protection with cybersecurity data as code is becoming really part of that next big thing. >>Yeah. I mean, if, if you look at anything that is interesting in the world today, uh, at the center of it is data, right? Whether it's the blockchain or the defi or the AI drug discovery, or the autonomous vehicles or whatever you wanna do, there's data in, in, in the middle of that. And of course with that data, well, you've gotta manage it. You, you need compute engines, right? You need to be able to protect it, secure it. And, you know, that's kind of what we do, and we're not going to create all those solutions, but we are gonna be an enabling layer to allow that data to be accessed no matter, you know, where, where it is. And, and, and of course, you know, leading in storage continues to be a super important part of our business. Number one, larger than number two than number three, number four, combined, and, and most of number five as well, and, and growing share. And, and you saw today, the software defined innovations, allowing that, you know, data layer to exist across the edge, the colos, the OnPrem, and the public clouds >>Throughout a stat yesterday. I can't remember if it was a keynote of the analyst round table, but it was 9 million cell towers. And if I heard, right, you kinda look at those as potential data centers talk about that's >>Right. It it's actually 7 million, but, but probably will be 9 million and not, not too long, I don't have the update, but so yeah, the public clouds all together is about 600 data centers. They're about 7 million cellular base stations in the world. Every single one of those is becoming a, you know, multi access, edge compute node. And what are they putting in there? They're putting many data centers of compute and GPS and storage. And, you know, 5g is not about, uh, connecting people that was 4g and before 5g is about connecting things. And there are way more things than there are people, right? And, uh, you know, this, this, this edge is, is rapidly developing. You'll also have private 5g and you'll have, you know, again, embedded intelligence I believe is gonna be in everything this next decade is going to be about that intelligent, connected future, taking that data, turning it into useful outsides in insights and outcomes. And, you know, lots of new businesses will be existing. Businesses will be transformed and also disrupted. >>Yeah. I mean, I think that's so right on and not to pat ourselves on the back day, but we called that edge distributed computing a couple years ago on the cube. And that's, what's turning into the home with COVID you saw that become a workplace, basically compute center, these compute nodes, tying it together as we, what everyone's talking about right now. So as customers say, okay, I want to keep my operations steady, steady, and secure. How do I glue it together? How do I bring these compute node together? That seems to be the top question on, on top of people's minds. And they want it to be cloud native, which means they want it to run cloud-like and they want to connect these compute node together. That's a big discussion point. What's your view on, >>Well, you know, if you, if you sort of have a, a cloud here, a cloud there cloud everywhere, and you, you know, have lots of different Kubernetes frameworks, uh, and you've got, you know, everything is, is spread out, it's a disaster, right? And, and, and it's, it's a, it's a, it's a real challenge to manage all that. So what people are trying to do is create ruthless standardization. It's like, how do you drive cost out and get speed? It's ruthless standardization create consistent environments where you can operate the across all the different domains that, that you want. And so, uh, you know, this is what we're bringing together in, in, in the capabilities that we're delivering. >>And that chaos is great opportunity for you. Um, how are you feeling about VMware these days, new team, uh, give us the update there. >>Yeah. The team is doing well. You know, I think the tons message is resonating. You know, people want Kubernetes and, and, and container based apps, for sure. That's the main, you know, growth in, in, in, in, in new, in new workloads. Uh, but they also want it to work with what they have. Yeah. And they don't want it to be locked into one particular infrastructure. So software finding everything, making it run in all the public clouds, you know, we've had a great success with VxRail, you know, that, that absolutely continues. We have, uh, 200,000 plus nodes, 15,000 customers and growing, we have edge satellite nodes and we continue to work together in SD wan in software defined networking in VMware cloud foundation, uh, you know, expressed, uh, in, in, in all locations. >>You know, one of the things that we've been seeing with the trend towards, um, future of work, which is a big theme, here is a lot of managed services are popping up where the complexity is so ha high that customers want to manage services. Uh, and also the workforce of it's kind of changing. You got a younger generation coming in, how do you see that future of the workforce? The next level? It's not gonna be like, yesterday's it, it's gonna be distributed computing dashboard based. And then you've got these managed services, you know, need to have the training and expertise maybe to run something at scale. How do, how do you see that connecting? Cuz that seems to be another big trend people are talking about, Hey, it's complex someone manage it for me. And I want ease of views. I want the easy button in it. >>Yeah. Well we we've all been at this a while. So we can remember, you know, the beginnings of converged infrastructure and then hyperconverged, which wasn't that long go. And now we have consumption based business models. These are all along the trajectory of the easy button that you're talking about and customers really thinking about the value line, where are the things that really differentiate and add value for their business. And it's not below the value line in those infrastructure areas are creating that easy button with appliances, with consumption based models and allowing them to deploy the scarce resources. They have to the things that really drive their unique differe. And you know, if you look at our managed services flex on demand, all the sort of ancestors and predecessors of apex, those have been great businesses for us. And now with apex, we're kind of industrializing this and, and making it, you know, at scale for all >>Customers, you know, the three of us, we go back, we, we, our first interactions with you separately, we're in the nine. And then we reconnected in the 2012. I think it was Tarkin Mayer had a little breakout session with CIOs. You brought us to early on a Dell tech world in Austin. And of course it was, >>It was just Dell world. Then Dell >>Four, we had Dell tech, you and then EMC world in 2010 was our first cube. And now that's all come together here in Las Vegas. So, you know, it's been great. Uh, the three of us come together and so really appreciate that. Yeah. >>Awesome. Absolutely awesome. >>Well, you know, really appreciate you guys being here, the wonderful work you do in thank you in, you know, bringing out the, the, the stories and, and showing off and helping us show off the innovations that, you know, our team has been working on. You know, during the past year >>It's been great in conversations and, and on a personal note, it's been great to have, uh, chat with all the top people and your company. Appreciate it. Um, someone told me to ask you this question, I want to ask you, you, we've all seen waves of innovation cycles up and down. We're kind of on one. Now you're seeing an inflection point, this next gen, uh, computing and, and web three cultural shit F with workforces and distributed computing decentralization. You mentioned that DFI earlier, how do you see this wave coming? Cause we've seen cycles come and go.com. Bubble kind of looks the same as the web three NFTs and stuff. Now it seems to be Look different, but how do you see this next wave? Cuz looking back on all the other ones that you you have lived through and you rode >>Well. So, you know, the, the way I see it is is, uh, to some extent, these are like foundational layers that have to be built for the next phase to occur. And if you look at the sort of new companies that are being founded today, and we see a lot of those, you, you, you, you see'em, we invest in a bunch of 'em, you know, they're, they're not going and, and kind of redoing the old foundational layers, they're going deeply into vertical businesses and, and disrupting and adding value on top of those. And I think that's, that's really the, the point of, of technology, right? It's enabling human progress us in, in all fields, it's making us healthier. It's making us safer. It's making us more successful in everything that, that we as humans do. And so all these layers of technology are enabling further progress and I think it's absolutely gonna continue. It's all been super exciting. Yeah. You know, so far for the first several decades, but as I, as I believe it, it's, it's just a pre-game show. >>And it's clear your strategy is, is, is really building that foundation of a layer, hardening it, but making it flexible enough, anybody read your book, you're a technology, visionary. A lot of people put you in a, you know, finance bucket, but you can, you can see that you can connect the dots. And that's what you're doing with your foundation of layers. You that's where you're making the bets, isn't it? Uh, you don't can't predict the future. You've said that many times, but you can sort of see where it's going and be prepared for >>It. Well, you, you, you know, you think about any company in, in the industry or any public sector organization, right? Uh, they're, they're, they're wanting to evolve more quickly and transform more quick, more quickly. Right. And we can give them an infrastructure or set of tools, a set of capabilities to help them go faster. >>Yeah. And the other one thing in the eighties, when you started Dell and we were in college, there was no open source really then if look at the growth of open source, talk about those layers, open source, better Silicon GPS, faster, cheap >>More now and now we even have, uh, open source instruction sets for processors. So I mean the whole world's changing. It's exciting. You have people around the world working together. I mean, when you see our development teams, uh, whether they're in Israel or Ireland or Bangalore or Singapore, Hopton Austin, Silicon valley, you know, Taiwan, they're, they're all, they're all collaborating together and, you know, driving, driving innovation and, and, and our business is not that dissimilar from our customers >>Like great to have you in the queue. Great. To have a physical event. People are excited. I'm talking to people, Hey, haven't been back in Vegas in two years. Thanks for having this event. Great to see you. Thanks for coming on the cube. >>Absolutely. Thank you guys. >>Michael Dell here in the cube CEO of Dell technologies. I'm John far, Dave Volante. We'll be right back, more live coverage here at Dell tech world.
SUMMARY :
I'm John furry host of the cube with Dave Alon here with Michael Hey, John, Dave, great to be with you guys. Thank God you had all that supply for the, for the people having the remote remote end to work now Well, it turns out that, you know, What's the story that you cite to the CIOs saying, we're here to help you with that new architecture cloud Well, you know, customers want to go faster, right? And you said, well, why don't you just buy the company? And you know, as we make this shift to outcomes, And number one, number one, share in PCs, by the way, congratulations, again, hit that milestone. all the performance is enabling photo realistic, uh, you know, uh, And now with VMware as a separate company, do you feel like you have built up the kind of capabilities they want, whether it's, you know, snowflake now working with the on premise and how's that fitting into the solutions and the role of data, obviously data protection with cybersecurity And, and, and of course, you know, And if I heard, right, you kinda look at those as potential data centers talk about of those is becoming a, you know, multi access, And that's, what's turning into the home with COVID you saw that And so, uh, you know, this is what we're bringing together Um, how are you feeling about VMware these days, everything, making it run in all the public clouds, you know, How do, how do you see that connecting? So we can remember, you know, the beginnings of converged infrastructure Customers, you know, the three of us, we go back, we, we, our first interactions with you separately, It was just Dell world. So, you know, it's been great. Well, you know, really appreciate you guys being here, the wonderful work you do in thank you in, Cuz looking back on all the other ones that you you have And if you look at the sort of new companies that are being founded today, you know, finance bucket, but you can, you can see that you can connect the dots. And we can give them an source really then if look at the growth of open source, talk about those layers, open source, you know, driving, driving innovation and, and, and our business is not that dissimilar from our Like great to have you in the queue. Thank you guys. Michael Dell here in the cube CEO of Dell technologies.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Volante | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
Israel | LOCATION | 0.99+ |
Michael Dell | PERSON | 0.99+ |
Vegas | LOCATION | 0.99+ |
Ireland | LOCATION | 0.99+ |
2010 | DATE | 0.99+ |
Dave Alon | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Austin | LOCATION | 0.99+ |
Singapore | LOCATION | 0.99+ |
101 billion | QUANTITY | 0.99+ |
Bangalore | LOCATION | 0.99+ |
John far | PERSON | 0.99+ |
7 million | QUANTITY | 0.99+ |
2012 | DATE | 0.99+ |
9 million | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
41 | QUANTITY | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
three | QUANTITY | 0.99+ |
11 million pixels | QUANTITY | 0.99+ |
Michael Dell | PERSON | 0.99+ |
15,000 customers | QUANTITY | 0.99+ |
two years ago | DATE | 0.99+ |
five | QUANTITY | 0.99+ |
both companies | QUANTITY | 0.99+ |
more than a hundred billion dollars | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
200,000 plus nodes | QUANTITY | 0.99+ |
Silicon valley | LOCATION | 0.98+ |
EMC | ORGANIZATION | 0.98+ |
two years | QUANTITY | 0.98+ |
last year | DATE | 0.98+ |
eighties | DATE | 0.98+ |
Taiwan | LOCATION | 0.98+ |
about 600 data centers | QUANTITY | 0.97+ |
first cube | QUANTITY | 0.97+ |
today | DATE | 0.97+ |
Four | QUANTITY | 0.97+ |
Tarkin Mayer | PERSON | 0.97+ |
next decade | DATE | 0.96+ |
telco | ORGANIZATION | 0.95+ |
about 7 million | QUANTITY | 0.94+ |
one | QUANTITY | 0.92+ |
apex | ORGANIZATION | 0.92+ |
ING | ORGANIZATION | 0.92+ |
Dell Technologies | ORGANIZATION | 0.91+ |
U S a | ORGANIZATION | 0.9+ |
couple years ago | DATE | 0.9+ |
one thing | QUANTITY | 0.89+ |
one company | QUANTITY | 0.88+ |
first interactions | QUANTITY | 0.87+ |
nine | QUANTITY | 0.87+ |
9 million cell | QUANTITY | 0.86+ |
past year | DATE | 0.85+ |
pandemic | EVENT | 0.85+ |
wave | EVENT | 0.84+ |
single | QUANTITY | 0.84+ |
17% revenue | QUANTITY | 0.83+ |
first several decades | QUANTITY | 0.82+ |
COVID | TITLE | 0.81+ |