Image Title

Search Results for Kata 1.0:

Sirisha Kadamalakalva, DataRobot | AWS Marketplace Seller Conference 2022


 

>>Welcome back to the cubes coverage here in Seattle for AWS marketplace seller conference, the combination of the Amazon partner network, combined with the marketplace from the AWS partner organization, the APO and John Forer host of the queue, bringing you all the action and what it all means. Our next guest is Trisha kata, Malva, chief strategy officer at DataRobot. Great to have you. Thanks for coming on. >>Thank you, John. Great to be here. >>So DataRobot obviously in the big data business data is the big theme here. A lot of companies are in the marketplace selling data solutions. I just ran into snowflake person. I ran into another data analyst company, lot of, lot of data everywhere. You're seeing security. You're seeing insights a lot more going on with data than ever before. It's one of the most popular categories in the marketplace. Talk about DataRobot what you guys are doing. What's your product in there? Yeah, >>Absolutely. John. So we are an artificial intelligence machine learning platform company. We have been around for 10 years. This is this year marks our 10th anniversary and we provide a platform for data scientists and also citizen data scientists. So essentially wanna be data scientists on the business side to rapidly experiment with data and to get insights and then productionize ML models. So the 100% workflow that goes into identifying the data that you need for machine learning and then building models on top of that and operationalizing a, >>How big is the company, roughly employee count? What's the number in >>General general, about a thousand employees. And we have customers all over the world. Our biggest verticals are financial services, insurance, manufacturing, healthcare pharma, all the highly regulated, as well as our tech presence is also growing. And we have people spread across multiple geographies and I can't disclose a customer number, but needless to say, we have hundreds of customers across the >>World. A lot of customers. Yeah, yeah. You guys are well known in the industry have been following some of the recent news lately as well. Yeah. Obviously data's exploding. What in the marketplace are you guys offering? What's the pitch, someone hits the marketplace that wants to buy DataRobot what's the pitch. >>The pitch is if you're looking to get real value from your data science, personal investments and your data, then you have DataRobot that you can download from your AWS marketplace. You can do a free trial and essentially get from, get value from data in a matter of minutes and not months or quarters, that's generally associated with IML. And after that, if you want to purchase you, it's a private offer on, in the marketplace. So you need to call DataRobot representative, but AWS marketplace offers a fantastic distribution channel for us. >>Yeah. I mean, one of the things I heard Chris say, who's now heading up the marketplace and the partner network was the streamlining, a lot of the benefits for the sellers and for the buyers to have a great experience buyers. Clearly we see this as a macro trend, that's gonna only get stronger in terms of self-service buying bundling, having the console on AWS for low level services like infrastructure. But now you've got other business applications that like analytics applies to. You're seeing that work. Now he said things like than the keynote, I wanna get your reaction to like, we're gonna make this more like a C I C D pipeline. We're gonna have more native services built into AWS. What that means to me is that sounds like, oh, if I have a solution, like DataRobot, that can be more native into AWS level services. How do you see that working out for you guys is that play well for your strategy and your customers? What's the, what's the what's resonating with the >>Customers. It plays extremely well with the strategy. So I call this as a win, win, win strategy, win for DataRobot win for customers and win for AWS, which is our partner. And it's a win for DataRobot because the amount of people, the number of eyeballs that look at AWS marketplace, a significantly higher than, than the doors that we can go knock on. So it's a distribution multiplier for us. And the integration into AWS services that you're talking about. It is very important because in this day and age, we need to be interoperable with cloud player services that they offer, whether it is with SageMaker or Redshift, we support all of those. And it's a win for customers because customers, it is a very important growing buyer persona for DataRobot. Yeah. And they already have pre-committed spend with AWS and they can use the, those spend dollars for DataRobot to procure DataRobot. So it eases their procurement life cycle as >>Well. It's a forced multiplier on, on the revenue side, correct? I mean, as well as, as on the business front cost of sales, go down the cost of order dollar. Correct. This is good. Goodness. >>It's it's definitely sorry, just to finish my thought on the win for the partner for AWS. It's great win for them because they're getting the consumption from the partner side, to your point on the force multiplier. Absolutely. It is a force multiplier on the revenue side, and it's great for customers and us, because for us, we have seen that the deal size increases when there is the cloud commit that we can draw down for, for our customers, the procurement cycle shortens. And also we have multiple constituencies within the customers working together in a very seamless fashion. >>How has the procurement going through AWS helped your customers? What specific things are you seeing that are popping out as benefits to the customer? >>So from a procurement standpoint, we, we are early in our marketplace journey. We got listed about a year ago, but the amount of revenue that has gone through marketplace is pretty significant at DataRobot. We experienced like just in, by, I think this quarter until this quarter, we got like about 20 to 30 transactions that went through AWS marketplace. And that is significant within just a year of us operating on the marketplace. And the procurement becomes easier for our customers. Yeah. Because they trust AWS and we can put our legal paperwork through the AWS machine as well, which we haven't done yet. But if we do that, that'll be a further force multiplier because that's the, the less friction there is. >>I like how you say that it's a machine. Yeah. And if you think about the benefits too, like one of the things that I see happening, and I love to get your thoughts because I think this is what's happening here. Infrastructure services, I get that IAS done hardware I'm oversimplifying, but all the, all the goodness, but as customers have business apps and vertical market solutions, you got more AI involved. You need more data that's specialized for that use case. Or you need a business application. Those, you don't hear words like let's provision that app. I mean, your provision hardware and, and infrastructure, but the, the new net cloud native is that you provision turn on the apps. So you're seeing the wave of building apps are composing Lego blocks, if you will. So it seems like the customers are starting to assemble the solution, almost like deploying a service, correct. And just pressing a button. And it happens. This seems to be where the, the business apps are going. >>Yeah, absolutely. You agree for us? We are, we are a data science platform and for us being very close to the data that the customers have is very important. And where if, if the customer's data is in Redshift, we are close to there. So being very close to the hyperscale or ecosystem in that entire C I C D pipeline, and also the data platform pipeline is very important. >>You know, what's interesting is, is the data is such a big part of, I mean, DevOps infrastructure has code has been the movement for decade. Yeah. So throw security in there. It's dev SecOps. Yeah. That is the developer now. Yeah. They're running essentially what used to be it now the new ops is security and data. Yeah. You see, in those teams really level up to be highly high velocity data meshes, semantic layer. These are words I'm hearing in the industry around the big waves of data, having this mesh. Yeah. Having it connected. So you're starting to see data availability become more pervasive. And, and we see this as a way that's powering this next gen data science revolution where it's like the business person is now the data science person. >>That's exactly. That is, that is what DataRobot does the best. We were founded with the vision that we wanted to democratize the access to AI within enterprises. It shouldn't be restricted to a small group of people don't get me wrong. Data scientists also love DataRobot. They use DataRobot. But the mission is to enhance many, many hundreds of people within an organization to use data science, like how you use Tableau on a regular basis, how you use Microsoft Excel on a regular basis. We want to democratize AI. And when you want to democratize AI, you need to democratize access to data, which is, which could be stored in data marketplaces, which could be stored in data warehouses and push all the intelligence that we grab from that data into the E R P into the apps layer. Because at the end of the day, business users, customers consume predictions through applications layer. >>You know, it's interesting, you mentioned that comment about, you know, trying not to, to offend data scientists, it's actually a rising tide that the tsunami of data is actually making that population bigger too. Right. So correct. You also have data engineering, which has come out of the woodwork. We covered a lot on the cube, which is, you know, we call data as code. So infrastructure as code kind of a spoof on that. But the reality is that there's a lot more data engineering. I call that the smallest population. Those are the, those are the alphas, the alpha geeks. Yeah. Hardcore data operating systems, kind of education, data science, big pool growing. And then the users yeah. Are the new data science practitioners. Correct? Exactly. So kind of a, the landscape is you see that picture too, right? >>For sure. I mean, we, we have presence in all of those, right? Like data engineers are very important. Data scientists. Those are core users of DataRobot like, how can you develop thousands and hundreds of thousands of models without having to hand code? If you have to hand code, it takes months and years to solve one problem for one customer in one location. I mean, see how fast the microeconomic conditions are moving. And data engineers are very important because at the end of the day, yes, you do. You create the model, but you need to operationalize that model. You need to monitor that model for data drift. You need to monitor how the model is performing and you need to productionize the insights that you gain. And for that engineering effort is very important behind the scenes. Yeah. And the users at the end of the day, they are the ones who consume the predictions. >>Yeah. I mean the volume and, and the scale and scope of the data requires a lot of automation as well. Correct. Cause you had that on top of it. You gotta have a platform that's gonna do the heavy lifting. >>Correct. Exactly. The platform is we call it as an augmented platform. It augments data scientists by eliminating the tedious work that they don't want to do in their everyday life, which some of which is like feature engineering, right? It's a very high value add work. However, it takes like multiple iterations to understand which features in your data actually impact the outcome. >>This is where the SAS platform is a service is evolved and we call that super cloud, right. This new model where people can scale it out and up. So horizontally, scalable cloud, but vertically integrated into the applications. It's an integrator dilemma. Not so much correct innovators dilemma, as we say in the queue. Yeah. So I have to ask you, I'm a, I'm a buyer I'm gonna come to the marketplace. I want DataRobot why should they buy DataRobot what's in it for them? What's the key features of DataRobot for a company to hit the subscribe, buy button. >>Absolutely. Do you want to scale your data science to multiple projects? Do you want to be ahead of your competition? Do you want to make AI real? That is our pitch. We are not about doing data science for the sake of data science. We are about generating business value out of data science. And we have done it for hundreds of customers in multiple different verticals across the world, whether it is investment banks or regional banks or insurance companies or healthcare companies, we have provided real value out of data for them. And we have the knowhow in how to solve, whether it is your supply chain, forecasting, problem, demand, forecasting problem, whether it is your foreign exchange training problem, how to solve all these use cases with AI, with DataRobot. So if you want to be in the business of using your data and being ahead of your competitors, DataRobot is your tool log choice. >>Sure. Great to have you on the cube as a strategy officer, you gotta look at the chess board, right. And we're kind of in the mid game, I call it the cloud opening game was, you know, happened. Now we're in the mid game of cloud computing where you're seeing a lot of refactoring of opportunities where technologies and data is the key to success, being things secure and operationally, scalable, etcetera, et cetera. What's the key right now for the ecosystem as a strategy, look at the chessboard for data robots. Obviously marketplace is important strategy. Yeah. And bet for, for DataRobot. What else do you see for your company to be successful? And you could share with, with customers watching. >>Yeah. For us, we are in the intelligence layer, the data, the layer below us is the data layer. The layer about us is the applications and the engagement layer. DataRobot I mean, interoperability and ecosystem is important for every company, but for DataRobot it's extra important because we are in that middle of middle layer of intelligence. And we, we have to integrate with all different data warehouses out there enable our customers to pull the data out in a very, very faster way and then showcase all the predictions into, into their tool of choice. And from a chessboard perspective, I like your phrase of we are in the mid cycle of the cloud revolution. Yeah. And every cloud player has a data science platform, whether it is simple one or more complex one, or whether it has been around for quite some time or it's been latent features. And it is important for us that we have complimentary value proposition with all of them, because at the end of the day, we want to maximize our customer's choice. And DataRobot wants to be a neutral platform in supporting all the different vendors out there from a complementary standpoint, because you don't want to have a vendor lock in for your customers. So you create models in SageMaker. For example, you monitor those in DataRobot or you create models in DataRobot and monitor those in AWS so that you have to provide like a very flexible >>That's a solution architecture. >>Correct? Exactly. You have to provide a very flexible tech stack for your customers. >>Yeah. That's the choice. That's the choice. It's all good. Thank you for coming on the cube, sharing the data robot. So I really appreciate it. Thank >>You for coming. Thank you very much for the opportunity. >>Okay. Breaking it all down with the partners here, the marketplace, it's the future, obviously where people are gonna buy the buyers and sellers coming together, the partner network and marketplace, the big news here at 80 seller conference. I'm John ferry with the cube will be right back with more coverage after this short break.

Published Date : Sep 21 2022

SUMMARY :

AWS partner organization, the APO and John Forer host of the queue, bringing you all the action and So DataRobot obviously in the big data business data is the big theme here. So the 100% workflow that goes into identifying the data a customer number, but needless to say, we have hundreds of customers across the What in the marketplace are you guys offering? And after that, if you want to purchase you, it's a private offer on, out for you guys is that play well for your strategy and your customers? a significantly higher than, than the doors that we can go knock on. cost of sales, go down the cost of order dollar. It is a force multiplier on the revenue side, And the procurement becomes easier for our customers. So it seems like the customers are starting to assemble the solution, if the customer's data is in Redshift, we are close to there. That is the developer now. But the mission is to enhance So kind of a, the landscape is you see that picture too, right? at the end of the day, yes, you do. You gotta have a platform that's gonna do the heavy lifting. It augments data scientists by eliminating the tedious What's the key features of DataRobot for a company to hit the subscribe, So if you want to be in the business of using your data and being ahead of your competitors, the mid game, I call it the cloud opening game was, you know, happened. because at the end of the day, we want to maximize our customer's choice. You have to provide a very flexible tech stack for your customers. That's the choice. Thank you very much for the opportunity. I'm John ferry with the cube will be right back with more coverage after this short break.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AWSORGANIZATION

0.99+

JohnPERSON

0.99+

Sirisha KadamalakalvaPERSON

0.99+

ChrisPERSON

0.99+

SeattleLOCATION

0.99+

APOORGANIZATION

0.99+

thousandsQUANTITY

0.99+

100%QUANTITY

0.99+

AmazonORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

DataRobotORGANIZATION

0.99+

Trisha kataPERSON

0.99+

MalvaPERSON

0.99+

hundreds of customersQUANTITY

0.99+

one problemQUANTITY

0.99+

one customerQUANTITY

0.99+

one locationQUANTITY

0.99+

10th anniversaryQUANTITY

0.98+

TableauTITLE

0.98+

DataRobotTITLE

0.98+

hundredsQUANTITY

0.98+

a yearQUANTITY

0.98+

30 transactionsQUANTITY

0.97+

oneQUANTITY

0.96+

John ferryPERSON

0.96+

about 20QUANTITY

0.95+

hundreds of thousands of modelsQUANTITY

0.94+

SageMakerTITLE

0.93+

this quarterDATE

0.93+

John ForerPERSON

0.92+

about a thousand employeesQUANTITY

0.92+

aboutDATE

0.89+

decadeQUANTITY

0.87+

LegoORGANIZATION

0.86+

monthsQUANTITY

0.86+

DevOpsTITLE

0.86+

a year agoDATE

0.85+

Marketplace Seller ConferenceEVENT

0.84+

80 seller conferenceEVENT

0.83+

this yearDATE

0.78+

10 yearsQUANTITY

0.78+

yearsQUANTITY

0.76+

SecOpsTITLE

0.71+

of dataQUANTITY

0.68+

peopleQUANTITY

0.67+

RedshiftTITLE

0.67+

IASTITLE

0.65+

big waves of dataEVENT

0.63+

DataRobotCOMMERCIAL_ITEM

0.62+

devTITLE

0.59+

2022DATE

0.57+

Protect Your Data & Recover from Cyberthreats & Ransomware in Minutes


 

>>Welcome back to the cubes coverage of H P S. Green Lake announcement. We've been following Green Lake and the cadence of announcements making. Now we're gonna talk about ransomware, ransomware become a household term. But what people really don't understand is that virtually any bad actor can become a ransomware criminal by going on the dark web hiring a ransomware as a service sticking, putting a stick into a server and taking a piece of the action and that is a really insidious threat. Uh, the adversaries are extremely capable, so we're going to dig into that with Omar assad, who's the storage platform, lead cloud data services at H P E and Deepak verma vice president of product Zito, which is now an H P E company Gentlemen, welcome to the cube. Good to see you. Thank you. >>Thank you. Welcome. Pleasure to be here. So >>over you heard my little narrative upfront. How does the Xarelto acquisition fit into that discourse? >>Thank you. Dave first of all, we're extremely excited to welcome Sir toe into the HP family. Uh, the acquisition of Puerto expands the Green Lake offerings from H P E uh, into the data protection as a service and ransomware protection as a service capabilities and it at the same time accelerates the transformation that the HP storage businesses going through as it transforms itself into more of a cloud native business, which sort of follows on from the May 4th announcements that you helped us cover. Uh, this enables the HP sales teams to now expand the data protection perimeter and to start offering data protection as a service and ransomware as a service with the best in class technologies uh, from a protection site as well as from ransomware recovery side of the house. And so we're all the way down already trying to integrate uh, you know, the little offerings as part of the Green lake offerings and extending support through our services organization. And the more of these announcements are gonna roll out later in the month. >>And I think that's what you want to see from it as a service offering. You want to see a fast cadence of new services that are not a box by a box that are applying. No, it's services that you want to access. So let's, let's talk about before we get into the tech, can we talk about how you're helping customers deal with ransomware? Maybe some of the use cases that you're seeing. >>First of all, extremely excited to be part of the HP family now. Um, Quick history and that we've been around for about 11 years. We've had about 9000 plus customers and they all benefit from essentially the same technology that we invented 11 years ago. First and foremost, one of the use cases has been continuous data protection. So were built on the CdP platform, which means extremely low RTO S and R P O S for recovery. I'll give you example there um, United Airlines is an application that cost them $1 million dollars for every hour that they're down. They use traditional approaches. That would be a lot of loss with Zito, we have that down two seconds of loss in case and the application goes down. So that's kind of core and fundamental to our plaque. The second uh critical use case that for us has been simplicity. A lot of customers have said we make the difficult, simple. So DRS is a complex uh process. Um, give you an example there. Hcea Healthcare Consolidated four different disaster recovery platforms into a single platform in Puerto and saved about $10 million dollars a year. So it's making that operations of having disaster recovery process is much simpler. Um the third kind of critical use case for us as uh, the environment has evolved as the landscape has involved has been around hybrid cloud. So being able to take customers to the platforms that they want to go to that's critical for us And for our customers an example, there is Kingston technology's so Kingston tried some competitive products to move to Azure, it would take them about 24 hours to recover 30 VMS or so with zero technology. They will get about all their 1000 VMS up in Azure instantaneously. So these are three use cases that were foundational. Built. Built the company in the tech. >>Nice. Thank you. Thank you for that. So simple works well these days, especially with all this complexity we have to deal with. Can we get into the secret sauce a little bit. I mean CdP has been around forever. What do you guys do that? That's different. Maybe you can talk about that. Sure. >>Um it's cdp based, I think we've perfected the technology. It's less about being able to just copy the data. It's more about what you do when things go bump. We've made it simpler with driven economies of scale lower and being platform agnostic. We've really brought that up across to whatever platforms once upon a time it was moving from physical to virtual or even across different virtualization platforms and then being able to move across to whatever cloud platform customer may want or or back >>to cbP continuous data protection by the way for the audience that may not know that go ahead. And >>one of the additional points that I want to add to the box comment over here is the the basics of platform independence is what really drew uh hp technologists into the technology because you know, one of the things we have many, we have the high end platform with the H B electra nine Kv of the electro six kids the midrange platform. Then we have a bunch of file and object offerings on the side. What zero does it University universally applies to all those technologies and along with, you know, as you pair them up with our computer offerings to offer a full stack but now the stack is disaster recovery capable. Natively with the integration of certo, you know, one of the things that, you know, Deepak talked about about the as your migrations that a lot of the customers are talking about cloud is also coming up as a D our use case for a lot of our customers, customers, you know, you know, as we went through thousands of customers interviews one of the, one of the key things that came back was investing in a D our data center which is just waiting there for a disaster to happen. It's a very expensive insurance policy. So absurd. Oh, through its native capabilities allows customers to do is to just use public cloud as a D our target and and as a service, it just takes care of all the format conversions and recoveries and although that's completely automated inside the platform and and we feel that, you know, when you combine this either at the high end of data center storage offering or the middle age offering with this replication, D. R. And ransomware protection built into the same package, working under the same hood, it just simplifies and streamlines the customers deployment. >>Come here a couple of things. So first of all historically, if you wanted to recover to appoint within let's say, you know, 10 seconds, five seconds you have to pay up. Big time. Number one. Number two is you couldn't test your D. R. It was too risky. So people just had it in, they had a checkbox on compliance but they actually couldn't really test it because they were afraid they were going to lose data. So it sounds like you're solving both of those problems or >>or you know we remember the D. R. Test where it was a weekend. It was an event right? It was the event and at the end of july that the entire I. T. Organizing honey >>it's not gonna be home this weekend. Exactly what >>we've changed. That is a click of a button. You can D. R. Test today if you want to you can have disaster recovery still running. You can D. R. Test in Azure bring up your environment an isolated network bubble, make sure everything's running and bring it and bring it down. The interesting thing is the technology was invented back when our fear in the industry was losing a data center was losing power was catastrophic, natural disasters. But the technology has lent itself very well to the new threats which which are very much around ransomware as you mentioned because it's a type of disaster. Somebody's going after your data. Physical servers are still around but you still need to go back to a point in time and you need to do that very quickly. So the technology has really just found itself uh appealing to new challenges. >>If a customer asks you can I really eliminate cyber attacks, where should I put my my if I had 100 bucks to spend. Should I spend it on you know layers and defense should I spend it on recovery. Both, what would you tell them? >>I think it's a balanced answer. I think prevention is 100% impossible. Uh It's really I'd say spend it in in thirds. You want to spend a third of it and and prevention a third of it maybe in detection and then a third of it in uh recovery. So it's really that balancing act that means you can't leave the front door open but then have a lot of recovery techniques invested in. It has to be it has to be a balance and it's also not a matter of if it's a matter of when so we invest in all three areas. Hopefully two of them will work to your advantage. >>You dave you you should always protect your perimeter. I mean that that goes without saying but then as you invest in other aspects of the business, as Deepak mentioned, recovery needs to be fast and quick recovery whether from your recovering from a backup disaster. Are you covering from a data center disaster a corrupted file or from a ransomware attack. A couple of things that zero really stitches together like journal based recovery has been allowed for a while but making journal based recovery platform independent in a seamless fashion with the click of a button within five seconds go back to where your situation was. That gives you the peace of mind that even if the perimeter was breached, you're still protected, you know, five minutes into the problem And, and that's the peace of mind, which along with data protection as a service, disaster recovery as a service and now integrating this, you know, recovery from ransomware along with it in a very simple, easy to consume package is what drew us into the >>more you can do this you said on the use the cloud as a target. I could use the cloud as an air gap if I wanted to. It sounds like it's cloud Native, correct? Just wrap your stack in kubernetes and shove it in the cloud and have a host and say we're cloud to No, really I'm serious. So >>absolutely, we we looked at that approach and that that's where the challenge comes in, Right? So I give you the example of Kingston technology just doesn't scale, it's not fast enough. What we did was developed a platform for cloud Native. We consume cloud services where necessary in order to provide that scalability. So one example in Azure is being able to use scale set. So think about a scenario where you just declare a disaster, you've got 1000 VMS to move over, we can spin up the workers that need to do the work to get 1000 VMS spin them down. So you're up and running instantaneously and that involves using cloud Native uh tools and technologies, >>can we stay on that for a minute, So take take us through an example of what life was like would be like without zero trying to recover and what it's like with Puerto resources, complexity time maybe you could sort of paint a picture. Sure. >>Let me, I'll actually use an example from a customer 10 Kata. They uh develop defensive fabrics, especially fabric. So think about firefighters, think about our men and women abroad that need protective clothing that developed the fibers behave. They were hit by ransomware by crypto locker. That this was before zero. Unfortunately it took they took about a two week uh data loss. It took them weeks to recover that environment, bring it back up and the confidence was pretty low. They invested in, they looked at our technology, they invested in the technology and then they were hit with a different variant of crypto locker immediately. The the IT administrators and the ITS folks there were relieved right, they had a sense of confidence to say yes we can recover. And the second time around they had data loss of about 10 seconds, they could recover within a few minutes. So that's the before and after picture giving customers that confidence to say yep, a breach happened, we tried our best but now it's up to recovery and I can recover without having to dig tapes out from some vault and hopefully have a good copy of data sitting there and then try that over and over again and there's a tolerance right before a time before which business will not be able to sustain itself. So what we want to do is minimize that for businesses so that they can recover as quickly as possible with as little data loss as possible. >>Thank you for that. So, Omar, there's a bigger sort of cyber recovery agenda that you have as part of, of green lake, I'm sure. What, what should we expect, what's next? Where do you want to take this? >>So uh excellent question point in the future day. So one of the things that you helped us, uh you know, unveil uh in May was the data services. Cloud console. Data services. Cloud console was the first uh sort of delivery as we took the storage business as it is and start to transform into more of a cloud native business. We introduced electra uh which is the cloud native hardware with the customers buy for persistent storage within their data center. But then data services, cloud console truly cemented that cloud operational model. Uh We separated the management from, from the devices itself and sort of lifted it up as a sas service into the public, public cloud. So now what you're gonna see is, you know, more and more data and data management services come up on the data services. Cloud console and and zero is going to be one of the first ones. Cloud physics was another one that we we talked about, but zero is the is the true data management service that is going to come up on data services, cloud console as part of the Green Lake services agenda that that HP has in the customer's environ and then you're gonna see compliance as a service. You're going to see data protection as a service. You're gonna see disaster recovery as a service. But the beautiful thing about it is, is choice with simplicity as these services get loaded up on data services, clown console. All our customers instantly get it. There's nothing to install, there's nothing to troubleshoot uh, there's nothing to size. All those capabilities are available on the console, customers go in and just start consuming Xarelto capabilities from a management control plane, Disaster recovery control plan are going to be available on the data services, cloud console, automatically detecting electro systems, rian Bear systems, container based systems, whichever our customers have deployed and from there is just a flip of a button. Another way to look at it is it sort of gives you that slider that you have data protection or back up on one side, you've got disaster recovery on one side, you've got ransomware protection on on the extreme right side, you can just move a slider across and choose the service level that you want without worrying about best practices, installation, application integration. All of that just takes control from the data services, cloud concepts. >>Great, great summary because historically you would have to build that right now. You can buy it as a service. You can programmatically, you know, deploy it and that's a game changer. Have to throw it over the fence to some folks. That's okay. Now, you know, make it make it work and then they change the code and you come back a lot of finger pointing. It's now it's your responsibility. >>Absolutely. Absolutely. We're excited to provide Zito continue provides the desert of customers but also integrate with the Green Green Lake platform and let the rest of Green Lake customers experience some of the sort of technology and really make that available as a service. >>That's great. This is a huge challenge for customers. I mean they do, I pay their ransom. Do not pay the ransom. If I pay the ransom the FBI is going to come after me. But if I don't pay the ransom, I'm not gonna get the crypto key. So solutions like this are critical. You certainly see the president pushing for that. The United States government said, hey, we got to do a better job. Good job guys, Thanks for for sharing your story in the cube and congratulations. Thank >>you. Thank you David. >>All right. And thank you for watching everybody. Uh this is the, I want to tell you that everything that you're seeing today as part of the Green Lake announcement is going to be available on demand as part of the HP discover more. So you got to check that out. Thank you. You're watching the cube. >>Mhm mm.

Published Date : Sep 28 2021

SUMMARY :

Uh, the adversaries are extremely capable, so we're going to dig into that with Omar assad, Pleasure to be here. over you heard my little narrative upfront. itself into more of a cloud native business, which sort of follows on from the May 4th announcements that you And I think that's what you want to see from it as a service offering. First and foremost, one of the use cases has been Thank you for that. It's more about what you do when things go bump. to cbP continuous data protection by the way for the audience that may not know that go ahead. technologists into the technology because you know, one of the things we have many, we have the high end platform with So first of all historically, if you wanted to recover to appoint within let's say, or you know we remember the D. R. Test where it was a weekend. it's not gonna be home this weekend. back to a point in time and you need to do that very quickly. Both, what would you tell them? So it's really that balancing act that means you can't leave the front door You dave you you should always protect your perimeter. more you can do this you said on the use the cloud as a target. So think about a scenario where you just declare a disaster, you've got 1000 VMS to move over, complexity time maybe you could sort of paint a picture. So that's the before and after picture giving customers that confidence to Thank you for that. So one of the things that you You can programmatically, you know, deploy it and that's a game changer. of the sort of technology and really make that available as a service. If I pay the ransom the FBI is going to come after me. Thank you David. So you got to check that out.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

OmarPERSON

0.99+

100 bucksQUANTITY

0.99+

FBIORGANIZATION

0.99+

1000 VMSQUANTITY

0.99+

H P EORGANIZATION

0.99+

DeepakPERSON

0.99+

May 4thDATE

0.99+

10 secondsQUANTITY

0.99+

HPORGANIZATION

0.99+

100%QUANTITY

0.99+

five minutesQUANTITY

0.99+

United AirlinesORGANIZATION

0.99+

five secondsQUANTITY

0.99+

KingstonORGANIZATION

0.99+

two secondsQUANTITY

0.99+

MayDATE

0.99+

second timeQUANTITY

0.99+

FirstQUANTITY

0.99+

$1 million dollarsQUANTITY

0.99+

Omar assadPERSON

0.99+

BothQUANTITY

0.99+

AzureTITLE

0.99+

three use casesQUANTITY

0.99+

firstQUANTITY

0.99+

oneQUANTITY

0.98+

DavePERSON

0.98+

todayDATE

0.98+

secondQUANTITY

0.98+

about 10 secondsQUANTITY

0.98+

30 VMSQUANTITY

0.98+

11 years agoDATE

0.98+

PuertoLOCATION

0.98+

thousands of customersQUANTITY

0.97+

Hcea HealthcareORGANIZATION

0.97+

bothQUANTITY

0.97+

ZitoORGANIZATION

0.97+

zeroQUANTITY

0.97+

about 11 yearsQUANTITY

0.97+

United States governmentORGANIZATION

0.96+

about 24 hoursQUANTITY

0.96+

thirdQUANTITY

0.96+

Green LakeORGANIZATION

0.96+

third kindQUANTITY

0.95+

D. R. TestEVENT

0.94+

Green LakeLOCATION

0.94+

H P S. Green LakeORGANIZATION

0.93+

about a two weekQUANTITY

0.93+

threeQUANTITY

0.93+

about 9000 plus customersQUANTITY

0.93+

about $10 million dollars a yearQUANTITY

0.93+

single platformQUANTITY

0.92+

Cloud physicsTITLE

0.91+

one sideQUANTITY

0.89+

XareltoTITLE

0.89+

one exampleQUANTITY

0.88+

10 KataORGANIZATION

0.86+

minutesQUANTITY

0.85+

end of julyDATE

0.84+

Deepak vermaPERSON

0.83+

two ofQUANTITY

0.82+

first onesQUANTITY

0.82+

LakeORGANIZATION

0.81+

PuertoORGANIZATION

0.81+

Green Green LakeORGANIZATION

0.78+

Number twoQUANTITY

0.78+

a minuteQUANTITY

0.78+

Joe Fernandes, Red Hat | Red Hat Summit 2020


 

>> From around the globe, it's the CUBE with digital coverage of Red Hat Summit 2020 brought to you by Red Hat. >> Hi, I'm Stu Miniman, and this is the CUBE's coverage of a Red Hat Summit 2020 happening digitally. We're connecting with Red Hat executives, thought leaders, practitioners, wherever they are around the globe, bringing them remotely into this online event. Happy to welcome back to the program, Joe Fernandez, who's the Vice President and General Manager, of Core Cloud Platforms with Red Hat. Joe, thanks so much for joining us. >> Yeah, thanks for having me. Glad to be here. >> All right, so, Joe, you know, Cloud, of course, has been a conversation we've been having for a lot of years. When I went to Red Hat Summit last year, when I went to IBM, I think last year, there was discussion of moving from kind of chapter one, if you will, to chapter two. Some of the labels that we put on things back in the early days, like Hybrid Cloud and Multicloud, they're coming into a little bit clearer picture. So, let's just give a high level, what you're seeing from your customers when they talk about Hybrid and Multicloud environment? What does that mean to your customers? And therefore, how is Red Hat meeting them where they are? >> Yeah, sure. So, Red Hat obviously, serves an enterprise customer base. And what we've seen in that customer base, really since the start and it's really informed our strategy, is the fact that all their applications aren't going to run in one place, right? So they're really employing a hybrid class strategy, a Hybrid and Multicloud strategy, that spans from their data centers out to a public cloud, typically then out to multiple public clouds as their cloud investments grow, as they move more applications. And now, even out to the edge for many of those customers. So that's the newest footprint that we're getting asked about. So really we think of that as the open hybrid cloud. And you know, our goal is really to provide a consistent platform for applications regardless of where they run across all those environments. >> Yeah. Let's get down a second on that because we've had consistency for quite a while. You look at the largest cloud provider out there, they said, hybrid environment, will give you the exact same hardware that we're running in the public cloud of your bet. You know, that in your environment. Of course, Red Hat's a software company. You've lived across lots of platforms. We're going to Red Hat's entire existence. So, you know, where is that consistency needed? How do you, well, think about how Red Hat does things? Maybe the same and a little different than some of the other players that are then, positioning and even repositioning their hybrid story over the last year or so. >> Yeah. So, we're really excited to see a lot of folks in the industry, including all the major public cloud providers are now talking about Hybrid and talking about these types of initiatives that we've been talking about for quite some time. But yeah, it's a little bit different when we talk about Hybrid Cloud, when we talk about Multicloud, we're talking about being able to run not just in one public cloud and then in a non-premise clients that mirrors that cloud. We're really talking about being able to run across multiple clouds. So having that consistency across, running in, say Amazon to Azure to Google, and then carrying that into your on-premise environments, whether that's on Bare Metal, on VMware, on OpenStack, and then, like I said, out out to the edge, right? So that consistency is important for people who are concerned about how their applications are going to operate in these different environments. Because otherwise, they'd have to manage those differences themselves. I'm speaking as part of Red Hat, right? This is what the company was built on, right? In 20 years ago, it was all about Linux bringing consistency for enterprise applications running across x86 hardware, right? So regardless of who your OEM vendor was, as long as you're building to the x86 standard and leveraging Linux as a base, Red Hat Enterprise Linux became that same consistent operating environment for applications, which is important for our software vendors, but also more importantly for customers themselves as they yep those apps into production. >> Yeah, I guess, you know, last question I have for kind of just the landscape out there. We've been talking for a number of years. When you talk to practitioners, they don't get caught up in the labels that we use in the industry. Do they have a cloud strategy? Yes, most companies have a cloud strategy, and if you ask them is their cloud strategy same today, as it was a quarter ago or a year ago, they say, of course not. Everything's changed. We know in today's day and age, what I was doing a month ago is probably very different from what I am doing today. So, I know you've got a survey that was done of enterprise users. I saw it when it came out a month ago. And, you know, some good data in there. So, you know, where are we? And what data do you have to share with us on kind of the customer adoption with (mumbles). >> Yeah, so I think, you know, we put out a survey not too long ago and we started as, I think, over 60% of customers were adopting a hybrid cloud strategy exactly as I described. Thinking about their applications in terms of, in an environment that spans multiple cloud infrastructures, as well as on-premise footprints. And then, you know, going beyond that, we think that number will grow based on what we saw in that survey. That just mirrors the conversations that I've had with customers, that many of us here at Red Hat have been having with those same customers over the years. Because everybody's in a different spot in terms of their transformation efforts, in terms of their adoption of cloud technologies and what it means for their business. So we need to meet customers where they're at, understand that everybody's at a different spot and then make sure that we can help them make that transition. And it's really an evolution, as opposed to , I think, some people in the past might've thought of as a revolution where all the data centers are going to shut down and everything's going to move all at once. And so helping customers evolve. And that transition is really what Red Hat is all about. >> Yeah. And, so often, Joe, when I talk to some of the vendors out there, when you talk about Hybrid, you talk about Multicloud, it's talking about something you mentioned, it's a box, it's a place, it's, you know, the infrastructure discussion. But when I've been having conversations with a lot of your peers of these interviews for Red Hat Summit. We know that, it's the organization and it's the applications that are hugely important as these changes go and happen. So talk a little bit about that. What's happening to the organization? How are you helping the infrastructure team keep up and the app dev team move forward? >> Yeah, so first, I'll start with, that on the technology side, right? One of the things that that has enabled this type of consistency and portability has been sort of the advent of Linux containers as a standard packaging format that can span across all these different (mumbles), right? So we know that Linux runs in all these different footprints and Linux containers, as a portable packaging format, enables that. And then Kubernetes enables customers to orchestrate containers at scale. So that's really what OpenShift is focused on, is delivering an enterprise Kubernetes platform. Again, spanning all these environments that leverages container-based packaging, provides enterprise Kubernetes orchestration and management, to manage in all those environments. What that then also does on the people front is bring infrastructure and operations teams together, right? Because Kubernetes containers represents the agility for both sides, right? Or application developers, it represents the ability to pay their application and all their dependencies. And know that when they run it in one environment, it will be consistent with how it runs in other environments. So eliminating that problem of, works on my machine, but it doesn't work, you know, in prod or what have you. So it brings consistency for developers. Infrastructure teams, it gives them the ability to basically make decisions around where the best places to run these applications without having to think about that from a technology perspective, but really from things that should matter more, like cost and convenience to customers and performance and so forth. So, I think we see those teams coming together. That being said, it is an evolution in people and process and culture. So we've done a lot of work. We launched a global transformation office. We had previously launched a Red Hat open innovation labs and have done a lot of work with our consulting services and our partners as well, to help with, sort of, people in process evolutions that need to occur to adopt these types of technologies as well as, to move towards a more cloud native approach. >> All right. So Joe, what one of the announcements that made it the show, it is talking about how OpenShift is working with virtualization. So, I think back to the earliest container days, there was a discussion of, "oh, you know, Docker and containers, "it kills VM." Or you know, Cloud of course. Some Cloud services run on VMs, other run on containers, they're serverless. So there's a lot of confusion out there as to. >> Yep. >> What happened, we know in IT, no technology ever dies, everything's always additive. It's figuring out the right solutions and the right bet. So, help us understand what Red Hat is doing when it comes to virtualization in OpenShift and Kubernetes and, how is your approach different than some of what we've already seen in the marketplace? >> Yeah, so definitely we've seen just explosive adoption of containers technology, right? Which has driven the OpenShift business and Red Hat's business overall. So, we expect that to continue, right? More applications moving towards that container-based, packaging and deployment model and leveraging Kubernetes and OpenShift to manage those environments. That being said, as you mentioned, virtualization has been around for a really long time, right? And, predominantly, most applications, today, are running virtualized. And so some of them have made the transition to containers or were built a container native from the start. But many more are still running in VM based environments and may never make that switch. So, what we were looking at is, how do we manage this sort of hybrid environment from the application perspective where you have some applications running in containers, other applications running in VMs? We have platforms like Red Hat, OpenStack, Red Hat Virtualization that leveraged the KVM hypervisor and Red Hat Enterprise Linux to serve apps running in a VM based environment. What we did with Kubernetes is, instead, how could we innovate to have convergence on the orchestration and management fund? And we leveraged the fact that, KVM, you know, a chosen hypervisor, is actually a Linux process that can itself be containerized. And so by running the hypervisor in a container, we can then span VMs that could be managed on that same platform as the containers run. So what you have in OpenShift Virtualization is the ability to use Kubernetes to manage containerized workloads, as well as, standard VM based workloads. And these are full VMs. These aren't micro VMs or, you know, things like Firecracker Kata Container. These are standard VMs that could be, well, Windows guests or Linux guests, running inside those VMs. And so it helps you basically, manage that type of environment where you may be moving to containers and more cloud native approach, but those containers need to interact or work with applications that are still in a VM based deployment environment. And we think it's really exciting, we've demoed it at the last Red Hat Summit. We're going to talk about it even more here, in terms of how we're going to bring those products to market and enable customers. >> Okay, yeah, Joe, let me make sure I understand this because as you said, it is a different approach. So, number one, if I'm moving towards a (mumbles) management solution, this is going to fit natively into what I'm doing. It's not taking some of my traditional management tools and saying, "oh, I also get some visibility containers." There's more, you know, here's my Kubernetes solution. And just some of those containers happen to be virtualized. Did I get that piece right? >> Yeah, I think it's more like... so we know that Kubernetes is going to be in in the environment because we know that, yeah, people are moving application workloads to standard Linux containers. But we also know that virtual machines are going to still exist in that environment. So you can think about it as, how would we enable Kubernetes to manage a virtual machine in the same way that it manages a Linux container? And, what we do there, is we actually, put the VM inside the container, right? So because the VM, specifically with (mumbles) is just a Linux process, and that's what a Linux container is. It's a Linux process, right? So you can run the hypervisor, span the virtual machines, inside of containers. But those virtual machines, are just like any other VM that would run in OpenStack or Red Hat Virtualization or what have you. And you could, vSphere for example. So those are traditional virtual machines, that are now being managed in a Kubernetes environment. And what we're seeing is sort of, this evolution of Kubernetes to take on these new types of workloads. VMs is just one example, of something that you can now manage with Kubernetes. >> Okay. And, help me understand what this means to really the app dev in my application portfolio. Because you know, the original promise of virtualization was, I can just stick my application in a VM and I never need to think about it ever again. And well, that was super helpful when windows NT was going end of life. In 2020, we do find that most companies do want to update their applications, and they are talking about, do I refactor them? Do I make them microservices architecture? I don't want to have that iceberg of an application that I'm just dragging along slowly into the new world. So. >> Yeah. >> What is this virtualization integration with Kubernetes? You mean for the AppDev and the applications? >> Yeah, sure, so what we see customers doing, what we see the application development team is doing is modernizing a lot of their existing applications, right? So they're taking traditional monolithic applications or end tier, like the applications that may run in a VM based environment and they're moving them towards more of a distributed architecture leveraging microservices based approach. But that doesn't happen all at once either, right? So, oftentimes what you see is your microservices, are still connected to VM based applications. Or maybe you're breaking down a monolithic application. The core is still running in a VM, but some of those business functions have now been carved out and containerized. So, you're going to end up in a hybrid environment from the application perspective in terms of how these applications are packaged, and deployed. The question is, what does that mean for your deployment architecture? Does it mean you always have to run a virtualization platform and a container platform together? That's how it's done today, right? OpenShift and Kubernetes run on top of vSphere, they run on top of Amazon and Azure and Google bands, and on top of OpenStack. But what if you could actually just run Kubernetes directly on Bare Metal and manage those types of workloads? That's really sort of the idea. A whole bunch of virtualization solution was based on is, let's just merge VMs natively with Kubernetes in the same way that we manage containers. And then, it can facilitate for the application developer. This evolution of apps that are running in one environment towards apps that are running essentially, in a hybrid environment from how they're packaged and deployed. >> Yeah, absolutely, something I've been hearing for the last year or so, that hybrid deployment, pulling apart application, sometimes it's even, the core piece as you said, is on premises and then I might have some of the more transactional pieces happening in the public cloud. So really interesting. So, how long has Red Hat been working on this? My (mumbles), something, you know, I'm familiar with in the CNCF. I believe it has been around for a couple of years. >> Yeah. >> So talk to us about just kind of how long it took to get here and, fully support stateful applications now. What's the overall roadmap look like? >> Yeah, so, so (mumbles) as a open source project was launched more than two years ago now. As you know, Red Hat really drives all of our development upstream in the open source community. So we launched (mumbles) project. We've been collaborating with other vendors and even customers on that. But then, you know, over time we then decided, how do we bring these technologies to market, which technologies make sense to bring the market? So, (mumbles) is the open source project. OpenShift and OpenShift Virtualization, which is what this feature is referred to commercially, is the product that then we would ship and support for running this in production environments. The capabilities, right. So, I think, those have been evolving as well. So, virtual machines have a specific requirements in terms of not only how they're deployed and managed, but how they connect to storage, how they connect networking, how do you do things like fencing and all sorts of live migration and that type of thing. We've been building out those types of capabilities. They're certainly still more to do there. But it's something that we're really excited about, not just from the perspective of running VMs, but just even more broadly from the perspective of how Kubernetes is expanding to take on new workloads, right? Because Kubernetes has moved far beyond just running, cloud native applications, today, you can run stateful services in containers. You can run things like AI and machine learning and analytics and IoT type services. But it hasn't come for free, right? This has come through a lot of hard work in the Kubernetes community, in the various associated communities, the container communities, communities like (mumbles). But it's all kind of trying to leverage that same automation, that same platform to just do more things. The cool thing is, it'll not just be Red Hat talking about it, but you'll see that from a lot of customers that are doing sessions at our summit this year and beyond. Talking about how, what it means to them. >> Yeah, that's great. Always love hearing the practitioner viewpoint. All right, Joe, I want to give you the final word when it comes to this whole space things kind of move pretty fast, but also we remember it when we first saw it. So, tell us what the customers who were kind of walking away from Red Hat Summit 2020 should be looking at and understanding that they might not have thought about if they were looking at Kubernetes, a year or two ago? >> Yeah, I think a couple of things. One is, yeah, Kubernetes and this whole container ecosystem is continuing to evolve, continuing to add capabilities and continue to expand the types of workloads, that it can run. Red Hat is right in the center of it. It's all happening in open source. Red Hat as a leading contributor to Kubernetes and open source in general, is driving a lot of this innovation. We're working with some great customers and partners, other vendors, who are working side by side with us as well. And I think the most important thing is we understand that it's an evolution for customers, right? So this evolution towards moving applications to the public cloud, adopting a hybrid cloud approach. This evolution in terms of expanding the types of workloads, and how you run and manage them. And that approach is something that we've always helped customers do and we're doing that today as they move out towards embracing a cloud native. >> All right, well, Joe Fernandez, thank you so much for the updates. Congratulations on the launch of OpenShift Virtualization. I definitely look forward to talking to some the customers in finding out that helping them along their hybrid cloud journey. All right. Lots more coverage from the CUBE at Red Hat Summit. I'm Stu Miniman ,and thank you for watching the CUBE.

Published Date : Apr 29 2020

SUMMARY :

brought to you by Red Hat. and General Manager, of Core Cloud Platforms with Red Hat. Glad to be here. What does that mean to your customers? is the fact that all their applications aren't going to run So, you know, where is that consistency needed? and then, like I said, out out to the edge, right? And what data do you have And that transition is really what Red Hat is all about. and it's the applications that are hugely important and management, to manage in all those environments. So, I think back to the earliest container days, It's figuring out the right solutions and the right bet. is the ability to use Kubernetes And just some of those containers happen to be virtualized. of something that you can now manage with Kubernetes. that I'm just dragging along slowly into the new world. in the same way that we manage containers. sometimes it's even, the core piece as you said, So talk to us about just kind of is the product that then we would All right, Joe, I want to give you the final word and continue to expand the types of workloads, Congratulations on the launch of OpenShift Virtualization.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Joe FernandezPERSON

0.99+

Joe FernandesPERSON

0.99+

Stu MinimanPERSON

0.99+

JoePERSON

0.99+

last yearDATE

0.99+

Red HatORGANIZATION

0.99+

2020DATE

0.99+

AmazonORGANIZATION

0.99+

LinuxTITLE

0.99+

a month agoDATE

0.99+

GoogleORGANIZATION

0.99+

a year agoDATE

0.99+

both sidesQUANTITY

0.99+

one exampleQUANTITY

0.99+

vSphereTITLE

0.99+

KubernetesTITLE

0.99+

Red Hat Summit 2020EVENT

0.99+

Red Hat SummitEVENT

0.98+

a yearDATE

0.98+

todayDATE

0.98+

OpenShiftTITLE

0.98+

firstQUANTITY

0.98+

a quarter agoDATE

0.98+

over 60%QUANTITY

0.97+

MulticloudORGANIZATION

0.97+

Red Hat VirtualizationTITLE

0.97+

Red Hat EnterpriseTITLE

0.97+

one placeQUANTITY

0.97+

Red HatTITLE

0.97+

oneQUANTITY

0.96+

WindowsTITLE

0.96+

FirecrackerTITLE

0.96+

this yearDATE

0.96+

OneQUANTITY

0.95+

Red Hat Enterprise LinuxTITLE

0.95+

IBMORGANIZATION

0.93+

one environmentQUANTITY

0.93+

two agoDATE

0.92+

windows NTTITLE

0.92+

x86TITLE

0.91+

Clayton Coleman, Red Hat | Red Hat Summit 2020


 

>>from around the globe. It's the Cube with digital coverage of Red Hat. Summit 2020 Brought to you by Red Hat. >>Hi, I'm stupid, man. And this is the Cube's coverage of the Red Hat Summit 2020 course. The event this year is digital. We're talking to Red Hat executives, partners and customers where they are around the globe, pulling them in remotely happy to welcome back to the program. One of our Cube alumni on a very important topic, of course, that red hat open shift and joining me is Clayton Coleman. Who's the open shift chief architect with Red Hat. Clayton, thanks so much for joining us. Thank you >>for having me today. >>All right, So before we get into the product, it's probably worthwhile that we talked about you know what's happening in the community and talking specifically, you know, kubernetes the whole cloud, native space. Normally we would have gotten together. I would have seen you at Cube Con Ah, you know, at the end of March. But instead, here we are at the end of April. Looking out, you know, more CN cf events later this year, but first Red Hat Summit is a great open source event and broad community. So would really love your viewpoint as to what's happening in that ecosystem. >>It's been a really interesting year, obviously. Ah, with an open source community, you know, we react to this. Um, like we always react to all the things that go on in open source. People come to the community and sometimes they have more time, and sometimes they have less time. I think just from a community perspective, there's been a lot of people you know. It's reaching out to their colleagues outside of their companies, to their friends and coworkers and all of the different participants in the community. And there's been a lot of people getting together for a little bit of extra time trying todo, you know, connect virtually where they can't connect physically. And it's been it's been great to at least see where we've come this year. We haven't had Cube con and that'll be coming up later this year. But Kubernetes just had the 1 18 release, and I think Kubernetes is moving into that phase where it's a mature, open source project. We've got a lot of the processes down. I'm really happy with the work that the steering committee, um, has gone through. We handed off the last of the bootstrap Steering Committee members hand it off to the new, fully elected steering committee last year, and it's gone absolutely smoothly, which has been phenomenal on the The core project is trying to be a little bit more stable and to focus on closing out those loose ends being a little bit more conservative to change. And at the same time, the ecosystem has really exploded in a number of directions, as as Kubernetes becomes more of a bedrock technology for, um, enterprises and individuals and startups and everything in between. We've really seen a huge amount of of innovation in the space, and every year it just gets bigger and bigger. There's a lot of exciting projects that >>I >>have never even talk to somebody on the Kubernetes project. But they have made and build and, uh, and solve problems for their environments without us ever having to be involved, which I think it's success. >>Yeah, Clayton, you know, one of the challenges when you talk to practitioners out there is just keeping up with the pace of change. Can really be challenging. Something we really saw acutely was Docker was rolling out updates every six weeks. Most customers aren't going to be able to change fast enough to keep up with things you love your view point both is toe really what the CN CF says, as well as how Red Hat thinks of products. So you talked about you know, kubernetes 1.18. My understanding, even Google isn't yet packaging and offering that version there. So there's a lag between things. And as we start talking about managing across lots of clusters, how does Red Hat think of this? How should customers think about this? How do we make sure that we're, you know, staying secure and keeping updated on things without getting run over by the constant treadmill of >>change? That the interesting part about kubernetes Is it so much more than just that core project? You know, no matter what any of us in the in the core kubernetes project or in the products that red hat that build around open shift and layers on top, there's a There's a whole ecosystem of components that most people think of this fundamental to accomplishing building applications deploying them, running them, Whether it's their continuous integration pipelines or it's their monitoring stacks, we really as communities has become a little bit more conservative. >>Um, I >>think we really nail down our processes for taking that change from the community, testing it. You know, we run tens of thousands of automation tests a week on the latest and greatest kubernetes code, given time to soak, and we did it together with all those pieces of the ecosystem and then make sure that they work well together. And I've noticed over the last two years that the rate of oops we missed that in KUBERNETES 1 17 that by the time someone saw it, people are already using that that started to go down for us, it really hasn't been about the pace of keeping up with the upstream. But it's about making sure that we can responsibly pull together all the other ecosystem components that are still have much newer and a little bit. How do we say, Ah, they are then the exciting phase of their development while still giving ah predictable, reliable update stream. I would say that the challenges that most people are going to see is how they bring together all those pieces. And that's something that, on open shift, we think of as our goal is to help pull together all the pieces of this ecosystem, Um, and to make some choices for customers that makes sense and to give them flexibility where it's not clear yet what the right choice might be or where different people could reasonably disagree. And I'm really excited. I feel like we've got our We have a release cadence down and we're shipping the latest Cube after it's had time to quickly review, and I think we've gotten better and better at that. So I'm really proud of the team on Red Hat and how they've worked within the community so that everybody benefits from that in that testing of that stability. >>Great. I'd like to teach here, you dig in a little bit on the application side what's happening from the work loads that customers are using? Ah, what other innovations happening around that space? And how is Red Hat really helping? Really, The the infrastructure team and the developer team work even closer together, like Red Hat has done for a long time. >>This is This is a great question. I say There's two key, um, two key groups coming together. People are bringing substantial important critical production workloads, and they expect things both to just work, but also to be able to understand it. And they're making the transition. Ah, lot of folks I talked to were making the transition from previous systems they've got. They've been running open shift for a while, or they've been running kubernetes for a while, and they're getting ready to move, um, a significant portion of their applications over. And so, you know, in the early days of any project, you get the exciting Greenfield development and you get to go play with new technologies. But as you start moving your 1st 1 and then 10 and then 100 of your core business applications from the EMS or from bare metal into containers, you're taking advantage of that technology in a responsible way. And so the the expectations on us as engineers and community members is to really make sure that we're closing out the little stuff. You know, no bug is too small, but it can't trip up someone's production applications. So seeing a lot of that whether it's something new and exciting like, Um uh, model is a service or ai workloads or whether it's traditional big enterprise transaction processing. APS on the other side on that development, um, model I think we're starting to see phase to our community is 2.0, in the community, which is people are really leveraging the flexibility and the power of containers, things that aren't necessarily new to people who had. We got into containers early and had a chance to go through a couple of iterations. But now people are starting to find patterns that up level development teams, so being able to run applications the same way on a local machine as in a production environment. Well, most production environments are there now, and so people are really having toe. They're having to go through all of their tools and saying, Well, does this process that works for an individual developer also work when I want to move it there, my production or staging environments to production, and so on. New projects like K native and tectonic, which are kubernetes native, that's just one part of the ecosystem around development. On top of kubernetes, there's tons of exciting projects out there from companies that have adopted the full stack of kubernetes. They built it into their mindset, this idea of flexible infrastructure, and we're seeing this explosion of new ways where kubernetes is really just a detail, and containers are just the detail and the fact that it's running this little thing called Docker down at the heart of it. Nobody talks about anymore, and so that that transition has been really exciting. I think there's a lot that we're trying to do to help developers and administrators see eye to eye. And a lot of it's learning from the customers and users out there who really paved the way the which is the open source way. It's learning from others and helping others benefit from that. >>Yeah, I think you bring up a really important point we've been saying for a couple of years. Now that you know KUBERNETES should get to the point where it's boring and boring in a way also cause it's gonna be baked in everywhere we saw from basically customers just taking the code, really spending a lot of their own things by building the stack to, of course, lots of customers have used open shift over the year to If I'm adopting Public Cloud more and more, they're using those services from that standpoint. Can you talk a bit about how Red Hat is really integrating with public clouds? And you know your architectural technical philosophy on that? And how might that be? Differ from some other companies that you might call a little bit more, you know, Cloud of Jason, as opposed to being deeply integrated with the public cloud. >>The interesting thing about Kubernetes is that while it was developed on top of the clouds, it wasn't really built from Day one assuming a cloud underneath it. And I think that was an opportunity that we really missed. And to be fair, we had to make the thing work first before we depended on these unreliable clouds. You know, when we started, the clouds were really hitting their stride on stability and reliability, and people were it was the hot was becoming the obvious choice to some of what we've tried to do is take flexible infrastructure is a given, um, assume that the things that the cloud provides should be programmed for the for the benefit of the developer and the application, and I think that's a that's a key trend is we're not using the cloud because our administration teams want us. We're using the cloud because it makes us more powerful developers. That enables new scenarios. It shortens the the time between idea reality. What we have done in open shift is we've really built around The idea of open shift running on a cloud should take advantage of that cloud to an extreme degree, which is infrastructure could be flexible. The machines in that cluster need to come and go according to the demands of the applications on top of it. So giving a little bit more power to the cluster and taking a little bit of way from the cloud I'm. But that benefits. That also needs to benefit that those who are running on premise because I think, as you noted, our goal is you want this ubiquitous kubernetes environment everywhere, and the operations teams and the development teams and the Dev Ops teams in between need to have a consistent environment and so you can do this on the cloud. But you don't have that flexibility on premise. You've lost something. And so what we've tried to do as well is to think about those ideas that are what we think of as quote unquote cloud native that starts with a mutable operating systems. It starts with everything being declarative and working backwards from, you know, I wanna have 15 machines and then the cluster or controllers on the cluster say, Oh, well, you know, one of the machines has gone bad. Let's replace it on the cloud. You ask for a new I'm cloud infrastructure provider or you ask the the cloud a p i for a new machine, and then you replace it automatically, and no one knows any better on premise. We'd love to do the same thing with both bare metal virtualization on top of kubernetes. So we have that flexibility to say you may not have all of the options, but we should certainly be able to say, Oh, well, this hardware is bad or the machine stopped, so let's reboot it, and there's a lot of that same mindset that could be applied. We think that'll, um if you need virtualization, you can always use it. But virtualization is a layer on top benefits from some of the same things that all the other extensions and applications on top of kubernetes competitive trump. So trying to pay that layer and make sure that you have flexible, reliable storage on premise through our SEF and red hat storage products, which are built on top of the cluster exactly like virtualization, is both on top of the cluster. So you get cloud native storage mixed in working with those teams toe. Take those operational best practices. You know there's well, I think one of the things that interests me is no. 1 20 years ago, who was running an early version of SEF wouldn't have some approach to run these very large things that scales organizations like CERN have been using SEF for over a decade at extremely large scales. Some of what our mindset is we think it's time to bake some of that knowledge actually into our software for a very long time. We've kind of been building out and adding more and more software, but we always left the automation and the the knowledge about how that software supposed to be run to the side. And so by taking that and we talked about operators. Kubernetes really enshrines. This principle is taking that idea, taking some of that operational knowledge into the software we ship. Um, though that software can rely on kubernetes open shift tries to hide the details of the infrastructure underneath and our goal. I think in the long run it will just make everybody's lives easier. I shouldn't have to ship you a SEF admin for you to be successful. And we think we think there's a lot more room here that's really gonna improve how operations teams work, that the software that they use day to day. >>So Clinton you mentioned virtualization is one of the topics in there. Of course, virtualization is very prevalent in a customer's data center environment today. Red Hat open shift, oftentimes in data centers, is sitting on BM ware environments. Of course. Recently, VM Ware announced that they have kubernetes baked into the solution, and red hat has open shift with red hat virtualization. Maybe, you know, without going into too much depth, and you probably have breakouts and white papers on this. But you know what kind of decision point should customers be thinking about when they're deciding? Do I do this in bare metal. Do I do it in virtualization? What are some of the, you know, just high level trade offs there when they need to make those decisions, >>I think it's, um I think the 1st 1 is Virtualization is a mature technology. It's a known quantity for many organizations, and so those who are comfortable with virtualization, I'd say, like any responsible, uh, architecture engineering team. You don't want to stop using something that's working well just because you can. And a lot of what I would see as the transition that companies on is for some organizations without a big investment in virtualization. They don't see the need for it anymore, except as maybe a technical detail of how they isolate insecure workloads. One of the great things about virtualization technology that we're all aware of over the last couple years is it creates a boundary between work loads and the underlying environment. That doesn't mean that the underlying environment and containers can't be as secure or benefit from those same techniques. And so we're starting to see that in the community, this kind of spectrum of virtualization all the way from the big traditional virtualization to very streamlined, stripped down virtualization wrappers around containers. Um, like some of the cloud providers use for their application environments. So I'm really excited about the open source. Community is touching each of these points on the spectrum. Some of our goals are if you're happy with your infrastructure provider, we want to work well with, and that's kind of the pragmatic of everyone's on a different step in that journey. The benefit of containers is no matter how fast you make of VM, it's never gonna be quite as fast, is it containers. And it's never gonna be quite as easy for a developer to run on their laptop. And I think working through this is there's still a lot of work that we as a community to do around, making it easier for developers to build containers and test them locally in smaller environments. But all of that flexibility can still benefit from virtualization under later or virtualization used as an isolation technology. So projects like Kata and some of the work that's being done in the open source community around projects like firecracker taking the same, um, open source ideas and remixing them a different points gives us a lot of flexibility. So I would say, um, I'm actually less interested in virtualization then all of the other technologies that are application centric and at the heart of it, a VM isn't really a developer centric idea. It's specifically an administrative concept that benefits the administrator, and developers can take advantage of it. But I think all of the capabilities that you think of when you think about building an application like scaling out and making sure patches are applied, being able to roll back separating your configuration on then all of the hundreds of other levels of complexity that will add around that like service MASH and the ability to gracefully tolerate failures in your database. These were where I think, um, virtualization needs to work with the platform rather than being something that dominates how we think about the platform. It's application first, not being first. >>Yeah, no, you're absolutely right that the critique I've always given, you know for a number of years now is if you look at virtualization, the promise was, let's take that old application that probably should have been updated and just shove it in a VM and never think about it again. That's not doing good things for the user. So if I look at that at one end of the spectrum away at the other end of the spectrum, trying not to think about infrastructure, you mentioned K native s 01 of the things that you know I've been digging in tryingto learn more about at Red Hat Summit has really been the open shift server lists. So give us the update on that piece. Um, you know, that's obviously very different discussion than what we were just having from a virtualization standpoint. Eso How does open shift look at server lists? How does that tie into what? You know, if I'm doing server, listen, Amazon versus you know some of the other open source options for serverless. How should I be thinking about that? >>There's a lot of great choices on the spectrum out there. I think one of the interesting things and I love the word spectrum here because cane native kind of sits in a spot where it tries to be, as the name says, it tries to be as kubernetes native as possible, which lets you tap into some of those additional capabilities when you need it. And one of the things I've always appreciate it is the more restrictive framework is usually the better. It is doing that one thing and doing it really well. We learned this with rails. We learned this with no Js. And as people have built over the years, the idea of simple development platforms. The core function idea is a great simple idea, but sometimes you need to break out of that. You need extra flexibility or your application needs to run longer or slow Start is actually an issue. One of the things I think is most interesting about K native and I see comers and user. I think this way it's a good point. Um, that gives you some of the flexibility of kubernetes and a lot of the simplicity of, um, the functions is a service, but I think that there's going to be an inevitable set of use cases that tie into that which are simpler where open organization has a very opinionated way of running applications, and I think that flexibility will really benefit K native. Whereas some of the more opinionated remarks around server lists lose a little bit of that. So that's one dimension that I still think a native is well positioned to kind of capture the broadest possible audience, which for kubernetes and Containers was kind of our mindset. We wanted to solve enough of the problems that you can solve. You can run all your software. We don't have to solve all those problems to such a level that there's endless complexity, although we've been accused of having endless complexity and Cooper days before, but just trying to think through what are the problems that everyone's going to have to give them a way out? I'm at the same time for us, when we think about prioritization functions is service about integration. It's about taking applications and connecting them, connecting them through kubernetes. And so it really depends on identity and access to data and tying that into your cloud environment. If you're running on top of a cloud or tying it into your back end databases, if your on premise, >>I >>think that is where the ecosystem is still working to bring together and standardize some of those pieces in kubernetes or on top of Kubernetes. What I'm really excited about is the team as much. You know, there's been this core community effort to get a native to a G, a quality. Alongside that, the open shift serverless team has been trying to make it a dramatically simpler action. If you have kubernetes and open shift, it's a one click action to get started with, Um Kay native and just like any other technology. How accessible it is determines how easy users find it to get started and to build the applications they need. So for us, it's not just about the core technology. It's about someone who's not familiar with Serverless or not familiar with kubernetes. Bring up an editor and build a function and then deploy it on top of open shift. See it scale out like a normal kubernetes application, not having to know about pods or persistent volumes or notes. And so these air, these are some of the steps. I've been really proud that the team's done. I think there's a huge amount of innovation that will happen this year and next year, as the maturity of the kubernetes ecosystem really grows up, we'll start to see standardized technologies, for I'm sharing identity across multiple clouds across multiple environments. It's no good if you've got these applications on the cloud that need to tie into your corporate L dap. But you can't connect your corporate held up to the cloud. And so your applications need 1/3 identity system. Nobody wants 1/3 identity system. And so, working through some of this thing where the challenges I think that hybrid organizations are already facing and our job is just to work with them in the open source communities and with the cloud providers partner with them and open source so that the technologies in kubernetes fit very well into whatever environment they run it. Alright, >>well, Clayton, really appreciate all the updates there. I know the community is definitely looking forward to digging through some of the breakout sessions reading all the new announcements. And, of course, we look forward to seeing you on the team participating in many of the kubernetes related events happening later this >>year. That's right. It's ah, gonna be a good year. >>All right. Thanks so much for joining us. I'm still Minuteman and as always thank you for watching you. >>Yeah, yeah, yeah, yeah

Published Date : Apr 29 2020

SUMMARY :

Summit 2020 Brought to you by Red Hat. Who's the open shift chief architect with Red Hat. All right, So before we get into the product, it's probably worthwhile that we talked about you We handed off the last of the bootstrap Steering Committee members hand it off to the new, have never even talk to somebody on the Kubernetes project. going to be able to change fast enough to keep up with things you love your view point both in the products that red hat that build around open shift and layers on top, there's it really hasn't been about the pace of keeping up with the upstream. I'd like to teach here, you dig in a little bit on the application side what's And a lot of it's learning from the customers and users out there who really And you know your architectural technical philosophy on that? on the cluster say, Oh, well, you know, one of the machines has gone bad. What are some of the, you know, just high level trade offs the ability to gracefully tolerate failures in your database. the things that you know I've been digging in tryingto learn more about at Red Hat Summit has really the functions is a service, but I think that there's going to be an inevitable and open source so that the technologies in kubernetes fit very well into I know the community is definitely looking forward to digging It's ah, gonna be a good year. I'm still Minuteman and as always thank you for watching

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
ClaytonPERSON

0.99+

15 machinesQUANTITY

0.99+

Red HatORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

ClintonPERSON

0.99+

CERNORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

100QUANTITY

0.99+

last yearDATE

0.99+

red hatORGANIZATION

0.99+

Clayton ColemanPERSON

0.99+

10QUANTITY

0.99+

next yearDATE

0.99+

two key groupsQUANTITY

0.99+

VM WareORGANIZATION

0.99+

one clickQUANTITY

0.99+

two keyQUANTITY

0.99+

CubeORGANIZATION

0.99+

Summit 2020EVENT

0.99+

end of AprilDATE

0.98+

Red Hat SummitEVENT

0.98+

SEFTITLE

0.98+

bothQUANTITY

0.98+

oneQUANTITY

0.98+

OneQUANTITY

0.98+

firstQUANTITY

0.98+

end of MarchDATE

0.97+

this yearDATE

0.97+

one partQUANTITY

0.97+

Red Hat Summit 2020EVENT

0.97+

one dimensionQUANTITY

0.97+

later this yearDATE

0.96+

todayDATE

0.96+

eachQUANTITY

0.93+

KubernetesTITLE

0.93+

Day oneQUANTITY

0.93+

hundredsQUANTITY

0.92+

KayPERSON

0.92+

one endQUANTITY

0.91+

20 years agoDATE

0.91+

one thingQUANTITY

0.91+

KataTITLE

0.91+

1st 1QUANTITY

0.91+

red hatTITLE

0.89+

CN CFORGANIZATION

0.87+

over a decadeQUANTITY

0.86+

tens of thousands of automation testsQUANTITY

0.85+

last two yearsDATE

0.84+

MinutemanPERSON

0.82+

KubernetesORGANIZATION

0.82+

CubeCOMMERCIAL_ITEM

0.82+

every six weeksQUANTITY

0.81+

1/3QUANTITY

0.79+

cfEVENT

0.75+

Steering CommitteeORGANIZATION

0.75+

last couple yearsDATE

0.74+

K native sORGANIZATION

0.74+

a weekQUANTITY

0.73+

BMORGANIZATION

0.68+

kubernetesTITLE

0.66+

later thisDATE

0.63+

Mark Shuttleworth, Canonical | KubeCon + CloudNativeCon EU 2019


 

>> Live from Barcelona, Spain, It's theCUBE. Covering KubeCon + CloudNativeCon Europe 2019. Brought to you by Red Hat, the Cloud Native Computing Foundation, and ecosystem partners. >> Welcome back, to theCUBE coverage here at KubeCon + CloudNativeCon. I'm Stu Miniman, my co-host is Corey Quinn. And happy to welcome back to the program Mark Shuttleworth who os the CEO of Canonical. Of course, the orange shirts of Ubuntu, are seen all throughout the show. Mark, thank you so much for joining us, great so see you. >> Great to see you. >> All right, so for years, actually, we've had these conversations at the OpenStack Summit. It's interesting that, every time you mention it around this show you get snark online, as like, it is dead, Kubernetes killed it and it's like wait, no, no, you know we're talking about, a couple of open-source projects. I've been talking to people, especially in the telco space, that's like, oh yeah, well no, we just run OpenStack underneath and Kubernetes on top and put all things together. Give us a little bit of your broad view of some of these big trends, and open-source monoliths and microservices and all these pieces, all kind of fly together. >> Yeah, I think if your in the Reddit SubChannels, then you know it can feel a bit like turf war, and gangster-type, free software riffing, right. But the reality is, OpenStack solves business problems for people. They want large scale, virtualized infrastructure, that's cheaper than VMware. We are deploying OpenStacks in enterprise environments at double the scale and double the speed, in other words, like twice as many every month, as we were a year ago. I think people have gotten comfortable with the idea that Kubernetes is an application operations construct. I think we will see virtualization blur into the Kubernetes lives, but mainly for security reasons. So I want deeper isolation of applications that come from third-party vendors, for example. And I'm willing to trade performance for isolation, in circumstances where I am bringing in third-party code into my private infrastructure. After we see a couple of significant security compromises, I mean, we saw the GitHub compromise. If you shave that Yak, it gets to a very uncomfortable place of, what are we actually running as root all over our data centers with Docker and Docker Hub. So, people are going to want that kind of isolation of containers, the Kata Containers work is going to bring that. But that's very different to the proposition of, essentially, give me large scale, machine virtualization which OpenStack addresses. OpenStack hasn't done itself any favors, don't need to go into that here. But nonetheless, as far as we're concerned, it's straight forward to deliver large scale, low cost, enterprise virtualization infrastructure for telco's or IT use cases. >> Let's get into this ecosystem here. I want to say the Cloud Native ecosystem, and I say that specifically because there are some that look at this and they say, oh, there's dozens of projects now, Kubernetes is a platform against platform. Somebody even mentioned the word big tent once. We've seen some projects merging, we've seen some various pieces. >> I saw making a bigger tent on the keynote and I was like, not my favorite choice of words. >> I seem to remember a certain article that you wrote poking a whole in the big tent thing. What's the same, what's different? What's your take on this? Is it an ecosystem? Is it Kubernetes and friends, as Corey has liked to say here? What's your take? >> Look, I think we're still trying to figure out what are the appropriate labels to attach to this kind of forum, it is a forum, right. There is a tremendous amount of value attached to being here, to the ideas that are getting bounced about. But I wouldn't call it a simple community in the sort of, traditional open-source sense. The reality is there's very serious money behind every, sort of project that's been framed as a community project. This is a new kind of consortium. And that brings with it certain, delicate, political posturing and so on. But, nonetheless, it's a valuable place to be. It's definitely staking out important concepts and operational platforms, ideas, regimes, whatever you want to call it. This is going to be a fun week. >> I started off my career in the Linux world as a grumpy Unix administrator because there really wasn't any other kind. Then I started dipping my toes into the Linux world and something struck me, almost immediately, about Ubuntu. Was how welcoming everyone was in the community. There was no such thing as a stupid question. I asked the kind of questions you would expect from someone working on a computer, wearing a suit. People were very eager to embrace newcomers into that. It was one of the absolute best things that I saw coming out of Canonical, in addition to the software itself. I love that you're here as a part of this. What is the larger picture? What do you see in the Cloud Native ecosystem that's resonating with what Canonical's doing? >> So, the big thing that we do is, essentially, try to figure out where, what's possible with open-source that's hard to do. And then make it really straight forward so that more people can do the important stuff easily. That doesn't stop people from doing all the crazy stuff at the periphery that you can do with Ubuntu. It's generally easier with Ubuntu than any other platform. But we try to make the really most important things really easy for everybody. That's the first thing. The second thing is, we're a little non-judgemental about the fact that there are different perspectives on the same stuff. In the Ubuntu ecosystem, we make a point of saying that GNOME guys, and the KDE guys, and the LXQt, and the MATE guys. The Ubuntu ecosystem is where they actually meet to hash out how they can do stuff in a way that means users get a real choice between those. There's a very similar role for us to play in an environment like this. It's kind of acronym soup out there. Like 50 new projects every KubeCon. They're all interesting, they're all important, there's a lot of overlap between them. There's work for us to do in figuring out which ones are going to be really more important in the tent. We did that very effectively with OpenStack. The people who rode the OpenStack wave with us haven't had to abandon their OpenStacks. Because the stuff that we really chose to make central and easy, turned out to be the stuff that was the important poles in the tent. And we'll do exactly the same stuff here with Kubernetes. So, to put that into context, it's been real fun to be on the booth. We had, just tons, of people coming up and saying thank you for Microk8s. Microk8s is a single package of Kubernetes, that works in lots Linux distributions. It gives you, in about a minute, it gives you a standard Kubernetes environment, that's pure upstream. That, for a developer, just let's you get productive immediately. Figure out these new development application operations, constructs. You can use it on an airplane, you can use it on a train. Of course, it's compatible with all of the public clouds so that's the second thing that we're doing. We work with Amazon, with the EKS team, I spoke at their event on Monday. We work with Azure, the AKS team, we work with Google, we work with Oracle, we work with IBM. Essentially making sure that all of them offer Ubuntu worker nodes for their Kubernetes, SaaS offerings. That means that the developer who's doing stuff on their workstation with Microk8s can take those containers straight to any other public clouds. So, we're not trying to force people to use a particular solution, we're saying, in all of those environments, there are going to be choices people have. We want to make that as easy as possible for them. We want to avoid unnecessary friction in that process. That kind of underlining culture is coming through in this forum, as well. >> We've had many conversations about how you've always tried to make the job of that developer really easy. One of the things we always look at on this show is how much of it is the infrastructure people, or the platform underneath and the developer, and how much are they coming together. Anything different about this ecosystem? >> Very much so, yeah. >> Or your customers here that you can share? >> Kubernetes is an application construct. You can think of it as a next generation message bus. It's how components of an application find each other, communicate with each other, essentially, coordinate with each other. That makes it very tightly woven in to the developer experience. By contrast, you can be sitting writing a Java application inside a bank and not know or care whether it's going to be running on a physical machine, a virtual machine or an OpenStack cloud. You just don't know, you don't care. It's too far away from the application. Kubernetes is right there. I think that's one of the really interesting things is that it's bringing those infrastructure brains together with the application, app dev brains, in a very interesting way. It's going to be challenging. I wouldn't underestimate it, there are a lot of people, sort of, wondering around here, feeling a little confused, but that's okay. Do you know what I mean, the stuff shakes out. >> So, something that's been a recurring theme here has been the idea of going in a multi-cloud direction. Where people are talking about wanting to build workloads that they can seamlessly deploy across different providers. People talk about that, periodically, as a strategic goal but I'm not seeing people do it very often in the real world. You're in a much better position than a lot of us, to see that. Is that something you're seeing people moving towards as an adoption? >> Well, yes. Because we work with all of the major public clouds to optimize Ubuntu there, in a way that I don't think any other Linux does. You get an optimized Amazon Ubuntu on Amazon. You get an optimized Azure Ubuntu on Azure, and so on. >> Going very deep in the Amazon ecosystem. Most of my customers are using Ubuntu far ahead of anything else out there. >> That's right. >> And it's the right answer for what they're doing. >> That's right. It gives them, essentially, the best of what Amazon's offering, it still gives them the ability to feel like if they want to go somewhere else, they can. And that actually works well for Amazon. In the early days, I think there was a little tension between us and the cloud guys, because they were saying, look, if people use Ubuntu then they can go somewhere else. Yes, but in a sense, that makes them more likely to be more relaxed about starting wherever they choose to start. We don't advise enterprises as to which cloud to use. We advise them to engage with those clouds and figure out their differences, they are different. Amazon's really good at some things that are different, to what Microsoft is good at. Oracle is really good at some things which are different too. And what we're starting to see is the level of maturity in the enterprise governance process. They know they want to work with multiple clouds. They initially thought that was a straight kind of commodity exchange, competition thing. They now realize that it's a bit richer than that. That there are actually business reasons to have deeper relationships with particular clouds, based on what those clouds are prioritizing, and what they are prioritizing. So, we're not going to say you should use this cloud, you should use that cloud. Obviously, we can draw a distinction between the clouds where we're deeply engaged and the clouds where, you know, where you just don't have the benefit of that. But, more importantly we can say, you know, here are the set of practices that you can adopt internally that will give you comfort that your getting the best out of those clouds, the ones that you've chosen. And you have the portability that you really need. The key turns out be, enabling your developers, to use multiple clouds and challenging the developers to do different phases of the development life cycle on different clouds. Develop on your private cloud or your work station, use Microk8s, for example. Do tests on one cloud. Do staging and production on a different cloud. Now you already know that that whole, seamless ecosystem works. If you want to go use a high value, proprietary function, effectively on a cloud, that's a business decision and it's not a bad business decision. There's some spectacular capabilities from Amazon that are unique to Amazon. Or from Microsoft that are unique, or from Oracle that are unique to Oracle. They're spectacular. Those are business decisions to use them. There's other stuff that effectively you can give yourself optionality on. I wouldn't be black and white about that, put yourself in a position to make smart choices. And our best customers are getting are getting there. PayPal, they're operating on Ubuntu in a very sophisticated way, across multiple public clouds and private infrastructure. >> All right, so Mark we're five years into Kubernetes now. We've seen adoption grow, people feel there's a certain level of maturity here. There's always that concern that we've reached that peak and we're about to fall off the cliff. What do we need to worry about? What does the ecosystem need to do to make sure we continue along the stability and security that customers are looking for. >> There will be an over shoot regardless. I don't think there's any sort of leadership or governance approach that could avoid that. It's a little bit like, if your stock is going crazy. On the one hand, you're kind of happy. On the other hand, if you feel it's over valued it's a difficult sort of thing to say. You need to say, guys, you know what I mean, we're humans too. We've got our challenges to work through. And no one likes volatility, but too a certain extent, there's always speculation and over shoot, and over-enthusiasm, and hype. Kubernetes will over shoot. There's a bunch of emperors walking around here that, frankly, have no clothes. My job, our job, is very calmly, to sort through the wheat from the chaff. Make sure that it's possible for people to experiment with everything. But, that the stuff that we think has legs, effectively, is nicely integrated for people, that they have that for the long term, they won't regret things. We have a good track record of doing that. We've done it in the Linux desktop. We did it in OpenStack, we're doing it in public cloud. We've done it here in the Cloud Native world. I'd say things like AI are going in the same direction. Again, tons of complexity, tons of new options. Helping people effectively navigate through that is what we do very well. >> Yeah, one of the questions that I started to see as well, as we look at the way that these technologies continue to evolve, has been that, for better or worse, when developers are writing applications now and even infrastructure people are working with a lot of the things they care about. What operating system, let alone what distribution they're using, is increasingly slipping beneath the waves. People don't think about that as a primary area of focus anymore. And as, I guess, of the foundational Linux vendors in this space, how are you seeing that evolving? And how does Canonical remain relevant in a world where suddenly, people in a serverless future, I just throw some code over somewhere else and it runs is the limit of where most companies get involved. >> Yes, of course, we can point to the servers. And on the servers, we can point to the operating systems and inside the containers, we can point to the operating systems and underneath the serverless code, we can point to the language runtimes. So, the reality is that those things matter less and less to the developer. >> Yes. >> They still matter to the institution. So, I'm super comfortable with the language that says, the OS doesn't matter. What it means is that that whole tangle is getting professionalized and abstracted. But to be confident in the abstractions, someone needs to do a lot of work. I know how much work we do with Google, with Amazon, with Microsoft, with Oracle, with IBM, to make sure that nobody else has to feel like the OS matters. That that stuff essentially just works. You can extend that out to what we do with VMware, what we do, essentially, on bare-metal, what we do on developer workstations, what we do with the Windows crowd, effectively, and Windows subsystem for Linux, so that developers really can just build on Windows subsystem for Linux, Ubuntu, effectively, and ship that container straight to Amazon EKS and have it just work. There are a ton of little lies that have to line up. Containers are all kind of a fiction. The fiction breaks if those pieces don't line up. So, being Ubuntu, effectively and being being able to be consistent in all of those places, is a ton of work to enable it not to matter for anybody upstairs. That's allowing developers to go faster. It's allowing them to be more productive. It's allowing them to be more heroic. And it's allowing the people who do worry about the middleware to have far fewer nights scratching their heads as to, why didn't this version of this library tie up to that driver with that kernel. All of those things are still there. When you drop that container onto Amazon, we've got to connect the GPGPU in the hardware, through the hypervisor, to the guest OS, up into the container. And there's code getting injected all the way up. It's only the fact that we can typically have Ubuntu everywhere there that, essentially, allows those pieces to line up without some spectacular fireworks. It satisfies me when people say they don't have to worry about that. >> It's a victory condition. >> Mark, I want to give you the final word. What should we be looking for, from Canonical, through the rest of the year? >> So, for us, this has been a big year in terms of visibility in the enterprise. In terms of penetration, Ubuntu's everywhere in the Fortune 500, everywhere in the Global 2000. What's changed this year, is the CIO suddenly is seeing Ubuntu on their desk. For two reasons, one is IBM Red Hat. The CIO suddenly wants to know, okay, what does this mean? What else are we running? Where else can we get 24/7 SLAs? Where else can we get long term commitments to Linux and so on? And the fact is Ubuntu's already in the building so that's one, sort of, easy connect. The other thing is, there's really interesting, new workloads that Ubuntu leads in the enterprise. Obviously the container story, the multi-cloud story, edge. It's not just telcos. Every retailer, every logistics company, anybody that has physical distribution is now trying to say, well how can I automate compute in my physical world, effectively. So, edge is super interesting and IoT beyond that. People transforming businesses through taking a Raspberry Pi with Ubuntu and putting a snap on it is really, really cool. Which of those is going to drive the biggest headlines or the scariest headlines, I can't tell you. We're just trying to take care of security, performance and operations across all of them. >> All right, well, Mark Shuttleworth, always a pleasure to catch up, thank you so much for the updates. >> Great to see you. >> All right, for Corey Quinn, I'm Stu Miniman. We'll be back with lots more coverage here from KubeCon + CloudNativeCon 2019 in Barcelona, Spain. Thanks for watching theCUBE. (upbeat music)

Published Date : May 22 2019

SUMMARY :

Brought to you by Red Hat, And happy to welcome back to the program Mark Shuttleworth I've been talking to people, especially in the telco space, of containers, the Kata Containers work is going to bring that. and I say that specifically because there are some on the keynote and I was like, I seem to remember a certain article that you wrote This is going to be a fun week. I asked the kind of questions you would expect of saying that GNOME guys, and the KDE guys, One of the things we always look at on this show is It's going to be challenging. in the real world. to optimize Ubuntu there, in a way that I don't think in the Amazon ecosystem. and the clouds where, you know, What does the ecosystem need to do But, that the stuff that we think has legs, effectively, that these technologies continue to evolve, And on the servers, we can point to the operating systems You can extend that out to what we do with VMware, Mark, I want to give you the final word. Which of those is going to drive the biggest headlines always a pleasure to catch up, We'll be back with lots more coverage here

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
IBMORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

Mark ShuttleworthPERSON

0.99+

Corey QuinnPERSON

0.99+

OracleORGANIZATION

0.99+

Stu MinimanPERSON

0.99+

GoogleORGANIZATION

0.99+

CanonicalORGANIZATION

0.99+

Cloud Native Computing FoundationORGANIZATION

0.99+

MarkPERSON

0.99+

five yearsQUANTITY

0.99+

Red HatORGANIZATION

0.99+

MondayDATE

0.99+

AKSORGANIZATION

0.99+

PayPalORGANIZATION

0.99+

WindowsTITLE

0.99+

second thingQUANTITY

0.99+

KubeConEVENT

0.99+

EKSORGANIZATION

0.99+

UbuntuTITLE

0.99+

Barcelona, SpainLOCATION

0.99+

JavaTITLE

0.99+

50 new projectsQUANTITY

0.99+

LinuxTITLE

0.99+

first thingQUANTITY

0.99+

oneQUANTITY

0.99+

OpenStackTITLE

0.98+

KubernetesTITLE

0.98+

two reasonsQUANTITY

0.98+

OpenStack SummitEVENT

0.98+

a year agoDATE

0.98+

CloudNativeConEVENT

0.97+

twiceQUANTITY

0.97+

OneQUANTITY

0.97+

one cloudQUANTITY

0.96+

CoreyPERSON

0.96+

OpenStacksTITLE

0.96+

dozens of projectsQUANTITY

0.96+

GitHubORGANIZATION

0.96+

LXQtTITLE

0.95+

VMworld Day 1 General Session | VMworld 2018


 

For Las Vegas, it's the cube covering vm world 2018, brought to you by vm ware and its ecosystem partners. Ladies and gentlemen, Vm ware would like to thank it's global diamond sponsors and it's platinum sponsors for vm world 2018 with over 125,000 members globally. The vm ware User Group connects via vmware customers, partners and employees to vm ware, information resources, knowledge sharing, and networking. To learn more, visit the [inaudible] booth in the solutions exchange or the hemoglobin gene vm village become a part of the community today. This presentation includes forward looking statements that are subject to risks and uncertainties. Actual results may differ materially as a result of various risk factors including those described in the 10 k's 10 q's and k's vm ware. Files with the SEC. Ladies and Gentlemen, please welcome Pat Gelsinger. Welcome to vm world. Good morning. Let's try that again. Good morning and I'll just say it is great to be here with you today. I'm excited about the sixth year of being CEO. When it was on this stage six years ago were Paul Maritz handed me the clicker and that's the last he was seen. We have 20,000 plus here on site in Vegas and uh, you know, on behalf of everyone at Vm ware, you know, we're just thrilled that you would be with us and it's a joy and a thrill to be able to lead such a community. We have a lot to share with you today and we really think about it as a community. You know, it's my 23,000 plus employees, the souls that I'm responsible for, but it's our partners, the thousands and we kicked off our partner day yesterday, but most importantly, the vm ware community is centered on you. You know, we're very aware of this event would be nothing without you and our community and the role that we play at vm wares to build these cool breakthrough innovations that enable you to do incredible things. You're the ones who take our stuff and do amazing things. You altogether. We have truly changed the world over the last two decades and it is two decades. You know, it's our anniversary in 1998, the five people that started a vm ware, right. You know, it was, it was exactly 20 years ago and we're just thrilled and I was thinking about this over the weekend and it struck me, you know, anniversary, that's like old people, you know, we're here, we're having our birthday and it's a party, right? We can't have a drink yet, but next year. Yeah. We're 20 years old. Right. We can do that now. And I'll just say the culture of this community is something that truly is amazing and in my 38 years, 38 years in tech, that sort of sounds like I'm getting old or something, but the passion, the loyalty, almost a cult like behavior that we see in this team of people to us is simply thrilling. And you know, we put together a little video to sort of summarize the 20 years and some of that history and some of the unique and quirky aspects of our culture. Let's watch that now. We knew we had something unique and then we demonstrated that what was unique was also some reasons that we love vm ware, you know, like the community out there. So great. The technology I love it. Ware is solid and much needed. Literally. I do love Vmr. It's awesome. Super Awesome. Pardon? There's always someone that wants to listen and learn from us and we've learned so much from them as well. And we reached out to vm ware to help us start building. What's that future world look like? Since we're doing really cutting edge stuff, there's really no better people to call and Bmr has been known for continuous innovation. There's no better way to learn how to do new things in it than being with a company that's at the forefront of technology. What do you think? Don't you love that commitment? Hey Ashley, you know, but in the prep sessions for this, I thought, boy, what can I do to take my commitment to the next level? And uh, so, uh, you know, coming in a couple days early, I went to down the street to bad ass tattoo. So it's time for all of us to take our commitment up level and sometimes what happens in Vegas, you take home. Thank you. Vm Ware has had this unique role in the industry over these 20 years, you know, and for that we've seen just incredible things that have happened over this period of time and it's truly extraordinary what we've accomplished together. And you know, as we think back, you know, what vm ware has uniquely been able to do is I'll say bridge across know and we've seen time and again that we see these areas of innovation emerging and rapidly move forward. But then as they become utilized by our customers, they create this natural tension of what business wants us flexibility to use across these silos of innovation. And from the start of our history, we have collectively had this uncanny ability to bridge across these cycles of innovation. You know, an act one was clearly the server generation. You know, it may seem a little bit, uh, ancient memory now, but you remember you used to walk into your data center and it looked like the loove the museum of it passed right? You know, and you had your old p series and your z series in your sparks and your pas and your x86 cluster and Yo, it had to decide, well, which architecture or am I going to deploy and run this on? And we bridged across and that was the magic of Esx. You don't want to just changed the industry when that occurred. And I sort of called the early days of Esx and vsphere. It was like the intelligence test. If you weren't using it, you fail because Yup. Servers, 10 servers become one months, become minutes. I still have people today who come up to me and they reflect on their first experience of vsphere or be motion and it was like a holy moment in their life and in their careers. Amazing and act to the Byo d, You know, can we bridge across these devices and users wanted to be able to come in and say, I have my device and I'm productive on it. I don't want to be forced to use the corporate standard. And maybe more than anything was the power of the iphone that was introduced, the two, seven, and suddenly every employee said this is exciting and compelling. I want to use it so I can be more productive when I'm here. Bye. Jody was the rage and again it was a tough challenge and once again vm ware helped to bridge across the surmountable challenge. And clearly our workspace one community today is clearly bridging across these silos and not just about managing devices but truly enabling employee engagement and productivity. Maybe act three was the network and you know, we think about the network, you know, for 30 years we were bound to this physical view of what the network would be an in that network. We are bound to specific protocols. We had to wait months for network upgrades and firewall rules. Once every two weeks we'd upgrade them. If you had a new application that needed a firewall rule, sorry, you know, come back next month we'll put, you know, deep frustration among developers and ceos. Everyone was ready to break the chains. And that's exactly what we did. An NSX and Nice Sierra. The day we acquired it, Cisco stock drops and the industry realizes the networking has changed in a fundamental way. It will never be the same again. Maybe act for was this idea of cloud migration. And if we were here three years ago, it was student body, right to the public cloud. Everything is going there. And I remember I was meeting with a cio of federal cio and he comes up to me and he says, I tried for the last two years to replatform my 200 applications I got to done, you know, and all of a sudden that was this. How do I do cloud migration and the effective and powerful way. Once again, we bridged across, we brought these two worlds together and eliminated this, uh, you know, this gap between private and public cloud. And we'll talk a lot more about that today. You know, maybe our next act is what we'll call the multicloud era. You know, because today in a recent survey by Deloitte said that the average business today is using eight public clouds and expected to become 10 plus public clouds. And you know, as you're managing different tools, different teams, different architectures, those solution, how do you, again bridge across, and this is what we will do in the multicloud era, we will help our community to bridge across and take advantage of these powerful cycles of innovation that are going on, but be able to use them across a consistent infrastructure and operational environment. And we'll have a lot more to talk about on this topic today. You know, and maybe the last item to bridge across maybe the most important, you know, people who are profit. You know, too often we think about this as an either or question. And as a business leader, I'm are worried about the people or the And Milton Friedman probably set us up for this issue decades ago when he said, planet, right? the sole purpose of a business is to make profits. You want to create a multi-decade dilemma, right? For business leaders, could I have both people and profits? Could I do well and do good? And particularly for technology, I think we don't have a choice to think about these separately. We are permeating every aspect of business. And Society, we have the responsibility to do both and have all the things that vm ware has accomplished. I think this might be the one that I'm most proud of over, you know, w we have demonstrated by vsphere and the hypervisor alone that we have saved over 540 million tons of co two emissions. That is what you have done. Can you believe that? Five hundred 40 million tons is enough to have 68 percent of all households for a year. Wow. Thank you for what you have done. Thank you. Or another translation of that. Is that safe enough to drive a trillion miles and the average car or you could go to and from Jupiter just in case that was in your itinerary a thousand times. Right? He was just incredible. What we have done and as a result of that, and I'll say we were thrilled to accept this recognition on behalf of you and what you have done. You know, vm were recognized as number 17 in the fortune. Change the world list last week. And we really view it as accepting this honor on behalf of what you have done with our products and technology tech as a force for good. We believe that fundamentally that is our opportunity, if not our obligation, you know, fundamentally tech is neutral, you know, we together must shape it for good. You know, the printing press by Gutenberg in 1440, right? It was used to create mass education and learning materials also can be used for extremist propaganda. The technology itself is neutral. Our ecosystem has a critical role to play in shaping technology as a force for good. You know, and as we think about that tomorrow, we'll have a opportunity to have a very special guest and I really encourage you to be here, be on time tomorrow morning on the stage and you know, Sanjay's a session, we'll have Malala, Nobel Peace Prize winner and fourth will be a bit of extra security as you come in and you understand that. And I just encourage you not to be late because we see this tech being a force for good in everything that we do at vm ware. And I hope you'll enjoy, I'm quite looking forward to the session tomorrow. Now as we think about the future. I like to put it in this context, the superpowers of tech know and you know, 38 years in the industry, you know, I am so excited because I think everything that we've done over the last four decades is creating a foundation that allows us to do more and go faster together. We're unlocking game, changing opportunities that have not been available to any people in the history of humanity. And we have these opportunities now and I, and I think about these four cloud, you have unimaginable scale. You'll literally with your Amex card, you can go rent, you know, 10,000 cores for $100 per hour. Or if you have Michael's am ex card, we can rent a million cores for $10,000 an hour. Thanks Michael. But we also know that we're in many ways just getting started and we have tremendous issues to bridge across and compatible clouds, mobile unprecedented scale. Literally, your application can reach half the humans on the planet today. But we also know that five percent, the lowest five percent of humanity or the other half of humanity, they're still in the lower income brackets, less than five percent penetrated. And we know that we have customer examples that are using mobile phones to raise impoverished farmers in Africa, out of poverty just by having a smart phone with proper crop, the information field and whether a guidance that one tool alone lifting them out of poverty. Ai knows, you know, I really love the topic of ai in 1986. I'm the chief architect of the 80 46. Some of you remember what that was. Yeah, I, you know, you're, you're my folk, right? Right. And for those of you who don't, it was a real important chip at the time. And my marketing manager comes running into my office and he says, Pat, pat, we must make the 46 a great ai chip. This is 1986. What happened? Nothing an AI is today, a 30 year overnight success because the algorithms, the data have gotten so much bigger that we can produce results, that we can bring intelligence to everything. And we're seeing dramatic breakthroughs in areas like healthcare, radiology, you know, new drugs, diagnosis tools, and designer treatments. We're just scratching the surface, but ai has so many gaps, yet we don't even in many cases know why it works. Right? And we'll call that explainable ai and edge and Iot. We're connecting the physical and the digital worlds was never before possible. We're bridging technology into every dimension of human progress. And today we're largely hooking up things, right? We have so much to do yet to make them intelligent. Network secured, automated, the patch, bringing world class it to Iot, but it's not just that these are super powers. We really see that each and each one of them is a super power in and have their own right, but they're making each other more powerful as well. Cloud enables mobile conductivity. Mobile creates more data, more data makes the AI better. Ai Enables more edge use cases and more edge requires more cloud to store the data and do the computing right? They're reinforcing each other. And with that, we know that we are speeding up and these superpowers are reshaping every aspect of society from healthcare to education, the transportation, financial institutions. This is how it all comes together. Now, just a simple example, how many of you have ever worn a hardhat? Yeah, Yo. Pretty boring thing. And it has one purpose, right? You know, keep things from smacking me in the here's the modern hardhat. It's a complete heads up display with ar head. Well, vr capabilities that give the worker safety or workers or factory workers or supply people the ability to see through walls to understand what's going on inside of the equipment. I always wondered when I was a kid to have x Ray Vision, you know, some of my thoughts weren't good about why I wanted it, but you know, I wanted to. Well now you can have it, you know, but imagine in this environment, the complex application that sits behind it. You know, you're accessing maybe 50 year old building plants, right? You're accessing HVAC systems, but modern ar and vr capabilities and new containerized displays. You'll think about that application. You know, John Gage famously said the network is the computer pat today says the application is now a network and pretty typically a complicated one, you know, and this is the vm ware vision is to make that kind of environment realizable in every aspect of our business and community and we simply have been on this journey, any device, any application, any cloud with intrinsic security. And this vision has been consistent for those of you who have been joining us for a number of years. You've seen this picture, but it's been slowly evolving as we've worked in piece by piece to refine and extend this vision, you know, and for it, we're going to walk through and use this as the compass for our discussion today as we walk through our conversation. And you know, we're going to start by a focus on any cloud. And as we think about this cloud topic, you know, we see it as a multicloud world hybrid cloud, public cloud, but increasingly seeing edge and telco becoming clouds in and have their own right. And we're not gonna spend time on it today, but this area of Telco to the is an enormous opportunity for us in our community. You know, data centers and cloud today are over 80 percent virtualized. The Telco network is less than 10 percent virtualized. Wow. An industry that's almost as big as our industry entirely unvirtualized, although the technologies we've created here can be applied over here and Telco and we have an enormous buildout coming with five g and environments emerging. What an opportunity for us, a virgin market right next to us and we're getting some early mega winds in this area using the technologies that you have helped us cure rate than the So we're quite excited about this topic area as well. market. So let's look at this full view of the multicloud. Any cloud journey. And we see that businesses are on a multicloud journey, you know, and today we see this fundamentally in these two paths, a hybrid cloud and a public cloud. And these paths are complimentary and coexisting, but today, each is being driven by unique requirements and unique teams. Largely the hybrid cloud is being driven by it. And operations, the public cloud being driven more by developers and line of business requirements and as some multicloud environment. So how do we deliver upon that and for that, let's start by digging in on the hybrid cloud aspect of this and as we think about the hybrid cloud, we've been talking about this subject for a number of years and I want to give a very specific and crisp definition. You're the hybrid cloud is the public cloud and the private cloud cooperating with consistent infrastructure and consistent operations simply put seamless path to and from the cloud that my workloads don't care if it's here or there. I'm able to run them in a agile, scalable, flexible, efficient manner across those two environments, whether it's my data center or someone else's, I can bring them together to make that work is the magic of the Vm ware Cloud Foundation. The vm ware Cloud Foundation brings together computer vsphere and the core of why we are here, but combines with that networking storage delivered through a layer of management and automation. The rule of the cloud is ruthlessly automate everything. We laid out this vision of the software defined data center seven years ago and we've been steadfastly working on this vision and vm ware. Cloud Foundation provides this consistent infrastructure and operations with integrated lifecycle management automation. Patching the m ware cloud foundation is the simplest path to the hybrid cloud and the fastest way to get vm ware cloud foundation is hyperconverged infrastructure, you know, and with this we've combined integrated then validated hardware and as a building block inside of this we have validated hardware, the v Sand ready environments. We have integrated appliances and cloud delivered infrastructure, three ways that we deliver that integrate integrated hyperconverged infrastructure solution. And we have by far the broadest ecosystem of partners to do it. A broad set of the sand ready nodes from essentially everybody in the industry. Secondly, we have integrated appliances, the extract of vxrail that we have co engineered with our partners at Dell technology and today in fact Dell is releasing the power edge servers, a major step in blade servers that again are going to be powering vxrail and vxrack systems and we deliver hyperconverged infrastructure through a broader set of Vm ware cloud partners as well. At the heart of the hyperconverged infrastructure is v San and simply put, you know, be San has been the engine that's just been moving rapidly to take over the entire integration of compute and storage and expand to more and more areas. We have incredible momentum over 15,000 customers for v San Today and for those of you who joined us, we say thank you for what you have done with this product today. Really amazing you with 50 percent of the global 2000 using it know vm ware. V San Vxrail are clearly becoming the standard for how hyperconverge is done in the industry. Our cloud partner programs over 500 cloud partners are using ulv sand in their solution, you know, and finally the largest in Hci software revenue. Simply put the sand is the software defined storage technology of choice for the industry and we're seeing that customers are putting this to work in amazing ways. Vm Ware and Dell technologies believe in tech as a force for good and that it can have a major impact on the quality of life for every human on the planet and particularly for the most underdeveloped parts of the world. Those that live on less than $2 per day. In fact that this moment 5 billion people worldwide do not have access to modern affordable surgery. Mercy ships is working hard to change the global surgery crisis with greater than 400 volunteers. Mercy ships operates the largest NGO hospital ship delivering free medical care to the poorest of the poor in Africa. Let's see from them now. When the ship shows up to port, literally people line up for days to receive state of the art life, sane changing life saving surgeries, tumor site limbs, disease blindness, birth defects, but not only that, the personnel are educating and training the local healthcare providers with new skills and infrastructure so they can care for their own. After the ship has left, mercy ships runs on Vm ware, a dell technology with VX rail, Dell Isilon data protection. We are the it platform for mercy ships. Mercy ships is now building their next generation ship called global mercy, which were more than double. It's lifesaving capacity. It's the largest charity hospital ever. It will go live in 20 slash 20 serving Africa and I personally plan on being there for its launch. It is truly amazing what they are doing with our technology. Thanks. So we see this picture of the hybrid cloud. We've talked about how we do that for the private cloud. So let's look over at the public cloud and let's dig into this a little bit more deeply. You know, we're taking this incredible power of the Vm ware Cloud Foundation and making it available for the leading cloud providers in the world and with that, the partnership that we announced almost two years ago with Amazon and on the stage last year, we announced their first generation of products, no better example of the hybrid cloud. And for that it's my pleasure to bring to stage my friend, my partner, the CEO of aws. Please welcome Andy Jassy. Thank you andy. You know, you honor us with your presence, you know, and it really is a pleasure to be able to come in front of this audience and talk about what our teams have accomplished together over the last, uh, year. Yo, can you give us some perspective on that, Andy and what customers are doing with it? Well, first of all, thanks for having me. I really appreciate it. It's great to be here with all of you. Uh, you know, the offering that we have together customers because it allows them to use the same software they've been using to again, where cloud and aws is very appealing to manage their infrastructure for years to be able to deploy it an aws and we see a lot of customer momentum and a lot of customers using it. You see it in every imaginable vertical business segment in transportation. You see it with stagecoach and media and entertainment. You see it with discovery communications in education, Mit and Caltech and consulting and accenture and cognizant and dxc you see in every imaginable vertical business segment and the number of customers using the offering is doubling every quarter. So people were really excited about it and I think that probably the number one use case we see so far, although there are a lot of them, is customers who are looking to migrate on premises applications to the cloud. And a good example of that is mit. We're there right now in the process of migrating. In fact, they just did migrate 3000 vms from their data centers to Vm ware cloud native us. And this would have taken years before to do in the past, but they did it in just three months. It was really spectacular and they're just a fun company to work with and the team there. But we're also seeing other use cases as well. And you're probably the second most common example is we'll say on demand capabilities for things like disaster recovery. We have great examples of customers you that one in particular, his brakes, right? Urban in those. The brings security trucks and they all armored trucks coming by and they had a critical need to retire a secondary data center that they were using, you know, for Dr. so we quickly built to Dr Protection Environment for $600. Bdms know they migrated their mission critical workloads and Wallah stable and consistent Dr and now they're eliminating that site and looking for other migrations as well. The rate of 10 to 15 percent. It was just a great deal. One of the things I believe Andy, he'll customers should never spend capital, uh, Dr ever again with this kind of capability in place. That is just that game changing, you know, and you know, obviously we've been working on expanding our reach, you know, we promised to make the service available a year ago with the global footprint of Amazon and now we've delivered on that promise and in fact today or yesterday if you're an ozzie right down under, we announced in Sydney, uh, as well. And uh, now we're in US Europe and in APJ. Yeah. It's really, I mean it's very exciting. Of course Australia is one of the most virtualized places in the world and, and it's pretty remarkable how fast European customers have started using the offering to and just the quarter that's been out there and probably have the many requests customers has had. And you've had a, probably the number one request has been that we make the offering available in all the regions. The aws has regions and I can tell you by the end of 2019 will largely be there including with golf clubs and golf clap. You guys have been, that's been huge for you guys. Yeah. It's a government only region that we have that a lot of federal government workloads live in and we are pretty close together having the offering a fedramp authority to operate, which is a big deal on a game changer for governments because then there'll be able to use the familiar tools they use and vm ware not just to run their workloads on premises but also in the cloud as well with the data privacy requirements, security requirements they need. So it's a real game changer for government too. Yeah. And this you can see by the picture here basically before the end of next year, everywhere that you are and have an availability zone. We're going to be there running on data. Yup. Yeah. Let's get with it. Okay. We're a team go faster. Okay. You'll and you know, it's not just making it available, but this pace of innovation and you know, you guys have really taught us a few things in this respect and since we went live in the Oregon region, you know, we've been on a quarterly cadence of major releases and two was really about mission critical at scale and we added our second region. We added our hybrid cloud extension with m three. We moved the global rollout and we launched in Europe with m four. We really add a lot of these mission critical governance aspects started to attack all of the industry certifications and today we're announcing and five right. And uh, you know, with that, uh, I think we have this little cool thing you know, two of the most important priorities for that we're doing with ebs and storage. Yeah, we'll take, customers, our cost and performance. And so we have a couple of things to talk about today that we're bringing to you that I think hit both of those on a storage side. We've combined the elasticity of Amazon Elastic Block store or ebs with ware is Va v San and we've provided now a storage option that you'll be able to use that as much. It's very high capacity and much more cost effective and you'll start to see this initially on the Vm ware cloud. Native us are five instances which are compute instances, their memory optimized and so this will change the cost equation. You'll be able to use ebs by default and it'll be much more cost effective for storage or memory intensive workloads. Um, it's something that you guys have asked for. It's been very frequently requested it, it hits preview today. And then the other thing is that we've worked really hard together to integrate vm ware's Nsx along with aws direct neck to have a private even higher performance conductivity between on premises and the cloud. So very, very exciting new capabilities to show deep integration between the companies. Yeah. You know, in that aspect of the deep integration. So it's really been the thing that we committed to, you know, we have large engineering teams that are working literally every day. Right on bringing together and how do we fuse these platforms together at a deep and intimate way so that we can deliver new services just like elastic drs and the c and ebs really powerful, uh, capabilities and that pace of innovation continue. So next maybe. Um, maybe six. I don't know. We'll see. All right. You know, but we're continuing this toward pace of innovation, you know, completing all of the capabilities of Nsx. You'll full integration for all of the direct connect to capabilities. Really expanding that. You're only improving licensed capabilities on the platform. We'll be adding pks on top of for expanded developer a capabilities. So just. Oh, thank you. I, I think that was formerly known as Right, and y'all were continuing this pace of storage Chad. So anyway. innovation going forward, but I think we also have a few other things to talk about today. Andy. Yeah, I think we have some news that hopefully people here will be pretty excited about. We know we have a pretty big database business and aws and it's. It's both on the relational and on the nonrelational side and the business is billions of dollars in revenue for us and on the relational side. We have a service called Amazon relational database service or Amazon rds that we have hundreds of thousands of customers using because it makes it much easier for them to set up, operate and scale their databases and so many companies now are operating in hybrid mode and will be for a while and a lot of those customers have asked us, can you give us the ease of manageability of those databases but on premises. And so we talked about it and we thought about and we work with our partners at Vm ware and I'm excited to announce today, right now Amazon rds on Vm ware and so that will bring all the capabilities of Amazon rds to vm ware's customers for their on premises environments. And so what you'll be able to do is you'll be able to provision databases. You'll be able to scale the compute or the memory or the storage for those database instances. You'll be able to patch the operating system or database engines. You'll be able to create, read replicas to scale your database reads and you can deploy this rep because either on premises or an aws, you'll be able to deploy and high high availability configuration by replicating the data to different vm ware clusters. You'll be able to create online backups that either live on premises or an aws and then you'll be able to take all those databases and if you eventually want to move them to aws, you'll be able to do so rather easily. You have a pretty smooth path. This is going to be available in a few months. It will be available on Oracle sql server, sql postgresql and Maria DB. I think it's very exciting for our customers and I think it's also a good example of where we're continuing to deepen the partnership and listen to what customers want and then innovate on their behalf. Absolutely. Thank you andy. It is thrilling to see this and as we said, when we began the partnership, it was a deep integration of our offerings and our go to market, but also building this bi-directional hybrid highway to give customers the capabilities where they wanted cloud on premise, on premise to the cloud. It really is a unique partnership that we've built, the momentum we're feeling to our customer base and the cool innovations that we're doing. Andy, thank you so much for you Jordan Young, rural 20th. You guys appreciate it. Yeah, we really have just seen incredible momentum and as you might have heard from our earnings call that we just finished this. We finished the last quarter. We just really saw customer momentum here. Accelerating. Really exciting to see how customers are starting to really do the hybrid cloud at scale and with this we're just seeing that this vm ware cloud foundation available on Amazon available on premise. Very powerful, but it's not just the partnership with Amazon. We are thrilled to see the momentum of our Vm ware cloud provider program and this idea of the vm ware cloud providers has continued to gain momentum in the industry and go over five years. Right. This program has now accumulated more than 4,200 cloud partners in over 120 countries around the globe. It gives you choice, your local provider specialty offerings, some of your local trusted partners that you would have in giving you the greatest flexibility to choose from and cloud providers that meet your unique business requirements. And we launched last year a program called Vm ware cloud verified and this was saying you're the most complete embodiment of the Vm ware Cloud Foundation offering by our cloud partners in this program and this logo you know, allows you to that this provider has achieved the highest standard for cloud infrastructure and that you can scale and deliver your hybrid cloud and partnering with them. It know a particular. We've been thrilled to see the momentum that we've had with IBM as a huge partner and our business with them has grown extraordinarily rapidly and triple digits, but not just the customer count, which is now over 1700, but also in the depth of customers moving large portions of the workload. And as you see by the picture, we're very proud of the scope of our partnerships in a global basis. The highest standard of hybrid cloud for you, the Vm ware cloud verified partners. Now when we come back to this picture, you know we, you know, we're, we're growing in our definition of what the hybrid cloud means and through Vm Ware Cloud Foundation, we've been able to unify the private and the public cloud together as never before, but we're also seeing that many of you are interested in how do I extend that infrastructure further and farther and will simply call that the edge right? And how do we move data closer to where? How do we move data center resources and capacity closer to where the data's being generated at the operations need to be performed? Simply the edge and we'll dig into that a little bit more, but as we do that, what are the things that we offer today with what we just talked about with Amazon and our VCP p partners is that they can consume as a service this full vm ware Cloud Foundation, but today we're only offering that in the public cloud until project dimension of project dimension allows us to extend delivered as a service, private, public, and to the edge. Today we're announcing the tech preview, a project dimension Vm ware cloud foundation in a hyperconverged appliance. We're partnered deeply with Dell EMC, Lenovo for the first partners to bring this to the marketplace, built on that same proven infrastructure, a hybrid cloud control plane, so literally just like we're managing the Vm ware cloud today, we're able to do that for your on premise. You're small or remote office or your edge infrastructure through that exact same as a service management and control plane, a complete vm ware operated end to end environment. This is project dimension. Taking the vcf stack, the full vm ware cloud foundation stack, making an available in the cloud to the edge and on premise as well, a powerful solution operated by BM ware. This project dimension and project dimension allows us to have a fundamental building block in our approach to making customers even more agile, flexible, scalable, and a key component of our strategy as well. So let's click into that edge a little bit more and we think about the edge in the following layers, the compute edge, how do we get the data and operations and applications closer to where they need to be. If you remember last year I talked about this pendulum swinging of centralization and decentralization edge is a decentralization force. We're also excited that we're moving the edge of the devices as well and we're doing that in two ways. One with workspace, one for human optimized devices and the second is project pulse or Vm ware pulse. And today we're announcing pulse two point zero where you can consume it now as a service as well as with integrated security. And we've now scaled pulse to support 500 million devices. Isn't that incredible, right? I mean this is getting a scale. Billions and billions and finally networking is a key component. You all that. We're stretching the networking platform, right? And evolving how that edge operates in a more cloud and that's a service white and this is where Nsx St with Velo cloud is such a key component of delivering the edge of network services as well. Taken together the device side, the compute edge and rethinking and evolving the networking layer together is the vm ware edge strategy summary. We see businesses are on this multicloud journey, right? How do we then do that for their private of public coming together, the hybrid cloud, but they're also on a journey for how they work and operate it across the public cloud and the public cloud we have this torrid innovation, you'll want Andy's here, challenges. You know, he's announcing 1500 new services or were extraordinary innovation and you'll same for azure or Google Ibm cloud, but it also creates the same complexity as we said. Businesses are using multiple public clouds and how do I operate them? How do I make them work? You know, how do I keep track of my accounts and users that creates a set of cloud operations problems as well in the complexity of doing that. How do you make it work? Right? And your for that. We'll just see that there's this idea cloud cost compliance, analytics as these common themes that of, you know, keep coming up and we're seeing in our customers that are new role is emerging. The cloud operations role. You're the person who's figuring out how to make these multicloud environments work and keep track of who's using what and which data is landing where today I'm thrilled to tell you that the, um, where is acquiring the leader in this space? Cloudhealth technologies. Thank you. Cloudhealth technologies supports today, Amazon, azure and Google. They have some 3,500 customers, some of the largest and most respected brands in the, as a service industry. And Sasa business today rapidly span expanding feature sets. We will take cloudhealth and we're going to make it a fundamental platform and branded offering from the um, where we will add many of the other vm ware components into this platform, such as our wavefront analytics, our cloud, choreo compliance, and many of the other vm ware products will become part of the cloudhealth suite of services. We will be enabling that through our enterprise channels as well as through our MSP and BCPP partners as well know. Simply put, we will make cloudhealth the cloud operations platform of choice for the industry. I'm thrilled today to have Joe Consella, the CTO and founder. Joe, please stand up. Thank you joe to your team of a couple hundred, you know, mostly in Boston. Welcome to the Vm ware family, the Vm ware community. It is a thrill to have you part of our team. Thank you joe. Thank you. We're also announcing today, and you can think of this, much like we had v realize operations and v realize automation, the compliment to the cloudhealth operations, vm ware, cloud automation, and some of you might've heard of this in the past, this project tango. Well, today we're announcing the initial availability of Vm ware, cloud automation, assemble, manage complex applications, automate their provisioning and cloud services, and manage them through a brokerage the initial availability of cloud automation services, service. Your today, the acquisition of cloudhealth as a platform, the aware of the most complete set of multicloud management tools in the industry, and we're going to do so much more so we've seen this picture of this multicloud journey that our customers are on and you know, we're working hard to say we are going to bridge across these worlds of innovation, the multicloud world. We're doing many other things. You're gonna hear a lot at the show today about this year. We're also giving the tech preview of the Vm ware cloud marketplace for our partners and customers. Also today, Dell technologies is announcing their cloud marketplace to provide a self service, a portfolio of a Dell emc technologies. We're fundamentally in a unique position to accelerate your multicloud journey. So we've built out this any cloud piece, but right in the middle of that any cloud is the network. And when we think about the network, we're just so excited about what we have done and what we're seeing in the industry. So let's click into this a little bit further. We've gotten a lot done over the last five years. Networking. Look at these numbers. 80 million switch ports have been shipped. We are now 10 x larger than number two and software defined networking. We have over 7,500 customers running on Nsx and maybe the stat that I'm most proud of is 82 percent of the fortune 100 has now adopted nsx. You have made nsx these standard and software defined networking. Thank you very much. Thank you. When we think about this journey that we're on, we started. You're saying, Hey, we've got to break the chains inside of the data center as we said. And then Nsx became the software defined networking platform. We started to do it through our cloud provider partners. Ibm made a huge commitment to partner with us and deliver this to their customers. We then said, boy, we're going to make a fundamental to all of our cloud services including aws. We built this bridge called the hybrid cloud extension. We said we're going to build it natively into what we're doing with Telcos, with Azure and Amazon as a service. We acquired the St Wagon, right, and a Velo cloud at the hottest product of Vm ware's portfolio today. The opportunity to fundamentally transform branch and wide area networking and we're extending it to the edge. You're literally, the world has become this complex network. We have seen the world go from the old defined by rigid boundaries, simply put in a distributed world. Hardware cannot possibly work. We're empowering customers to secure their applications and the data regardless of where they sit and when we think of the virtual cloud network, we say it's these three fundamental things, a cloud centric networking fabric with intrinsic security and all of it delivered in software. The world is moving from data centers to centers of data and they need to be connected and Nsx is the way that we will do that. So you'll be aware of is well known for this idea of talking but also showing. So no vm world keynote is okay without great demonstrations of it because you shouldn't believe me only what we can actually show and to do that know I'm going to have our CTL come onstage and CTL y'all. I used to be a cto and the CTO is the certified smart guy. He's also known as the chief talking officer and today he's my demo partner. Please walk, um, Vm ware, cto ray to the stage. Right morning pat. How you doing? Oh, it's great ray, and thanks so much for joining us. Know I promised that we're going to show off some pretty cool stuff here. We've covered a lot already, but are you up to the task? We're going to try and run through a lot of demos. We're going to do it fast and you're going to have to keep me on time to ask an awkward question. Slow me down. Okay. That's my fault if you run along. Okay, I got it. I got it. Let's jump right in here. So I'm a CTO. I get to meet lots of customers that. A few weeks ago I met a cio of a large distribution company and she described her it infrastructure as consisting of a number of data centers troll to us, which he also spoke of a large number of warehouses globally, and each of these had local hyperconverged compute and storage, primarily running surveillance and warehouse management applications, and she pulls me four questions. The first question she asked me, she says, how do I migrate one of these data centers to Vm ware cloud on aws? I want to get out of one of these data centers. Okay. Sounds like something andy and I were just talking exactly, exactly what you just spoke to a few moments ago. She also wanted to simplify the management of the infrastructure in the warehouse as themselves. Okay. He's age and smaller data centers that you've had out there. Her application at the warehouses that needed to run locally, butter developers wanted to develop using cloud infrastructure. Cloud API is a little bit late. The rds we spoken with her in. Her final question was looking to the future, make all this complicated management go away. I want to be able to focus on my application, so that's what my business is about. So give me some new ways of how to automate all of this infrastructure from the edge to the cloud. Sounds pretty clear. Can we do it? Yes we can. So we're going to dive right in right now into one of these demos. And the first demo we're going to look at it is vm ware cloud on aws. This is the best solution for accelerating this public cloud journey. So can we start the demo please? So what you were looking at here is one of those data centers and you should be familiar with this product. It's a familiar vsphere client. You see it's got a bunch of virtual machines running in there. These are the virtual machines that we now want to be able to migrate and move the VMC on aws. So we're going to go through that migration right now. And to do that we use a product that you've seen already atx, however it's the x has been, has got some new cool features since the last time we download it. Probably on this stage here last year, I wanted those in particular is how do we do bulk migration and there's a new cool thing, right? Whole thing we want to move the data center en mass and his concept here is cloud motion with vsphere replication. What this does is it replicates the underlying storage of the virtual machines using vsphere replication. So if and when you want to now do the final migration, it actually becomes a vmotion. So this is what you see going on right here. The replication is in place. Now when you want to touch you move those virtual machines. What you'll do is a vmotion and the key thing to think about here is this is an actual vmotion. Those the ends as room as they're moving a hustler, migrating remained life just as you would in a v motion across one particular infrastructure. Did you feel complete application or data center migration with no dying town? It's a Standard v motion kind of appearance. Wow. That is really impressive. That's correct. Wow. You. So one of the other things we want to talk about here is as we are moving these virtual machines from the on prem infrastructure to the VMC on aws infrastructure, unfortunately when we set up the cloud on VMC and aws, we only set up for hosts, uh, that might not be, that'd be enough because she is going to move the whole infrastructure of that this was something you guys, you and Andy referred to briefly data center. Now, earlier, this concept of elastic drs. what elastic drs does, it allows the VMC on aws to react to the workloads as they're being created and pulled in onto that infrastructure and automatically pull in new hosts into the VMC infrastructure along the way. So what you're seeing here is essentially the MC growing the infrastructure to meet the needs of the workloads themselves. Very cool. So overseeing that elastic drs. we also see the ebs capabilities as well. Again, you guys spoke about this too. This is the ability to be able to take the huge amount of stories that Amazon have, an ebs and then front that by visa you get the same experience of v Sign, but you get this enormous amount of storage capabilities behind it. Wow. That's incredible. That's incredible. I'm excited about this. This is going to enable customers to migrate faster and larger than ever before. Correct. Now she had a series of little questions. Okay. The second question was around what about all those data centers and those age applications that I did not move, and this is where we introduce the project which you've heard of already tonight called project dementia. What this does, it gives you the simplicity of Vm ware cloud, but bringing that out to the age, you know what's basically going on here, vmc on aws is a service which manages your infrastructure in aws. We know stretch that service out into your infrastructure, in your data center and at the age, allowing us to be able to manage that infrastructure in the same way. Once again, let's dive down into a demo and take a look at what this looks like. So what you've got here is a familiar series of services available to you, one of them, which is project dimension. When you enter project dimension, you first get a view of all of the different infrastructure that you have available to you, your data centers, your edge locations. You can then dive deeply into one of these to get a closer look at what's going on here. We're diving into one of these The problem is there's a networking problem going on in this warehouse. warehouses and we see it as a problem here. How do we know? We know because vm ware is running this as a managed service. We are directly managing or sorry, monitoring your infrastructure or we discover there's something going wrong here. We automatically create the ASR, so somebody is dealing with this. You have visibility to what's going on, but the vm ware managed service is already chasing the problem for you. Oh, very good. So now we're seeing this dispersed infrastructure with project dementia, but what's running on it so well before we get with running out, you've got another problem and the problem is of course, if you're managing a lot of infrastructure like this, you need to keep it up to date. And so once again, this is where the vm ware managed service kicks in. We manage that infrastructure in terms of patching it and updating it for you. And as an example, when we released a security patch, here's one for the recent l, one terminal fault, the Vmr managed service is already on that and making sure that your on prem and edge infrastructure is up to date. Very good. Now, what's running? Okay. So what's running, uh, so we mentioned this case of this software running at the edge infrastructure itself, and these are workloads which are running locally in those age, uh, those edge locations. This is a surveillance application. You can see it here at the bottom it says warehouse safety monitor. So this is an application which gathers images and then stores those images He said my sql database on top there, now this is where we leverage the somewhere and it puts them in a database. technology you just learned about when Andy and pat spoke about disability to take rds and run that on your on prem infrastructure. The block of virtual machines in the moment are the rds components from Amazon running in your infrastructure or in your edge location, and this gives you the ability to allow your developers to be able to leverage and operate against those Apis, but now the actual database, the infrastructure is running on prem and you might be doing just for performance reasons because of latency, you might be doing it simply because this data center is not always connected to the cloud. When you take a look into under the hood and see what's going on here, what you actually see this is vsphere, a modified version of vsphere. You see this new concept of my custom availability zone. That is the availability zone running on your infrastructure which supports or ds. What's more interesting is you flip back to the Amazon portal. This is typically what your developers are going to do. Once again, you see an availability zone in your Amazon portal. This is the availability zone running on your equipment in your data center. So we've truly taken that already as infrastructure and moved it to the edge so the developer sees what they're comfortable with and the infrastructure sees what they're comfortable with bridging those two worlds. Fabulous. Right. So the final question of course that we got here was what's next? How do I begin to look to the future and say I am going to, I want to be able to see all of my infrastructure just handled in an automated fashion. And so when you think about that, one of the questions there is how do we leverage new technologies such as ai and ml to do that? So what you've got here is, sorry we've got a little bit later. What you've got here is how do I blend ai in a male and the power of what's in the data center itself. Okay. And we could do that. We're bringing you the AI and ml, right? And fusing them together as never before to truly change how the data center operates. Correct. And it is this introduction is this merging of these things together, which is extremely powerful in my mind. This is a little bit like a self driving vehicle, so thinking about a car driving down the street is self driving vehicle, it is consuming information from all of the environment around it, other vehicles, what's happening, everything from the wetter, but it also has a lot of built in knowledge which is built up to to self learning and training along the way in the kids collecting lots of that data for decades. Exactly. And we've got all that from all the infrastructure that we have. We can now bring that to bear. So what we're focusing on here is a project called project magna and project. Magna leverage is all of this infrastructure. What it does here is it helps connect the dots across huge datasets and again a deep insight across the stack, all the way from the application hardware, the infrastructure to the public cloud, and even the age and what it does, it leverages hundreds of control points to optimize your infrastructure on Kpis of cost performance, even user specified policies. This is the use of machine language in order to fundamentally transform. I'm sorry, machine learning. I'm going back to some. Very early was here, right? This is the use of machine learning and ai, which will automatically transform. How do you actually automate these data centers? The goal is true automation of your infrastructure, so you get to focus on the applications which really served needs of your business. Yeah, and you know, maybe you could think about that as in the past we would have described the software defined data center, but in the future we're calling it the self driving data center. Here we are taking that same acronym and redefining it, right? Because the self driving data center, the steep infusion of ai and machine learning into the management and automation into the storage, into the networking, into vsphere, redefining the self driving data center and with that we believe fundamentally is to be an enormous advance and how they can take advantage of new capabilities from bm ware. Correct. And you're already seeing some of this in pieces of projects such as some of the stuff we do in wavefront and so already this is how do we take this to a new level and that's what project magnet will do. So let's summarize what we've seen in a few demos here as we work in true each of these very quickly going through these demos. First of all, you saw the n word cloud on aws. How do I migrate an entire data center to the cloud with no downtime? Check, we saw project dementia, get the simplicity of Vm ware cloud in the data center and manage it at the age as a managed service check. Amazon rds and Vm ware. Cool Demo, seamlessly deploy a cloud service to an on premises environment. In this case already. Yes, we got that one coming in are in m five. And then finally project magna. What happens when you're looking to the future? How do we leverage ai and ml to self optimize to virtual infrastructure? Well, how did ray do as our demo guy? Thank you. Thanks. Thanks. Right. Thank you. So coming back to this picture, our gps for the day, we've covered any cloud, let's click into now any application, and as we think about any application, we really view it as this breadth of the traditional cloud native and Sas Coobernetti is quickly maybe spectacularly becoming seen as the consensus way that containers will be managed and automate as the framework for how modern APP teams are looking at their next generation environment, quickly emerging as a key to how enterprises build and deploy their applications today. And containers are efficient, lightweight, portable. They have lots of values for developers, but they need to also be run and operate and have many infrastructure challenges as well. Managing automation while patch lifecycle updates, efficient move of new application services, know can be accelerated with containers. We also have these infrastructure problems and you know, one thing we want to make clear is that the best way to run a container environment is on a virtual machine. You know, in fact, every leader in public cloud runs their containers and virtual machines. Google the creator and arguably the world leader in containers. They runs them all in containers. Both their internal it and what they run as well as G K, e for external users as well. They just announced gke on premise on vm ware for their container environments. Google and all major clouds run their containers and vms and simply put it's the best way to run containers. And we have solved through what we have done collectively the infrastructure problems and as we saw earlier, cool new container apps are also typically some ugly combination of cool new and legacy and existing environments as well. How do we bridge those two worlds? And today as people are rapidly moving forward with containers and Coobernetti's, we're seeing a certain set of problems emerge. And Dan cone, right, the director of CNCF, the Coobernetti, uh, the cloud native computing foundation, the body for Coobernetti's collaboration and that, the group that sort of stewards the standardization of this capability and he points out these four challenges. How do you secure them? How do you network and you know, how do you monitor and what do you do for the storage underneath them? Simply put, vm ware is out to be, is working to be is on our way to be the dial tone for Coobernetti's. Now, some of you who were in your twenties might not know what that means, so we know over to a gray hair or come and see me afterward. We'll explain what dial tone means to you or maybe stated differently. Enterprise grade standard for Cooper netties and for that we are working together with our partners at Google as well as pivotal to deliver Vm ware, pks, Cooper netties as an enterprise capability. It builds on Bosh. The lifecycle engine that's foundational to the pivotal have offerings today, uh, builds on and is committed to stay current with the latest Coobernetti's releases. It builds on Nsx, the SDN container, networking and additional contributions that were making like harbor the Vm ware open source contribution for the container registry. It packages those together makes them available on a hybrid cloud as well as public cloud environments with pks operators can efficiently deploy, run, upgrade their coopernetties environments on SDDC or on all public clouds. While developers have the freedom to embrace and run their applications rapidly and efficiently, simply put, pks, the standard for Coobernetti's in the enterprise and underneath that Nsx you'll is emerging as the standard for software defined networking. But when we think about and we saw that quote on the challenges of Kubernetes today, we see that networking is one of the huge challenge is underneath that and in a containerized world, things are changing even more rapidly. My network environment is moving more quickly. NSX provides the environment's easily automate networking and security for rapid deployment of containerized environments that fully supports the MRP chaos, fully supports pivotal's application service, and we're also committed to fully support all of the major kubernetes distribution such as red hat, heptio and docker as well Nsx, the only platform on the planet that can address the complexity and scale of container deployments taken together Vm Ware, pks, the production grade computer for the enterprise available on hybrid cloud, available on major public clouds. Now, let's not just talk about it again. Let's see it in action and please walk up to the stage. When di Carter with Ray, the senior director of cloud native marketing for Vm ware. Thank you. Hi everybody. So we're going to talk about pks because more and more new applications are built using kubernetes and using containers with vm ware pts. We get to simplify the deploying and the operation of Kubernetes at scale. When the. You're the experts on all of this, right? So can you take as true the scenario of how pks or vm ware pts can really help a developer operating the Kubernedes environment, developed great applications, but also from an administrator point of view, I can really handle things like networking, security and those configurations. Sounds great. I love to dive into the demo here. Okay. Our Demo is. Yeah, more pks running coubernetties vsphere. Now pks has a lot of cool functions built in, one of which is Nsx. And today what I'm going to show you is how NSX will automatically bring up network objects as quick Coobernetti's name spaces are spun up. So we're going to start with the fees per client, which has been extended to Ron pks, deployed cooper clusters. We're going to go into pks instance one, and we see that there are five clusters running. We're going to select one other clusters, call application production, and we see that it is running nsx. Now a cluster typically has multiple users and users are assigned namespaces, and these namespaces are essentially a way to provide isolation and dedicated resources to the users in that cluster. So we're going to check how many namespaces are running in this cluster and more brought up the Kubernetes Ui. We're going to click on namespace and we see that this cluster currently has four namespaces running wire. We're going to do next is bringing up a new name space and show that Nsx will automatically bring up the network objects required for that name space. So to do that, we're going to upload a Yammel file and your developer may actually use Ku Kata command to do this as well. We're going to check the namespace and there it is. We have a new name space called pks rocks. Yeah. Okay. Now why is that guy now? It's great. We have a new name space and now we want to make sure it has the network elements assigned to us, so we're going to go to the NSX manager and hit refresh and there it is. PKS rocks has a logical robber and a logical switch automatically assigned to it and it's up and running. So I want to interrupt here because you made this look so easy, right? I'm not sure people realize the power of what happened here. The developer, winton using Kubernetes, is api infrastructure to familiar with added a new namespace and behind the scenes pks and tardy took care of the networking. It combination of Nsx, a combination of what we do at pks to truly automate this function. Absolutely. So this means that if you are on the infrastructure operation, you don't need to worry about your developer springing up namespaces because Nsx will take care of bringing the networking up and then bringing them back down when the namespace is not used. So rate, but that's not it. Now, I was in operations before and I know how hard it is for enterprises to roll out a new product without visibility. Right, so pks took care of those dates, you operational needs as well, so while it's running your clusters, it's also exporting Meta data so that your developers and operators can use wavefront to gain deep visibility into the health of the cluster as well as resources consumed by the cluster. So here you see the wavefront Ui and it's showing you the number of nodes running, active parts, inactive pause, et cetera. You can also dive deeper into the analytics and take a look at information site, Georgia namespace, so you see pks rocks there and you see the number of active nodes running as well as the CPU utilization and memory consumption of that nice space. So now pks rocks is ready to run containerized applications and microservices. So you just get us a very highlight of a demo here to see a little bit what pks pks says, where can we learn more? So we'd love to show you more. Please come by the booth and we have more cool functions running on pks and we'd love to have you come by. Excellent. Thank you, Lindy. Thank you. Yeah, so when we look at these types of workloads now running on vsphere containers, Kubernedes, we also see a new type of workload beginning to appear and these are workloads which are basically machine learning and ai and in many cases they leverage a new type of infrastructure, hardware accelerators, typically gps. What we're going to talk about here is how in video and Vm ware have worked together to give you flexibility to run sophisticated Vdi workloads, but also to leverage those same gpu for deep learning inference workloads also on vsphere. So let's dive right into a demo here. Again, what you're seeing here is again, you're looking at here, you're looking at your standard view realized operations product, and you see we've got two sets of applications here, a Vdi desktop workload and machine learning, and the graph is showing what's happening with the Vdi desktops. These are office workers leveraging these desktops everyday, so of course the infrastructure is super busy during the daytime when they're in the office, but the green area shows this is not been used very heavily outside of those times. So let's take a look. What happens to the machine learning application in this case, this organization leverages those available gpu to run the machine learning operations outside the normal working hours. Let's take a little bit of a deeper dive into what the application it is before we see what we can do from an infrastructure and configuration point of view. So this machine learning application processes a vast number of images and it clarify or sorry, it categorizes these images and as it's doing so, it is moving forward and putting each of these in a database and you can see it's operating here relatively fast and it's leveraging some gps to do that. So typical image processing type of machine learning problem. Now let's take a dive in and look at the infrastructure which is making this happen. First of all, we're going to look only at the Vdi employee Dvt, a Vdi infrastructure here. So I've got a bunch of these applications running Vdi applications. What I want to do is I want to move these so that I can make this image processing out a application run a lot faster. Now normally you wouldn't do this, but pot insisted that we do this demo at 10:30 in the morning when the office workers are in there, so we're going to move older Vdi workloads over to the other cluster and that's what you're seeing is going on right now. So as they move over to this other cluster, what we are now doing is freeing up all of the infrastructure. The GPU that Vdi workload was using here. We see them moving across and now you've freed up that infrastructure. So now we want to take a look at this application itself, the machine learning application and see how we can make use of that. Now freed up infrastructure we've got here is the application is running using one gpu in a vsphere cluster, but I've got three more gpu is available now because I've moved the Vdi workloads. We simply modify the application, let it know that these are available and you suddenly see an increase in the processing capabilities because of what we've done here in terms of making the flexibility of accessing those gps. So what you see here is the same gps that youth for Vdi, which you probably have in your infrastructure today, can also be used to run sophisticated machine learning and ai type of applications on your vsphere infrastructure. So let's summarize what we've seen in the various demos here in this section. First of all, we saw how the MRPS simplifies the deployment and operating operation of Kubernetes at scale. What we've also seen is that leveraging the Nvidia Gpu, we can now run the most demanding workloads on vsphere. When we think about all of these applications and these new types of workloads that people are running. I want to take one second to speak to another workload that we're seeing beginning to appear in the data center. And this is of course blockchain. We're seeing an increasing number of organizations evaluating blockchains for smart contract and digital consensus solutions. So this tech, this technology is really becoming or potentially becoming a critical role in how businesses will interact each other, how they will work together. We'd project concord, which is an open source project that we're releasing today. You get the choice, performance and scale of verifiable trust, which you can then bring to bear and run in the enterprise, but this is not just another blockchain implementation. We have focused very squarely on making sure that this is good for enterprises. It focuses on performance, it focuses on scalability. We have seen examples where running consensus algorithms have taken over 80 days on some of the most common and widely used infrastructure in blockchain and we project conquered. You can do that in two and a half hours. So I encourage you to check out this project on get hub today. You'll also see lots of activity around the whole conference. Speaking about this. Now we're going to dive into another section which is the anti device section. And for that I need to welcome pat back up there. Thank you pat. Thanks right. So diving into any device piece of the puzzle, you and as we think about the superpowers that we have, maybe there are no more area that they are more visible than in the any device aspect of our picture. You know, and as we think about this, the superpowers, you know, think about mobility, right? You know, and how it's enabling new things like desktop as a service in the mobile area, these breadth of smartphones and devices, ai and machine learning allow us to manage them, secure them and this expanding envelope of devices in the edge that need to be connected and wearables and three d printers and so on. We've also seen increasing research that says engaged employees are at the center of business success. Engaged employees are the critical ingredient for digital transformation. And frankly this is how I run vm ware, right? You know, I have my device and my work, all my applications, every one of my 23,000 employees is running on our transformed workspace one environment. Research shows that companies that, that give employees ready anytime access are nearly three x more likely to be leaders in digital transformation. That employees spend 20 percent of their time today on manual processes that can be automated. The way team collaboration and speed of division decisions increases by 16 percent with engaged employees with modern devices. Simply put this as a critical aspect to enabling your business, but you remember this picture from the silos that we started with and each of these environments has their own tribal communities of management, security automation associated with them, and the complexity associated with these is mind boggling and we start to think about these. Remember the I'm a pc and I'm a Mac. Well now you have. I'm an Ios. I'm a droid and other bdi and I'm now a connected printer and I'm a connected watch. You remember citrix manager and good is now bad and sccm a failed model and vpns and Xanax. The chaos is now over at the center of that is vm ware, workspace one, get it out of the business of managing devices, automate them from the cloud, but still have the mentor price. Secure cloud based analytics that brings new capabilities to this critical topic. You'll focus your energy on creating employee and customer experiences. You know, new capabilities to allow like our airlift, the new capability to help customers migrate from their sccm environment to a modern management, expanding the use of workspace intelligence. Last year we announced the chromebook and a partnership with HP and today I'm happy to announce the next step in our partnerships with Dell. And uh, today we're announcing that Dell provisioning for Vm ware, workspace one as part of Dell's ready to work solutions Dallas, taking the next leap and bringing workspace one into the core of their client to offerings. And the way you can think about this as Literally a dell drop ship, lap pops showing up to new employee. day one, productivity. You give them their credential and everything else is delivered by workspace one, your image, your software, everything patched and upgraded, transforming your business, right beginning at that device experience that you give to your customer. And again, we don't want to talk about it. We want to show you how this works. Please walk to the stage with re renew the head of our desktop products marketing. Thank you. So we just heard from pat about how workspace one integrated with Dell laptops is really set up to manage windows devices. What we're broadly focused on here is how do we get a truly modern management system for these devices, but one that has an intelligence behind it to make sure that we're kept with a good understanding of how to keep these devices always up to date and secure. Can we start the demo please? So what we're seeing here is to be the the front screen that you see of workspace one and you see you've got multiple devices a little bit like that demo that patch assured. I've got Ios, android, and of course I've got windows renewal. Can you please take us through how workspace one really changes the ability of somebody an it administrator to update and manage windows into our environment? Absolutely. With windows 10, Microsoft has finally joined the modern management body and we are really excited about that. Now. The good news about modern management is the frequency of ostp updates and how quickly they come out because you can address all those security issues that are hitting our radar on a daily basis, but the bad news about modern management is the frequency of those updates because all of us in it admins, we have to test each and every one of our applications would that latest version because we don't want to roll out that update in case of causes any problems with workspace one, we saw that we simply automate and provide you with the APP compatibility information right out of the box so you can now automate that update process. Let's take a quick look. Let's drill down here further into the windows devices. What we'll see is that only a small percentage of those devices are on that latest version of operating system. Now, that's not a good thing because it might have an important security fix. Let's scroll down further and see what the issue is. We find that it's related to app compatibility. In fact, 38 percent of our devices are blocked from being upgraded and the issue is app compatibility. Now we were able to find that not by asking the admins to test each and every one of those, but we combined windows analytics data with APP intelligent out of the box and be provided that information right here inside of the console. Let's dig down further and see what those devices and apps look like. So knew this is the part that I find most interesting. If I am a system administrator at this point I'm looking at workspace one is giving me a key piece of information. It says if you proceed with this update, it's going to fail 84, 85 percent at a time. So that's an important piece of information here, but not alone. Is it telling me that? It is telling me roughly speaking why it thinks it's going to fail. We've got a number of apps which are not ready to work with this new version, particularly the Mondo card sales lead tracker APP. So what we need to do is get engineering to tackle the problems with this app and make sure that it's updated. So let's get fixing it in order to fix it. What we'll do is create an automation and we can do this right out of the box in this automation will open up a Jira ticket right from within the console to inform the engineers about the problem, not just that we can also flag and send a notification to that engineering manager so that it's top of mine and they can get working on this fixed right away. Let's go ahead and save that automation right here, ray UC. There's the automation that we just So what's happening here is essentially this update is now scheduled meeting. saved. We can go and update oldest windows devices, but workspace one is holding the process of proceeding with that update, waiting for the engineers to update the APP, which is going to cause the problem. That's going to take them some time, right? So the engineers have been working on this, they have a fixed and let's go back and see what's happened to our devices. So going back into the ios updates, what we'll find is now we've unblocked those devices from being upgraded. The 38 percent has drastically dropped down. It can rest in peace that all of the devices are compliant and on that latest version of operating system. And again, this is just a snapshot of the power of workspace one to learn more and see more. I invite you all to join our EOC showcase keynote later this evening. Okay. So we've spoken about the presence of these new devices that it needs to be able to manage and operate across everything that they do. But what we're also seeing is the emergence of a whole new class of computing device. And these are devices which are we commonly speak to have been at the age or embedded devices or Iot. And in many cases these will be in factories. They'll be in your automobiles, there'll be in the building, controlling, controlling, uh, the building itself, air conditioning, etc. Are quite often in some form of industrial environment. There's something like this where you've got A wind farm under embedded in each of these turbines. This is a new class of computing which needs to be managed, secured, or we think virtualization can do a pretty good job of that in new virtualization frontier, right at the edge for iot and iot gateways, and that's gonna. That's gonna, open up a whole new realm of innovation in that space. Let's dive down and taking the demo. This spaces. Well, let's do that. What we're seeing here is a wind turbine farm, a very different than a data center than what we're used to and all the compute infrastructure is being managed by v center and we see to edge gateway hose and they're running a very mission critical safety watchdog vm right on there. Now the safety watchdog vm is an fte mode because it's collecting a lot of the important sensor data and running the mission critical operations for the turbine, so fte mode or full tolerance mode, that's a pretty sophisticated virtualization feature allowing to applications to essentially run in lockstep. So if there's a failure, wouldn't that gets to take over immediately? So this no sophisticated virtualization feature can be brought out all the way to the edge. Exactly. So just like in the data center, we want to perform an update, so as we performed that update, the first thing we'll do is we'll suspend ft on that safety watchdog. Next, we'll put two. Oh, five into maintenance mode. Once that's done, we'll see the power of emotion that we're all familiar with. We'll start to see all the virtual machines vmotion over to the second backup host. Again, all the maintenance, all the update without skipping a heartbeat without taking down any daily operations. So what we're seeing here is the basic power of virtualization being brought out to the age v motion maintenance mode, et cetera. Great. What's the big deal? We've been doing that for years. What's the, you know, come on. What's the big deal? So what you're on the edge. So when you get to the age pack, you're dealing with a whole new class of infrastructure. You're dealing with embedded systems and new types of cpu hours and process. This whole demo has been done on an arm 64. Virtualization brought to arm 64 for embedded devices. So we're doing this on arm on the edge, correct. Specifically focused for embedded for age oems. Okay. Now that's good. Okay. Thank you ray. Actually, we've got a summary here. Pat, just a second before you disappear. A lot to rattle off what we've just seen, right? We've seen workspace one cross platform management. What we've also seen, of course esx for arm to bring the power of vfx to edge on 64, but are in platforms will go no. Okay. Okay. Thank you. Thanks. Now we've seen a look at a customer who is taking advantage of everything that we just saw and again, a story of a customer that is just changing lives in a fundamental way. Let's see. Make a wish. So when a family gets the news that a child is sick and it's a critical illness, it could be a life threatening illness. The whole family has turned upside down. Imagine somebody comes to you and they say, what's the one thing you want that's in your heart? You tell us and then we make that happen. So I was just calling to give you the good news that we're going to be able to grant jackson a wish make, which is the largest wish granting organizations in the United States. English was featured in the cbs 60 minutes episode. Interestingly, it got a lot of hits, but uh, unfortunately for the it team, the whole website crashed make a wish is going through a program right now where we're centralizing technology and putting certain security standards in place at our chapters. So what you're seeing here, we're configuring certain cloud services to make sure that they always are able to deliver on the mission whether they have a local problem or not is we continue to grow the partnership and work with vm ware. It's enabling us to become more efficient in our processes and allows us to grant more wishes. It was a little girl. She had a two year old brother. She just wanted a puppy and she was forthright and I want to name the puppy in my name so my brother would always have me to list them off a five year old. It's something we can't change their medical outcome, but we can change their spiritual outcome and we can transform their lives. Thank you. Working together with you truly making wishes come true. The last topic I want to touch on today, and maybe the most important to me personally is security. You got to fundamentally, when we think about this topic of security, I'll say it's broken today and you know, we would just say that the industry got it wrong that we're trying to bolt on or chasing bad, and when we think about our security spend, we're spending more and we're losing more, right? Every day we're investing more in this aspect of our infrastructure and we're falling more behind. We believe that we have to have much less security products and much more security. You know, fundamentally, you know, if you think about the problem, we build infrastructure, right? Generic infrastructure, we then deploy applications, all kinds of applications, and we're seeing all sorts of threats launched that as daily tens of millions. You're simple virus scanner, right? Is having tens of millions of rules running and changing many times a day. We simply believe the security model needs to change. We need to move from bolted on and chasing bad to an environment that has intrinsic security and is built to ensure good. This idea of built in security. We are taking every one of the core vm ware products and we are building security directly into it. We believe with this, we can eliminate much of the complexity. Many of the sensors and agents and boxes. Instead, they'll directly leverage the mechanisms in the infrastructure and we're using that infrastructure to lock it down to behave as we intended it to ensure good, right on the user side with workspace one on the network side with nsx and microsegmentation and storage with native encryption and on the compute with app defense, we are building in security. We're not chasing threats or adding on, but radically reducing the attack surface. When we look at our applications in the data center, you see this collection of machines running inside of it, right? You know, typically running on vsphere and those machines are increasingly connected. Through nsx and last year we introduced the breakthrough security solution called app defense and app defense. Leverages the unique insight we get into the application so that we can understand the application and map it into the infrastructure and then you can lock down, you could take that understanding, that manifest of its behavior and then lock those vms to that intended behavior and we do that without the operational and performance burden of agents and other rear looking use of attack detection. We're shrinking the attack surface, not chasing the latest attack vector, you know, and this idea of bolt on versus chasing bad. You sort of see it right in the network. Machines have lots of conductivity, lots of applications running and something bad happens. It basically has unfettered access to move horizontally through the data center and most of our security is north, south. MosT of the attacks are eastwest. We introduced this idea of microsegmentation five years ago, and by it we're enabling organizations to secure some networks and separate sensitive applications and services as never before. This idea isn't new, that just was never practical before nsx, but we're not standing still. Our teams are innovating to leap beyond 12. What's next beyond microsegmentation, and we see this in three simple words, learn, imagine a system that can look into the applications and understand their behavior and how they should operate. we're using machine learning and ai instead of chasing were to be able to ensure good where that that system can then locked down its behavior so the system consistently operates that way, but finally we know we have a world of increasing dynamic applications and as we move to more containerize the microservices, we know this world is changing, so we need to adapt. We need to have more automation to adapt to the current behavior. Today I'm very excited to have two major announcements that are delivering on this vision. The first of those vsphere platinum, our flagship vm ware vsphere product now has app defense built right in platinum will enable virtualization teams. Yeah, go ahead. Yeah, let's use it. Platinum will enable virtualization teams you to give an enormous contribution to the security profile of your enterprise. You could see whatever vm is for its purpose, its behavior until the system. That's what it's allowed to do. Dramatically reducing the attack surface without impact. On operations or performance, the capability is so powerful, so profound. We want you to be able to leverage it everywhere, and that's why we're building it directly into vsphere, vsphere platinum. I call it the burger and fries. You know, nobody leaves the restaurant without the fries who would possibly run a vm in the future without turning security on. That's how we want this to work going forward. Vsphere platinum and as powerful as microsegmentation has been as an idea. We're taking the next step with what we call adaptive microsegmentation. We are fusing Together app defense and vsphere with nsx to allow us to align the policies of the application through vsphere and the network. We can then lock down the network and the compute and enable this automation of the microsegment formation taken together adaptive microsegmentation. But again, we don't want to just tell you about it. We want to show you. Please welcome to the stage vj dante, who heads our machine learning team for app dispense. Vj a very good vj. Thanks for joining us. So, you know, I talked about this idea right, of being able to learn, lock and adapt. Uh, can you show it to us? Great. Yeah. Thank you. With vc a platinum, what we have done is we have put in everything you need to learn, lock and adapt, right with the infrastructure. The next time you bring up your wifi at line, you'll actually see a difference right in there. Let's go with that demo. There you go. And when you look at our defense there, what you see is that all your guests, virtual machines and all your host, hundreds of them and thousands of virtual machines enabling for that difference. It's in there. And what that does is immediately gets you visibility into the processes running on those virtual machines and the risk for the first time. Think about it for the first time. You're looking at the infrastructure through the lens of an application. Here, for example, the ecommerce application, you can see the components that make up that application, how they interact with each other, the specific process, a specific ip address on a specific board. That's what you get, but so we're learning the behavior. Yes. Yeah, that's very good. But how do you make sure you only learn good behavior? Exactly. How do we make sure that it's not bad? We actually verify me insured. It's all good. We ensured that everybody these reputation is verified. We ensured that the haven is verified. Let's go to svc host, for example. This process can exhibit hundreds of behaviors across numerous. Realize what we do here is we actually verify that failure saw us. It's actually a machine learning models that had been trained on millions of instances of good, bad at you said, and then automatically verify that for okay, so we said, you. We learned simply, learn now, lock. How does that work? Well, once you learned the application, locking it is as simple as clicking on that verify and protect button and then you can lock both the compute and network and it's done. So we've pushed those policies into nsx and microsegmentation has been established actually locked down the compute. What is the operating system is exactly. Let's first look at compute, protected the processes and the behaviors are locked down to exactly what is allowed for that application. And we have bacon policies and program your firewall. This is nsx being configured automatically for you, laurie, with one single click. Very good. So we said learn lock. Now, how does this adapt thing work? Well, a bad change is the only constant, but modern applications applications change on a continuous basis. What we do is actually pretty simple. We look at every change as it comes in determinant is good or bad. If it's good, we say allow it, update the policies. That's bad. We denied. Let's look at an example as asco dxc. It's exhibiting a behavior that they've not seen getting the learning period. Okay? So this machine has never behave this This hasn't been that way. But. way. But again, our machine learning models had seen thousands of instances of this process. They know this is normal. It talks on three 89 all the time. So what it's done to the few things, it's lowered the criticality of the alarm. Okay, so false positive. Exactly. The bane of security operations, false positives, and it has gone and updated. Jane does locks on compute and network to allow for that behavior. Applications continues to work on this project. Okay, so we can learn and adapt and action right through the compute and the network. What about the client? Well, we do with workplace one, intelligence protect and manage end user endpoint, but what's one intelligence? Nsx and actually work together to protect your entire data center infrastructure, but don't believe me. You can watch it for yourself tomorrow tom cornu keynote. You want to be there, at 1:00 PM, be there or be nowhere. I love you. Thank you veejay. Great job. Thank you so much. So the idea of intrinsic security and ensuring good, we believe fundamentally changing how security will be delivered in the enterprise in the future and changing the entire security industry. We've covered a lot today. I'm thrilled as I stand on stage to stand before this community that truly has been at the center of changing the world of technology over the last couple of decades. In it. We've talked about this idea of the super powers of technology and as they accelerate the huge demand for what you do, you know in the same way we together created this idea of the virtual infrastructure admin. You'll think about all the jobs that we are spawning in the discussion that we had today, the new skills, the new opportunities for each one of us in this room today, quantum program, machine learning engineer, iot and edge expert. We're on the cusp of so many new capabilities and we need you and your skills to do that. The skills that you possess, the abilities that you have to work across these silos of technology and enabled tomorrow. I'll tell you, I am now 38 years in the industry and I've never been more excited because together we have the opportunity to build on the things that collective we have done over the last four decades and truly have a positive global impact. These are hard problems, but I believe together we can successfully extend the lifespan of every human being. I believe together we can eradicate chronic diseases that have plagued mankind for centuries. I believe we can lift the remaining 10 percent of humanity out of extreme poverty. I believe that we can reschedule every worker in the age of the superpowers. I believe that we can give modern ever education to every child on the planet, even in the of slums. I believe that together we could reverse the impact of climate change. I believe that together we have the opportunity to make these a reality. I believe this possibility is only possible together with you. I asked you have a please have a wonderful vm world. Thanks for listening. Happy 20th birthday. Have a great topic.

Published Date : Aug 28 2018

SUMMARY :

of devices in the edge that need to be

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmazonORGANIZATION

0.99+

AndyPERSON

0.99+

IBMORGANIZATION

0.99+

MichaelPERSON

0.99+

1998DATE

0.99+

TelcoORGANIZATION

0.99+

1986DATE

0.99+

TelcosORGANIZATION

0.99+

EuropeLOCATION

0.99+

Paul MaritzPERSON

0.99+

DellORGANIZATION

0.99+

BostonLOCATION

0.99+

Andy JassyPERSON

0.99+

LenovoORGANIZATION

0.99+

10QUANTITY

0.99+

DeloitteORGANIZATION

0.99+

JoePERSON

0.99+

SydneyLOCATION

0.99+

Joe ConsellaPERSON

0.99+

AfricaLOCATION

0.99+

Pat GelsingerPERSON

0.99+

OregonLOCATION

0.99+

20 percentQUANTITY

0.99+

AshleyPERSON

0.99+

16 percentQUANTITY

0.99+

VegasLOCATION

0.99+

JupiterLOCATION

0.99+

Last yearDATE

0.99+

last yearDATE

0.99+

first questionQUANTITY

0.99+

LindyPERSON

0.99+

telcoORGANIZATION

0.99+

John GagePERSON

0.99+

10 percentQUANTITY

0.99+

MicrosoftORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

Dan conePERSON

0.99+

68 percentQUANTITY

0.99+

200 applicationsQUANTITY

0.99+

GoogleORGANIZATION

0.99+

50 percentQUANTITY

0.99+

Vm Ware Cloud FoundationORGANIZATION

0.99+

1440DATE

0.99+

30 yearQUANTITY

0.99+

HPORGANIZATION

0.99+

38 percentQUANTITY

0.99+

38 yearsQUANTITY

0.99+

$600QUANTITY

0.99+

20 yearsQUANTITY

0.99+

one monthsQUANTITY

0.99+

firstQUANTITY

0.99+

todayDATE

0.99+

windows 10TITLE

0.99+

hundredsQUANTITY

0.99+

yesterdayDATE

0.99+

80 millionQUANTITY

0.99+

five percentQUANTITY

0.99+

second questionQUANTITY

0.99+

JodyPERSON

0.99+

TodayDATE

0.99+

tomorrowDATE

0.99+

SanjayPERSON

0.99+

23,000 employeesQUANTITY

0.99+

five peopleQUANTITY

0.99+

sixth yearQUANTITY

0.99+

82 percentQUANTITY

0.99+

five instancesQUANTITY

0.99+

tomorrow morningDATE

0.99+

CoobernettiORGANIZATION

0.99+

Dave Buckley, Paddy Power Betfair | OpenStack Summit 2018


 

(upbeat electronic music) >> Announcer: Live from Vancouver, Canada, it's theCUBE, covering OpenStack Summit North America 2018. Brought to you by Red Hat, the OpenStack Foundation, and its ecosystem partners. >> Welcome back to The Cube's coverage of OpenStack Summit 2018 in Vancouver. I'm Stu Miniman with my co-host John Troyer. Happy to welcome back a company we've spoken to a few times at events, Paddy Power Betfair. First time guest coming to us from across the pond, Dave Buckley who is the automation engineer with Paddy Power Betfair, thanks for joining us. >> Thank you for having me. >> Alright, so first of all, you've been to a couple summits and we've talked to Paddy Power about OpenStack. Before we get into your specific implementation, tell us about your experience here this week and any compare, contrast to previous years. >> Yeah so I'm very lucky, I got to come to the previous two summits in North America. I guess what I've enjoyed this week, it's kind of a slight tilt towards, it's away from being purely OpenStack, kind of towards this open infrastructure kind of thing, 'cause like I said, especially last year in Boston, Q and NEs was becoming a big thing. Yeah, and kind of, the OpenStack Foundation becoming kind of more, not that it wasn't before, but more community-based and being part of the ecosystem. So, yeah, I think it's been quite interesting seeing that. >> Not to put words in your mouth but, it was even, the last year or two, it's more aware of some of the complimentary things and adding pieces. You know, we had, one of the interviews we did this week was person who's the SIC lead for the Kubernetes stuff, that sits under another Foundation, things like that. Yeah, exactly. It's been quite interesting this week, I guess, sort the Kata Container project, which wasn't something I'd been aware of before Monday morning basically. I remember we were sitting in the keynotes, and they were like, you can have this container-like thing which has all the speed of a container, but it's as secure as a BM. And you're thinking, how, how is that even possible? So I've really enjoyed, I got to go to one of the sessions yesterday, one of the technical introductions on that. >> Yeah, I always love, there's certain things where, okay, this is what I'm going to do with my schedule, and turns into, this got announced, or I didn't know about this, and you knew, blow up my schedule, let me change everything else. Yeah, exactly, I think you always, you can't, you have to be flexible, right? Adaptable, and as the week goes on you just go to what you think is interesting. >> John: So Dave, you and your company have been working with OpenStack for quite a while. >> Dave: Yeah. >> And you obviously run a system that needs to be stable. Right, needs to, you take care of betting and people's money. >> Dave: Exactly. >> So that needs to be solid. But I understand you recently went though an upgrade and have some experiences talking about that? Can you talk a little bit about where you are with your OpenStack implementation and that sort of migration? >> Sure. So, I guess it's about three years ago, it was Betfair at the time, so this was before the merger of the two companies. So Betfair started using OpenStack, and I think it was actually the last time the summit was here, in Vancouver. So a couple of my colleagues who were kind of the technical leads at the time. Steve Armstrong and Steve Perera, they flew out here, to kind of get a feel for OpenStack, what it was, talk to people who'd had experiences with it. I actually think that conference back then was very informative of what the platform today now looks like. So some of the conversations they had there with people like New Age Networks and Arista, which we used for the switching, but conversations they had there kind of ended up being now what we're using in production. I guess over the past couple of years, so the big thing that happened obviously was this merger between Paddy Power and Betfair, following that they had an exercise which they called the single customer platform, which is annoyingly, for a sys-admin guy, kind of like me, they, it's always been abbreviated to SCP, but you have to ignore that. So that was to kind of consolidate and integrate the Paddy Power and Betfair co-bases and put it on a single platform, which was our OpenStack and Nuage platform. So that kind of completed in January this year, so that's live, so basically the Paddy Power sports book has an entirely new website, all running on OpenStack. A lot quicker and more efficient then the previous version. So that's been a real success. And as part of that, I should say that stability is really vital, so kind of in our business. If the site is down we don't make any money, and if it happens during a big sporting event you have a big problem. >> Do you have a metric around that? What a minute or an hour of down time would be? >> So I guess it always depends, so the nature of our traffic is very spikey. So obviously when you have a big, it's on a Saturday in Europe, the football, soccer, maybe I should say, is like a very big deal. >> We have a global audience, football's okay. >> I'll stick with football then. >> We were all watching the royal wedding. >> I don't want to talk about that. The football, if you, we just get peak traffic on that day. And, even within the year, there's a thing called the Grand National, which is a big event in the UK, big horse racing, I guess like the Kentucky Derby. It's kind of when we get our maximum traffic in the year. Yeah, you always need to be prepared for that. So one of the things as you mentioned, we kind of look into upgrade OpenStack from Kilo to Newton. So we've been on Kilo from the start. We're using Red Hat's distribution of OpenStack, so what Red Hat offer is this, they have like every three releases I think it is? They have this long release life-cycle. So that's kind of the reason we're going to Newton, cause we have kind of the, then the support will go to 2021. [Stu] - But if I remember, it's Red Hat the OpenStack Platform 10. >> Dave: Yeah. >> And 13 is going to be queened as their next one that's going to be released. >> Exactly, so I think they just announced that this week, right? So I think at some point in the next year or two we'd be going to queens. >> How do you determine when you make that jump and anything around the upgrade process, you know, good and bad that you could share. >> Dave: Yeah, so I guess going from, we were overdue an upgrade in this case, Kilos, you know, pretty old now. What we're lucky that we can do is because we have Nuage, it's like an external SDM provider, so the entire data plane is controlled by Nuage, and you can kind of plug as many OpenStacks as you like really into Nuage, and you offload all the networking to Nuage. So what's that's allowed us to do is basically we'd have had a lot of trouble if we'd had to do an in place upgrade, so I've actually been to one of the groups this week, quite a lot of people were talking about upgrades and just like all the nightmares it's caused. I know it's getting better as like the releases come out, but what we were able to do is kind of building new, an entirely new OpenStack cloud on the side of, so we've kind of turned it kind of an immutable OpenStack, so your OSB 7 cloud is there, we built this new OSB 10. But they're both circ into the same networking, so the same Nuage SDN. And the way our developers deploy their applications, I guess you want to see this in more detail, we've done presentations at these summits in the past, but kind of in short, every deployment we do immutable deployments as well, so for every deployment we'll create a new subnet within Nuage, and kind of do rolling update of your VMs that are on that new subnet into like a VIP which is kind of where the constant is, so all the traffic's come in to that VIP then you just flip things in and out below it when you do a deployment, so what that basically means is from a developers point of view, when they're migrating from OSB 7 to OSB 10 they'll essentially spin up new networks and new VMs in OSB 10 and that deployment pipeline will kind of just seamlessly, everything else will stay the same because the networking doesn't change. So we don't have to have any downtime on the data plane or the control plane. Which is really beneficial for us 'cause the way, I guess this is I'll just describe the way developers do deployments like we rely heavily on the OpenStack API being available. You pay a cost in that you, so you need extra hardware to do that I guess, but yeah we found it is something that's worked for us. >> John: Anything else with the networking and specifically that you all are running, the load balancing or resiliency that you need to have for your apps? >> Dave: Yeah so one of the things was, so it's kind of another problem there were trying to solve with this whole project, this new OpenStack platform is that historically Betfair, as it was at the time, had always run out of a single data-center. But we had another site, but it was mainly kind of a development environments right in there. So the company thought why don't we just have, we should just have both DCs for resiliency, try and run things in like an active-active configuration. Which is fine for external customer facing applications where we've had an external load balance server that can point traffic between the two DCs. But then the question is what do you do with internal apps? So this is what led us to use Avi Networks, which is kind of a cloud native load balancing technology, so we've been using to provide like GSLB internal laps, so basically we'll load balance traffic between the two data-centers so it gets deployed within your OpenStack environment, has a really neat integration with Nuage, the Nuage SDN layer, and will resolve you to whichever data-center is appropriate at that time. So if you have a full data-center outtage, you should be able to go "Okay, point stuff over there". >> John: So it makes you and the networking team or the IT team into the heroes not the villains, you're usually the people saying "No" or "We can't do that". >> I guess so, I guess so yeah you're probably right. It's cool technology though. I guess that we're very lucky and that we're given the opportunity by the people at the company to experiment with new things, so even though we're about stability but we're also about trying to push things forward in terms of what technology to use. >> Stu: Dave I'm curious how kind of the hybrid or multi-cloud type of environments fit into what you're doing today, give us the update there. >> Dave: Yeah so that's something very in our radar at the moment I guess it's, yeah it's what everybody's doing, looking to how you can have this hybrid cloud model. So I think, going back three years again, at that time, being like an online betting company, it's a highly regulated business and only at that point it was really possible to kind of put some of this stuff into the public cloud, it seems like things have come a long way, so it's something we're looking at at the moment, we're evaluating different solutions, different vendors like the Googles, AWSs, and seeing or even like some OpenStack public clouds and seeing maybe how could we migrate some workloads out into the public cloud, how do we want to that, to give us more resiliency, and also as I was saying about our spiky traffic, it just makes a lot of sense to be able to say burst out into whichever public cloud vendor on a Saturday when the football's on to deal with that peak load. So it's something we're very much looking at at the moment. But yeah no formal decisions as of yet. Unless they've done something while I've been away. >> John: With containers here at the show, lots of different threads right? Containers, Edge, the OpenDev track, things like that. Anything else, we've talked about Kata, anything else that came up that was interesting here that you just watch Kubernetes and container track as well? >> Dave: So I guess in terms of containers it's, sitting in the Keynotes on Monday you would, if you weren't watching if you were just listening, you probably wouldn't know you were at an OpenStack Summit right since there's as much Kubernetes container stuff as there is OpenStack. It's interesting so we've kind of been doing... Again, similar to the public cloud conversation, it's something that's very relevant to us at the moment, we've done kind of a few proof-of-concept ideas, evaluating different solutions, so we have like running a Cube cluster ourself, obviously we have a strong relationship with Red Hat that we've kind of explored to using OpenShift maybe, and then come the networking layer you can integrate with Nuage which would be really cool for us so it'll allow us to do kind of the all the networking, access control mechanisms as we do for our virtual machines. And again this is also something in the whole public cloud conversation is well if wanted to containers in the public cloud as well like you have all the different offerings, would we want to run our own, in like an AWS or something? Or maybe go to someone like Google where you have that supported self-service model I suppose. But yeah at the moment it's kind of at those stages so I think Steve did a presentation on the Kubernetes stuff like a PCO we done at the last Summit. But yeah still at the moment still want to make some firm decisions about which direction we're going to go but a lot of the developers a very keen for this and obviously for guys like us we all know the value of it so I think at the moment because we had that focus on stability we should now have a period of time where we're able to kind of look at all this stuff a bit more, hopefully get some container solutions into production which would be awesome. >> Stu: Dave Buckley we really appreciate you giving us the update, love to be able to do some of those longitudinal case studies as to where you've been where you're going, what you're thinking about. Be sure to check out thecube.net, you can actually search for Patty Power Betfair, see some of those previous interviews from Dave's peers. Loads more interviews there as well as all the shows we're going to be at in the future where hope you come by and say "Hi". For John Troyer I'm Stu Miniman, thanks so much for watching theCUBE. >> (electro-dance music) >> (soft piano)

Published Date : May 24 2018

SUMMARY :

Brought to you by Red Hat, the OpenStack Foundation, First time guest coming to us from across the pond, and any compare, contrast to previous years. Yeah, and kind of, the OpenStack Foundation and they were like, you can have this Adaptable, and as the week goes on you just John: So Dave, you and your company And you obviously run a system that needs to be stable. So that needs to be solid. So some of the conversations they had there So obviously when you have a big, So one of the things as you mentioned, And 13 is going to be queened as their next one So I think at some point in the next year or two and anything around the upgrade process, you know, the traffic's come in to that VIP then you just flip the Nuage SDN layer, and will resolve you to whichever John: So it makes you and the networking team given the opportunity by the people at the company Stu: Dave I'm curious how kind of the hybrid doing, looking to how you can have this hybrid cloud that came up that was interesting here that you just the public cloud as well like you have all the different in the future where hope you come by and say "Hi".

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
StevePERSON

0.99+

Dave BuckleyPERSON

0.99+

DavePERSON

0.99+

John TroyerPERSON

0.99+

JohnPERSON

0.99+

VancouverLOCATION

0.99+

BetfairORGANIZATION

0.99+

Stu MinimanPERSON

0.99+

New Age NetworksORGANIZATION

0.99+

Steve PereraPERSON

0.99+

Steve ArmstrongPERSON

0.99+

Red HatORGANIZATION

0.99+

EuropeLOCATION

0.99+

two companiesQUANTITY

0.99+

AWSsORGANIZATION

0.99+

GooglesORGANIZATION

0.99+

Paddy PowerORGANIZATION

0.99+

North AmericaLOCATION

0.99+

UKLOCATION

0.99+

OpenStack FoundationORGANIZATION

0.99+

AristaORGANIZATION

0.99+

StuPERSON

0.99+

January this yearDATE

0.99+

GoogleORGANIZATION

0.99+

BostonLOCATION

0.99+

OSB 10TITLE

0.99+

2021DATE

0.99+

yesterdayDATE

0.99+

Vancouver, CanadaLOCATION

0.99+

bothQUANTITY

0.99+

next yearDATE

0.99+

OpenStackTITLE

0.99+

OpenStack Summit 2018EVENT

0.99+

MondayDATE

0.99+

KubernetesTITLE

0.99+

two data-centersQUANTITY

0.99+

last yearDATE

0.99+

First timeQUANTITY

0.99+

Kentucky DerbyEVENT

0.98+

this weekDATE

0.98+

OpenShiftTITLE

0.98+

Avi NetworksORGANIZATION

0.98+

oneQUANTITY

0.98+

single platformQUANTITY

0.97+

Paddy Power BetfairORGANIZATION

0.97+

todayDATE

0.97+

OpenStack Summit North America 2018EVENT

0.97+

OSB 7TITLE

0.97+

thecube.netOTHER

0.97+

OpenStackORGANIZATION

0.96+

three yearsQUANTITY

0.96+

Grand NationalEVENT

0.96+

NuageORGANIZATION

0.96+

OpenStack Platform 10TITLE

0.94+

single data-centerQUANTITY

0.93+

two DCsQUANTITY

0.93+

Radhesh Balakrishnan, Red Hat | OpenStack Summit 2018


 

(upbeat music) >> Narrator: Live from Vancouver, Canada, It's theCUBE. Covering OpenStack Summit, North America, 2018. Brought to you by Red Hat, the OpenStack Foundation, and its ecosystem partners. >> Welcome back to theCUBE's coverage of OpenStack Summit 2018, here in Vancouver. Three days wall-to-wall coverage. I'm Stu Miniman with my cohost, John Troyer. Happy to welcome back to the program, Radhesh Balakrishnan, who is the general manager of OpenStack with Red Hat. Radhesh, great to see you. It's been a week since John talked to you, and always good to have you on at the show. >> Great to be on. Good to be here talking about OpenStack at OpenStack Summit. >> Yeah so, look, OpenStack is in the title of your job. I believe, did we have a birthday cake and a party celebrating a certain milestone? >> That is indeed true; so it's the fifth anniversary of that fact that we've had a product, Red Hat OpenStack Platform, on the market. And so, we've been doing a little bit of a look back at how far we have come in the last five years as well as looking ahead at, you know, how does the next three to five years shape as well. >> Yeah, Radhesh, I'm going to date myself and when I think back to, gosh it was 18 years ago, I was working with Linux, and there were kernels all over the place and things like that. And then, I worked for an enterprise storage company and was like, ugh, like keeping up with Chrome.org was a pain in the neck. There came out this thing called Red Hat Advanced Server that was like, oh wait, we can glom onto this, we can support this with our customers, and that eventually turned into RHEL, which, of course, kind of becomes the main standard for how to do Linux. I feel like we have a lot of similarities. >> Radhesh: Absolutely, absolutely. >> In how we did. RHOSP, I believe, is the acronym, so. >> That's exactly right, and we like to have long names. >> Which are very descriptive, but Red Hat OpenStack Platform, fundamentally, to your point brings the same valid proposition that RHEL brought to Linux, to OpenStack, with the twist that, it's not just curated OpenStack, but it's a co-engineered solution of Linux and Cavium and OpenStack. And along the way we learned that, look, it's not just OpenStack and the infrastructure solution. It's done in conjunction with the software-defined storage solution or it's done in conjunction with software-defined networking. Or, fast-forward all the way now, it's being done in conjunction with cloud-native applications running on top of it, right? But regardless, in five years we've been able to grow to address these different demands being placed at infrastructure level, and at the same time evolved to address new-use cases as well; Telco is an example of that. >> Radhesh, let's spend a couple of minutes, though, on the OpenStack Platform itself from Red Hat. Some of the things, guys, that you were bringing to market, I know we talked about, here at the show, fast-forward upgrades, for instance were, they were just introducing, and maybe some other things in the Queens release that you all are bringing forward and have engineered. >> Yeah, thanks for that question, very topical, in the sense that yesterday we launched OSP 13, which is the latest and greatest version based on Queens release. If you look at the innovation packed in that it fundamentally falls in three buckets. One is the bread part that you talked about, whereby, anybody who is standing on OSP 10, which was the prior, long-release lifecycle product, over to 13, how do you kind of get over there in a graceful manner is the first area that we have addressed. The second area is around security, because how do you make sure that OpenStack-based clouds are secure by default, from the day you roll out all the way to until you retire it, right? I don't know if there's going to be a retirement, but that's the intent of all the security enablements that we have in the product as well. And the third one, how do we make sure that containers in OpenStack can come together in a nice manner. >> Yeah, the container piece is something else that, so a lot of effort, here at the show. They announced Kata containers, which, trying to give the security of a VM, lightweight VM. How does Red Hat look at Kata containers? I know Red Hat, you know Linux's containers, you know, very strong position, fill us in on that. >> Yeah, to maybe pull back a little bit and then look at the larger picture of there is the notion of infrastructure or the open infrastructure that you need and OpenStack is a good starting point for that. And then, you overlay on top of that an application deployment management configuration, lifecycle management solution that's the container platform called OpenShift, right. These are the two centers of gravity for the stack. Now, aspects such as Kata containers or Hubbard, which is for again, similar concept of addressing how do you use virtualization in addition to containers to bring some of the value around security et cetera, right? So we are continuing to engage in all these upstream projects, but we'll be careful and methodical in bringing those technologies into our products as we go along. >> Okay, how about Edge is the other kind of major topic that we're having here, I know I've interviewed some Red Hat customers looking at NFV solutions, so some of the big telcos you know specifically that use various pieces. What do you hear from your customers and help us kind of draw that line between the NFV to the Edge. Yeah, so Edge has become the center is kind of the new joke in the sense that, from an NFV perspective, customers have already effectively addressed the CORD errors and the challenges, now it's about how do you scale that and deploy that on a massive scale, right? That's a good problem to have. Now the goodness of virtualization can be brought all the way down to the radio Edge so that a programmable network becomes the reality that a telco or a carrier can get into. So in that context, Edge becomes a series of use cases. You know, it's not just one destination. Another way to say it is there is Edge an objective and there is Edge as a noun. Edge as the objective is a set of technologies that are enabling Edge, Edge networking, right. Edge management, for example, and then Edge as a destination where you have a series of Edge locations starting from CORD error center going all the way to radio. Now, the technology answers for all these are just being figured out right now. So you can say, you know, put crudely, KBM, OpenStack, containers, and Ansible will be all good elements that will come into the picture when it comes to a solution for all these footprints. >> Nice. Radhesh, maybe let's switch over to talk about the summit here, and the people here, filled with people being productive with OpenStack, right? Either looking at it, upgrading it, inheriting it. We talked to people in a bunch of different scenarios Red Hat, huge installed base, and you are good at helping and supporting, and uplifting, and upskilling a set of operators who started with Linux and now have to be responsible for an entire cloud infrastructure. Plus, now, at this conference, we've been talking about containers, we've been talking about open dev, right. That's again broadening the scope of what an operator might have to deal with. How does Red Hat look at that? How are you and your team helping upskill and enhance the role of the operator? >> Yeah, so I think it comes down to, how do we make sure that we are understanding the journey that the operator himself or herself is taking from a career perspective, right, the skill set of evolving from Linux and core automation-related skills to going to being able to understand what does it mean to live with cloud implementation on a day-to-day basis. What does it mean to live with network function virtualization as the way in which new services are going to be deployed. So, our course curriculum has evolved to be able to address all these needs today. That's one dimension, the other dimension is how do we make sure that the product itself is so easy that the journey is getting to a point where the infrastructure is invisible, and the focus is on the application platform on top. So I think we have multiple areas of focus to get to the point where it's so relevant that it's invisible, if that paradox makes sense. That's what we're trying to make happen with OpenStack. >> Radhesh, Red Hat has a very large presence at the show here; we were noting in the keynote the underlying infrastructure didn't get a lot of discussion because it is more mature, and therefore, we can talk about everything like VGPUs and containers, and everything like that. But Red Hat has a lot in the portfolio that helps in some of those underlying pieces. So maybe you can give us some of the highlights there. >> Absolutely, so we aren't looking at OpenStack as the be-all end-all destination for customers, but rather an essential ingredient in the journey to a hybrid cloud. So when you have that lens it becomes natural to you that a portfolio of our offerings, which are either first-party or in conjunction with our partners --we have over 400 partners with whom we have joint solutions as well -- so you naturally take a holistic view and then say, "How do you optimize the experience of ceph plus OpenStack for example." So we were talking about Edge recently, right, in the context of Edge we realize that there is a particular use-case for hyperconverged infrastructure whereby you need to collocate, compute, and store it in a way that the footprint is so small and easy to manage plus you want to have one life-cycle both for OpenStack and ceph right, so to address that we announced, right at hypercloud infrastructure for cloud, as an offering that is co-engineered between ceph team, or our storage team, and the OpenStack team. Right, that's just an example of how, by bringing the rest of the portfolio, we're able to address needs being expressed by our customers today. Or you look forward in terms of use-case, one thing that we are hearing from all our large customers, such as the Amadeus's of the world is, make the experience of OpenShift on OpenStack, easy to deploy and manage, as well as reduce the penalty of running containers on VMs. Because we understand the benefits of security and all of that, but we want to be able to get that without having any penalty of using a virtual infrastructure so that's why we're heavily focused on OpenShift, on OpenStack, as the form-factor for delivering that while continuing to work on things such as Kata containers as well as, you know, Kuryrs, as technology is evolving to make communities much richer as well as the infrastructure management at OpenStack level richer. >> You brought up an interesting point, we spoke a little bit yesterday with John Allessio and Margaret Dawson, about really that kind of multi-cloud world out there, because pieces like Kubernetes and Ansible, aren't just in the data center with this one stack, it's spanning across multiple environments and when we talk to customers, they do cloud, and cloud is multiple things in multiple places and changing all the time. So I'd love to get your viewpoint on what you hear from customers, how Red Hat's helping them across all those environments. >> Absolutely, so the key differentiation we see in being able to provide to our customers is that unlike some of the other providers out there, they're where they are stitching you with a particular private cloud, with the particular public cloud, and then saying, "Hey, this is sort of the equivalent of the AOL walled gardens, if you will, right, that's being created for a particular private and public cloud. What we're saying is fundamentally three things. First is, the foundation of Linux skills from RHEL that you have is going to be what you can build on to innovate for today and tomorrow, that's number one. Secondly, you can invest in infrastructure that is 100% open using OpenStack so that you can use commodity hardware, bring in multiple use-cases which are bleeding it, such as data lags, big data, Apache Spark, or going all the way to cloud-native application, development on top of OpenStack. And then last but not least, when you are embarking on a multi-cloud journey it is important that you're not tied to innovation speed of one particular public cloud provider, or even a private cloud provider, for that matter, so being able to get to a container platform, which is OpenShift, that can run pretty much everywhere, either on PREM or on a public cloud, and give you that single pane of consistency for your application, which is where business and IT alignment is the focus right now, then I think you've got the best of all the worlds. You know, freedom from vendor-lock in, and a future-proof infrastructure and application platform that can take you to where you need to go, right. So pretty excited to be able to deliver on that consistently as of today, as well as in the coming years. >> All right, we just want to give you the final word, for people out there that ... you know, often they get their opinion based on when they first heard of something. OpenStack's been around for a number of years, five years now, with your platform. Give us the takeaway for 2018 here from OpenStack Summit as to how they should be thinking about OpenStack, in that larger picture. >> The key takeaway is that OpenStack is rock-solid, that you can bring into your environment, not just to power your virtual machine infrastructure, but also baremetal infrastructure on which you can bring in containers as well. So if you're thinking about an infrastructure fabric, either to power your telco network or to power your private cloud in its entirety OpenStack is the only place that you need to be looking at and our OpenStack platform from end to end delivers that value proposition. Now the second aspect to think about is, OpenStack is a step in the journey of a hybrid future destination that you can get to. Red Hat not only has the set of surround products and technologies to round-up the solution, but also have the largest partner ecosystem to offer you choice. So what's your excuse from getting to a hybrid cloud today if not tomorrow? >> Well, Radhesh Balakrishnan, thank you for all the updates appreciate catching up with you once again. For John Troyer, I'm Stu Minimam, getting near the end of three days wall-to-wall coverage here in Vancouver, thank you so much for watching theCUBE. (upbeat music)

Published Date : May 23 2018

SUMMARY :

Brought to you by Red Hat, the OpenStack Foundation, and always good to have you on at the show. Great to be on. Yeah so, look, OpenStack is in the title of your job. how does the next three to five years shape as well. the main standard for how to do Linux. RHOSP, I believe, is the acronym, so. and at the same time evolved to address in the Queens release that you all are all the way to until you retire it, right? Yeah, the container piece is something else that, or the open infrastructure that you need and the challenges, now it's about how do you scale that That's again broadening the scope that the journey is getting to a point where at the show here; we were noting in the keynote that the footprint is so small and easy to manage Kubernetes and Ansible, aren't just in the data center of the AOL walled gardens, if you will, right, All right, we just want to give you the final word, OpenStack is the only place that you need to be looking at getting near the end of three days wall-to-wall coverage

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Radhesh BalakrishnanPERSON

0.99+

John TroyerPERSON

0.99+

JohnPERSON

0.99+

VancouverLOCATION

0.99+

RadheshPERSON

0.99+

five yearsQUANTITY

0.99+

Margaret DawsonPERSON

0.99+

100%QUANTITY

0.99+

Stu MinimanPERSON

0.99+

Stu MinimamPERSON

0.99+

Red HatORGANIZATION

0.99+

OpenStack FoundationORGANIZATION

0.99+

TelcoORGANIZATION

0.99+

RHELTITLE

0.99+

John AllessioPERSON

0.99+

FirstQUANTITY

0.99+

2018DATE

0.99+

AOLORGANIZATION

0.99+

third oneQUANTITY

0.99+

second aspectQUANTITY

0.99+

LinuxTITLE

0.99+

fifth anniversaryQUANTITY

0.99+

Three daysQUANTITY

0.99+

tomorrowDATE

0.99+

second areaQUANTITY

0.99+

over 400 partnersQUANTITY

0.99+

OpenShiftTITLE

0.98+

EdgeTITLE

0.98+

yesterdayDATE

0.98+

Vancouver, CanadaLOCATION

0.98+

firstQUANTITY

0.98+

OpenStackTITLE

0.98+

OpenStack Summit 2018EVENT

0.98+

OpenStack SummitEVENT

0.98+

North AmericaLOCATION

0.97+

first areaQUANTITY

0.97+

OSP 10TITLE

0.97+

18 years agoDATE

0.97+

singleQUANTITY

0.97+

bothQUANTITY

0.97+

OneQUANTITY

0.97+

SecondlyQUANTITY

0.97+

theCUBEORGANIZATION

0.97+

AmadeusORGANIZATION

0.96+

AnsibleORGANIZATION

0.95+

oneQUANTITY

0.95+

OpenStackORGANIZATION

0.95+

todayDATE

0.94+

three daysQUANTITY

0.94+

13TITLE

0.94+

OSP 13TITLE

0.94+

OpenStack Summit & Ecosystem Analysis | OpenStack Summit 2018


 

>> Narrator: Vancouver, Canada. It's theCUBE, covering OpenStack Summit North America, 2018. Brought to you by Red Hat, the OpenStack Foundation and its Ecosystem partners. (soft music) >> Hi, and you're watching SiliconANGLE Medias coverage of theCUBE, here at OpenStack Summit 2018 in beautiful Vancouver. I'm Stu Miniman with my co-host John Troyer. We've been here, this is now the third day of coverage, John. We've done a couple dozen interviews already. We've got one more day of coverage. We had some kind of perceptions coming in and I have some interesting differing viewpoints as to where we are for OpenStack the project, where this show itself is going. First of all John, give me your impressions overall. Vancouver, your first time here, city I fell in love with last time I came here, and let's get into the show itself, too. >> Sure, sure, I mean the show's a little bit smaller this year than it had been in past years. Some of that is because they pulled some of the technical stuff out last year, or a couple years ago. By being a little bit smaller, and being in a place like Vancouver, I get good energy off of the crowd. The folks we've talked to, the folks that have been going to sessions, have said they've been very good. The people here are practitioners. They are running OpenStack, or about to run OpenStack, or upgrading their OpenStack, or other adjacent technologies. They're real people doing real work. As we talk to folks and sponsors, the conversations have been productive. So, I'd say in general, this kind of a small venue and a beautiful city allows for a really productive community-oriented event, so that's been great. >> Alright, so John come on, on the analysis segment we're not allow to pull any punches. Attendance, absolutely is down. Three years ago when we were here it was around 5500. Mark Collier, on our opening segment, said there was about 2600. But two-year point, I've not talked to a single vendor or attendee here that was like, "Oh boy, nobody's here, "it's not goin' on." Yes, the Expo Hall is way smaller and people flowing through the Expo Hall isn't great all the time, but why is that? Because the people that are here, they're in sessions. They have 40 sessions about Edge Computing. Hot topic, we've talked a bunch about that. Interesting conversations. There is way more in Containers. Containers for more than three years, been a topic conversation. There's so many other sessions of people digging in. The line you've used a couple a time is the people here are people that have mortgages. In a good way, it means these are jobs, these are not them, "Oh, I heard about "this cool new thing, and I'm going to "go check out beautiful Vancouver." Now, yes, we've brought our spouses or significant others, and checking out the environment because yeah, this place is awesome, but there's good energy at the show. There's good technical conversation. Many of the people we've talked to, even if they're not the biggest OpenStack fans, they're like, "But our customers are using this in a lot of different ways." Let's talk about OpenStack. Where is it, where isn't it? What's your take from what you've heard from the customers and the vendors? >> Sure, I definitely think the conversation is warranted. As we came in, from outside the community there was a lot of conversation, even backchannel, like why are you going to OpenStack Summit? What's going on there, is it still alive? Which is kind of a perception of maybe it's an indication of where the marketing is on this project, or where it is on the hype cycle. In terms of where it is and where it isn't, it's built into everything. At this point OpenStack, the infrastructure management, open infrastructure management solution, seems to be mature. Seems to be inside every Telco, every cable company, every transportation company, every bank. People who need private resources and have the smarts and power to do that have leveraged OpenStack now. That seems stable. What was interesting here is, that that doesn't speak to the health overall, and the history of, or the future of the project itself, the foundation, the Summit, I think those are separate questions. You know, the infrastructure and projects seem good. Also here, like we've talked about, this show is not just about OpenStack now. It's about Containers, it's broadening the scope of these people informally known as infrastructure operators, to the application level as well. >> Yeah, if you want to hear a little bit more, some two great interviews we did yesterday. Sean Michael Kerner, who's a journalist. Been here for almost every single one of the OpenStack shows. He's at eWeek, had some really good discussion. He said private cloud, it doesn't exist. Now, he said what does he mean by that? There are companies that are building large scalable cloud with OpenStack but it's like if some of the big China Telecom, big China cloud companies. Oracle and IBM have lots of OpenStack, in what they do, and yes there are, as you mentioned, the telcos are a big used case. We had some Canonical customers talking about Edge as in a used case for a different type of scalability. Lots of nodes but not one massive infrastructure as a service piece. If I talk, kind of the typical enterprise, or definitely going the SNE piece of the market, this is not something that they go and use. They will use services that have OpenStack. It might be part of the ecosystem that they're playing, but people saying, "Oh, I had my VMware environment "and I want to go from virtualization "to private cloud" OpenStack is not usually the first choice, even though Red Hat has some customers that kind of fit into some of the larger sides of that, and we'll be talking to them more about that today. Randy Bias is the other one, take a look. Randy was one of the early, very central to a lot of stuff happening in the Foundation. He's in the networking space now, and he says even though he's not a cheerleader for OpenStack, he's like, "Why am I here? "That's where my customers are." >> Right, right. I mean, I do think it's interesting that public cloud is certainly mentioned. AWS, Google, et cetera, but it's not top of mind for a lot of these folks, and it's mentioned in very different ways depending on, kind of, the players. I think it's very different from last week at Red Hat Summit. Red Hat, with their story, and OpenShift on top of OpenStack, definitely talked public cloud for folks. Then they cross-cloud, hybrid-cloud. I think that was a much different conversation than I've been hearing this week. I think basically, kind of maybe, depends on the approach of the different players in the market, Stu. I know you've been talkin' to different folks about that. >> Yeah, absolutely. So like, Margaret Dawson at Red Hat helped us talk about how that hybrid-cloud works because here, I hate to say it's, some oh yeah, public cloud, that's too expensive. You're renting, it's always going to be more. It's like, well no, come on, let's understand. There's lot of applications that are there and customers, it's an and message for almost all of them. How does that fit together, I have some critiques as to how this goes together. You brought up another point though John, OpenStack Foundation is more than just OpenStack projects. So, Kata Containers, something that was announced last year, and we're talking about there's Edge, there's a new CI/CD tool, Zuul, which is now fully under the project. Yes, joke of the week, there is no OpenStack, there's only Zuul. There are actually, there's another open-source project named Zuul too, so boy, how many CI/CD tools are out there? We've got two different, unrelated, projects with the same name. John, you look at communities, you look at foundations, if this isn't the core knitting of OpenStack, what is their role vis-a-vis the cloud native and how do they compare to say, the big player in this space is Linux Foundation which includes CNCF. >> That's a good one. I mean, in some sense like all organic things, things are either growing or shrinking. Just growing or dying. On the other hand, in technology, nothing ever truly dies. I think the project seems mature and healthy and it's being used. The Foundation is global in scope and continues to run this. I do wonder about community identity and what it means to be an OpenStack member. It's very community-oriented, but what's at the nut of it here if we're really part of this cloud-native ecosystem. CNCF, you know, it's part of Linux Foundation, all these different foundations, but CNCF, on the other hand, is kind of a grab-bag of technology, so I'm not sure what it means to be a member of CNCF either. I think both of these foundations will continue to go forward with slightly different identities. I think for the community as a whole, the industry as a whole, they are talking and they better be talking, and it's good that they're talking now and working better together. >> Yeah, great discussion we had with Lisa-Marie Namphy who is an OpenStack Ambassador. She holds the meat up in Silicon Valley and when she positions it, it's about cloud-native and its about all these things. So like, Kubernetes is front and center whereas some of the OpenStack people are saying, "Oh no, no, we need to talk more about OpenStack." That's still the dynamic here was, "Oh, we go great together." Well, sometimes thou dost protest too much. Kubranetes doesn't need OpenStack, OpenStack absolutely must be able to play in this Container, cloud-native Kubranetes world. There's lots of other places we can learn about Kubranetes. It is an interesting dynamic that have been sorting out, but it is not a zero-sum game. There's absolutely lots, then we have, I actually was real impressed how many customers we got to speak with on the air this time. Nice with three days of programming, we had a little bit of flexibility, and not just people that were on the keynote stage. Not just people that have been coming for years, but a few of the interviews we had are relatively new. Not somebody that have been on since very early in the alphabet, now we're at queens. >> Right. >> Anything more from the customers or that Container, Kubranetes dynamic that you want to cover? >> Sure, well I mean just that, you know, Containers at least, Containers are everywhere here. So, I think that kind of question has been resolved in some sense. It was a little more contentious last year than this year. I'm actually more bullish on OpenStack as a utility project, after this week, than before. I think I can constantly look people in the eye and say that. The interesting thing for me though, coming from Silicon Valley, is you're so used to thinking about VCs and growth, and new startups, and where's the cutting edge that it's kind of hard to talk about this, maybe this open source business model where the customer basis is finite. It's not growing at 100% a year. Sometimes the press has a hard time covering that. Analysts have a hard time covering that. And if you wanted to give advice to somebody to get into OpenStack, I'm not sure who should if they're not in it already, there's definitely defined use cases, but I think maybe those people have already self-identified. >> Alright, so yeah, the last thing I wanted to mention is yeah. Big thank you to our sponsors to help get us here. The OpenStack Foundation, really supportive of us for years. Six years of us covering it. Our headline sponsor, Red Hat, had some great customers. Talked about this piece, and kind of we talk about it's practically Red Hat month on theCUBE for John with Red Hat Summit and OpenStack. Canonical, Contron, Nuage Networks, all helping us to be able to bring this content to you. Be sure to check out theCUBE.net for all the coverage in the past as well as where we'll be. Hit John Troyer, J. Troyer, on Twitter or myself, Stu, on Twitter if you ever have any questions, people we should be talking to, viewpoints, whether you agree or disagree with what we're talking about. Big thanks to all of our crew here. Thank you to the wonderful people of Vancouver for being so welcoming of this event and of all of us. Check out all the interviews. For John Troyer, I'm Stu Miniman. Thanks for watching theCUBE. (soft upbeat music)

Published Date : May 23 2018

SUMMARY :

Brought to you by Red Hat, the OpenStack Foundation and let's get into the show itself, too. the folks that have been going to sessions, Many of the people we've talked to, and have the smarts and power to do that but it's like if some of the big China Telecom, in the market, Stu. Yes, joke of the week, but CNCF, on the other hand, but a few of the interviews we had are relatively new. in the eye and say that. for all the coverage in the past

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
CNCFORGANIZATION

0.99+

John TroyerPERSON

0.99+

Mark CollierPERSON

0.99+

IBMORGANIZATION

0.99+

Margaret DawsonPERSON

0.99+

Sean Michael KernerPERSON

0.99+

JohnPERSON

0.99+

RandyPERSON

0.99+

Stu MinimanPERSON

0.99+

TelcoORGANIZATION

0.99+

Red HatORGANIZATION

0.99+

OracleORGANIZATION

0.99+

Silicon ValleyLOCATION

0.99+

VancouverLOCATION

0.99+

AWSORGANIZATION

0.99+

Randy BiasPERSON

0.99+

J. TroyerPERSON

0.99+

40 sessionsQUANTITY

0.99+

CanonicalORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

last yearDATE

0.99+

Linux FoundationORGANIZATION

0.99+

Lisa-Marie NamphyPERSON

0.99+

bothQUANTITY

0.99+

last weekDATE

0.99+

StuPERSON

0.99+

Six yearsQUANTITY

0.99+

third dayQUANTITY

0.99+

OpenStack FoundationORGANIZATION

0.99+

first timeQUANTITY

0.99+

more than three yearsQUANTITY

0.99+

twoQUANTITY

0.99+

yesterdayDATE

0.99+

Three years agoDATE

0.99+

Vancouver, CanadaLOCATION

0.99+

three daysQUANTITY

0.98+

OpenStackORGANIZATION

0.98+

first choiceQUANTITY

0.98+

one more dayQUANTITY

0.98+

Nuage NetworksORGANIZATION

0.98+

OpenStackTITLE

0.98+

ContronORGANIZATION

0.98+

this weekDATE

0.98+

China TelecomORGANIZATION

0.98+

this yearDATE

0.98+

two-yearQUANTITY

0.97+

theCUBE.netOTHER

0.97+

OpenStack SummitEVENT

0.97+

todayDATE

0.97+

OpenStack Summit 2018EVENT

0.97+

oneQUANTITY

0.95+

FirstQUANTITY

0.95+

about 2600QUANTITY

0.94+

100% a yearQUANTITY

0.93+

two great interviewsQUANTITY

0.93+

single vendorQUANTITY

0.92+

SiliconANGLE MediasORGANIZATION

0.91+

Red Hat SummitEVENT

0.9+

eWeekORGANIZATION

0.89+

Anne Bertucio, OpenStack Foundation | OpenStack Summit 2018


 

>> Announcer: Live from Vancouver, Canada it's theCUBE covering OpenStack Summit North America 2018. Brought to you by Red Hat, the OpenStack Foundation and its ecosystem partners. >> Welcome back to theCUBE here at OpenStack Summit 2018 in Vancouver. I'm Stu Miniman with co-host this week is John Troyer. I'm happy to welcome to the program, first time guest. It's Anne Bertucio, who is the Kata Containers Community Manager with the OpenStack Foundation. Thanks so much for joining us. >> Thank you for having me. >> All right, it's our pleasure and the containers has been a discussion we've been having for a few years now. I remember when we were last year in Vancouver, three years ago that the joke was it was Docker, Docker, Docker year. Tell us a little bit first your role, how long you've been with the foundation, and what you're covering there. >> Absolutely, I've been with the foundation for going on three years at this point. The Kata Containers Project we announced in December. It's come up and come in there as a community manager helping them figure out since December to the launch now, in less than six months we had to figure out how are we going to work together. How are we going to merge two code bases and we have to create a new open source project and new community. So leading that has been a big part of my work. >> So there's a whole track on Containers now. Give us a little bit of flavor for our audience that couldn't be sitting in the keynote and attend all the sessions. What were they missing? >> I think the major theme was security. Mia, she's the PM of security at Google. She opened it up saying containers don't contain. And I almost wished we'd been on a game show. Like containers don't contain. That was the theme of the day and we talked about where did Kata come from? Kata came from how do we answer that question. I think people got so excited about performance and portability about containers. We forgot about security a little bit and now we're seeing some of the ramifications and it's time to make this the year of security. >> So you talk about bringing two code basis together. Can you talk a little bit about what some of the ingredients are here to get to our dish that we finally call Kata Containers Projects? >> Yeah, absolutely, so we have ren-V from Hyper and we had Clear Containers from Intel. And they both looked at things a little differently like Hyper has a fracty implementation that was really critical to their customers. Clear Containers are becoming a little bit from runC Vert containers. And what we arrive at for 1.0 is the OCI compatible runtime is going to put a lightweight VM around your container, and we're thrilled to look beyond 1.0 and to things like supporting hardware accelerators. >> So it may be just to raise it up one level before we go on. How do containers in some sense, let's repeat maybe what you said, see if I get it right. >> Anne: Yeah. >> It's wrapping a container and a lightweight VM. And that gives us the isolation and security that's traditionally associated with a virtual machine with all the APIs and flexibility and performance, and all the other goodness of a container. One container in one VM is the first implementation. >> Yeah, I think the easy way to think about, you're talking about Docker Docker Docker. So in Kata, really instead of using runC as your runtime, we would just say Kata runtime, and now we have our Docker containers but they're wrapped in these light weight VMs each with their own kernel. >> I think back to the early days when we were trying to figure out what these whole containers were and was that the death of virtualization? It was like VMs, gosh they take minutes to spin up, and container is super fast. Security, oh VMs yeah, there's security there but we need to move fast, fast, fast. So explain how this helps bring together the peanut butter and chocolate, if we will? >> Absolutely, oh I love peanut butter and chocolate but that's really what it is. Like you were saying virtualization, yes. Super secure, slow. I think I have a clip art chart with a sad turtle on it. A little bit slower. The container is super fast, we're getting a little nervous about security. I think we maybe see groups and name spaces are good, but people who are enterprise environments. They've been putting full blown VMs around their containers 'cause they were saying well it's not enough. And I need two isolation boundaries, not just one. >> Right, in terms of some of the use cases then. I imagine multitenancy would be one and then perhaps even, I think some of the newest trend defense in depth of even an individual app putting different zones in different components or different risk zones in their own containers, their own VMs. Even inside an individual app just making sure that the different components can only talk to each other in ways that they're suppose to. >> Absolutely, I think it's anytime where you're running untrusted code, or you have questions about what's going on there or you just want a heightened security. Kata is an easy used case then. >> Sure, I guess my VMware call it microsegmentation would be their buzz word on it. >> Oh I got to think about what mine is going to be. >> Or we can all use the same words, it's good. >> So Anne, Intel Clear Containers was a piece of this. Of course Intel partners with everyone there. Give us a little bit also the ecosystem and the team that makes this up. Is this, people out there will be like, oh, well but Docker has their solution and VMware has their solution. How does this fit into the broader ecosystem? >> Our team is incredibly diverse. I've just been thrilled with 1.0. We had 40 contributors from a good diversity of companies. Our architecture committee, it's Google, it's Huawei, Hyper, Intel and Microsoft and I think we've, I was saying in the other note the other day. I was on a call for a architecture committee and we had AMD, ARM and Intel all talking about the same solution. So it's the beauty of open source that we've brought all of these groups together. >> One of the things that also struck us especially if we've been here. The diversity of the show is always really good. The main keynote, it's not oh, did they brought up some people of diversities. Oh no, these are the project leads and therefore they're doing this. Can you touch on some of the diversity and activities at the show itself? >> In terms of technologies, we're looking at or? >> No, I just, so there is, I'm just saying you talked about the community, the diversity of companies as well, the diversity of people. So we've got lots of the women inclusion. >> Oh sure. >> Things like that. >> Yeah, I know we had the executive producer of Chasing Grace was here and I know she's been, Jennifer Clower, is that correct? >> Stu: Yes, Jennifer Clower. We actually interviewed her last week at a different show. >> Oh fantastic. Yeah her document has been incredibly well received. I know she's making the rounds to get the word out there about what's going on with Women in Tech. And we were more than thrilled to host her and have her here and be apart of conversation. >> Clear Community is a big part of OpenStack, the OpenStack Summit and care of the OpenStack Foundation. In terms of Kata Containers, you work for the OpenStack Foundation. Is Kata officially then part of the OpenStack or does that have a different governance model? >> That's a great question. This is an area of confusion because it's the first time the foundation is broken out and there's the OpenStack Project, and there's Kata Containers the Project, but we both live at the OpenStack Foundation. >> John: Okay. >> I think the guiding principles though, and it's really helped us over the last four months is that the OSF, OpenStack Foundation, we believe in open source, open design, open development and open community. And Kata, we were like that's a great home. We believe in that as well. >> Any customers that are yet talking about their early usage of Kata that you can share? >> I think we have a lot of customers from runV and Clear Containers and Kata is going to be their next path forward. So with 1.0 out yesterday, I'm excited to see. We should see some upgrades real soon here. >> What's the path for them to get from where they are to the 1.0? Is that pretty straightforward? >> It should be, yeah, we think so. And they have their support from Intel and from Hyper to help them out with that as well. >> Stu: Okay. >> I was going to ask is Kata Containers, is it integrated in an API or is OpenStack necessary for it or is it independent of, from an infrastructure perspective, OpenStack, the stack? >> Yeah, it's completely independent, but it's also compatible. >> John: Okay. >> You can run on Azure, Google, OpenStack, agnostic of the infrastructure underneath it. >> John: Great. >> Anne, want to give you a final word. Takeaways from the show that you'd want people to have. >> Absolutely, I think the final word is containers are fantastic, it's probably time to take a look at your container architecture. Think about it from a security perspective, and I would encourage everyone to go check out Kata Containers and see if that's the solution for them. >> Anne Bertucio, really appreciate you joining and sharing with us everything happening. It can work with or without the OpenStack Containers. Absolutely a big trend, but security absolutely top of mind from everyone we've talked to. If it's not top of mind of a company, I'm always a little bit worried about them. For John Troyer, I'm Stu Miniman. We'll be back with lots more coverage here from OpenStack Summit 2018 in Vancouver. Thanks for watching theCUBE. (uptempo techno music)

Published Date : May 23 2018

SUMMARY :

and its ecosystem partners. I'm happy to welcome to the program, first time guest. and the containers has been a discussion and we have to create a new open source project and attend all the sessions. and it's time to make this the year of security. to get to our dish that we finally and we had Clear Containers from Intel. So it may be just to raise it up one level and all the other goodness of a container. and now we have our Docker containers the peanut butter and chocolate, if we will? I think we maybe see groups and name spaces are good, that the different components can only talk to each other Absolutely, I think it's anytime would be their buzz word on it. and the team that makes this up. and we had AMD, ARM and Intel all talking and activities at the show itself? the diversity of companies as well, We actually interviewed her last week at a different show. I know she's making the rounds to get the word out there the OpenStack Summit and care of the OpenStack Foundation. This is an area of confusion because it's the first time and it's really helped us over the last four months and Clear Containers and Kata is going to be What's the path for them to get and from Hyper to help them out with that as well. but it's also compatible. agnostic of the infrastructure underneath it. Takeaways from the show that you'd want people to have. Kata Containers and see if that's the solution for them. and sharing with us everything happening.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jennifer ClowerPERSON

0.99+

HuaweiORGANIZATION

0.99+

JohnPERSON

0.99+

John TroyerPERSON

0.99+

GoogleORGANIZATION

0.99+

Anne BertucioPERSON

0.99+

OpenStack FoundationORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

Stu MinimanPERSON

0.99+

DecemberDATE

0.99+

OSFORGANIZATION

0.99+

VancouverLOCATION

0.99+

AnnePERSON

0.99+

AMDORGANIZATION

0.99+

Red HatORGANIZATION

0.99+

ARMORGANIZATION

0.99+

40 contributorsQUANTITY

0.99+

HyperORGANIZATION

0.99+

three yearsQUANTITY

0.99+

yesterdayDATE

0.99+

MiaPERSON

0.99+

last weekDATE

0.99+

IntelORGANIZATION

0.99+

Vancouver, CanadaLOCATION

0.99+

three years agoDATE

0.99+

less than six monthsQUANTITY

0.99+

last yearDATE

0.99+

OpenStackTITLE

0.99+

StuPERSON

0.98+

two isolation boundariesQUANTITY

0.98+

KataTITLE

0.98+

one levelQUANTITY

0.98+

OpenStackORGANIZATION

0.98+

OpenStack Summit 2018EVENT

0.98+

two code basesQUANTITY

0.98+

first implementationQUANTITY

0.98+

One containerQUANTITY

0.98+

first timeQUANTITY

0.97+

Kata ContainersTITLE

0.97+

OneQUANTITY

0.97+

OpenStackEVENT

0.97+

oneQUANTITY

0.96+

bothQUANTITY

0.96+

OpenStack Summit North America 2018EVENT

0.95+

Docker Docker DockerTITLE

0.94+

this weekDATE

0.92+

Kata ContainersORGANIZATION

0.91+

AzureTITLE

0.88+

two codeQUANTITY

0.88+

eachQUANTITY

0.87+

DockerORGANIZATION

0.86+

theCUBEORGANIZATION

0.85+

KataPERSON

0.84+

one VMQUANTITY

0.84+

last four monthsDATE

0.8+

1.0OTHER

0.8+

ClearORGANIZATION

0.79+

Sean Michael Kerner, eWeek | OpenStack Summit 2018


 

>> Announcer: Live from Vancouver, Canada. It's theCUBE covering OpenStack Summit North America 2018, brought to you by Red Hat, the OpenStack Foundation and its ecosystem partners. >> Welcome back, I'm Stu Miniman and my cohost John Troyer and you're watching theCUBE, the worldwide leader in tech coverage and this is exclusive coverage from OpenStack Summit 2018 in Vancouver. Usually this time of year it is a little bit overcast, but for the second time the OpenStack Summit has been here, the sun is shining. It has been gorgeous weather but we are in here really digging in and understanding it One of the people I have gotten to know through this community especially, is our wrap up guest today, Sean Michael Kerner, who is a senior editor with eWeek, amongst other bi-lines that you have. Pleasure to see you. >> Great, good seeing you too Stu. >> Alright, so we let you keep on the Toronto Bluejays hat >> Thank you, there we go. >> We have had quite a few Canadians on our program here. >> Well, seeing as how you're here in Canada, it's not all that surprising. >> It's lovely. They have you working on Victoria Day. >> Yeah, that's unfortunate but I will take Memorial Day off in a week, so it works out. >> Excellent. So Sean, for our audience that might not know you, give us a little bit about your background. You've been to umpteen of these shows. >> Sure. I have been with the same publication roughly, I guess 15-16 years at this point. I've been writing before there was cloud, core living and Opensource stuff, networking. And then through the magic of technology, I shifted a little bit to security, which is a core focus for me. I have been to every OpenStack Summit since the San Diego Summit, I guess, 2011. Somebody can correct me afterwards. I did miss the Sydney Summit for various reasons, but yeah, I've been to a bunch of these things, so interesting to see how things have shifted over the years from nothing to certain heights to where we are now. >> Alright, so bring us up to that, as to where we are now. Attendance is down a little bit. They haven't been talking a lot about it but quality I guess is here. Sessions, they've broadened down a bit of the scope. We have been digging into it, but want to get your take so far. >> Yeah, well it's like anything else, there are standard hype cycles, as it were and there's a trough of disillusionment. I wouldn't call this a trough of disillusionment, but when you get to a certain plateau, people just, there'sn't as much interest. In the early days, I remember the San Diego Summit I went to. They didn't schedule it properly. They didn't know how many people they were going to have, and they had to line up around the corner and stuff. That was six years ago, but that is when OpenStack was new. There was no such thing as the Foundation, and everyone was trying to figure out what was what. And, there was no clue at this point. Cloud is a well understood thing. There are competitive efforts or complimentary efforts, as the Foundation would probably like to put it; whether it's CNCF, there's the public cloud and it's different. There is, with all respect to the OpenStack Foundation and its member projects, there's not as much excitement. This in now a stable, mature ecosystem and because of that, I don't think there's as much of a draw. When something is brand new and shiny, you get more of a draw. If they would have put the name Blockchain somewhere, maybe, maybe they would have had a few more. They put Kubernetes in there, which is fine, but no machine learning or artificial intelligence quite yet, though that's a topic somewhere in there too. >> Yeah, John, you've been making a lot of comments this week talking about we've matured and the lower layer pieces just work a bit more. Give us your take about that. >> Sure. That's the way it seems. There wasn't a whole lot of talk about the release, news release, and all the different components, even the keynotes. But, the people we have talked to, both on the vendor and the customer side, they have working production OpenStack environments. They're very large. They require very few admins. They work. They're embedded in telecom and banking, et cetera. It's here and it's working. >> Yeah, that's so something that happened, maybe three cycles ago at this point, because they used to have the release the same time as the Summit and the Design Summit. It was together, so, there was essentially a celebration of the release. People would talk about the release and then they desegrigated that. I think that took a lot of steam out of the reason why you got developers to attend. So, when you don't have the Design Summit, there's this separate open endeavor, there's the forum, I don't quite understand how that works here now. There isn't as much momentum. Yeah, I agree with you. There has been very little talk about Queens. In each of the project update sessions I have been to, and I have been to a couple, there has always been a slight on Rocky, what's coming. I think we are on the second milestone of Rocky, at this point, so there's some development, but at this point it is incremental featurettes. There is no whiz bang. OK, we're going to have flying cars, you know send a Tesla to outer space kind of Earth shattering kind of news, literally, because that's not where it's at. It's just incremental tuck in features in stability and that kind of thing. >> Alright, you talk space and thinks like that and it brings to mind a certain attendee of the program that has actually been to outer space and maybe one of the more notable moments of the show so far. Give us your take on Mr. Shuttleworth. >> Well, I'm a big fan of Mr. Shuttleworth, top to bottom. Hey Mark. Big fan, always have been. He has his own opinion on things of course. Usually in a keynote you don't tend to take direct aim at competitors and he chose to do that. It made some people a little uncomfortable. I happened to be sitting in the front row, where I like to sit, and there was some Red Hat people, and there were some frantic emails going back and forth. And people were trying to see what was going on et cetera. I think, for me, a little bit of drama is okay. You guys go to more shows than I do, and sometimes you get these kind of sales kind of things. But in an open community, there's almost an unwritten rule, which perhaps will be written after this conference, that whether or not everybody is a business competitor or not, is that this is neutral territory as it were and everybody is kind of friendly. In the exhibit hall, you can say this and that, we are better, whatever, but on the stage you don't necessarily do that, so there was some drama there. Some of my peers wrote about that and I will be writing about it as well. It's a, I prefer to write about technology and not necessarily drama. Whether somebody is faster, better, stronger than others, you let the number prove them out. When we talk about Opensource, Opensource Innovation without Canonical, there probably wouldn't have been an OpenStack. All the initial OpenStack reference and limitations are on Canonical. They got a number of large public clouds, as does Red Hat. I think they both have their tactical merits and I'm sure on some respects Red Hat's better and on some respects Canonical is better, but him standing up there and beating on the competition was something that across the 13 summits I have been I have never seen before. One guy I talked to my first OpenStack Summit was in San Diego and the CTO of VMware at the time came up to, VMware was not an OpenStack contributor at the time, they were thinking about it, and he was fielding questions about how it was competitive or not and he was still complimentary. So there has always been that kind of thing. So, a little bit of an interesting shift, a little bit of drama, and gives this show something memorable, because you and I and others will be able to talk about this five years from now, et cetera. >> You talked about something you would write up. I mean part of your job is to take things back to the readers at eWeek. >> Yeah. >> What are the things, highlights you're going to be covering? >> The highlights for me, Stu and I talked about this at one point off the camera, this is not an OpenStack Summit necessarily, they're calling it Open Infrastructure. I almost thought that they would change, we almost thought that they would change the name of the entire organization to the Open Infrastructure Foundation. That whole shift, and I know the foundation has been talking about that since Sydney last year, that they're going to shift to that, but, that's the take away. The platform itself is not the only thing. Enabling the open infrastructure is nice. They're going to try and play well and where it fits within the whole stack. That gets very confusing because talking about collaboration is all fine and nice, but that is not necessarily news. That is how the hot dog is made and that's nice. But, people want to know what's in that dog and how it is going to work. I think it's a tougher show for me to cover than it has been in past years, because there has been less news. There's no new release. There was Kata 1.0 release and there was the Zuul project coming out on its own. Zuul project, they said it was 3.0, it was actually March was Zuul 3.03. Kata Container project, okay, interesting, we'll see how it goes. But a tougher project, tougher event for me to cover for that reason. Collaboration is all fine and nice. But, the CNCF CloudNativeCon KubeCon event two weeks ago, or three weeks ago, had a little bit more news and a lot it's same kind of issues come up here. So, long winded answer, tough to come up with lessons learned out of this, other than everyone wants to be friends, well some people want to be. And, collaboration is the way forward. But that is not necessarily a new message. >> When I think about Kubernetes, we are talking about the multi cloud world and that's still, the last few years, where it's been. Where does OpenStack really fit in that multi cloud world? One of the things I have been a little disappointed actually, is most of the time, when I'm having a conversation, it's almost the, yeah, there's public cloud, but we are going to claw things back and I need it for governments, and I need all of these other things. When I talk to customers, it is I'm going to choose what I put in my data center. I'm going to choose how I use probably multiple public cloud finders. It is not an anti-public cloud message, and it feels a little bit on the anti-public cloud mass. I want to work with what you're hearing when you >> talk to users? >> When I talk to users, vast majority of people, unless it's something, where there's regulatory issues or certain legacy issues or private cloud, public cloud period. The private cloud idea is gone or mostly gone. When I think about private clouds, it's really VmWare. We have virtualized instances that sitting there. >> What's OpenStack? >> OpenStack is fine, but how many are running OpenStack as a private cloud premise? >> Yeah, so what's OpenStack then? >> When I think of OpenStack, Oracles public cloud. Oracle is not here surprisingly. Oracle's public cloud, Larry Ellison, who I know you guys have spoken to more than once on theCUBE at various points on Oracle World and other things. Oracle's public cloud, they want to compete against AWS. That's all. OpenStack IBM cloud, all OpenStack. The various big providers out of China are OpenStack based. OEH is here. So that's where it fits in is that underlying infrastructure layer. Walmart uses it. Bestbuy, all these other places, Comcast, et cetera; ATT. But individual enterprises, not so much. I have a hard time finding individual enterprises that will tell me we are running our own private cloud as OpenStack. They will tell me they're running VmWare, they will tell me they're running REV or even some flavor of Citrix end server, but not a private cloud. They may have some kind of instances and they will burst out, but it's not, I don't think private cloud for mid tier enterprises ever took off the way some people thought five years ago. >> That's interesting. Let's go meta for a second. You talked about things you do and don't write about, you don't necessarily write the VC's are not here necessarily, but you don't write about necessarily financial stuff. >> Sometimes. There was actually at the Portland summit, I did a panel with press and analysts at the time and afterwards there might have been four different VC's that came up to me and asked me what I thought about different companies. They were looking at different things where they would invest. And I remember, we looked at the board and one VC who shall remain nameless, and I said you know what, we'll look at this board with all these companies and five years from now, three quarters of them will not be here. I think I was probably wrong because it is more than that. There are so many. I wrote a story, I don't remember the exact name of it, but I wrote a story not that long ago about OpenStack deadpool. There are so, multiple companies that raised funding that disappeared. In the networking space, there were things like Plumgrid, they mminorly acquired for assets by Vmware, if I'm not mistaken. There was Pivotal, Joshua McKenzie, one of the co-founders of OpenStack itself, got acquired by Cisco. But they would have collapsed perhaps otherwise. Nebula Computing is perhaps, it still shocks me. They raised whatever it was 50 odd million, someone will correct me afterward. Chris Kemp, CTO of NASA who helped start it. Gone. So, there has been tremendous consolidation. I think when VC's lose money, they lose interest really fast. The other thing you have to think about, from the VC side, they don't write too much on the financial. My good friend Fredrick, who didn't make it, Where are you, Fredrick, where are you? Does more on that funding side. But has there been a big exit for an OpenStack company? Not really, not really. And without that kind of thing, without that precedence it's a tough thing, especially for a market that is now eight years old, give or take. >> Even the exits that had a decent exit, you know that got bought into the say IBM's, Cisco's of the world, and when you look a couple of years later, there's not much left of those organizations. >> Yeah. It's also really hard. People really don't want to compete against, well, some people want to compete against AWS. But, if you're going to try to go toe to toe with them, it's a challenge. >> Okay, so what brings you back here every year? You're speaking at the show. You're talking to people. >> What brings me back here is regardless of the fact that momentum has probably shifted, it's not in that really hype stage, OpenStack's core infrastructure, literally, core infrastructure that runs important assets. Internet assets, whether it certain public cloud vendors, large Fortune 500 companies, or otherwise. So it's an important piece of the stack, whether it's in the hype cycle or not, so that brings me back, because it's important. It brings me back because I have a vested interest. I have written so much about it so I'm curious to see how it continues to evolve. Specifically, I'm speaking here on Thursday doing a panel on defending Cloud Counsel Security as a core competence, a core interest for me. With all these OpenStack assets out there, how they're defended or not is a critical interest. In the modern world, cyber attacks are a given. Everybody should assume they're always under a constant state of attack and how that security works is a core area of interest and why I will keep coming back. I will also keep coming back because I expect there to be another shift. I don't think we have heard the end of the OpenStack story yet. I think the shift towards open infrastructure will evolve a little bit and will come to an interesting conclusion. >> Alright, last thing is what's your favorite question you're asking at this show. Any final things you want to ask us as we wrap? >> Yeah, my favorite, well, I want to ask you guys, what the most interesting answer you got from all the great people you interviewed because I'm sure some of it was negative and you got mostly positive as well. >> Well, we aren't used to answering the questions Stu. >> I'm used to being on the other side here, right. >> Well, I do say we got a lot of stuff about some interesting and juicy cases, like I say, the practitioners I talked to were real. I was always impressed by how few administrators it takes to run a huge OpenStack based cloud once it's set up. I think that's something interesting to me. You asked some folks about a public cloud a lot. >> Yeah, so it has been interesting. For me, it's, we've reached that certain maturity level. I was looking at technology. What's kind of the watermark that this is going to come to? We had said years ago, I don't think you're going to have somebody selling a billion dollars worth of distribution on OpenStack. So, that story with how Kubernetes and Containers and everything fits in, OpenStack is part of the picture, and it might not be the most exciting thing, but then again, if you watch Linux as long as most of us have, Red Hat took a really long time to get a billion dollars and it was much more than just Linux that got them there. This still has the opportunity to be tooling inside the environment. We have talked to a number of users that use it. It's in there. It's not that the flagpole, we're an OpenStack company anymore because there really aren't many companies saying that that is the core of their mission, but that is still an important piece of the overall fabric of what we are covering. >> Exactly right. >> Alright, we on that note, Sean Michael Kerner, we really appreciate you joining us. Please support good technology journalism because it is people like him that help us understand the technology. I read his stuff all the time and always love chatting with him off the record and dragged him on here and Fredrick from Techron Show we are disappointed you could not join us, but we'll get you next time. For Jon Troyer, I'm Stu Miniman, be sure to join us for the third day tomorrow of three days of wall to wall live coverage here from OpenStack Summit 2018 in Vancouver. And once again, thank you for watching theCUBE. (upbeat music)

Published Date : May 23 2018

SUMMARY :

brought to you by Red Hat, the OpenStack Foundation One of the people I have gotten to know through this it's not all that surprising. They have you working on Victoria Day. Yeah, that's unfortunate but I will take Memorial Day off You've been to umpteen of these shows. I have been to every OpenStack Summit since We have been digging into it, but want to get and they had to line up around the corner and stuff. Give us your take about that. But, the people we have talked to, both on the vendor and a celebration of the release. more notable moments of the show so far. In the exhibit hall, you can say this and that, the readers at eWeek. That is how the hot dog is made and that's nice. actually, is most of the time, when I'm having When I talk to users, have spoken to more than once on theCUBE at various You talked about things you do and don't write about, In the networking space, there were things like Even the exits that had a decent exit, you know some people want to compete against AWS. You're speaking at the show. of the OpenStack story yet. Any final things you want to ask us as we wrap? the great people you interviewed because I'm I talked to were real. This still has the opportunity to be I read his stuff all the time and always love chatting

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
FredrickPERSON

0.99+

Sean Michael KernerPERSON

0.99+

Jon TroyerPERSON

0.99+

CanadaLOCATION

0.99+

JohnPERSON

0.99+

Chris KempPERSON

0.99+

MarkPERSON

0.99+

IBMORGANIZATION

0.99+

OracleORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

Stu MinimanPERSON

0.99+

ComcastORGANIZATION

0.99+

CanonicalORGANIZATION

0.99+

WalmartORGANIZATION

0.99+

NASAORGANIZATION

0.99+

ThursdayDATE

0.99+

Larry EllisonPERSON

0.99+

John TroyerPERSON

0.99+

San DiegoLOCATION

0.99+

Red HatORGANIZATION

0.99+

ChinaLOCATION

0.99+

Open Infrastructure FoundationORGANIZATION

0.99+

OpenStack FoundationORGANIZATION

0.99+

AWSORGANIZATION

0.99+

OpenStackORGANIZATION

0.99+

SeanPERSON

0.99+

Joshua McKenziePERSON

0.99+

StuPERSON

0.99+

VancouverLOCATION

0.99+

VmwareORGANIZATION

0.99+

Vancouver, CanadaLOCATION

0.99+

EarthLOCATION

0.99+

three daysQUANTITY

0.99+

OraclesORGANIZATION

0.99+

ATTORGANIZATION

0.99+

BestbuyORGANIZATION

0.99+

six years agoDATE

0.99+

last yearDATE

0.99+

OpenStack Summit 2018EVENT

0.99+

two weeks agoDATE

0.99+

three weeks agoDATE

0.99+

MarchDATE

0.99+

LinuxTITLE

0.99+

VMwareORGANIZATION

0.99+

five years agoDATE

0.99+

OpenStackTITLE

0.99+

firstQUANTITY

0.99+

SydneyLOCATION

0.98+

second milestoneQUANTITY

0.98+

OpenStack SummitEVENT

0.98+

Sydney SummitEVENT

0.98+

bothQUANTITY

0.98+

OpenStack Summit 2018EVENT

0.98+

San Diego SummitEVENT

0.98+

second timeQUANTITY

0.98+

Kata 1.0TITLE

0.98+

CNCFORGANIZATION

0.98+

ShuttleworthPERSON

0.98+

eachQUANTITY

0.98+

PlumgridORGANIZATION

0.98+

eight yearsQUANTITY

0.98+

oneQUANTITY

0.98+

13 summitsQUANTITY

0.97+

Victoria DayEVENT

0.97+

this weekDATE

0.96+

OneQUANTITY

0.95+

KubernetesTITLE

0.95+

Chris Hoge, OpenStack Foundation | OpenStack Summit 2018


 

>> Narrator: Live from Vancouver, Canada it's theCUBE covering OpenStack Summit North America 2018. Brought to you by Red Hat, the OpenStack Foundation, and its ecosystem partners. >> Welcome back to theCUBE, I'm Stu Miniman, with my cohost John Troyer, and happy to welcome to the program, fresh off the container keynote, Chris Hodge, who's the senior strategic program manager with the OpenStack Foundation. Thanks so much for joining us. >> Oh yeah, thanks so much for having me. >> Alright, so short trip for you, then John's coming from the Bay Area, I'm coming from the east coast. You're coming up from Portland, which is where it was one of the attendees at the Portland OpenStack Summit, they said, "OpenStack has arrived, theCUBE's there." So, shout out to John Furrier and the team who were there early. I've been to all the North America ones since. You've been coming here for quite a while and it's now your job. >> I've been to every OpenStack Summit since then. And to the San Francisco Summit prior to that, so it was, yeah, I've been a regular. >> Okay so for those people that might not know, what's a Foundation member do these days? Other than, you know, you're working on some of the tech, you're giving keynotes, you know, what's a day in the life? >> Yeah, I mean, I mean for me, I feel like I'm really lucky because the OpenStack Foundation, you know, has you know, kind of given me a lot of freedom to go interact with other communities and that's been one of my primary tasks, to go out and work with adjacent communities and really work with them to build integrations between OpenStack and right now, particularly, Kubernetes and the other applications that are being hosted by the CNCF. >> Yeah, so I remember, and I've mentioned it a few times this week, three years ago we were sitting in the other side of the convention center, with theCUBE and it was Docker, Docker, Docker. The container sessions were overflowing and then a year later it was, you know, oh my gosh, Kubernetes. >> Chris: Yeah. (chuckles) >> This wave of, does one overtake the other, how do they fit together, and you know, in the keynotes yesterday and I'm sure your keynote today, talked a lot a bit about you know, the various ways that things fit together, because with open source communities in general and tech overall, it's never binary, it's always, it depends, and there's five different ways you could put things together depending on your needs. So, what are you seeing? >> I mean it's almost, yeah, I mean saying that it's one or the other and that one has to win and the other has to lose is actually kind of, it's kind of silly, because when we talk about Kubernetes and we talk about Docker, we're generally talking about applications. And, you know, and, with Kubernetes, when you're very focused on the applications you want to have existing infrastructure in place. I mean, this is what it's all about. People talk about, "I'm going to run my Kubernetes application "on the cloud, and the cloud has infrastructure." Well, OpenStack is infrastructure. And in fact, it is open source, it's an open source cloud. And so, so for me it feels like it's a very natural match, because you have your open application delivery system and then it integrates incredibly well with an open source cloud and so whether you're looking for a public cloud running on OpenStack or you're hosting a private cloud, you know, to me it's a very natural pairing to say that you have an OpenStack cloud, you have a bunch of integrations into Kubernetes and that the two work together. >> I think this year that that became a lot clearer, both in the keynotes and some of the sessions. The general conversation we've had with folks about the role of Kubernetes or an orchestration or the cloud layer, the application layer, the application deployment layer say, and the infrastructure somebody's got to manage the compute the network storage down here. At least, in this architectural diagram with my hands but, you can also, a couple of demos here showed deploying Kubernetes on bare metal alongside OpenStack, with that as the provider. Can you talk a little bit about that architectural pattern? It makes sense, I think, but then, you know, it's a apparent contradiction, wait a minute so now the Kubernetes is on the bare metal? So talk about that a little bit. >> So, I think, I think one of the ways you can think about resolving the contradiction is OpenStack is a bunch of applications. When you go and you install OpenStack we have all of these microsurfaces that are, some are user facing and some are controlling the architecture underneath. But they're applications and Kubernetes is well-suited for application delivery. So, say that you're starting with bare metal. You're starting with a bare metal cloud. Maybe managed by OpenStack, so you have OpenStack there at the bottom with Ironic, and you're managing your bare metal. You could easily install Kubernetes on that and that would be at your infrastructure layer, so this isn't Kubernetes that you're giving to your users, it's not Kubernetes that you're, you know, making world facing, this is internally for your organization for managing your infrastructure. But, you want OpenStack to provide that cloud infrastructure to all of your users. And since OpenStack is a big application with a lot of moving parts, Kubernetes actually becomes a very powerful tool, or any other container orchestration scheme becomes a very powerful tool for saying that you drop OpenStack on top of that and then all of a sudden you have a public cloud that's available for, you know, for the users within your organization, or you could be running a public cloud and providing those services for other people. And then suddenly that becomes a great platform for hosting Kubernetes applications on, and so the layers kind of interleave with one another. But even if you're not interested in that. Let's say you're running Kubernetes as bare metal and you're just, you want to have Kubernetes here providing some things. There's still things that OpenStack provides that you may already have existing in your infrastructure. >> Kubernetes kind of wants, it wants to access some storage. >> It wants to consume storage for example, and so we have OpenStack Cinder, which right now it supports you know, somewhere between, you know over 70 storage drivers, like these drivers exist and the nice thing about it is... You have one API to access this and we have two drivers within that, two Cinder drivers, you can either choose the, the flex volume storage or the container storage interface, the CSI storage interface. And Cinder just provides that for you. And that means if you have mixed storage within your data center, you put it all behind a Cinder API and you have one interface to your Kubernetes. >> So Chris, I believe that's one of the pieces of I believe it's called the Cloud Provider OpenStack. You talked about in the keynote. Maybe walk us through with that. >> Cloud Provider OpenStack is a project that is hosted within the, within the Kubernetes community. And it's... The owner of that code is the SIG OpenStack community inside of Kubernetes. I'm one of the three leads, one of the three SIG leads of that group and, that code does a number of things. The first is there's a cloud manager interface that is a consistent interface for Kubernetes to access infrastructure information in clouds. So information about a node, when a node joins a system, Kubernetes will know about it. Ways to attach storage, ways to provision load balancers. The cloud manager interface allows Kubernetes to do this on any cloud, whether it be Azure or GCE or Amazon. Also OpenStack. Cloud Provider OpenStack is the specific code that allows us to do that, and in fact we were, OpenStack was one of the first providers that existed in upstream Kubernetes you know, so it's kind of, we've been there since the very beginning, like this has been a, you know, an effort that's happened from the beginning. >> Somewhat non-ironically, right? A lot of that you've talked about, the OpenStack Foundation and this OpenStack Summit, a lot of the things talked about here are not OpenStack per se, the components, they are containers, there's the OpenDev Conference here, colocated. Is there confusion, there doesn't, I'm getting it straight in my head, Is there, was there, did you sense any confusion of folks here or is that, if you're in it you understand what's going on and why all these different threads are flowing together in kind of an open infrastructure conversation. It seems like the community gets it and understand it and is broadened because of it. >> Yeah, I mean, to me I've seen a tremendous shift over the last year in the general understanding of the community of the role all of these different applications play. And I think it's really, it's actually a testament to the success of all of these projects, in particular, we're building open APIs, we're building predictable behavior, and once you have that, and you have many people, many different organizations that are able to provide that, they're all able to communicate with one another and leverage the strengths of the other projects. >> All of a sudden, a standard interface, low and behold, right? A thousand flowers bloom on top. >> You know, it essentially allows you to build new things on top of that, new more interesting things. >> Alright, Chris, any interesting customer stories out of the keynote that we should share with the audience? >> I mean, there are so many fantastic stories that you can talk about, I mean, of course we saw the CERN keynote, where they're running managed Kubernetes on top of OpenStack. They have over 250 Kubernetes clusters doing research that are managed by OpenStack Magnum. I mean that's just, to me that's just tremendous. That this is being used in production, it's being used in science, and it's not just across one cloud, it's across many clouds and, You know, we also have AT&T, which has been working very hard on combining OpenStack and Kubernetes to manage their next generation of, of teleco infrastructure. And so, they've been big drivers along with SK Telecom on using Kubernetes as an infrastructure layer and then putting OpenStack on top of that, and then delivering applications with that. And so those are, you know we, the OpenStack Foundation just published on Monday a new white paper about OpenStack, how OpenStack works with containers and these are just a couple of the case studies that we actually have listed in that white paper. >> Chris, you're at the interface between OpenStack, which has become more mature and more stable, and containers, which, although it is maturing is still a little bit, is moving fast, right? Containers and Kubernetes both, a lot of development. Every summit, a lot of new projects, lot of new ways of installing, lot of new components, lot of new snaps. All sorts of things. What are you looking forward to now over the next year in terms of container maturity and how that's going to help us? >> So... People are talking so much now about security with containers and this is another really exciting thing that's coming out of our work because, you know, during one of the container keynotes, one of the things that was kind of driven home was containers don't contain. But, we're actually, at the OpenStack Foundation, we're kind of taking that on, and we, and my colleague Anne Bertucio has been leading a project, you know, has been community manager for a product called Kata Containers, which is, you know, you could almost call it containers that do contain. So I think that this is going to be really exciting in the next year as we talk more and more about we're building more generic interfaces and allowing all sorts of new approaches to solving complex problems, be it in security, be it in performance, be it in logging and monitoring. And so, I think, so the tools that are coming out of this and you know, creating these abstractions and how people are creatively innovating on top of those is pretty exciting. >> The last thing I'm hoping you can help connect the dots for us on is, when we talk Kubernetes, we're talking about multi-cloud. One of the big problems about Kubernetes, you know, came out of Google from you know, if you just say, "Why would Google do this?" It's like, well, there's that one really big cloud out there and if I don't have some portability and be able to move things, that one cloud might just continue to dominate. So, help connect OpenStack to how it lives in this multi-cloud world. Kubernetes is a piece of that, but you know, maybe, would love your viewpoint. >> Yeah, so. This is happening on so many levels. We see lots of large organizations who want to take back control of the cost of cloud and the cost of their cloud infrastructure and so they're starting to pull away from the big public clouds and invest more in private infrastructure. We see this with companies like eBay, we see it with companies like AT&T and Walmart, where they're investing heavily in OpenStack clouds. So that they have more control over the cost and how their applications are delivered. But you're also seeing this in a lot of... Like especially municipalities outside of the United States, you know, different governments that have data restrictions, restrictions on where data lives and how it's accessed, and we're seeing more governments and more businesses overseas that are turning to OpenStack as a way to have cloud infrastructure that is on their home soil, that you know, kind of meets the requirements that are necessary, you know that are necessary for them. And then kind of the third aspect of all of this is sometimes you just, sometimes you need to have lots of availability across, you know, many clouds. And you can have a private cloud, but possibly, in order to serve your customers, you might need public cloud resources, and federation across, across this, both in OpenStack and Kubernetes is improving at such an incredible pace that it becomes very easy to say that I have two, three, four, five clouds, but we're able to, we're able to combine them all and make them all look like one. >> Alright, well Chris Hodge, we really appreciate the updates on OpenStack and Kubernetes in all the various permutations. >> Yeah, it was great talking about it. This is, I mean this is the work that I love and I'm excited about, and this is, you know, I'm looking forward to it, I have fun with it and I keep looking forward to everything that's coming. >> Awesome, well we love to be able to share these stories, the technologists, the customers and everything going on in the industry. For John Troyer, I'm Stu Miniman, back with more coverage here from OpenStack Summit 2018 in beautiful Vancouver, British Columbia. Thanks for watching theCUBE. (tech music)

Published Date : May 22 2018

SUMMARY :

Brought to you by Red Hat, the OpenStack Foundation, to the program, fresh off the container keynote, I'm coming from the east coast. And to the San Francisco Summit prior to that, because the OpenStack Foundation, you know, has a year later it was, you know, oh my gosh, Kubernetes. and there's five different ways you could and the other has to lose is actually kind of, and the infrastructure somebody's got to manage and so the layers kind of interleave with one another. a Cinder API and you have one interface to your Kubernetes. I believe it's called the Cloud Provider OpenStack. The owner of that code is the and is broadened because of it. and once you have that, and you have many people, All of a sudden, a standard interface, You know, it essentially allows you to build new things that you can talk about, I mean, of course Containers and Kubernetes both, a lot of development. and you know, creating these abstractions and Kubernetes is a piece of that, but you know, that is on their home soil, that you know, in all the various permutations. and I'm excited about, and this is, you know, stories, the technologists, the customers and everything

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Chris HodgePERSON

0.99+

Anne BertucioPERSON

0.99+

WalmartORGANIZATION

0.99+

ChrisPERSON

0.99+

John TroyerPERSON

0.99+

Stu MinimanPERSON

0.99+

Chris HogePERSON

0.99+

SK TelecomORGANIZATION

0.99+

AT&TORGANIZATION

0.99+

OpenStack FoundationORGANIZATION

0.99+

PortlandLOCATION

0.99+

Red HatORGANIZATION

0.99+

MondayDATE

0.99+

twoQUANTITY

0.99+

John FurrierPERSON

0.99+

United StatesLOCATION

0.99+

two driversQUANTITY

0.99+

North AmericaLOCATION

0.99+

eBayORGANIZATION

0.99+

Bay AreaLOCATION

0.99+

GoogleORGANIZATION

0.99+

OpenStackTITLE

0.99+

a year laterDATE

0.99+

JohnPERSON

0.99+

threeQUANTITY

0.99+

Vancouver, CanadaLOCATION

0.99+

oneQUANTITY

0.99+

firstQUANTITY

0.99+

OpenStack Summit 2018EVENT

0.99+

Vancouver, British ColumbiaLOCATION

0.99+

three years agoDATE

0.99+

yesterdayDATE

0.99+

last yearDATE

0.98+

next yearDATE

0.98+

fourQUANTITY

0.98+

KubernetesTITLE

0.98+

CNCFORGANIZATION

0.98+

theCUBEORGANIZATION

0.98+

todayDATE

0.98+

three leadsQUANTITY

0.98+

bothQUANTITY

0.98+

this weekDATE

0.98+

fiveQUANTITY

0.98+

AmazonORGANIZATION

0.97+

this yearDATE

0.97+

over 70 storage driversQUANTITY

0.97+

one interfaceQUANTITY

0.97+

OpenStackORGANIZATION

0.97+

third aspectQUANTITY

0.96+

over 250 KubernetesQUANTITY

0.96+

one cloudQUANTITY

0.96+

San Francisco SummitEVENT

0.96+

five different waysQUANTITY

0.95+

Dan Kohn, CNCF | KubeCon 2017


 

>> Narrator: Live from Austin, Texas, it's theCUBE, covering KubeCon and CloudNativeCon 2017, brought to you by Red Hat, the Linux Foundation, and theCUBE's ecosystem partners. >> Okay, welcome back, everyone. This is theCUBE's exclusive coverage live here in Austin, Texas for the CNCF's two conferences, CloudNativeCon, which was yesterday, and two days, today and tomorrow, KubeCon for Kubernetes' conference. This is theCUBE, of course, from SiliconANGLE Media. I'm John Furrier with my cohost, Stu Miniman. Our next guest, Dan Kohn, is the executive director of the CNCF, the man who put it all together. Congratulations. Welcome back to theCUBE. Good to see you. >> Oh, absolutely. Thrilled to have you guys back here again. >> So you kind of doing a victory lap here now, high fiving each other? >> Dan: Great hugs. >> John: Great event. >> Laughing: I'm glad it's a good event, and I am hearing fantastic feedback that folks are thrilled to be here. But we sort of describe this moment for the organization and the community as being the end of the beginning. >> John: Yeah. >> Where we now have all the major cloud vendors, all of the biggest enterprise software companies. We have a core group of 14 projects anchored by Kubernetes, but tons and tons of work in front of us. >> And tons of success, so I'm just going to read a couple of highlights from yesterday. There's a lot today. Baidu joins the CNCF, a lot of scaling production application examples, 31 new silver end-user members joined, Alibaba Cloud update to platinum, CoreDNS 1.0, Containerd, Fluentd, Jaeger, tons of news. Obviously, we've been pumping out the coverage. Today, again, more and more great goodness. But really interesting is that you guys have put a frame around this community to allow it to grow, to fertilize the open source vibe, which is all cloud but yet scaled. And you put up a slide I want to get your reaction to that I thought was compelling yesterday during your keynote. It was the flywheel, circle, and it said projects, products, profit. >> Dan: Right. >> And not that you're promoting profit, but you're not hiding the ball, either, saying, hey, you know what? There's a lot of commercial interest in cloud, obviously. We saw AWS' success last week. And that is if you create good products in this community framework, there's profit to be had. >> Right. So first of all, I should admit to plagiarizing that slide from Linux Foundation Executive Director Jim Zemlin. >> And similarly, I think you can look at a lot of aspects... >> It's an open source feature. >> Dan: Yes. >> Free for you to use. >> John: Right. >> Similarly, I think there's a lot of ways in which Kubernetes is trying to build on the success of Linux. And Jim even describes Kubernetes as the Linux of the cloud. >> John: Yeah. >> Stu: Yeah. >> John: That's a good point. >> Dan, one of the things we've been talking around Kubernetes is you talk about scale. >> Dan: Right. >> Talk about scale of the CNCF. You have 4 to 14 projects. People are a little worried when you get all the vendors around here and there's all these projects. It's a foundation thing, it's going to go off the rails. >> Dan: Yeah. >> Customers aren't going to have a voice. How do we make sure we kind of learn from some of the things that other projects have had challenges with in the past? >> And I think that's our advantage, which is the great thing about coming later than some of the other foundations, is we can look at where they had successes and where they had issues. And our aspiration for CNCF is to get to go make entirely new mistakes rather than replicating some of the issues that have come before. And so really from the beginning of CNCF, we had a somewhat unusual and frankly a little bit cumbersome charter where I describe it at times as a three-ring circus. We have a governing board made up of the vendors that are putting a lot of money into the community, but they don't get to run the projects and they don't even get to pick the projects. Instead, they appoint six of the nine members of an independent technical oversight committee, kind of like the Supreme Court. And then we have a third group in the end-user community that I'm thrilled to say is now up to 28 members in it. They appoint one of those folks. We finally got that working. We have Sam Lambert, the director of infrastructure at GitHub, who has just made a huge commitment to Kubernetes and is moving all their infrastructure over into it. Those seven appoint the last two. And so that body, and they just had their public meeting a couple hours ago. They feel very strongly about their independence, about their reputation, that they're trying to make very good judgments based on what they're seeing in the marketplace. >> That's interesting, the three-ring circle. I like how you put it. But let's talk about the end-user piece because I think that's critical. One of the things we were commenting earlier from the Lyft folks was you have a lot of end users who have built some large-scale systems out of their own sheer necessity. >> Dan: Definitely. >> And that is now being donated in. We saw Kubernetes come in with, you shepherded beautifully, went from Google, but you've got Lyft donating an amazing product convoy. >> This first convoy has a huge amount of excitement. And what was fun was, actually, on the same stage that they contributed back in LA in September, Uber contributed a separate project. Now, unlike Uber and Lyft, the two projects are in no way competitive- >> John: Yeah. >> Like Jaeger is really fantastic tracing one. But what they have in common is that they're companies that have had to grow from nothing to extremely high scale and then had problems that they solved. And they wanted to share that expertise with us. >> I want to get your thoughts on this. Because we've been speculating, on theCUBE, we've been kind of thinking, an editorial, but just that this is all good business. Now, that's pretty obvious, right? You're starting to see this kind of contribution, the gifts that keep on giving. These are significant code. >> Dan: Yeah. >> Not like, okay, let's start a little group and huddle and build something organically. You have real goodness coming in from Google, Uber, Lyft, and there's a million others. >> Dan: Right. >> How is that changing the game? Certainly accelerating it. That's really bringing goods to the table. >> Right. I think the whole... >> You have to manage it. >> Well, and for what it's worth, I don't actually manage the projects. And so we do provide a set of services- >> John: The community? >> -to them and we help them, we market them. But one of the unusual aspects of CNCF is that the projects do actually manage themselves. A little bit of guidance from the TOC, but we really are unusual in that sense. And that's one of the reasons the projects have been... >> And what's interesting is, to connect the dots, though, one step further, you're talking about a commercial entity donating massive intellectual property in the open for all the goodness of everyone else. But yet that flywheel is continuing. They're still using it. So it is inherently commercial dynamic. >> Right. And back to that circle, I think really the underlying concept is that companies agree that sharing key parts of their infrastructure has a huge amount of value to the whole ecosystem, to each other. And then they're absolutely eager to compete above that. And so you can look at it with the public clouds where we have now Amazon, Microsoft, Google, Alibaba, IBM, Oracle all at the table. They are absolutely fierce competitors. But they're saying that this specific software infrastructure layer isn't the area that they want to compete. They want to compete on all the value-added services, customer service, et cetera. >> Dan, I wonder if you can speak to how CNCF connects to some of the broader communities out there. Things like Kata containers got announced coming out of the OpenStack group. You've got a serverless track happening here, kind of extends some of where Kubernetes is going. How does CNCF fit into the broader... >> Sure. And it's definitely the case that all the innovation out there cannot happen in CNCF. Most obviously, everything that we do, almost everything depends on Linux. And so that's our parent organization, the Linux Foundation. But we've had a good collaboration with Jonathan Bryce from OverStack. They have two booths on the floor here at the show. And we've spoken to Clear Containers and RunV, the two predecessors in the past. But the part that I'm particularly pleased with for Kata containers is that it is an OCI-compliant runtime, that's another sister organization, and is really designed to work well for Kubernetes. And then they can pitch that and let the market go decide which container runtimes they find the most valuable. >> Obviously a lot of traction here in terms of the sentiment around service meshes and pluggable lock-in textures. That's been very cool. But security came up. So I want to get your thoughts around security, obviously storage and these older models around how to deal with storage and networking. Obviously, always in the action. >> Yeah. >> But security is top of mind for everyone. How is that being addressed? You know, talk is out there... >> Sure. I mean our philosophy on this is that moving to cloud-native and particularly the continuous integration and continuous development that goes along with that is the most important step that you can do to help secure your infrastructure. And Equifax is the example everyone always brings up. But there was a case where they were using known insecure software and they didn't have the processes up to place where instead of doing quarterly updates or monthly updates, you want to be doing dozens of updates per day. And a cloud-native infrastructure allows you to do that. >> What's next for you? Because you've got great traction with both community response, and the community has been absolutely amazing, the quality of people, level has been great, but also at the funding sponsors. You've got a lot of people that are involved. What's next? What happens next? What do you envision happening? What's the plan, and then how do you view that evolving? >> Well, I hate to fall into the buzzword implosion here, but if you go back to the crossing the chasm metaphor, I think we're still very much just in the early adopter phase. 2018 could very well be the moment that we jump over to the early majority. And I do feel like this whole community now has the velocity to do that and that we're on track for it. But as that happens, there's just far, far more people who need to be educated so they understand the projects and the options and how to work with them. And then hopefully they go from just being consumers of these technologies to contributors and that we can welcome them into our community and hopefully get the advantage of their expertise as well. >> I want to get your thoughts on a comment that Stu and I were talking about. Stu, you and I were talking about the notion of value creation above the stack, and then how Kubernetes, although some could say being commoditized, but it's also creating value because with that consistency of Kubernetes, you can now create value. So we believe, and I want to get your reaction to this, because we think a whole new ecosystem dynamic will emerge of a new kind of ecosystem. And if this new app developer combined with software engineering, which is really going on, you're talking about the cloud, the app developers will just build in value, that value creation will be rewarded. That's where monetization will be happening. >> And if I could build off that... >> John: Yeah. >> Dan, I loved one of your opening comments. You quoted, "exciting times for boring infrastructure, "maybe too exciting." So this week we've been teasing out there's a lot of work to make that infrastructure boring. You've got everybody on this floor, the CNCF board, lots of new projects making that. Where the action is and what this is going to create is that application monetization and the speed and agility of being able to create these cool new cloud-native applications out there. So it's interesting dynamic, spans broad pieces of this, layers of the stack there. >> Yeah. Well, I will point out that there was an odd level of unanimity of just a ton of different leaders in the community, in keynotes from Craig McLuckie and Chen Goldberg and others where they all agree that Kubernetes is not by any means the ultimate answer or the final answer. I think everybody now expects to see Kubernetes as a core aspect of the infrastructure for software for the next decade or more. But there's a belief that there's a whole ton of value that needs to be added above it, particularly to try and show for a regular application developer who just has a PHP app or no-GS microservices or anything else what's the easiest way to go from having a piece of software and deploying it effectively. >> Dan, so it's interesting. You watch the people on the outside. They're like, oh, look at Kubernetes. They're all holding hands and saying Kumbaya. We know there's some spirited debates that happen- >> Dan: Definitely. >> In the code, some projects that are sometimes competing up there. Why has the community come together, and where are some of the areas that we still need to work on and improve to help customers going forward? >> And again, I think they have the big advantage of having watched other communities that didn't value community and consensus and the ability to work through their issues. And so thankfully, we just have a ton of really capable engineers who also have some of those social or personal qualities that they care about working these things out. And to date, at least, I think most of those disagreements have been settled pretty amicably and in a positive direction. I think there's still huge swathes of this space that are still up in the air. Storage is an obvious one where there's a ton of work going on in a storage working group of CNCF. Serverless is another where I think everyone agrees that the application deployment model of AWS Lambda is really exciting and has things that people should replicate and should be brought over to Kubernetes. But how that should happen, what the software is, et cetera, there's still, in fact, we have our first serverless track today here at KubeCon where several different competing approaches are all talking about what they'd like to do. >> Awesome stuff. And you also announced some dates for next year, December 11 and 13 in Seattle. >> Dan: Yes. >> Okay. >> Dan: That's a year from now. >> November 14 and 15 in Shanghai. >> Now, you and I met in Hangzhou in the lobby, which was just amazing. But I certainly am hoping to convince you to go back to China with us. This will be our first event... >> I got a three-year visa. >> Good, yeah, that's the exactly right one. But this will be our first event in China, which I think is just a huge opportunity. We now have Baidu, Tencent, Huawai, ZTE, a number of startups. There's just so much excitement for this space over there that we're really excited to satisfy. >> Stu: And Copenhagen in May. >> And that's the last one. Thank you. May 2 to 4 in Copenhagen, and we're really excited for the event, to bring it to Europe and the rest of the world. >> Okay. So you've been working like a dog, you've been working hard. I've seen you in China. It's serendipitous. But it's not without being mentioned that this has been great effort by your team and the Linux Foundation and Jim and the whole team. But congratulations. Are you having a pinch me moment? I know it's too early to do a victory lap. >> But you've got to be pretty excited. >> Yeah. It really has been a great thing for the foundation that we sort of accomplished many of our 2018 and 2019 goals this year. But I'm sure we're going to find plenty of stuff to do next year. >> And your goal for the next 6 to 12 months, what's on your top three to-do's, continue the momentum? Share your API for... >> Yeah. What's great is that we really have plenty of members. We'd always like to add new ones and serve the ones we have better. But right now, the focus is really about providing better services to our projects. All of them feel overworked. They would love help on documentation, on marketing, on messaging about it, and some of them need help with testing development and other things. So that's really what we're buckling down on. >> Great community are going to test them, being here on the ground, personally present at creation. And I was standing there with J.J. and Lew Tucker, OpenStack three years ago, talking about Kubernetes. We were kind of ripping. We couldn't have imagined, then, obviously, they bolted it on last year with your event. Now second year here, huge community... >> But you have 4,100 folks here, is more than the previous four events combined. >> Yeah, awesome. >> So it really is exciting. >> TheCUBE, always on the ground. And sometimes the squirrel finds a nut. We found a cloud-native foundation, part of the Linux Foundation. CNCF, Cloud-Native Compute Foundation, really a new, growing, and relevant community for cloud and a new way to do software and reimagine the future from software engineering to full application development, a new way. This is theCUBE's coverage, and we are here live in Austin. More live coverage after this short break. We'll be right back. [Techno Music]

Published Date : Dec 7 2017

SUMMARY :

brought to you by Red Hat, the Linux Foundation, of the CNCF, the man who put it all together. Thrilled to have you guys back here again. for the organization and the community all of the biggest enterprise software companies. But really interesting is that you guys And that is if you create good products to plagiarizing that slide from Linux Foundation And Jim even describes Kubernetes as the Linux of the cloud. Dan, one of the things we've been talking all the vendors around here and there's all these projects. Customers aren't going to have a voice. And so really from the beginning of CNCF, One of the things we were commenting earlier And that is now being donated in. the two projects are in no way competitive- And they wanted to share that expertise with us. the gifts that keep on giving. and huddle and build something organically. How is that changing the game? I think the whole... I don't actually manage the projects. is that the projects do actually manage themselves. in the open for all the goodness of everyone else. isn't the area that they want to compete. coming out of the OpenStack group. And so that's our parent organization, the Linux Foundation. Obviously, always in the action. How is that being addressed? is the most important step that you can do What's the plan, and then how do you view that evolving? and the options and how to work with them. the app developers will just build in value, and the speed and agility of being able as a core aspect of the infrastructure We know there's some spirited debates that happen- In the code, some projects that are sometimes and the ability to work through their issues. And you also announced some dates But I certainly am hoping to convince you But this will be our first event in China, And that's the last one. and the Linux Foundation and Jim and the whole team. for the foundation that we sort of accomplished many And your goal for the next 6 to 12 months, and serve the ones we have better. being here on the ground, personally present at creation. is more than the previous four events combined. And sometimes the squirrel finds a nut.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmazonORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

IBMORGANIZATION

0.99+

Dan KohnPERSON

0.99+

AlibabaORGANIZATION

0.99+

HangzhouLOCATION

0.99+

Sam LambertPERSON

0.99+

sixQUANTITY

0.99+

ChinaLOCATION

0.99+

GoogleORGANIZATION

0.99+

CopenhagenLOCATION

0.99+

Jonathan BrycePERSON

0.99+

JimPERSON

0.99+

EuropeLOCATION

0.99+

JohnPERSON

0.99+

ZTEORGANIZATION

0.99+

John FurrierPERSON

0.99+

OracleORGANIZATION

0.99+

DanPERSON

0.99+

ShanghaiLOCATION

0.99+

SeattleLOCATION

0.99+

CNCFORGANIZATION

0.99+

AustinLOCATION

0.99+

Red HatORGANIZATION

0.99+

4QUANTITY

0.99+

LALOCATION

0.99+

Stu MinimanPERSON

0.99+

J.J.PERSON

0.99+

UberORGANIZATION

0.99+

SeptemberDATE

0.99+

Jim ZemlinPERSON

0.99+

Linux FoundationORGANIZATION

0.99+

GitHubORGANIZATION

0.99+

AWS'ORGANIZATION

0.99+

StuPERSON

0.99+

2018DATE

0.99+

4,100 folksQUANTITY

0.99+

TencentORGANIZATION

0.99+

LyftORGANIZATION

0.99+

Austin, TexasLOCATION

0.99+

Cloud-Native Compute FoundationORGANIZATION

0.99+

first eventQUANTITY

0.99+

next yearDATE

0.99+

13DATE

0.99+

two projectsQUANTITY

0.99+

2019DATE

0.99+

Chen GoldbergPERSON

0.99+

BaiduORGANIZATION

0.99+

Lew TuckerPERSON

0.99+

KubeConEVENT

0.99+

14 projectsQUANTITY

0.99+

yesterdayDATE

0.99+

15DATE

0.99+

CloudNativeConEVENT

0.99+

three years agoDATE

0.99+

tomorrowDATE

0.99+

oneQUANTITY

0.99+

November 14DATE

0.99+

4DATE

0.99+

Dan Walsh, Red Hat | KubeCon 2017


 

>> Announcer: Live from Austin Texas, it's theCUBE. Covering KubeCon and CloudNativeCon 2017. Brought to you by Red Hat, the Linux Foundation, and theCUBE's ecosystem partners. >> Welcome back, this is SiliconANGLE Media's live coverage wall to wall of KubeCon and CloudNativeCon here in Austin, Texas. Got the house banner rocking all day. I'm Stu Miniman, happy to be joined on the program, Dan Walsh who's a consulting engineering with Red Hat. Rocking the red hat, Dan thanks so much for joining us. >> Pleasure to be here. >> Alright so we've, you know Red Hat has a strong presence at the show, we had Clayton on yesterday, top contributor, won an award actually for all the contribution he's done here. Going through a lot of angles. Why don't you start with, tell us kind of your role, what you've been doing at Red Hat. >> So at Red Hat I'm a consulting engineer, which basically means I lead a team of about 20 engineers, and we work on the base operating system. Basically anything to do with containers from the operating system on down. So kernel engineers. But everything underneath Kubernetes. So traditionally for the last four and a half years I've been working on the Docker Project as well as other container type efforts. We've added things like file system support, Docker, lots of kernel changes, lots of, you know we're working forever on usernames base things like that. More recently though we've been working, we started to work on sort of one of the, well OpenShift and Kubernetes were built on top of Docker originally, and they found over time that the Docker base was changing in ways that were continuously breaking Kubernetes. So about a year and a half ago we start to work on a project called Crio. So a little history is if you go back, Kubernetes was originally built on top of Docker. But core OS came to Kubernetes and wanted to get rocket support into Kubernetes. And rather than add rocket support, Kubernetes decided to find this interface. Basically a CRI, container runtime interface, which is an API that they would call out to to run containers. So rocket could build a container runtime interface, they actually built a shim for the Docker API. But we decided at that time to basically build our own one, and we called it Crio. So it's container runtime interface for OCI images. So the plan was to build a very minimalist daemon that could support Kubernetes, and Kubernetes alone. So we don't support any other orchestrations or anything else. It's totally based on Kubernetes CRI. So our versioning matches up with Kubernetes. So Kubernetes one dot eight, you got Crio one dot eight. Kubernetes one dot nine, you got Crio one dot nine. >> So Dan we've been talking about this. You know Red Hat made a pretty strong bet on Kubernetes relatively early in there. Red Hat, very open, everything you do is 100% open source. Why for Crio, why only Kubernetes? There's other orchestrations out there that are open source. >> Well let's take a step back. So one of our goals in my group was to take, sort of what does it mean to run a container. So if you think about when I run a container, what do I need? I need a standard container image format, so there's the OCI image bundle format that defines that. The next thing I need is the ability to pull an image from a container registry to the host. So we built a library called containers image that actually implements all of the capabilities of moving containers back and forth around, but basically at a Command Line or a library level. We built a tool on top of that called Scopio, which allows us to basic Command Line, I can move from one container registry to another, I can move container registries to different kinds of storage. I can move directly from a container registry into a Docker daemon. So we have a, so the next step you need when you want to run a container is storage. So you need to take that container image and put in on disk. And in the case of containers you do that on top of what's called the copy and write file system. So you need to be able to have a layering file system. So we created another project called container storage that allows you to basically store those images on storage. The last step for running a container is actually to launch an OCI runtime. So we, OCI runtime specification and run c takes care of that. So we have the four building components for what it means to run a container separate. So we're building other tools around that, but we built one too that was focused on Kubernetes. And again, the reason Red Hat bet on Kubernetes is we felt that they had the best longterm potential, and judging by this show I think we made a sane bet. But we will work with others. I mean these are all fully open source projects. We actually have contributors coming in that are contributing at these low level tools. For instance pivotal is a major contributor in container image. And they're using it for pulling images into their base. We have other products that projects are using, and so it's just not Kubernetes. It's just Crio is a daemon for Kubernetes. >> Yeah Dan it's really interesting. You listen in Clayton's keynote this morning. He talked about one of the goals you have at Red Hat is making that underlying infrastructure boring so that everything about it can rely on it, and works on. There's a lot of work that goes on under there. So it's like, the plumbers and the mechanics down underneath making sure it all works. >> A lot of times when I give talks, the number one thing I'm always trying to teach people is that containers are not anything really significantly different. Containers are just processes on a Linux system. So if you booted up a regular REL system right now, and you looked at Pid One of a system. Let me take a step back, I define containers as being something that has, c groups associated with a resource constraints, it has some security constraints associated with it, and it has these things called name spaces, which is a virtualization layer that gives you a different view of the processes. If you looked at every process on a Linux system, they all c groups associated with them, they all have security constraints associated with them, and they all have name spaces associated with. So if you went to Pid One, if you went to slash proc Pid One slash NS you would see the name spaces associated with Pid One. So that means that every process on Linux is in a container. By the definition of a container being those three things. And all that happens on the system is you toggle those. So you can tighten them or change some of the name space and stuff like that, and that gives you the feel of the virtualization. But bottom line is they're all containers. So all the tools like Docker, rocket, Crio, run c, or any one of those tools are all just basically going into the kernel, configuring the kernel, and then launching the Pid One on the system. And from that point on it's just a kernel that's taking 'em. We at Red Hat has a t-shirt that we often wear that says Linux is containers and containers is Linux. And that actually proves the point. So bottom line is you know the operating system is key, and my team and the developers I work with, and the open source community is all about how can we make containers better? How can we further constrain these processes? How can we create new name spaces? How can we create new c groups, new stuff like that? So it's all low level stuff. >> Dan, you know give us some flavor as to some of the customer conversations you're having at the show here. Where are they? I mean we know it's a spectrum of where they are, but what are some of the commonalities that you're hearing? >> I mean at Red Hat our customers run the gamut. So you know we have customers who can barely get off a rel five which came out 12 years ago. Two sort of the leading edge customers. And the funny thing is a lot of these are in the some companies. So most of our customers at this point are just beginning to move into the container world. You know they might have a few containers running, or they had their developers insisting, hey this container stuff cool I want to start playing with it. But getting them from that step to the step of say Kubernetes, or to get them to step with OpenShift, is sort of a big leap. My fear with a lot of this is a lot of people are concentrating too much on the containers. You know the bottom line is what people need to do is develop applications. And secure applications. My history is very based in heavy security. So really we face a lot of customers who sort of have home grown environments. And their engineers come in and say oh I want to do a Docker build, or I want to talk to the Docker socket. And I always look at that and question, you know you're supposed to be building apps, you're building banking apps, or you're building military apps, you're building medical apps. They should be concentrating on that and not so much on the containers. And that's actually the beauty of OpenShift. You can set up OpenShift workloads in such a way that their interaction to build a container is just a Git check it. And it's not, you don't have to go out and understand what it means to build a container. You don't have to get the knowledge of what it means to be able to build a container and things like that. >> Dan you bring up a really good point. At this show most of the customers I'm talking about, it's really about the speed for them to be able to deliver on the applications. Yes there's the people building all the tooling, and the projects here, and there's many customers that are involved with it. But we've gone further up the stack where it's closer to the application, less to that underlying infrastructure. >> And the other thing customers are looking for, in my case, as I said I have a strong background in security, I did SE Linux for like 13 years. Most of my time talking to customers is about security, and how can we actually confine containers, how do we keep them under control, and especially when they go to multi tenancy. And some good things, I don't know if you're going to talk to Kata? Have you heard about the Kata project? >> So we've talked to a couple people, Kata coming out of the open-- >> Clear containers and-- >> Yeah clear container of the intel. >> Yeah and I think that those, getting to those levels of using hardware isolation, it really helps out in-- >> It's interesting because actually, you know when first looking at, it's like wait it's kind of a lightweight VM, it's a container. Where does that fit in? >> They're really just containers, 'cause they're not, a lightweight VM would be actually booting up like an init system and running logging and all these other things. So like a Kata container or, I'm more familiar with clear containers. A clear container is literally just running a very small init system and then it launches run c to run, actually start up the container. So it has almost no operating system inside of the lightweight VM. As opposed to running just regular virtual machines. >> Dan would love your take on, you know you talked about security. Security of containers, the role of security in the cloud native space. What are you seeing, and what do we need to work on even more as an industry? >> It's funny because my world view is at a much lower level than other security people that we talk to. There's other security people that'll be looking at sort of network isolation and role based access control inside of Kubernetes. I look at it as basically multi tenancy. So running multiple containers with different workloads, and what happens if one container gets hacked, how does that affect the other containers that are running and how do I protect the services? So over the years when we've been working with Docker, I got SE Linux support in, we've gotten Setcom support in. We're trying to take advantage of everything in the Linux kernel to further tighten the security. But the bottom line is a process inside of the container is talking to the real kernel on the host. Any vulnerability in the host kernel could lead to an escalation and a breakout. So that's why no matter what you say, a hyper, like a hyper shell, a separate container running inside of a VM is always going to be more secure. But that being, on the other hand, containers in a lot of cases you want to have some interaction. If you go all the way to VM you get really bad isolation. So you really have to cover the gamut. So a lot of times I'll tell people to look at containers as being, they're not a zero sum game. You don't have to throw away all your VMs to move to containers. I tell people the most secure way to run a application is separate physical hardware. The second most is on VM. So the third most is inside a container. And then you can go on to all down the line. But there's nothing to say that you can't run your containers inside of separate VMs, inside of separate physical machines. So you can set up your environment in such a way. Say you have your web front end sitting inside of VMs inside of (mumbles) zone on separate physical hardware you setup your databases or your credit card data on separate physical machines, separate VMs, and separate containers inside of it. So you can build up these really high levels of security based on containers, virtualization, and physical hardware. I can go on forever on this stuff. >> Dan Walsh, really appreciate sharing some of the ways that Red Hat's trying to help some of those underlying pieces become boring. So the customers won't have to worry about. >> That's really what it's about. If you know what's going on at the host level then I haven't done my job. So our goal is to basically take that host level, and make it disappear. And you can work with your higher level orchestration level. >> Well Dan, it's great to catch up with you, thanks so much for joining us. We'll be back with lots more coverage here from KubeCon 2017 in Austin, Texas. I'm Stu Miniman and you're watching theCUBE. (electronic music)

Published Date : Dec 7 2017

SUMMARY :

Brought to you by Red Hat, the Linux Foundation, Rocking the red hat, Dan thanks so much for joining us. presence at the show, we had Clayton on yesterday, So a little history is if you go back, So Dan we've been talking about this. So we have a, so the next step you need when you So it's like, the plumbers and the mechanics And all that happens on the system is you toggle those. some of the customer conversations you're having So you know we have customers who can barely get and the projects here, and there's many customers And the other thing customers are looking for, you know when first looking at, So it has almost no operating system inside of the Security of containers, the role of security So a lot of times I'll tell people to look at containers So the customers won't have to worry about. So our goal is to basically take that host level, Well Dan, it's great to catch up with you,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dan WalshPERSON

0.99+

DanPERSON

0.99+

Stu MinimanPERSON

0.99+

Red HatORGANIZATION

0.99+

ClaytonPERSON

0.99+

100%QUANTITY

0.99+

Linux FoundationORGANIZATION

0.99+

13 yearsQUANTITY

0.99+

Austin, TexasLOCATION

0.99+

KubeConEVENT

0.99+

TwoQUANTITY

0.99+

SiliconANGLE MediaORGANIZATION

0.99+

KubeCon 2017EVENT

0.99+

three thingsQUANTITY

0.99+

DockerTITLE

0.99+

CloudNativeConEVENT

0.98+

KubernetesTITLE

0.98+

LinuxTITLE

0.98+

Austin TexasLOCATION

0.98+

yesterdayDATE

0.98+

OpenShiftTITLE

0.98+

theCUBEORGANIZATION

0.98+

one containerQUANTITY

0.97+

about a year and a half agoDATE

0.97+

about 20 engineersQUANTITY

0.97+

oneQUANTITY

0.97+

firstQUANTITY

0.97+

KataTITLE

0.96+

thirdQUANTITY

0.96+

four building componentsQUANTITY

0.96+

12 years agoDATE

0.96+

Red HatTITLE

0.94+

Crio oneTITLE

0.94+

CloudNativeCon 2017EVENT

0.93+

secondQUANTITY

0.93+

Kubernetes oneTITLE

0.91+

nineTITLE

0.9+

CrioTITLE

0.9+

ScopioTITLE

0.88+

DockerORGANIZATION

0.86+

SE LinuxTITLE

0.81+

eightTITLE

0.81+

Lew Tucker, Cisco | KubeCon 2017


 

>> Announcer: Live from Austin Texas, it's theCUBE. Covering KubeCon and CloudNativeCon 2017. Brought to you by Red Hat, the Linux Foundation, and theCUBE's ecosystem partners. >> Welcome back everyone, this is theCUBE live in Austin, Texas for our exclusive coverage at the CloudNative Conference and KubeCon with Kubernetes via theCUBE. theCUBE which we're live, and 8 years running, I'm John Furrier, the founder of SiliconANGLE Media, my colleague, Stu Miniman. And I'm excited to have Cube alumni, and its distinguished industry legend, Lew Tucker, Vice President CT of Cloud Computing at Cisco Systems. Welcome back to theCUBE, great to see you. >> Great to be back, it's one of my favorite shows. >> Lou, we've had many conversations over the years, and it's always great to have you on because you're on the cutting-edge perspective, but you have a historical view as well, you've seen many waves of innovation. And obviously you own lots of property in the Computer's History museum, your resume goes on and on. But, you got to admire this community. Three years old, it was you, me and JJ we're sitting around at OpenStack in Vancouver three and a half years ago, having a beer after the event one of these days, and we were talking about Kubernetes, and we were really riffing on orchestration and kind of shooting the arrow forward, kind of reading the tea leaves. And we were predicting inter-clouding, inter-networking, Cisco core competency, the notion of application developers wanting infrastructure as code. We didn't actually say mircoservices but we were kind of describing a world that would be microservices, and this awesomeness that's going on with the Cloud. What a ... [Lew] You were right. You were right. >> We were right, it wasn't me, it was the community. This is how communities operate. >> It is. I think that what we're seeing, and particularly in these open source communities, you're getting the best ideas. And therefore, a lot of people are looking at this future space, and then we bring the kids out of the communities, get the projects that we work together on it, and that's how we move it forward. >> You've been a great leader in the community, just want to give you some props for that, you deserve it, but more importantly is just the momentum going on right now. And I want to get your take, you're squinting through the growth, you're looking at the innovation, looking at the big picture, certainly from a Cisco perspective, but also as an industry participant. Where's the action? Obviously containers grew, that tide came in, a lot of boats floated up. We saw microservices boom, then we now, Kubernetes' getting better and better, multiple versions, it's - some say commoditized, some would say more inter-operable. Really, that's the connection tissue for multi-cloud. >> Exactly right. >> Do you see the same thing? Where's the action? >> So, cloud computing is going everywhere now. And so it's natural that we see one of the next phases of this is in the area of multi-cloud. The customers, they are in public cloud, they have private data centers where they want to run similar applications. They don't want to have a completely different environment. What they really want to see is a consistent environment across which they can deploy applications. And that consistent environment also has to have security policies, authentication services, and a lot of these things. And to really drive the innovation, what I find interesting is that, the services that are coming now out of public cloud, whether it be an AI or server list, event-driven kind of programming models. Enterprises want to connect into them. And so one of the things I think that that leads to is that you're beginning to hear talk now, just beginning to hear it, which is this project called Istio. Which is a service mesh, because what that really allows -- >> John: What's the project name? >> It's called Istio? >> John: Istio. >> Lew: I-S-T-I-O. >> Okay. >> dot I-O. Everything is open source, it's a project that's contributed to by Google, and IBM, and Lyft, and now Cisco's getting involved in it, as well. And what it really plays into is this world of multi-cloud. That now we can actually access services in the public cloud from your own private data center, or from the public running applications in a public cloud, you can access services that are back in your data center. So it's really about this kind of application-level networking stack, that means that application developers can now off-load all of that heavy work to a service mesh, and therefore that'll accelerate application development. >> So it's interesting, I heard some talk about things like Envoy edge and service proxies, and service proxies have been a nice tool to kind of cobble together old legacy stuff, but now you're seeing stuff go to the next level. This data I heard in the keynote, I want to get your reaction 'cause this kind of jumps out at me. Lyft had created a mesh over hundreds of thousands of services over millions of transactions per second. Lyft. Uber's got some stuff on the monitoring side, Google's donated - This is large scale cloud guys who had to build their own stuff with open source, now contributing all this stuff back. This is the mesh you're talking about, correct? >> This is exactly right, yes. Because what we're seeing is, we've talked about micro services, and Kubernetes is about orchestration of containers. And that has accelerated application development and deploying it. But now the services, each one of those services still has all of this networking stuff they have to deal with. They have to deal with load balancing, they have to deal with retries, they have to deal with authentication. So instead, what is happening now, we're recognizing these common patterns, this is what the community does (mumbles). You see a common pattern, you abstract it, and you push that out into what is known as side cars now, so that the application developer doesn't have to -- the application doesn't get changed when you need to change, like, 'bring up a couple more services over here' 'put this on a different cloud'. The individual components now are unaffected by that, because all of that work has been offloaded into a service mesh. >> Lew, bring us inside a little bit. Dig into that next level of kind of networking. 'Cause you speak, kind of networking administrator, running around the data center, you get everything from pulling cables to zoning and everything like that. Now it's multi-cloud, multi-service, everything's faster. Through all the architect, the person running it, automation ... We don't have an hour, but give us a little bit about what it means to be a networking person these days. >> Well, it's interesting, because one of the things that we know application developers did not want to become, is to be a network engineer. And yet to do a lot of what they had to do, they had to learn a lot of those skills. And instead they would rather set things up by policy. For example, they would like to be able to say, 'if I'm deploying now the version two of my application', it's a classic thing we talk about in this deal, 'the next version we want to just direct' '5% of the traffic to it, make sure it's okay' 'before we turn over the whole thing.' You should be able to do that at the application level, and through a service mesh that is built in networking at the application level, the application guys can do it. Now the role of the network engineer is still the same, they have to provide the basic infrastructure to allow that to happen. And for example, a lot of the infrastructure now is extending the Cloud from public cloud through the cloud BPM services that they have back into the data center. So Cisco, for example, is putting technologies that are running at AWS and at Google, and Azure, that allows that to come back into the data center. So we can run Cisco virtual routers in the Cloud, connected back up in the data center. So their standard networking policy that the networking engineers really want to see enforced, they can be assured that that's enforced, and then Istio layers on top of it. >> And that's decoupled from the application. >> Right. Right. >> This is what we've been talking about since 2010, our eighth year of theCUBE, infrastructure as code. This is what DevOps was all about, and now it's evolving mainstream. >> Absolutely right. You really want infrastructure to be as boring as possible. And capable and then secure. And now give a lot more control over to the application developer. And we also know, right now it's really based largely on Kubernetes, it's a great example, but that will connect into virtual machines, it will connect into legacy services. So all of this has to do with connecting all of those pieces that are today in an enterprise, moving to a public cloud. And that transition doesn't happen wholesale. You move a couple over. >> Lew, one thing. I want you to look back, John talked about - We interviewed a bunch of years in OpenStack. What's your take on the role of OpenStack today, is there still a roll in OpenStack, and how's that kind of compare/contrast to what we're doing here? >> Happy to answer, because I actually am on both boards, I'm on the CNCF board and I'm on the OpenStack board, and I have contributors on my teams to both efforts across the board. And I think that the role that we're seeing of OpenStack is Openstack is evolving also, and it's becoming more embracive and it's becoming about open infrastructure. And it's really about, how do you create these open infrastructure plays. So it is about virtual machines, and containers, and bare metal, and setting up of those services. So Kubernetes works just great on top of OpenStack, and so now people get to have a choice, because one of the hard things I think for, mostly enterprise developers and everything else, is that the pace is changing so fast. So how do they try out some of the newer technologies that still can be connected back into the existing legacy systems? And that's why I think that we're seeing the role for OpenStack is to make that, you can put it with virtual machines, you can stand them up in there, and you can have the same virtual machines essentially running in the Cloud. >> So virtual machines versus other approaches has come up as a trade off, we heard in the keynote, between cost - I mean, speed, and security. Security's super important. So let me get your thoughts on how that plays out, because we've got the pluggable logger tech, which is another big theme we heard in the keynote, which is essentially just meaning, having a very focused, leverageable piece of code that can be connected into Kubernetes. But with VM's now, some are saying VM's are slow when you're trying to do security, but you want slow, boring when you need it, but you want speed and secure when you need it, too. How do you get both out of that? >> Without being too geeky in terms of, a virtual machine is emulating an entire computer. And so it looks like a computer, so you're running your traditional applications on top of a virtual machine. The same as they would if they were running on what we call, bare metal machine. So that is by necessity, much heavier. You're bringing around a whole operating system and things like that. Containers -- >> And there's a role for that, too. >> There's absolutely a role for that. >> Now containers? >> But containers, then, are really much more about, it's an application packaging exercise, so that you can say, 'I'm going to run this application, I just want all its dependencies packaged up.' I'll assume there's an operating system there. I'm going to count on the fact that there's a single operating system. So you can spin up containers, they're much more lightweight, much more quickly. And now there's even things such as Kata Containers that are coming out of Intel, which is now merging those technologies. >> Male: The clear containers. >> Clear containers, they came originally Clear Containers, and now it's merging, because we're saying, 'we want the security and the protection that you get' 'with a virtual machine, tied into, like the VTX' 'instruction set, in the hardware'. So you can get that level of security, assurances, but now you get the speed of containers. So, I think we're continuing to see the whole community evolving in this direction and making things easier for application developers, faster to do. They're increasing in scale, so management and orchestration - we talked about that three years ago, that that would be a big issue, and guess what? Of course it is. That's exactly what Kubernetes is addressing. >> And the role of the data is going to be critical, this is where a lot of people in the enterprise that we talked to, love the story, they love the narrative, but they're hearing things that they've never heard before and they kind of, slow down. So I'd like you to take a minute, Lew, and explain to the person watching, CIO, chief architect, network guy, whatever - what the hell is this Kubernetes hubbub about? What is Kubernetes, from your perspective? How would you wrap that up and describe the, what it is, and the impact to the customer? >> So, formally it's an orchestration of the container. So what that means is that, when you're developing an application, if you want it to be resilient, you want several instances of that application running, and you want traffic, then, to be low-balanced across it. Kubernetes provides that level of orchestration, to make sure there's always three running. If one fails, it can bring up another one. And it can do that completely automated. So it's a layer that really manages the deployment of containers. As an application developer, you still write your application, you package it up into a container, could be a doc or a container, and then you deploy it using Kubernetes in there. What is interesting, and I think that this is what we've recognized in this last year, I think, is that Kubernetes has a very simple networking model. Which is basically that of having a way to load-balance across multiple containers and keep them running. If you have anything more complicated about different services that you want to talk to from those containers, that may be different places in the universe, we don't have a mechanism for doing that. And everybody was having to write their own. So again, that's where the idea of a service mesh, STF -- >> John: That's where the meshing comes in. >> That's where the mesh ... >> Hundreds and hundreds of services. >> Lynkerd has been doing it for a while, Envoy. >> And Lyft and Uber, they had to do it because they had massive explosion of devices. >> Right, exactly right. And so that's why getting together the code from Lyft and Envoy, adding a control plane to it, which is what Istio really is about, brings that out, too. >> Sounds like an operating system to me, but Lew I one more question for you. You mentioned in, as you described it, Kubernetes, isn't that auto-scaling? If I'm familiar with AWS, isn't that just auto-scaling? Or is it auto-scaling for application instances? Or is auto-scaling more - defined differently? >> It does do the scaling part, it does the resiliency part, but it has a very simple model for that. And that's why you need to have other - but it's a beginning of that orchestration layer. >> Because at the container level, it has all those inherent problems. >> Right. And it can make sure to keep those containers alive and well, and manage the life cycle. >> John: And that's the difference. >> And that's the real difference. Whereas the auto-scaling from Amazon, as a service, is purely a networking capability then tied into bringing up new instances. >> So this is like auto-scaling on steroids. >> It is. But one of the differences also is that Kubernetes and what we're doing here is all open source. So you can run it anywhere. You don't get, a lot of people are very concerned about being locked in to, it used to be, you were locked into Oracle, or to Microsoft, or Java, on premise of things like that. >> Whatever proprietary operating system. >> And now they have concern being locked into these services that are in the public cloud providers. And what we're seeing now with Kubernetes and we're seeing in almost everything around here, by open sourcing them, the advantage is now the enterprise can run the same technology inside, without being locked into a vendor, as they do in the public cloud. >> Lew, so we spent a bunch of time talking about multi-cloud. Some of the more interesting pieces is what's happening at the edge, and IOT. We've heard Cisco talking about it for many years, networking of course important. What's your take, what are you working on, with regards to that these days. >> There's a couple new trends that we've been, IOT is actually now really getting realized, I think, because it is pushing a lot of the computing out to the edge, whether it be in cell phone towers or base stations, retail stores, that kind of edge. At the same time, we're seeing this multi-cloud that we want the big services. If I want to use a machine learning service, I want to use it up in the cloud, and I need to now connect it back to those devices. So multi-cloud is really about, addressing how do you develop applications that run across multiple, in the cloud, on the edge, in an IOT device. There's also, I think you've probably been hearing, server lists, and function as a service. These are, again, a lighter weight way to have kind of an event-driven model, so that if you have an IOT device and it just causes an event, you want to be able to spawn essentially a service, in the cloud, that only runs to process that one event, and then it goes away. So you're not paying to run instances of virtual machines or whatever, sitting there waiting for some event. You get a trigger, and you only pay - so it has this micro-billing capability as a part of it - so that you just can use only the resources. We finally realized the promise that we always had in cloud computing, which is that, pay for only what you need, for what you use. And so this is another way to do that. >> Lew, it's great to have you on theCUBE again, good to see you, great to get the update. I'd like to ask you one more final question to end the segment here. You always have your ear to the ground, reading the tea leaves, you have a unique skill to understand the tech at the root level. What's coming next? If we go back and we have these nice conversations where we're riffing on what's coming out in the next two, three years. It's unclear to some of the visionaries out there, so I got to ask you, what's going to be hot, what do you see emerging? As we saw Kubernetes and discussed, we couldn't have predicted this, I couldn't have. I knew it was going to be hot, I knew it was going to be big, but not this big, changing industry. What do you see out there? What would be the conversation you'd say, 'You know, we've got to watch this,' 'this is going to be a value creation opportunity,' 'enabling technology that's going to make a lot of things' 'flow nicely' - what kind of tech should ... >> Well, it may be a trite answer, 'cause I think a lot of people are seeing the same thing, is that we're actually laying the groundwork here, when we talk about multi-cloud, things that are distributed across multiple things. Accessing different services. I'm still a big believer in, it's going to be in the strength of those services. Whether they be speech-translation services, whether they be recommendation engine, whether it means big data services. Access to those services is what's going to be important. Three or four years from now, we're going to be talking about the intelligence -- >> Without a lot of heavy lifting to integrate it. >> Yes, that's exactly the point. We want it so that somebody can almost visually wire up these things, and take advantage of tremendously powerful machine-learning algorithms. That they don't want to have to hire the machine-learning experts to do it, they want to use that as a service. >> Slinging API, slinging services, wiring things up, sounds like it's an operating system to me. >> It's always an operating system at the end of the day. >> Lew Tucker, Vice President and CTO at Cisco Systems. Industry legend, on the board of CNCF, the fastest-growing organization, where projects equal products equals profit, and of course the OpenStack. Lew, thanks for coming on theCUBE, I'm John Furrier with Stu Miniman, back here live in Austin for more live coverage of CloudNativeCon and KubeCon, after this short break. >> Lew: Thank you.

Published Date : Dec 6 2017

SUMMARY :

Brought to you by Red Hat, the Linux Foundation, And I'm excited to have Cube alumni, and it's always great to have you on because This is how communities operate. communities, get the projects that we work together on it, just want to give you some props for that, you deserve it, And so one of the things I think that that leads to it's a project that's contributed to by Google, and IBM, This data I heard in the keynote, I want to get your so that the application developer doesn't have to -- Through all the architect, the person running it, And for example, a lot of the infrastructure now is Right. This is what we've been talking about since 2010, So all of this has to do with connecting kind of compare/contrast to what we're doing here? OpenStack is to make that, you can put it with boring when you need it, but you want speed and secure And so it looks like a computer, so you're running it's an application packaging exercise, so that you can say, So you can get that level of security, assurances, And the role of the data is going to be critical, So it's a layer that really manages the deployment Lynkerd has been doing it for a while, And Lyft and Uber, they had to do it because they had Envoy, adding a control plane to it, which is what Istio Sounds like an operating system to me, And that's why you need to have other - Because at the container level, it has all those And it can make sure to keep those containers And that's the real difference. But one of the differences also is that that are in the public cloud providers. Some of the more interesting pieces is because it is pushing a lot of the computing out to the Lew, it's great to have you on theCUBE again, I'm still a big believer in, it's going to be in the experts to do it, they want to use that as a service. sounds like it's an operating system to me. and of course the OpenStack.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JohnPERSON

0.99+

John FurrierPERSON

0.99+

UberORGANIZATION

0.99+

IBMORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

Stu MinimanPERSON

0.99+

Red HatORGANIZATION

0.99+

Lew TuckerPERSON

0.99+

LyftORGANIZATION

0.99+

AustinLOCATION

0.99+

AmazonORGANIZATION

0.99+

Linux FoundationORGANIZATION

0.99+

VancouverLOCATION

0.99+

MicrosoftORGANIZATION

0.99+

SiliconANGLE MediaORGANIZATION

0.99+

AWSORGANIZATION

0.99+

CNCFORGANIZATION

0.99+

Austin, TexasLOCATION

0.99+

Cisco SystemsORGANIZATION

0.99+

LewPERSON

0.99+

5%QUANTITY

0.99+

theCUBEORGANIZATION

0.99+

KubeConEVENT

0.99+

last yearDATE

0.99+

eighth yearQUANTITY

0.99+

8 yearsQUANTITY

0.99+

IntelORGANIZATION

0.99+

threeQUANTITY

0.99+

CloudNativeConEVENT

0.99+

EnvoyORGANIZATION

0.98+

oneQUANTITY

0.98+

both boardsQUANTITY

0.98+

2010DATE

0.98+

OracleORGANIZATION

0.98+

three years agoDATE

0.98+

Austin TexasLOCATION

0.98+

bothQUANTITY

0.97+

one eventQUANTITY

0.97+

one thingQUANTITY

0.97+

KubernetesTITLE

0.97+

three and a half years agoDATE

0.96+

CloudNative ConferenceEVENT

0.96+

CubeORGANIZATION

0.96+

JJPERSON

0.96+

OpenStackTITLE

0.96+

Hundreds andQUANTITY

0.95+

three yearsQUANTITY

0.94+

CloudNativeCon 2017EVENT

0.94+

todayDATE

0.94+

KubeCon 2017EVENT

0.94+

Cloud ComputingORGANIZATION

0.93+

version twoOTHER

0.91+

JavaTITLE

0.9+

single operating systemQUANTITY

0.9+

ThreeQUANTITY

0.9+

four yearsQUANTITY

0.9+

LouPERSON

0.9+

OpenStackORGANIZATION

0.89+

Vice PresidentPERSON

0.88+