theCUBE Previews Supercomputing 22
(inspirational music) >> The history of high performance computing is unique and storied. You know, it's generally accepted that the first true supercomputer was shipped in the mid 1960s by Controlled Data Corporations, CDC, designed by an engineering team led by Seymour Cray, the father of Supercomputing. He left CDC in the 70's to start his own company, of course, carrying his own name. Now that company Cray, became the market leader in the 70's and the 80's, and then the decade of the 80's saw attempts to bring new designs, such as massively parallel systems, to reach new heights of performance and efficiency. Supercomputing design was one of the most challenging fields, and a number of really brilliant engineers became kind of quasi-famous in their little industry. In addition to Cray himself, Steve Chen, who worked for Cray, then went out to start his own companies. Danny Hillis, of Thinking Machines. Steve Frank of Kendall Square Research. Steve Wallach tried to build a mini supercomputer at Convex. These new entrants, they all failed, for the most part because the market at the time just wasn't really large enough and the economics of these systems really weren't that attractive. Now, the late 80's and the 90's saw big Japanese companies like NEC and Fujitsu entering the fray and governments around the world began to invest heavily in these systems to solve societal problems and make their nations more competitive. And as we entered the 21st century, we saw the coming of petascale computing, with China actually cracking the top 100 list of high performance computing. And today, we're now entering the exascale era, with systems that can complete a billion, billion calculations per second, or 10 to the 18th power. Astounding. And today, the high performance computing market generates north of $30 billion annually and is growing in the high single digits. Supercomputers solve the world's hardest problems in things like simulation, life sciences, weather, energy exploration, aerospace, astronomy, automotive industries, and many other high value examples. And supercomputers are expensive. You know, the highest performing supercomputers used to cost tens of millions of dollars, maybe $30 million. And we've seen that steadily rise to over $200 million. And today we're even seeing systems that cost more than half a billion dollars, even into the low billions when you include all the surrounding data center infrastructure and cooling required. The US, China, Japan, and EU countries, as well as the UK, are all investing heavily to keep their countries competitive, and no price seems to be too high. Now, there are five mega trends going on in HPC today, in addition to this massive rising cost that we just talked about. One, systems are becoming more distributed and less monolithic. The second is the power of these systems is increasing dramatically, both in terms of processor performance and energy consumption. The x86 today dominates processor shipments, it's going to probably continue to do so. Power has some presence, but ARM is growing very rapidly. Nvidia with GPUs is becoming a major player with AI coming in, we'll talk about that in a minute. And both the EU and China are developing their own processors. We're seeing massive densities with hundreds of thousands of cores that are being liquid-cooled with novel phase change technology. The third big trend is AI, which of course is still in the early stages, but it's being combined with ever larger and massive, massive data sets to attack new problems and accelerate research in dozens of industries. Now, the fourth big trend, HPC in the cloud reached critical mass at the end of the last decade. And all of the major hyperscalers are providing HPE, HPC as a service capability. Now finally, quantum computing is often talked about and predicted to become more stable by the end of the decade and crack new dimensions in computing. The EU has even announced a hybrid QC, with the goal of having a stable system in the second half of this decade, most likely around 2027, 2028. Welcome to theCUBE's preview of SC22, the big supercomputing show which takes place the week of November 13th in Dallas. theCUBE is going to be there. Dave Nicholson will be one of the co-hosts and joins me now to talk about trends in HPC and what to look for at the show. Dave, welcome, good to see you. >> Hey, good to see you too, Dave. >> Oh, you heard my narrative up front Dave. You got a technical background, CTO chops, what did I miss? What are the major trends that you're seeing? >> I don't think you really- You didn't miss anything, I think it's just a question of double-clicking on some of the things that you brought up. You know, if you look back historically, supercomputing was sort of relegated to things like weather prediction and nuclear weapons modeling. And these systems would live in places like Lawrence Livermore Labs or Los Alamos. Today, that requirement for cutting edge, leading edge, highest performing supercompute technology is bleeding into the enterprise, driven by AI and ML, artificial intelligence and machine learning. So when we think about the conversations we're going to have and the coverage we're going to do of the SC22 event, a lot of it is going to be looking under the covers and seeing what kind of architectural things contribute to these capabilities moving forward, and asking a whole bunch of questions. >> Yeah, so there's this sort of theory that the world is moving toward this connectivity beyond compute-centricity to connectivity-centric. We've talked about that, you and I, in the past. Is that a factor in the HPC world? How is it impacting, you know, supercomputing design? >> Well, so if you're designing an island that is, you know, tip of this spear, doesn't have to offer any level of interoperability or compatibility with anything else in the compute world, then connectivity is important simply from a speeds and feeds perspective. You know, lowest latency connectivity between nodes and things like that. But as we sort of democratize supercomputing, to a degree, as it moves from solely the purview of academia into truly ubiquitous architecture leverage by enterprises, you start asking the question, "Hey, wouldn't it be kind of cool if we could have this hooked up into our ethernet networks?" And so, that's a whole interesting subject to explore because with things like RDMA over converged ethernet, you now have the ability to have these supercomputing capabilities directly accessible by enterprise computing. So that level of detail, opening up the box of looking at the Nix, or the storage cards that are in the box, is actually critically important. And as an old-school hardware knuckle-dragger myself, I am super excited to see what the cutting edge holds right now. >> Yeah, when you look at the SC22 website, I mean, they're covering all kinds of different areas. They got, you know, parallel clustered systems, AI, storage, you know, servers, system software, application software, security. I mean, wireless HPC is no longer this niche. It really touches virtually every industry, and most industries anyway, and is really driving new advancements in society and research, solving some of the world's hardest problems. So what are some of the topics that you want to cover at SC22? >> Well, I kind of, I touched on some of them. I really want to ask people questions about this idea of HPC moving from just academia into the enterprise. And the question of, does that mean that there are architectural concerns that people have that might not be the same as the concerns that someone in academia or in a lab environment would have? And by the way, just like, little historical context, I can't help it. I just went through the upgrade from iPhone 12 to iPhone 14. This has got one terabyte of storage in it. One terabyte of storage. In 1997, I helped build a one terabyte NAS system that a government defense contractor purchased for almost $2 million. $2 million! This was, I don't even know, it was $9.99 a month extra on my cell phone bill. We had a team of seven people who were going to manage that one terabyte of storage. So, similarly, when we talk about just where are we from a supercompute resource perspective, if you consider it historically, it's absolutely insane. I'm going to be asking people about, of course, what's going on today, but also the near future. You know, what can we expect? What is the sort of singularity that needs to occur where natural language processing across all of the world's languages exists in a perfect way? You know, do we have the compute power now? What's the interface between software and hardware? But really, this is going to be an opportunity that is a little bit unique in terms of the things that we typically cover, because this is a lot about cracking open the box, the server box, and looking at what's inside and carefully considering all of the components. >> You know, Dave, I'm looking at the exhibitor floor. It's like, everybody is here. NASA, Microsoft, IBM, Dell, Intel, HPE, AWS, all the hyperscale guys, Weka IO, Pure Storage, companies I've never heard of. It's just, hundreds and hundreds of exhibitors, Nvidia, Oracle, Penguin Solutions, I mean, just on and on and on. Google, of course, has a presence there, theCUBE has a major presence. We got a 20 x 20 booth. So, it's really, as I say, to your point, HPC is going mainstream. You know, I think a lot of times, we think of HPC supercomputing as this just sort of, off in the eclectic, far off corner, but it really, when you think about big data, when you think about AI, a lot of the advancements that occur in HPC will trickle through and go mainstream in commercial environments. And I suspect that's why there are so many companies here that are really relevant to the commercial market as well. >> Yeah, this is like the Formula 1 of computing. So if you're a Motorsports nerd, you know that F1 is the pinnacle of the sport. SC22, this is where everybody wants to be. Another little historical reference that comes to mind, there was a time in, I think, the early 2000's when Unisys partnered with Intel and Microsoft to come up with, I think it was the ES7000, which was supposed to be the mainframe, the sort of Intel mainframe. It was an early attempt to use... And I don't say this in a derogatory way, commodity resources to create something really, really powerful. Here we are 20 years later, and we are absolutely smack in the middle of that. You mentioned the focus on x86 architecture, but all of the other components that the silicon manufacturers bring to bear, companies like Broadcom, Nvidia, et al, they're all contributing components to this mix in addition to, of course, the microprocessor folks like AMD and Intel and others. So yeah, this is big-time nerd fest. Lots of academics will still be there. The supercomputing.org, this loose affiliation that's been running these SC events for years. They have a major focus, major hooks into academia. They're bringing in legit computer scientists to this event. This is all cutting edge stuff. >> Yeah. So like you said, it's going to be kind of, a lot of techies there, very technical computing, of course, audience. At the same time, we expect that there's going to be a fair amount, as they say, of crossover. And so, I'm excited to see what the coverage looks like. Yourself, John Furrier, Savannah, I think even Paul Gillin is going to attend the show, because I believe we're going to be there three days. So, you know, we're doing a lot of editorial. Dell is an anchor sponsor, so we really appreciate them providing funding so we can have this community event and bring people on. So, if you are interested- >> Dave, Dave, I just have- Just something on that point. I think that's indicative of where this world is moving when you have Dell so directly involved in something like this, it's an indication that this is moving out of just the realm of academia and moving in the direction of enterprise. Because as we know, they tend to ruthlessly drive down the cost of things. And so I think that's an interesting indication right there. >> Yeah, as do the cloud guys. So again, this is mainstream. So if you're interested, if you got something interesting to talk about, if you have market research, you're an analyst, you're an influencer in this community, you've got technical chops, maybe you've got an interesting startup, you can contact David, david.nicholson@siliconangle.com. John Furrier is john@siliconangle.com. david.vellante@siliconangle.com. I'd be happy to listen to your pitch and see if we can fit you onto the program. So, really excited. It's the week of November 13th. I think November 13th is a Sunday, so I believe David will be broadcasting Tuesday, Wednesday, Thursday. Really excited. Give you the last word here, Dave. >> No, I just, I'm not embarrassed to admit that I'm really, really excited about this. It's cutting edge stuff and I'm really going to be exploring this question of where does it fit in the world of AI and ML? I think that's really going to be the center of what I'm really seeking to understand when I'm there. >> All right, Dave Nicholson. Thanks for your time. theCUBE at SC22. Don't miss it. Go to thecube.net, go to siliconangle.com for all the news. This is Dave Vellante for theCUBE and for Dave Nicholson. Thanks for watching. And we'll see you in Dallas. (inquisitive music)
SUMMARY :
And all of the major What are the major trends on some of the things that you brought up. that the world is moving or the storage cards that are in the box, solving some of the across all of the world's languages a lot of the advancements but all of the other components At the same time, we expect and moving in the direction of enterprise. Yeah, as do the cloud guys. and I'm really going to be go to siliconangle.com for all the news.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Danny Hillis | PERSON | 0.99+ |
Steve Chen | PERSON | 0.99+ |
NEC | ORGANIZATION | 0.99+ |
Fujitsu | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Steve Wallach | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
NASA | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Steve Frank | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Seymour Cray | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Unisys | ORGANIZATION | 0.99+ |
1997 | DATE | 0.99+ |
Savannah | PERSON | 0.99+ |
Dallas | LOCATION | 0.99+ |
EU | ORGANIZATION | 0.99+ |
Controlled Data Corporations | ORGANIZATION | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
Penguin Solutions | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Tuesday | DATE | 0.99+ |
siliconangle.com | OTHER | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
21st century | DATE | 0.99+ |
iPhone 12 | COMMERCIAL_ITEM | 0.99+ |
10 | QUANTITY | 0.99+ |
Cray | PERSON | 0.99+ |
one terabyte | QUANTITY | 0.99+ |
CDC | ORGANIZATION | 0.99+ |
thecube.net | OTHER | 0.99+ |
Lawrence Livermore Labs | ORGANIZATION | 0.99+ |
Broadcom | ORGANIZATION | 0.99+ |
Kendall Square Research | ORGANIZATION | 0.99+ |
iPhone 14 | COMMERCIAL_ITEM | 0.99+ |
john@siliconangle.com | OTHER | 0.99+ |
$2 million | QUANTITY | 0.99+ |
November 13th | DATE | 0.99+ |
first | QUANTITY | 0.99+ |
over $200 million | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
more than half a billion dollars | QUANTITY | 0.99+ |
20 | QUANTITY | 0.99+ |
seven people | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
mid 1960s | DATE | 0.99+ |
three days | QUANTITY | 0.99+ |
Convex | ORGANIZATION | 0.99+ |
70's | DATE | 0.99+ |
SC22 | EVENT | 0.99+ |
david.vellante@siliconangle.com | OTHER | 0.99+ |
late 80's | DATE | 0.98+ |
80's | DATE | 0.98+ |
ES7000 | COMMERCIAL_ITEM | 0.98+ |
today | DATE | 0.98+ |
almost $2 million | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
20 years later | DATE | 0.98+ |
tens of millions of dollars | QUANTITY | 0.98+ |
Sunday | DATE | 0.98+ |
Japanese | OTHER | 0.98+ |
90's | DATE | 0.97+ |
Marcel Hild, Red Hat & Kenneth Hoste, Ghent University | Kubecon + Cloudnativecon Europe 2022
(upbeat music) >> Announcer: theCUBE presents KubeCon and CloudNativeCon Europe 2022, brought to you by Red Hat, the Cloud Native Computing Foundation, and its ecosystem partners. >> Welcome to Valencia, Spain, in KubeCon CloudNativeCon Europe 2022. I'm your host Keith Townsend, along with Paul Gillon. And we're going to talk to some amazing folks. But first Paul, do you remember your college days? >> Vaguely. (Keith laughing) A lot of them are lost. >> I think a lot of mine are lost as well. Well, not really, I got my degree as an adult, so they're not that far past. I can remember 'cause I have the student debt to prove it. (both laughing) Along with us today is Kenneth Hoste, systems administrator at Ghent University, and Marcel Hild, senior manager software engineering at Red Hat. You're working in office of the CTO? >> That's absolutely correct, yes >> So first off, I'm going to start off with you Kenneth. Tell us a little bit about the research that the university does. Like what's the end result? >> Oh, wow, that's a good question. So the research we do at university and again, is very broad. We have bioinformaticians, physicists, people looking at financial data, all kinds of stuff. And the end result can be very varied as well. Very often it's research papers, or spinoffs from the university. Yeah, depending on the domain I would say, it depends a lot on. >> So that sounds like the perfect environment for cloud native. Like the infrastructure that's completely flexible, that researchers can come and have a standard way of interacting, each team just use it's resources as they would, the Navana for cloud native. >> Yeah. >> But somehow, I'm going to guess HPC isn't quite there yet. >> Yeah, not really, no. So, HPC is a bit, let's say slow into adopting new technologies. And we're definitely seeing some impact from cloud, especially things like containers and Kubernetes, or we're starting to hear these things in HPC community as well. But I haven't seen a lot of HPC clusters who are really fully cloud native. Not yet at least. Maybe this is coming. And if I'm walking around here at KubeCon, I can definitely, I'm being convinced that it's coming. So whether we like it or not we're probably going to have to start worrying about stuff like this. But we're still, let's say, the most prominent technologies of things like NPI, which has been there for 20, 30 years. The Fortran programming language is still the main language, if you're looking at compute time being spent on supercomputers, over 1/2 of the time spent is in Fortran code essentially. >> Keith: Wow. >> So either the application itself where the simulations are being done is implemented in Fortran, or the libraries that we are talking to from Python for example, for doing heavy duty computations, that backend library is implemented in Fortran. So if you take all of that into account, easily over 1/2 of the time is spent in Fortran code. >> So is this because the libraries don't migrate easily to, distributed to that environment? >> Well, it's multiple things. So first of all, Fortran is very well suited for implementing these type of things. >> Paul: Right. >> We haven't really seen a better alternative maybe. And also it'll be a huge effort to re-implement that same functionality in a newer language. So, the use case has to be very convincing, there has to be a very good reason why you would move away from Fortran. And, at least the HPC community hasn't seen that reason yet. >> So in theory, and right now we're talking about the theory and then what it takes to get to the future. In theory, I can take that Fortran code put it in a compiler that runs in a container? >> Yeah, of course, yeah. >> Why isn't it that simple? >> I guess because traditionally HPC is very slow at adopting new stuff. So, I'm not saying there isn't a reason that we should start looking at these things. Flexibility is a very important one. For a lot of researchers, their compute needs are very picky. So they're doing research, they have an idea, they want you to run lots of simulations, get the results, but then they're silent for a long time writing the paper, or thinking about how to, what they can learn from the results. So there's lots of peaks, and that's a very good fit for a cloud environment. I guess at the scale of university you have enough diversity end users that all those peaks never fall at the same time. So if you have your big own infrastructure you can still fill it up quite easily and keep your users happy. But this busty thing, I guess we're seeing that more and more or so. >> So Marcel, talk to us about, Red Hat needing to service these types of end users. That it can be on both ends I'd imagine that you have some people still in writing in Fortran, you have some people that's asking you for objects based storage. Where's Fortran, I'm sorry, not Fortran, but where is Red Hat in providing the underlay and the capabilities for the HPC and AI community? >> Yeah. So, I think if you look at the user base that we're looking at, it's on this spectrum from development to production. So putting AI workloads into production, it's an interesting challenge but it's easier to solve, and it has been solved to some extent, than the development cycle. So what we're looking at in Kenneth's domain it's more like the end user, the data scientist, developing code, and doing these experiments. Putting them into production is that's where containers live and thrive. You can containerize your model, you containerize your workload, you deploy it into your OpenShift Kubernetes cluster, done, you monitor it, done. So the software developments and the SRE, the ops part, done, but how do I get the data scientist into this cloud native age where he's not developing on his laptop or on a machine, where he SSH into and then does some stuff there. And then some system admin comes and needs to tweak it because it's running out of memory or whatnot. But how do we take him and make him, well, and provide him an environment that is good enough to work in, in the browser, and then with IDE, where the workload of doing the computation and the experimentation is repeatable, so that the environment is always the same, it's reliable, so it's always up and running. It doesn't consume resources, although it's up and running. Where it's, where the supply chain and the configuration of... And the, well, the modules that are brought into the system are also reliable. So all these problems that we solved in the traditional software development world, now have to transition into the data science and HPC world, where the problems are similar, but yeah, it's different sets. It's more or less, also a huge educational problem and transitioning the tools over into that is something... >> Well, is this mostly a technical issue or is this a cultural issue? I mean, are HPC workloads that different from more conventional OLTP workloads that they would not adapt well to a distributed containerized environment? >> I think it's both. So, on one hand it's the cultural issue because you have two different communities, everybody is reinventing the wheel, everybody is some sort of siloed. So they think, okay, what we've done for 30 years now we, there's no need to change it. And they, so it's, that's what thrives and here at KubeCon where you have different communities coming together, okay, this is how you solved the problem, maybe this applies also to our problem. But it's also the, well, the tooling, which is bound to a machine, which is bound to an HPC computer, which is architecturally different than a distributed environment where you would treat your containers as kettle, and as something that you can replace, right? And the HPC community usually builds up huge machines, and these are like the gray machines. So it's also technical bit of moving it to this age. >> So the massively parallel nature of HPC workloads you're saying Kubernetes has not yet been adapted to that? >> Well, I think that parallelism works great. It's just a matter of moving that out from an HPC computer into the scale out factor of a Kubernetes cloud that elastically scales out. Whereas the traditional HPC computer, I think, and Kenneth can correct me here is, more like, I have this massive computer with 1 million cores or whatnot, and now use it. And I can use my time slice, and book my time slice there. Whereas this a Kubernetes example the concept is more like, I have 1000 cores and I declare something into it and scale it up and down based on the needs. >> So, Kenneth, this is where you talked about the culture part of the changes that need to be happening. And quite frankly, the computer is a tool, it's a tool to get to the answer. And if that tool is working, if I have a 1000 cores on a single HPC thing, and you're telling me, well, I can't get to a system with 2000 cores. And if you containerized your process and move it over then maybe I'll get to the answer 50% faster maybe I'm not that... Someone has to make that decision. How important is it to get people involved in these types of communities from a researcher? 'Cause research is very tight-knit community to have these conversations and help that see move happen. >> I think it's very important to that community should, let's say, the cloud community, HPC research community, they should be talking a lot more, there should be way more cross pollination than there is today. I'm actually, I'm happy that I've seen HPC mentioned at booths and talks quite often here at KubeCon, I wasn't really expecting that. And I'm not sure, it's my first KubeCon, so I don't know, but I think that's kind of new, it's pretty recent. If you're going to the HPC community conferences there containers have been there for a couple of years now, something like Kubernetes is still a bit new. But just this morning there was a keynote by a guy from CERN, who was explaining, they're basically slowly moving towards Kubernetes even for their HPC clusters as well. And he's seeing that as the future because all the flexibility it gives you and you can basically hide all that from the end user, from the researcher. They don't really have to know that they're running on top of Kubernetes. They shouldn't care. Like you said, to them it's just a tool, and they care about if the tool works, they can get their answers and that's what they want to do. How that's actually being done in the background they don't really care. >> So talk to me about the AI side of the equation, because when I talk to people doing AI, they're on the other end of the spectrum. What are some of the benefits they're seeing from containerization? >> I think it's the reproducibility of experiments. So, and data scientists are, they're data scientists and they do research. So they care about their experiment. And maybe they also care about putting the model into production. But, I think from a geeky perspective they are more interested in finding the next model, finding the next solution. So they do an experiment, and they're done with it, and then maybe it's going to production. So how do I repeat that experiment in a year from now, so that I can build on top of it? And a container I think is the best solution to wrap something with its dependency, like freeze it, maybe even with the data, store it away, and then come to it back later and redo the experiment or share the experiment with some of my fellow researchers, so that they don't have to go through the process of setting up an equivalent environment on their machines, be it their laptop, via their cloud environment. So you go to the internet, download something doesn't work, container works. >> Well, you said something that really intrigues me you know in concept, I can have a, let's say a one terabyte data set, have a experiment associated with that. Take a snapshot of that somehow, I don't know how, take a snapshot of that and then share it with the rest of the community and then continue my work. >> Marcel: Yeah. >> And then we can stop back and compare notes. Where are we at in a maturity scale? Like, what are some of the pitfalls or challenges customers should be looking out for? >> I think you actually said it right there, how do I snapshot a terabyte of data? It's, that's... >> It's a terabyte of data. (both conversing) >> It's a bit of a challenge. And if you snapshot it, you have two terabytes of data or you just snapshot the, like and get you to do a, okay, this is currently where we're at. So that's why the technology is evolving. How do we do source control management for data? How do we license data? How do we make sure that the data is unbiased, et cetera? So that's going more into the AI side of things. But at dealing with data in a declarative way in a containerized way, I think that's where currently a lot of innovation is happening. >> What do you mean by dealing with data in a declarative way? >> If I'm saying I run this experiment based on this data set and I'm running this other experiment based on this other data set, and I as the researcher don't care where the data is stored, I care that the data is accessible. And so I might declare, this is the process that I put on my data, like a data processing pipeline. These are the steps that it's going through. And eventually it will have gone through this process and I can work with my data. Pretty much like applying the concept of pipelines through data. Like you have these data pipelines and then now you have cube flow pipelines as one solution to apply the pipeline concept, to well, managing your data. >> Given the stateless nature of containers, is that an impediment to HPC adoption because of the very large data sets that are typically involved? >> I think it is if you have terabytes of data. Just, you have to get it to the place where the computation will happen, right? And just uploading that into the cloud is already a challenge. If you have the data sitting there on a supercomputer and maybe it was sitting there for two years, you probably don't care. And typically a lot of universities the researchers don't necessarily pay for the compute time they use. Like, this is also... At least in Ghent that's the case, it's centrally funded, which means, the researchers don't have to worry about the cost, they just get access to the supercomputer. If they need two terabytes of data, they get that space and they can park it on the system for years, no problem. If they need 200 terabytes of data, that's absolutely fine. >> But the university cares about the cost? >> The university cares about the cost, but they want to enable the researchers to do the research that they want to do. >> Right. >> And we always tell researchers don't feel constrained about things like compute power, storage space. If you're doing smaller research, because you're feeling constrained, you have to tell us, and we will just expand our storage system and buy a new cluster. >> Paul: Wonderful. >> So you, to enable your research. >> It's a nice environment to be in. I think this might be a Jevons paradox problem, you give researchers this capability you might, you're going to see some amazing things. Well, now the people are snapshoting, one, two, three, four, five, different versions of a one terabytes of data. It's a good problem to have, and I hope to have you back on theCUBE, talking about how Red Hat and Ghent have solved those problems. Thank you so much for joining theCUBE. From Valencia, Spain, I'm Keith Townsend along with Paul Gillon. And you're watching theCUBE, the leader in high tech coverage. (upbeat music)
SUMMARY :
brought to you by Red Hat, do you remember your college days? A lot of them are lost. the student debt to prove it. that the university does. So the research we do at university Like the infrastructure I'm going to guess HPC is still the main language, So either the application itself So first of all, So, the use case has talking about the theory I guess at the scale of university and the capabilities for and the experimentation is repeatable, And the HPC community usually down based on the needs. And quite frankly, the computer is a tool, And he's seeing that as the future What are some of the and redo the experiment the rest of the community And then we can stop I think you actually It's a terabyte of data. the AI side of things. I care that the data is accessible. for the compute time they use. to do the research that they want to do. and we will just expand our storage system and I hope to have you back on theCUBE,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul Gillon | PERSON | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
Kenneth | PERSON | 0.99+ |
Kenneth Hoste | PERSON | 0.99+ |
Marcel Hild | PERSON | 0.99+ |
Paul | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
two years | QUANTITY | 0.99+ |
Keith | PERSON | 0.99+ |
Marcel | PERSON | 0.99+ |
1 million cores | QUANTITY | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
50% | QUANTITY | 0.99+ |
20 | QUANTITY | 0.99+ |
Fortran | TITLE | 0.99+ |
1000 cores | QUANTITY | 0.99+ |
30 years | QUANTITY | 0.99+ |
two terabytes | QUANTITY | 0.99+ |
CERN | ORGANIZATION | 0.99+ |
2000 cores | QUANTITY | 0.99+ |
Ghent | LOCATION | 0.99+ |
Valencia, Spain | LOCATION | 0.99+ |
first | QUANTITY | 0.99+ |
Ghent | ORGANIZATION | 0.99+ |
one terabytes | QUANTITY | 0.99+ |
each team | QUANTITY | 0.99+ |
one solution | QUANTITY | 0.99+ |
KubeCon | EVENT | 0.99+ |
today | DATE | 0.99+ |
one terabyte | QUANTITY | 0.99+ |
Python | TITLE | 0.99+ |
Ghent University | ORGANIZATION | 0.99+ |
Kubernetes | TITLE | 0.98+ |
both | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
HPC | ORGANIZATION | 0.98+ |
two different communities | QUANTITY | 0.96+ |
terabytes of data | QUANTITY | 0.96+ |
both ends | QUANTITY | 0.96+ |
over 1/2 | QUANTITY | 0.93+ |
two | QUANTITY | 0.93+ |
Cloudnativecon | ORGANIZATION | 0.93+ |
CloudNativeCon Europe 2022 | EVENT | 0.92+ |
this morning | DATE | 0.92+ |
a year | QUANTITY | 0.91+ |
five | QUANTITY | 0.9+ |
theCUBE | ORGANIZATION | 0.89+ |
Fortran | ORGANIZATION | 0.88+ |
KubeCon | ORGANIZATION | 0.87+ |
two terabytes of data | QUANTITY | 0.86+ |
KubeCon CloudNativeCon Europe 2022 | EVENT | 0.86+ |
Europe | LOCATION | 0.85+ |
years | QUANTITY | 0.81+ |
a terabyte of data | QUANTITY | 0.8+ |
Navana | ORGANIZATION | 0.8+ |
200 terabytes of | QUANTITY | 0.79+ |
Kubecon + | ORGANIZATION | 0.77+ |
The University of Edinburgh and Rolls Royce Drive in Exascale Style | Exascale Day
>>welcome. My name is Ben Bennett. I am the director of HPC Strategic programs here at Hewlett Packard Enterprise. It is my great pleasure and honor to be talking to Professor Mark Parsons from the Edinburgh Parallel Computing Center. And we're gonna talk a little about exa scale. What? It means we're gonna talk less about the technology on Maura about the science, the requirements on the need for exa scale. Uh, rather than a deep dive into the enabling technologies. Mark. Welcome. >>I then thanks very much for inviting me to tell me >>complete pleasure. Um, so I'd like to kick off with, I suppose. Quite an interesting look back. You and I are both of a certain age 25 plus, Onda. We've seen these milestones. Uh, I suppose that the S I milestones of high performance computing's come and go, you know, from a gig a flop back in 1987 teraflop in 97 a petaflop in 2000 and eight. But we seem to be taking longer in getting to an ex a flop. Um, so I'd like your thoughts. Why is why is an extra flop taking so long? >>So I think that's a very interesting question because I started my career in parallel computing in 1989. I'm gonna join in. IPCC was set up then. You know, we're 30 years old this year in 1990 on Do you know the fastest computer we have them is 800 mega flops just under a getting flogged. So in my career, we've gone already. When we reached the better scale, we'd already gone pretty much a million times faster on, you know, the step from a tariff block to a block scale system really didn't feel particularly difficult. Um, on yet the step from A from a petaflop PETA scale system. To an extent, block is a really, really big challenge. And I think it's really actually related to what's happened with computer processes over the last decade, where, individually, you know, approached the core, Like on your laptop. Whoever hasn't got much faster, we've just got more often So the perception of more speed, but actually just being delivered by more course. And as you go down that approach, you know what happens in the supercomputing world as well. We've gone, uh, in 2010 I think we had systems that were, you know, a few 1000 cores. Our main national service in the UK for the last eight years has had 118,000 cores. But looking at the X scale we're looking at, you know, four or five million cores on taming that level of parallelism is the real challenge. And that's why it's taking an enormous and time to, uh, deliver these systems. That is not just on the hardware front. You know, vendors like HP have to deliver world beating technology and it's hard, hard. But then there's also the challenge to the users. How do they get the codes to work in the face of that much parallelism? >>If you look at what the the complexity is delivering an annex a flop. Andi, you could have bought an extra flop three or four years ago. You couldn't have housed it. You couldn't have powered it. You couldn't have afforded it on, do you? Couldn't program it. But you still you could have You could have bought one. We should have been so lucky to be unable to supply it. Um, the software, um I think from our standpoint, is is looking like where we're doing mawr enabling with our customers. You sell them a machine on, then the the need then to do collaboration specifically seems mawr and Maura around the software. Um, so it's It's gonna be relatively easy to get one x a flop using limb pack, but but that's not extra scale. So what do you think? On exa scale machine versus an X? A flop machine means to the people like yourself to your users, the scientists and industry. What is an ex? A flop versus >>an exa scale? So I think, you know, supercomputing moves forward by setting itself challenges. And when you when you look at all of the excess scale programs worldwide that are trying to deliver systems that can do an X a lot form or it's actually very arbitrary challenge. You know, we set ourselves a PETA scale challenge delivering a petaflop somebody manage that, Andi. But you know, the world moves forward by setting itself challenges e think you know, we use quite arbitrary definition of what we mean is well by an exit block. So, you know, in your in my world, um, we either way, first of all, see ah flop is a computation, so multiply or it's an ad or whatever on we tend. Thio, look at that is using very high precision numbers or 64 bit numbers on Do you know, we then say, Well, you've got to do the next block. You've got to do a billion billion of those calculations every second. No, a some of the last arbitrary target Now you know today from HPD Aiken by my assistant and will do a billion billion calculations per second. And they will either do that as a theoretical peak, which would be almost unattainable, or using benchmarks that stressed the system on demonstrate a relaxing law. But again, those benchmarks themselves attuned Thio. Just do those calculations and deliver and explore been a steady I'll way if you like. So, you know, way kind of set ourselves this this this big challenge You know, the big fence on the race course, which were clambering over. But the challenge in itself actually should be. I'm much more interesting. The water we're going to use these devices for having built um, eso. Getting into the extra scale era is not so much about doing an extra block. It's a new generation off capability that allows us to do better scientific and industrial research. And that's the interesting bit in this whole story. >>I would tend to agree with you. I think the the focus around exa scale is to look at, you know, new technologies, new ways of doing things, new ways of looking at data and to get new results. So eventually you will get yourself a nexus scale machine. Um, one hopes, sooner rather >>than later. Well, I'm sure you don't tell me one, Ben. >>It's got nothing to do with may. I can't sell you anything, Mark. But there are people outside the door over there who would love to sell you one. Yes. However, if we if you look at your you know your your exa scale machine, Um, how do you believe the workloads are going to be different on an extra scale machine versus your current PETA scale machine? >>So I think there's always a slight conceit when you buy a new national supercomputer. On that conceit is that you're buying a capability that you know on. But many people will run on the whole system. Known truth. We do have people that run on the whole of our archer system. Today's A 118,000 cores, but I would say, and I'm looking at the system. People that run over say, half of that can be counted on Europe on a single hand in a year, and they're doing very specific things. It's very costly simulation they're running on. So, you know, if you look at these systems today, two things show no one is. It's very difficult to get time on them. The Baroque application procedures All of the requirements have to be assessed by your peers and your given quite limited amount of time that you have to eke out to do science. Andi people tend to run their applications in the sweet spot where their application delivers the best performance on You know, we try to push our users over time. Thio use reasonably sized jobs. I think our average job says about 20,000 course, she's not bad, but that does mean that as we move to the exits, kill two things have to happen. One is actually I think we've got to be more relaxed about giving people access to the system, So let's give more people access, let people play, let people try out ideas they've never tried out before. And I think that will lead to a lot more innovation and computational science. But at the same time, I think we also need to be less precious. You know, we to accept these systems will have a variety of sizes of job on them. You know, we're still gonna have people that want to run four million cores or two million cores. That's absolutely fine. Absolutely. Salute those people for trying really, really difficult. But then we're gonna have a huge spectrum of views all the way down to people that want to run on 500 cores or whatever. So I think we need Thio broaden the user base in Alexa Skill system. And I know this is what's happening, for example, in Japan with the new Japanese system. >>So, Mark, if you cast your mind back to almost exactly a year ago after the HPC user forum, you were interviewed for Premier Magazine on Do you alluded in that article to the needs off scientific industrial users requiring, you know, uh on X a flop or an exa scale machine it's clear in your in your previous answer regarding, you know, the workloads. Some would say that the majority of people would be happier with, say, 10 100 petaflop machines. You know, democratization. More people access. But can you provide us examples at the type of science? The needs of industrial users that actually do require those resources to be put >>together as an exa scale machine? So I think you know, it's a very interesting area. At the end of the day, these systems air bought because they are capability systems on. I absolutely take the argument. Why shouldn't we buy 10 100 pattern block systems? But there are a number of scientific areas even today that would benefit from a nexus school system and on these the sort of scientific areas that will use as much access onto a system as much time and as much scale of the system as they can, as you can give them eso on immediate example. People doing chroma dynamics calculations in particle physics, theoretical calculations, they would just use whatever you give them. But you know, I think one of the areas that is very interesting is actually the engineering space where, you know, many people worry the engineering applications over the last decade haven't really kept up with this sort of supercomputers that we have. I'm leading a project called Asimov, funded by M. P S O. C in the UK, which is jointly with Rolls Royce, jointly funded by Rolls Royce and also working with the University of Cambridge, Oxford, Bristol, Warrick. We're trying to do the whole engine gas turbine simulation for the first time. So that's looking at the structure of the gas turbine, the airplane engine, the structure of it, how it's all built it together, looking at the fluid dynamics off the air and the hot gasses, the flu threat, looking at the combustion of the engine looking how fuel is spread into the combustion chamber. Looking at the electrics around, looking at the way the engine two forms is, it heats up and cools down all of that. Now Rolls Royce wants to do that for 20 years. Andi, Uh, whenever they certify, a new engine has to go through a number of physical tests, and every time they do on those tests, it could cost them as much as 25 to $30 million. These are very expensive tests, particularly when they do what's called a blade off test, which would be, you know, blade failure. They could prove that the engine contains the fragments of the blade. Sort of think, continue face really important test and all engines and pass it. What we want to do is do is use an exa scale computer to properly model a blade off test for the first time, so that in future, some simulations can become virtual rather than having thio expend all of the money that Rolls Royce would normally spend on. You know, it's a fascinating project is a really hard project to do. One of the things that I do is I am deaf to share this year. Gordon Bell Price on bond I've really enjoyed to do. That's one of the major prizes in our area, you know, gets announced supercomputing every year. So I have the pleasure of reading all the submissions each year. I what's been really interesting thing? This is my third year doing being on the committee on what's really interesting is the way that big systems like Summit, for example, in the US have pushed the user communities to try and do simulations Nowhere. Nobody's done before, you know. And we've seen this as well, with papers coming after the first use of the for Goku system in Japan, for example, people you know, these are very, very broad. So, you know, earthquake simulation, a large Eddie simulations of boats. You know, a number of things around Genome Wide Association studies, for example. So the use of these computers spans of last area off computational science. I think the really really important thing about these systems is their challenging people that do calculations they've never done before. That's what's important. >>Okay, Thank you. You talked about challenges when I nearly said when you and I had lots of hair, but that's probably much more true of May. Um, we used to talk about grand challenges we talked about, especially around the teraflop era, the ski red program driving, you know, the grand challenges of science, possibly to hide the fact that it was a bomb designing computer eso they talked about the grand challenges. Um, we don't seem to talk about that much. We talk about excess girl. We talk about data. Um Where are the grand challenges that you see that an exa scale computer can you know it can help us. Okay, >>so I think grand challenges didn't go away. Just the phrase went out of fashion. Um, that's like my hair. I think it's interesting. The I do feel the science moves forward by setting itself grand challenges and always had has done, you know, my original backgrounds in particle physics. I was very lucky to spend four years at CERN working in the early stage of the left accelerator when it first came online on. Do you know the scientists there? I think they worked on left 15 years before I came in and did my little ph d on it. Andi, I think that way of organizing science hasn't changed. We just talked less about grand challenges. I think you know what I've seen over the last few years is a renaissance in computational science, looking at things that have previously, you know, people have said have been impossible. So a couple of years ago, for example, one of the key Gordon Bell price papers was on Genome Wide Association studies on some of it. If I may be one of the winner of its, if I remember right on. But that was really, really interesting because first of all, you know, the sort of the Genome Wide Association Studies had gone out of favor in the bioinformatics by a scientist community because people thought they weren't possible to compute. But that particular paper should Yes, you could do these really, really big Continental little problems in a reasonable amount of time if you had a big enough computer. And one thing I felt all the way through my career actually is we've probably discarded Mawr simulations because they were impossible at the time that we've actually decided to do. And I sometimes think we to challenge ourselves by looking at the things we've discovered in the past and say, Oh, look, you know, we could actually do that now, Andi, I think part of the the challenge of bringing an extra service toe life is to get people to think about what they would use it for. That's a key thing. Otherwise, I always say, a computer that is unused to just be turned off. There's no point in having underutilized supercomputer. Everybody loses from that. >>So Let's let's bring ourselves slightly more up to date. We're in the middle of a global pandemic. Uh, on board one of the things in our industry has bean that I've been particularly proud about is I've seen the vendors, all the vendors, you know, offering up machine's onboard, uh, making resources available for people to fight things current disease. Um, how do you see supercomputers now and in the future? Speeding up things like vaccine discovery on help when helping doctors generally. >>So I think you're quite right that, you know, the supercomputer community around the world actually did a really good job of responding to over 19. Inasmuch as you know, speaking for the UK, we put in place a rapid access program. So anybody wanted to do covert research on the various national services we have done to the to two services Could get really quick access. Um, on that, that has worked really well in the UK You know, we didn't have an archer is an old system, Aziz. You know, we didn't have the world's largest supercomputer, but it is happily bean running lots off covert 19 simulations largely for the biomedical community. Looking at Druk modeling and molecular modeling. Largely that's just been going the US They've been doing really large uh, combinatorial parameter search problems on on Summit, for example, looking to see whether or not old drugs could be reused to solve a new problem on DSO, I think, I think actually, in some respects Kobe, 19 is being the sounds wrong. But it's actually been good for supercomputing. Inasmuch is pointed out to governments that supercomputers are important parts off any scientific, the active countries research infrastructure. >>So, um, I'll finish up and tap into your inner geek. Um, there's a lot of technologies that are being banded around to currently enable, you know, the first exa scale machine, wherever that's going to be from whomever, what are the current technologies or emerging technologies that you are interested in excited about looking forward to getting your hands on. >>So in the business case I've written for the U. K's exa scale computer, I actually characterized this is a choice between the American model in the Japanese model. Okay, both of frozen, both of condoms. Eso in America, they're very much gone down the chorus plus GPU or GPU fruit. Um, so you might have, you know, an Intel Xeon or an M D process er center or unarmed process or, for that matter on you might have, you know, 24 g. P. U s. I think the most interesting thing that I've seen is definitely this move to a single address space. So the data that you have will be accessible, but the G p u on the CPU, I think you know, that's really bean. One of the key things that stopped the uptake of GPS today and that that that one single change is going Thio, I think, uh, make things very, very interesting. But I'm not entirely convinced that the CPU GPU model because I think that it's very difficult to get all the all the performance set of the GPU. You know, it will do well in H p l, for example, high performance impact benchmark we're discussing at the beginning of this interview. But in riel scientific workloads, you know, you still find it difficult to find all the performance that has promised. So, you know, the Japanese approach, which is the core, is only approach. E think it's very attractive, inasmuch as you know They're using very high bandwidth memory, very interesting process of which they are going to have to, you know, which they could develop together over 10 year period. And this is one thing that people don't realize the Japanese program and the American Mexico program has been working for 10 years on these systems. I think the Japanese process really interesting because, um, it when you look at the performance, it really does work for their scientific work clothes, and that's that does interest me a lot. This this combination of a A process are designed to do good science, high bandwidth memory and a real understanding of how data flows around the supercomputer. I think those are the things are exciting me at the moment. Obviously, you know, there's new networking technologies, I think, in the fullness of time, not necessarily for the first systems. You know, over the next decade we're going to see much, much more activity on silicon photonics. I think that's really, really fascinating all of these things. I think in some respects the last decade has just bean quite incremental improvements. But I think we're supercomputing is going in the moment. We're a very very disruptive moment again. That goes back to start this discussion. Why is extra skill been difficult to get? Thio? Actually, because the disruptive moment in technology. >>Professor Parsons, thank you very much for your time and your insights. Thank you. Pleasure and folks. Thank you for watching. I hope you've learned something, or at least enjoyed it. With that, I would ask you to stay safe and goodbye.
SUMMARY :
I am the director of HPC Strategic programs I suppose that the S I milestones of high performance computing's come and go, But looking at the X scale we're looking at, you know, four or five million cores on taming But you still you could have You could have bought one. challenges e think you know, we use quite arbitrary focus around exa scale is to look at, you know, new technologies, Well, I'm sure you don't tell me one, Ben. outside the door over there who would love to sell you one. So I think there's always a slight conceit when you buy a you know, the workloads. That's one of the major prizes in our area, you know, gets announced you know, the grand challenges of science, possibly to hide I think you know what I've seen over the last few years is a renaissance about is I've seen the vendors, all the vendors, you know, Inasmuch as you know, speaking for the UK, we put in place a rapid to currently enable, you know, I think you know, that's really bean. Professor Parsons, thank you very much for your time and your insights.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ben Bennett | PERSON | 0.99+ |
1989 | DATE | 0.99+ |
Rolls Royce | ORGANIZATION | 0.99+ |
UK | LOCATION | 0.99+ |
500 cores | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
20 years | QUANTITY | 0.99+ |
Japan | LOCATION | 0.99+ |
Parsons | PERSON | 0.99+ |
1990 | DATE | 0.99+ |
Mark | PERSON | 0.99+ |
2010 | DATE | 0.99+ |
1987 | DATE | 0.99+ |
HP | ORGANIZATION | 0.99+ |
118,000 cores | QUANTITY | 0.99+ |
first time | QUANTITY | 0.99+ |
four years | QUANTITY | 0.99+ |
America | LOCATION | 0.99+ |
CERN | ORGANIZATION | 0.99+ |
third year | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
30 years | QUANTITY | 0.99+ |
2000 | DATE | 0.99+ |
four million cores | QUANTITY | 0.99+ |
two million cores | QUANTITY | 0.99+ |
Genome Wide Association | ORGANIZATION | 0.99+ |
two services | QUANTITY | 0.99+ |
Ben | PERSON | 0.99+ |
first systems | QUANTITY | 0.99+ |
two forms | QUANTITY | 0.99+ |
US | LOCATION | 0.99+ |
both | QUANTITY | 0.99+ |
IPCC | ORGANIZATION | 0.99+ |
three | DATE | 0.99+ |
today | DATE | 0.98+ |
Hewlett Packard Enterprise | ORGANIZATION | 0.98+ |
University of Cambridge | ORGANIZATION | 0.98+ |
five million cores | QUANTITY | 0.98+ |
a year ago | DATE | 0.98+ |
single | QUANTITY | 0.98+ |
Mark Parsons | PERSON | 0.98+ |
two things | QUANTITY | 0.98+ |
$30 million | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
Edinburgh Parallel Computing Center | ORGANIZATION | 0.98+ |
Aziz | PERSON | 0.98+ |
Gordon Bell | PERSON | 0.98+ |
May | DATE | 0.98+ |
64 bit | QUANTITY | 0.98+ |
Europe | LOCATION | 0.98+ |
One | QUANTITY | 0.97+ |
each year | QUANTITY | 0.97+ |
about 20,000 course | QUANTITY | 0.97+ |
Today | DATE | 0.97+ |
Alexa | TITLE | 0.97+ |
this year | DATE | 0.97+ |
HPC | ORGANIZATION | 0.96+ |
Intel | ORGANIZATION | 0.96+ |
Xeon | COMMERCIAL_ITEM | 0.95+ |
25 | QUANTITY | 0.95+ |
over 10 year | QUANTITY | 0.95+ |
1000 cores | QUANTITY | 0.95+ |
Thio | PERSON | 0.95+ |
800 mega flops | QUANTITY | 0.95+ |
Professor | PERSON | 0.95+ |
Andi | PERSON | 0.94+ |
one thing | QUANTITY | 0.94+ |
couple of years ago | DATE | 0.94+ |
over 19 | QUANTITY | 0.93+ |
U. K | LOCATION | 0.92+ |
Premier Magazine | TITLE | 0.92+ |
10 100 petaflop machines | QUANTITY | 0.91+ |
four years ago | DATE | 0.91+ |
Exascale | LOCATION | 0.91+ |
HPD Aiken | ORGANIZATION | 0.91+ |
Reliance Jio: OpenStack for Mobile Telecom Services
>>Hi, everyone. My name is my uncle. My uncle Poor I worked with Geo reminds you in India. We call ourselves Geo Platforms. Now on. We've been recently in the news. You've raised a lot off funding from one of the largest, most of the largest tech companies in the world. And I'm here to talk about Geos Cloud Journey, Onda Mantis Partnership. I've titled it the story often, Underdog becoming the largest telecom company in India within four years, which is really special. And we're, of course, held by the cloud. So quick disclaimer. Right. The content shared here is only for informational purposes. Um, it's only for this event. And if you want to share it outside, especially on social media platforms, we need permission from Geo Platforms limited. Okay, quick intro about myself. I am a VP of engineering a geo. I lead the Cloud Services and Platforms team with NGO Andi. I mean the geo since the beginning, since it started, and I've seen our cloud footprint grow from a handful of their models to now eight large application data centers across three regions in India. And we'll talk about how we went here. All right, Let's give you an introduction on Geo, right? Giorgio is on how we became the largest telecom campaign, India within four years from 0 to 400 million subscribers. And I think there are There are a lot of events that defined Geo and that will give you an understanding off. How do you things and what you did to overcome massive problems in India. So the slide that I want to talkto is this one and, uh, I The headline I've given is, It's the Geo is the fastest growing tech company in the world, which is not a new understatement. It's eggs, actually, quite literally true, because very few companies in the world have grown from zero to 400 million subscribers within four years paying subscribers. And I consider Geo Geos growth in three phases, which I have shown on top. The first phase we'll talk about is how geo grew in the smartphone market in India, right? And what we did to, um to really disrupt the telecom space in India in that market. Then we'll talk about the feature phone phase in India and how Geo grew there in the future for market in India. and then we'll talk about what we're doing now, which we call the Geo Platforms phase. Right. So Geo is a default four g lt. Network. Right. So there's no to geo three g networks that Joe has, Um it's a state of the art four g lt voiceover lt Network and because it was designed fresh right without any two D and three G um, legacy technologies, there were also a lot of challenges Lawn geo when we were starting up. One of the main challenges waas that all the smart phones being sold in India NGOs launching right in 2000 and 16. They did not have the voice or lt chip set embedded in the smartphone because the chips it's far costlier to embed in smartphones and India is a very price and central market. So none of the manufacturers were embedding the four g will teach upset in the smartphones. But geos are on Lee a volte in network, right for the all the network. So we faced a massive problem where we said, Look there no smartphones that can support geo. So how will we grow Geo? So in order to solve that problem, we launched our own brand of smartphones called the Life um, smartphones. And those phones were really high value devices. So there were $50 and for $50 you get you You At that time, you got a four g B storage space. A nice big display for inch display. Dual cameras, Andi. Most importantly, they had volte chip sets embedded in them. Right? And that got us our initial customers the initial for the launch customers when we launched. But more importantly, what that enabled other oh, EMS. What that forced the audience to do is that they also had to launch similar smartphones competing smartphones with voltage upset embedded in the same price range. Right. So within a few months, 3 to 4 months, um, all the other way EMS, all the other smartphone manufacturers, the Samsung's the Micromax is Micromax in India, they all had volte smartphones out in the market, right? And I think that was one key step We took off, launching our own brand of smartphone life that helped us to overcome this problem that no smartphone had. We'll teach upsets in India and then in order. So when when we were launching there were about 13 telecom companies in India. It was a very crowded space on demand. In order to gain a foothold in that market, we really made a few decisions. Ah, phew. Key product announcement that really disrupted this entire industry. Right? So, um, Geo is a default for GLT network itself. All I p network Internet protocol in everything. All data. It's an all data network and everything from voice to data to Internet traffic. Everything goes over this. I'll goes over Internet protocol, and the cost to carry voice on our smartphone network is very low, right? The bandwidth voice consumes is very low in the entire Lt band. Right? So what we did Waas In order to gain a foothold in the market, we made voice completely free, right? He said you will not pay anything for boys and across India, we will not charge any roaming charges across India. Right? So we made voice free completely and we offer the lowest data rates in the world. We could do that because we had the largest capacity or to carry data in India off all the other telecom operators. And these data rates were unheard off in the world, right? So when we launched, we offered a $2 per month or $3 per month plan with unlimited data, you could consume 10 gigabytes of data all day if you wanted to, and some of our subscriber day. Right? So that's the first phase off the overgrowth and smartphones and that really disorders. We hit 100 million subscribers in 170 days, which was very, very fast. And then after the smartphone faith, we found that India still has 500 million feature phones. And in order to grow in that market, we launched our own phone, the geo phone, and we made it free. Right? So if you take if you took a geo subscription and you carried you stayed with us for three years, we would make this phone tree for your refund. The initial deposit that you paid for this phone and this phone had also had quite a few innovations tailored for the Indian market. It had all of our digital services for free, which I will talk about soon. And for example, you could plug in. You could use a cable right on RCR HDMI cable plug into the geo phone and you could watch TV on your big screen TV from the geophones. You didn't need a separate cable subscription toe watch TV, right? So that really helped us grow. And Geo Phone is now the largest selling feature phone in India on it. 100 million feature phones in India now. So now now we're in what I call the geo platforms phase. We're growing of a geo fiber fiber to the home fiber toe the office, um, space. And we've also launched our new commerce initiatives over e commerce initiatives and were steadily building platforms that other companies can leverage other companies can use in the Jeon o'clock. Right? So this is how a small startup not a small start, but a start of nonetheless least 400 million subscribers within four years the fastest growing tech company in the world. Next, Geo also helped a systemic change in India, and this is massive. A lot of startups are building on this India stack, as people call it, and I consider this India stack has made up off three things, and the acronym I use is jam. Trinity, right. So, um, in India, systemic change happened recently because the Indian government made bank accounts free for all one billion Indians. There were no service charges to store money in bank accounts. This is called the Jonathan. The J. GenDyn Bank accounts. The J out off the jam, then India is one of the few countries in the world toe have a digital biometric identity, which can be used to verify anyone online, which is huge. So you can simply go online and say, I am my ankle poor on duh. I verify that this is indeed me who's doing this transaction. This is the A in the jam and the last M stands for Mobil's, which which were held by Geo Mobile Internet in a plus. It is also it is. It also stands for something called the U. P I. The United Unified Payments Interface. This was launched by the Indian government, where you can carry digital transactions for free. You can transfer money from one person to the to another, essentially for free for no fee, right so I can transfer one group, even Indian rupee to my friend without paying any charges. That is huge, right? So you have a country now, which, with a with a billion people who are bank accounts, money in the bank, who you can verify online, right and who can pay online without any problems through their mobile connections held by G right. So suddenly our market, our Internet market, exploded from a few million users to now 506 106 100 million mobile Internet users. So that that I think, was a massive such a systemic change that happened in India. There are some really large hail, um, numbers for this India stack, right? In one month. There were 1.6 billion nuclear transactions in the last month, which is phenomenal. So next What is the impact of geo in India before you started, we were 155th in the world in terms off mobile in terms of broadband data consumption. Right. But after geo, India went from one 55th to the first in the world in terms of broadband data, largely consumed on mobile devices were a mobile first country, right? We have a habit off skipping technology generation, so we skip fixed line broadband and basically consuming Internet on our mobile phones. On average, Geo subscribers consumed 12 gigabytes of data per month, which is one of the highest rates in the world. So Geo has a huge role to play in making India the number one country in terms off broad banded consumption and geo responsible for quite a few industry first in the telecom space and in fact, in the India space, I would say so before Geo. To get a SIM card, you had to fill a form off the physical paper form. It used to go toe Ah, local distributor. And that local distributor is to check the farm that you feel incorrectly for your SIM card and then that used to go to the head office and everything took about 48 hours or so, um, to get your SIM card. And sometimes there were problems there also with a hard biometric authentication. We enable something, uh, India enable something called E K Y C Elektronik. Know your customer? We took a fingerprint scan at our point of Sale Reliance Digital stores, and within 15 minutes we could verify within a few minutes. Within a few seconds we could verify that person is indeed my hunk, right, buying the same car, Elektronik Lee on we activated the SIM card in 15 minutes. That was a massive deal for our growth. Initially right toe onboard 100 million customers. Within our and 70 days. We couldn't have done it without be K. I see that was a massive deal for us and that is huge for any company starting a business or start up in India. We also made voice free, no roaming charges and the lowest data rates in the world. Plus, we gave a full suite of cloud services for free toe all geo customers. For example, we give goTV essentially for free. We give GOTV it'll law for free, which people, when we have a launching, told us that no one would see no one would use because the Indians like watching TV in the living rooms, um, with the family on a big screen television. But when we actually launched, they found that GOTV is one off our most used app. It's like 70,000,080 million monthly active users, and now we've basically been changing culture in India where culture is on demand. You can watch TV on the goal and you can pause it and you can resume whenever you have some free time. So really changed culture in India, India on we help people liver, digital life online. Right, So that was massive. So >>I'm now I'd like to talk about our cloud >>journey on board Animal Minorities Partnership. We've been partners that since 2014 since the beginning. So Geo has been using open stack since 2014 when we started with 14 note luster. I'll be one production environment One right? And that was I call it the first wave off our cloud where we're just understanding open stack, understanding the capabilities, understanding what it could do. Now we're in our second wave. Where were about 4000 bare metal servers in our open stack cloud multiple regions, Um, on that around 100,000 CPU cores, right. So it's a which is one of the bigger clouds in the world, I would say on almost all teams, with Ngor leveraging the cloud and soon I think we're going to hit about 10,000 Bama tools in our cloud, which is massive and just to give you a scale off our network, our in French, our data center footprint. Our network introduction is about 30 network data centers that carry just network traffic across there are there across India and we're about eight application data centers across three regions. Data Center is like a five story building filled with servers. So we're talking really significant scale in India. And we had to do this because when we were launching, there are the government regulation and try it. They've gotten regulatory authority of India, mandates that any telecom company they have to store customer data inside India and none of the other cloud providers were big enough to host our clothes. Right. So we we made all this intellectual for ourselves, and we're still growing next. I love to show you how we grown with together with Moran says we started in 2014 with the fuel deployment pipelines, right? And then we went on to the NK deployment. Pipelines are cloud started growing. We started understanding the clouds and we picked up M C p, which has really been a game changer for us in automation, right on DNA. Now we are in the latest release, ofem CPM CPI $2019 to on open stack queens, which on we've just upgraded all of our clouds or the last few months. Couple of months, 2 to 3 months. So we've done about nine production clouds and there are about 50 internal, um, teams consuming cloud. We call as our tenants, right. We have open stack clouds and we have communities clusters running on top of open stack. There are several production grade will close that run on this cloud. The Geo phone, for example, runs on our cloud private cloud Geo Cloud, which is a backup service like Google Drive and collaboration service. It runs out of a cloud. Geo adds G o g S t, which is a tax filing system for small and medium enterprises, our retail post service. There are all these production services running on our private clouds. We're also empaneled with the government off India to provide cloud services to the government to any State Department that needs cloud services. So we were empaneled by Maiti right in their ego initiative. And our clouds are also Easter. 20,000 certified 20,000 Colin one certified for software processes on 27,001 and said 27,017 slash 18 certified for security processes. Our clouds are also P our data centers Alsop a 942 be certified. So significant effort and investment have gone toe These data centers next. So this is where I think we've really valued the partnership with Morantes. Morantes has has trained us on using the concepts of get offs and in fries cold, right, an automated deployments and the tool change that come with the M C P Morantes product. Right? So, um, one of the key things that has happened from a couple of years ago to today is that the deployment time to deploy a new 100 north production cloud has decreased for us from about 55 days to do it in 2015 to now, we're down to about five days to deploy a cloud after the bear metals a racked and stacked. And the network is also the physical network is also configured, right? So after that, our automated pipelines can deploy 100 0 clock in five days flight, which is a massive deal for someone for a company that there's adding bear metals to their infrastructure so fast, right? It helps us utilize our investment, our assets really well. By the time it takes to deploy a cloud control plane for us is about 19 hours. It takes us two hours to deploy a compu track and it takes us three hours to deploy a storage rack. Right? And we really leverage the re class model off M C. P. We've configured re class model to suit almost every type of cloud that we have, right, and we've kept it fairly generous. It can be, um, Taylor to deploy any type of cloud, any type of story, nor any type of compute north. Andi. It just helps us automate our deployments by putting every configuration everything that we have in to get into using infra introduction at school, right plus M. C. P also comes with pipelines that help us run automated tests, automated validation pipelines on our cloud. We also have tempest pipelines running every few hours every three hours. If I recall correctly which run integration test on our clouds to make sure the clouds are running properly right, that that is also automated. The re class model and the pipelines helpers automate day to operations and changes as well. There are very few seventh now, compared toa a few years ago. It very rare. It's actually the exception and that may be because off mainly some user letter as opposed to a cloud problem. We also have contributed auto healing, Prometheus and Manager, and we integrate parameters and manager with our even driven automation framework. Currently, we're using Stack Storm, but you could use anyone or any event driven automation framework out there so that it indicates really well. So it helps us step away from constantly monitoring our cloud control control planes and clothes. So this has been very fruitful for us and it has actually apps killed our engineers also to use these best in class practices like get off like in France cord. So just to give you a flavor on what stacks our internal teams are running on these clouds, Um, we have a multi data center open stack cloud, and on >>top of that, >>teams use automation tools like terra form to create the environments. They also create their own Cuba these clusters and you'll see you'll see in the next slide also that we have our own community that the service platform that we built on top of open stack to give developers development teams NGO um, easy to create an easy to destroy Cuban. It is environment and sometimes leverage the Murano application catalog to deploy using heats templates to deploy their own stacks. Geo is largely a micro services driven, Um um company. So all of our applications are micro services, multiple micro services talking to each other, and the leverage develops. Two sets, like danceable Prometheus, Stack stone from for Otto Healing and driven, not commission. Big Data's tax are already there Kafka, Patches, Park Cassandra and other other tools as well. We're also now using service meshes. Almost everything now uses service mesh, sometimes use link. Erred sometimes are experimenting. This is Theo. So So this is where we are and we have multiple clients with NGO, so our products and services are available on Android IOS, our own Geo phone, Windows Macs, Web, Mobile Web based off them. So any client you can use our services and there's no lock in. It's always often with geo, so our sources have to be really good to compete in the open Internet. And last but not least, I think I love toe talk to you about our container journey. So a couple of years ago, almost every team started experimenting with containers and communities and they were demand for as a platform team. They were demanding community that the service from us a manage service. Right? So we built for us, it was much more comfortable, much more easier toe build on top of open stack with cloud FBI s as opposed to doing this on bare metal. So we built a fully managed community that a service which was, ah, self service portal, where you could click a button and get a community cluster deployed in your own tenant on Do the >>things that we did are quite interesting. We also handle some geo specific use cases. So we have because it was a >>manage service. We deployed the city notes in our own management tenant, right? We didn't give access to the customer to the city. Notes. We deployed the master control plane notes in the tenant's tenant and our customers tenant, but we didn't give them access to the Masters. We didn't give them the ssh key the workers that the our customers had full access to. And because people in Genova learning and experimenting, we gave them full admin rights to communities customers as well. So that way that really helped on board communities with NGO. And now we have, like 15 different teams running multiple communities clusters on top, off our open stack clouds. We even handle the fact that there are non profiting. I people separate non profiting I peoples and separate production 49 p pools NGO. So you could create these clusters in whatever environment that non prod environment with more open access or a prod environment with more limited access. So we had to handle these geo specific cases as well in this communities as a service. So on the whole, I think open stack because of the isolation it provides. I think it made a lot of sense for us to do communities our service on top off open stack. We even did it on bare metal, but that not many people use the Cuban, indeed a service environmental, because it is just so much easier to work with. Cloud FBI STO provision much of machines and covering these clusters. That's it from me. I think I've said a mouthful, and now I love for you toe. I'd love to have your questions. If you want to reach out to me. My email is mine dot capulet r l dot com. I'm also you can also message me on Twitter at my uncouple. So thank you. And it was a pleasure talking to you, Andre. Let let me hear your questions.
SUMMARY :
So in order to solve that problem, we launched our own brand of smartphones called the So just to give you a flavor on what stacks our internal It is environment and sometimes leverage the Murano application catalog to deploy So we have because it was a So on the whole, I think open stack because of the isolation
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
2015 | DATE | 0.99+ |
India | LOCATION | 0.99+ |
2014 | DATE | 0.99+ |
two hours | QUANTITY | 0.99+ |
$50 | QUANTITY | 0.99+ |
3 | QUANTITY | 0.99+ |
12 gigabytes | QUANTITY | 0.99+ |
three years | QUANTITY | 0.99+ |
Morantes | ORGANIZATION | 0.99+ |
70,000,080 million | QUANTITY | 0.99+ |
Andre | PERSON | 0.99+ |
three hours | QUANTITY | 0.99+ |
Samsung | ORGANIZATION | 0.99+ |
2000 | DATE | 0.99+ |
70 days | QUANTITY | 0.99+ |
Genova | LOCATION | 0.99+ |
five days | QUANTITY | 0.99+ |
2 | QUANTITY | 0.99+ |
zero | QUANTITY | 0.99+ |
0 | QUANTITY | 0.99+ |
170 days | QUANTITY | 0.99+ |
100 million subscribers | QUANTITY | 0.99+ |
Onda Mantis Partnership | ORGANIZATION | 0.99+ |
first phase | QUANTITY | 0.99+ |
100 million | QUANTITY | 0.99+ |
15 minutes | QUANTITY | 0.99+ |
10 gigabytes | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
16 | DATE | 0.99+ |
four years | QUANTITY | 0.99+ |
4 months | QUANTITY | 0.99+ |
one person | QUANTITY | 0.99+ |
49 p | QUANTITY | 0.99+ |
100 million customers | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
one billion | QUANTITY | 0.99+ |
Two sets | QUANTITY | 0.99+ |
155th | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
one key step | QUANTITY | 0.99+ |
last month | DATE | 0.99+ |
first country | QUANTITY | 0.98+ |
3 months | QUANTITY | 0.98+ |
around 100,000 CPU cores | QUANTITY | 0.98+ |
Joe | PERSON | 0.98+ |
100 | QUANTITY | 0.98+ |
27,001 | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
15 different teams | QUANTITY | 0.98+ |
Android IOS | TITLE | 0.98+ |
one month | QUANTITY | 0.98+ |
France | LOCATION | 0.98+ |
506 106 100 million | QUANTITY | 0.98+ |
Geo | ORGANIZATION | 0.98+ |
Elektronik Lee | ORGANIZATION | 0.98+ |
FBI | ORGANIZATION | 0.98+ |
one group | QUANTITY | 0.98+ |
1.6 billion nuclear transactions | QUANTITY | 0.98+ |
Andi | PERSON | 0.97+ |
Geo Mobile Internet | ORGANIZATION | 0.97+ |
five story | QUANTITY | 0.97+ |
Prometheus | TITLE | 0.97+ |
VxRail: Taking HCI to Extremes
>> Announcer: From the Cube studios in Palo Alto in Boston, connecting with thought leaders all around the world, this is theCube Conversation. >> Hi, I'm Stu Miniman. And welcome to this special presentation. We have a launch from Dell Technologies updates from the VxRail family. We're going to do things a little bit different here. We actually have a launch video Shannon Champion, of Dell Technologies. And the way we do things a lot of times, is, analysts get a little preview or when you're watching things. You might have questions on it. So, rather than me just wanting it, or you wanting yourself I actually brought in a couple of Dell Technologies expertS two of our Cube alumni, happy to welcome you back to the program. Jon Siegal, he is the Vice President of Product Marketing, and Chad Dunn, who's the Vice President of Product Management, both of them with Dell Technologies. Gentlemen, thanks so much for joining us. >> Good to see you Stu. >> Great to be here. >> All right, and so what we're going to do is we're going to be rolling the video here. I've got a button I'm going to press, Andrew will stop it here and then we'll kind of dig in a little bit, go into some questions when we're all done. We're actually holding a crowd chat, where you will be able to ask your questions, talk to the experts and everything. And so a little bit different way to do a product announcement. Hope you enjoy it. And with that, it's VxRail. Taking HCI to the extremes is the theme. We'll see what that means and everything. But without any further ado, let's let Shannon take the video away. >> Hello, and welcome. My name is Shannon Champion, and I'm looking forward to taking you through what's new with VxRail. Let's get started. We have a lot to talk about. Our launch covers new announcements addressing use cases across the Core, Edge and Cloud and spans both new hardware platforms and options, as well as the latest in software innovations. So let's jump right in. Before we talk about our announcements, let's talk about where customers are adopting VxRail today. First of all, on behalf of the entire Dell Technologies and VxRail teams, I want to thank each of our over 8000 customers, big and small in virtually every industry, who've chosen VxRail to address a broad range of workloads, deploying nearly 100,000 nodes today. Thank you. Our promise to you is that we will add new functionality, improve serviceability, and support new use cases, so that we deliver the most value to you, whether in the Core, at the Edge or for the Cloud. In the Core, VxRail from day one has been a catalyst to accelerate IT transformation. Many of our customers started here and many will continue to leverage VxRail to simply extend and enhance your VMware environment. Now we can support even more demanding applications such as In-Memory databases, like SAP HANA, and more AI and ML applications, with support for more and more powerful GPUs. At the Edge, video surveillance, which also uses GPUs, by the way, is an example of a popular use case leveraging VxRail alongside external storage. And right now we all know the enhanced role that IT is playing. And as it relates to VDI, VxRail has always been a great option for that. In the Cloud, it's all about Kubernetes, and how Dell Technologies Cloud platform, which is VCF on VxRail can deliver consistent infrastructure for both traditional and Cloud native applications. And we're doing that together with VMware. VxRail is the only jointly engineered HCI system built with VMware for VMware environments, designed to enhance the native VMware experience. This joint engineering with VMware and investments in software innovation together deliver an optimized operational experience at reduced risk for our customers. >> Alright, so Shannon talked a bit about, the important role of IT Of course right now, with the global pandemic going on. It's really, calling in, essential things, putting, platforms to the test. So, I really love to hear what both of you are hearing from customers. Also, VDI, of course, in the early days, it was, HCI-only-does-VDI. Now, we know there are many solutions, but remote work is putting that back front and center. So, Jon, why don't we start with you as the what is (muffled speaking) >> Absolutely. So first of all, Stu, thank you, I want to do a shout out to our VxRail customers around the world. It's really been humbling, inspiring, and just amazing to see The impact of our VxRail customers around the world and what they're having on on human progress here. Just for a few examples, there are genomics companies that we have running VxRail that have rolled out testing at scale. We also have research universities out in the Netherlands, doing the antibody detection. The US Navy has stood up a floating hospital to of course care for those in need. So we are here to help that's been our message to our customers, but it's amazing to see how much they're helping society during this. So just just a pleasure there. But as you mentioned, just to hit on the VDI comments, so to your points too, HCI, VxRail, VDI, that was an initial use case years ago. And it's been great to see how many of our existing VxRail customers have been able to pivot very quickly leveraging VxRail to add and to help bring their remote workforce online and support them with their existing VxRail. Because VxRail is flexible, it is agile, to be able to support those multiple workloads. And in addition to that, we've also rolled out some new VDI bundles to make it simpler for customers more cost effective cater to everything from knowlEdge workers to multimedia workers. You name it, you know from 250, desktops up to 1000. But again, back to your point VxRail, HCI, is well beyond VDI, it crossed the chasm a couple years ago actually. And VDI now is less than a third of the typical workloads, any of our customers out there, it supports now a range of workloads that you heard from Shannon, whether it's video surveillance, whether it's general purpose, all the way to mission critical applications now with SAP HAN. So, this has changed the game for sure. But the range of work loads and the flexibility of the actual rules which really helping our existing customers during this pandemic. >> Yeah, I agree with you, Jon, we've seen customers really embrace HCI for a number of workloads in their environments, from the ones that we sure all knew and loved back in the initial days of HCI. Now, the mission critical things now to Cloud native workloads as well, and the sort of the efficiencies that customers are able to get from HCI. And specifically, VxRail gives them that ability to pivot. When these, shall we say unexpected circumstances arise? And I think that that's informing their their decisions and their opinions on what their IP strategies look like as they move forward. They want that same level of agility, and ability to react quickly with their overall infrastructure. >> Excellent. Now I want to get into the announcements. What I want my team actually, your team gave me access to the CIO from the city of Amarillo, so maybe they can dig up that footage, talk about how fast they pivoted, using VxRail to really spin up things fast. So let's hear from the announcement first and then definitely want to share that that customer story a little bit later. So let's get to the actual news that Shannon's going to share. >> Okay, now what's new? I am pleased to announce a number of exciting updates and new platforms, to further enable IT modernization across Core, Edge and Cloud. I will cover each of these announcements in more detail, demonstrating how only VxRail can offer the breadth of platform configurations, automation, orchestration and Lifecycle Management, across a fully integrated hardware and software full stack with consistent, simplified operations to address the broadest range of traditional and modern applications. I'll start with hybrid Cloud and recap what you may have seen in the Dell Technologies Cloud announcements just a few weeks ago, related to VMware Cloud foundation on VxRail. Then I'll cover two brand new VxRail hardware platforms and additional options. And finally circle back to talk about the latest enhancements to our VxRail HCI system software capabilities for Lifecycle Management. Let's get started with our new Cloud offerings based on VxRail. VxRail is the HCI foundation for Dell Technologies, Cloud Platform, bringing automation and financial models, similar to public Cloud to On-premises environments. VMware recently introduced Cloud foundation for Delta, which is based on vSphere 7.0. As you likely know by now, vSphere 7.0 was definitely an exciting and highly anticipated release. In keeping with our synchronous release commitment, we introduced VxRail 7.0 based on vSphere 7.0 in late April, which was within 30 days of VMware's release. Two key areas that VMware focused on we're embedding containers and Kubernetes into vSphere, unifying them with virtual machines. And the second is improving the work experience for vSphere administrators with vSphere Lifecycle Manager or VLCM. I'll address the second point a bit in terms of how VxRail fits in in a moment for VCF 4 with Tom Xu, based on vSphere 7.0 customers now have access to a hybrid Cloud platform that supports native Kubernetes workloads and management, as well as your traditional VM-based workloads. So containers are now first class citizens of your private Cloud alongside traditional VMs and this is now available with VCF 4.0, on VxRail 7.0. VxRail's tight integration with VMware Cloud foundation delivers a simple and direct path not only to the hybrid Cloud, but also to deliver Kubernetes at Cloud scale with one complete automated platform. The second Cloud announcement is also exciting. Recent VCF for networking advancements have made it easier than ever to get started with hybrid Cloud, because we're now able to offer a more accessible consolidated architecture. And with that Dell Technologies Cloud platform can now be deployed with a four-node configuration, lowering the cost of an entry level hybrid Cloud. This enables customers to start smaller and grow their Cloud deployment over time. VCF and VxRail can now be deployed in two different ways. For small environments, customers can utilize a consolidated architecture which starts with just four nodes. Since the management and workload domains share resources in this architecture, it's ideal for getting started with an entry level Cloud to run general purpose virtualized workloads with a smaller entry point. Both in terms of required infrastructure footprint as well as cost, but still with a Consistent Cloud operating model. For larger environments where dedicated resources and role-based access control to separate different sets of workloads is usually preferred. You can choose to deploy a standard architecture which starts at eight nodes for independent management and workload domains. A standard implementation is ideal for customers running applications that require dedicated workload domains that includes Horizon, VDI, and vSphere with Kubernetes. >> Alright, Jon, there's definitely been a lot of interest in our community around everything that VMware is doing with vSphere 7.0. understand if you wanted to use the Kubernetes piece, it's VCF as that so we've seen the announcements, Dell, partnering in there it helps us connect that story between, really the VMware strategy and how they talk about Cloud and where does VxRail fit in that overall, Delta Cloud story? >> Absolutely. So first of all Stu, the VxRail course is integral to the Delta Cloud strategy. it's been VCF on VxRail equals the Delta Cloud platform. And this is our flagship on prem Cloud offering, that we've been able to enable operational consistency across any Cloud, whether it's On-prem, in the Edge or in the public Cloud. And we've seen the Dell tech Cloud Platform embraced by customers for a couple key reasons. One is it offers the fastest hybrid Cloud deployment in the market. And this is really, thanks to a new subscription offer that we're now offering out there where in less than 14 days, it can be still up and running. And really, the Dell tech Cloud does bring a lot of flexibility in terms of consumption models, overall when it comes to VxRail. Secondly, I would say is fast and easy upgrades. This is what VxRail brings to the table for all workloads, if you will, into especially critical in the Cloud. So the full automation of Lifecycle Management across the hardware and software stack across the VMware software stack, and in the Dell software and hardware supporting that, together, this enables essentially the third thing, which is customers can just relax. They can be rest assured that their infrastructure will be continuously validated, and always be in a continuously validated state. And this is the kind of thing that those three value propositions together really fit well, with any on-prem Cloud. Now you take what Shannon just mentioned, and the fact that now you can build and run modern applications on the same VxRail infrastructure alongside traditional applications. This is a game changer. >> Yeah, I love it. I remember in the early days talking with Dunn about CI, how does that fit in with Cloud discussion and the line I've used the last couple years is, modernize the platform, then you can modernize the application. So as companies are doing their full modernization, then this plays into what you're talking about. All right, we can let Shannon continue, we can get some more before we dig into some more analysis. >> That's good. >> Let's talk about new hardware platforms and updates. that result in literally thousands of potential new configuration options. covering a wide breadth of modern and traditional application needs across a range of the actual use cases. First up, I am incredibly excited to announce a brand new Dell EMC VxRail series, the D series. This is a ruggedized durable platform that delivers the full power of VxRail for workloads at the Edge in challenging environments or for space constrained areas. VxRail D series offers the same compelling benefits as the rest of the VxRail portfolio with simplicity, agility and lifecycle management. But in a lightweight short depth at only 20 inches, it's adorable form factor that's extremely temperature-resilient, shock resistant, and easily portable. It even meets milspec standards. That means you have the full power of lifecycle automation with VxRail HCI system software and 24 by seven single point of support, enabling you to rapidly react to business needs, no matter the location or how harsh the conditions. So whether you're deploying a data center at a mobile command base, running real-time GPS mapping on the go, or implementing video surveillance in remote areas, you can ensure availability, integrity and confidence for every workload with the new VxRail ruggedized D series. >> All right, Chad we would love for you to bring us in a little bit that what customer requirement for bringing this to market. I remember seeing, Dell servers ruggedized, of course, Edge, really important growth to build on what Jon was talking about, Cloud. So, Chad, bring us inside, what was driving this piece of the offering? >> Sure Stu. Yeah, yeah, having been at the hardware platforms that can go out into some of these remote locations is really important. And that's being driven by the fact that customers are looking for compute performance and storage out at some of these Edges or some of the more exotic locations. whether that's manufacturing plants, oil rigs, submarine ships, military applications, places that we've never heard of. But it's also about extending that operational simplicity of the the sort of way that you're managing your data center that has VxRails you're managing your Edges the same way using the same set of tools. You don't need to learn anything else. So operational simplicity is absolutely key here. But in those locations, you can take a product that's designed for a data center where definitely controlling power cooling space and take it some of these places where you get sand blowing or seven to zero temperatures, could be Baghdad or it could be Ketchikan, Alaska. So we built this D series that was able to go to those extreme locations with extreme heat, extreme cold, extreme altitude, but still offer that operational simplicity. Now military is one of those applications for the rugged platform. If you look at the resistance that it has to heat, it operates at a 45 degrees Celsius or 113 degrees Fahrenheit range, but it can do an excursion up to 55 C or 131 degrees Fahrenheit for up to eight hours. It's also resistant to heat sand, dust, vibration, it's very lightweight, short depth, in fact, it's only 20 inches deep. This is a smallest form factor, obviously that we have in the VxRail family. And it's also built to be able to withstand sudden shocks certified to withstand 40 G's of shock and operation of the 15,000 feet of elevation. Pretty high. And this is sort of like wherever skydivers go to when they want the real thrill of skydiving where you actually need oxygen to, to be for that that altitude. They're milspec-certified. So, MIL-STD-810G, which I keep right beside my bed and read every night. And it comes with a VxRail stick hardening package is packaging scripts so that you can auto lock down the rail environment. And we've got a few other certifications that are on the roadmap now for naval shock requirements. EMI and radiation immunity often. >> Yeah, it's funny, I remember when we first launched it was like, "Oh, well everything's going to white boxes. "And it's going to be massive, "no differentiation between everything out there." If you look at what you're offering, if you look at how public Clouds build their things, but I called it a few years or is there's a pure optimization. So you need to scale, you need similarities but you know you need to fit some, very specific requirements, lots of places, so, interesting stuff. Yeah, certifications, always keep your teams busy. Alright, let's get back to Shannon to view on the report. >> We are also introducing three other hardware-based additions. First, a new VxRail E Series model based on where the first time AMD EPYC processors. These single socket 1U nodes, offer dual socket performance with CPU options that scale from eight to 64 Cores, up to a terabyte of memory and multiple storage options making it an ideal platform for desktop VDI analytics and computer aided design. Next, the addition of the latest Nvidia Quadro RTX GPUs brings the most significant advancement in computer graphics in over a decade to professional work flows. Designers and artists across industries can now expand the boundary of what's possible, working with the largest and most complex graphics rendering, deep learning and visual computing workloads. And Intel Optane DC persistent memory is here, and it offers high performance and significantly increased memory capacity with data persistence at an affordable price. Data persistence is a critical feature that maintains data integrity, even when power is lost, enabling quicker recovery and less downtime. With support for Intel obtain DC persistent memory customers can expand in memory intensive workloads and use cases like SAP HANA. Alright, let's finally dig into our HCI system software, which is the Core differentiation for VxRail regardless of your workload or platform choice. Our joining engineering with VMware and investments in VxRail HCI system software innovation together deliver an optimized operational experience at reduced risk for our customers. Under the covers, VxRail offers best in class hardware, married with VMware HCI software, either vSAN or VCF. But what makes us different stems from our investments to integrate the two. Dell Technologies has a dedicated VxRail team of about 400 people to build market sell and support a fully integrated hyper converged system. That team has also developed our unique VxRail HCI system software, which is a suite of integrated software elements that extend VMware native capabilities to deliver seamless, automated operational experience that customers cannot find elsewhere. The key components of VxRail HCI system software shown around the arc here that include the extra manager, full stack lifecycle management, ecosystem connectors, and support. I don't have time to get into all the details of these elements today, but if you're interested in learning more, I encourage you to meet our experts. And I will tell you how to do that in a moment. I touched on the LCM being a key feature to the vSphere 7.0 earlier and I'd like to take the opportunity to expand on that a bit in the context of VxRail Lifecycle Management. The LCM adds valuable automation to the execution of updates for customers, but it doesn't eliminate the manual work still needed to define and package the updates and validate all of the components prior to applying them. With VxRail customers have all of these areas addressed automatically on their behalf, freeing them to put their time into other important functions for their business. Customers tell us that Lifecycle management continues to be a major source of the maintenance effort they put into their infrastructure, and then it tends to lead to overburden IT staff, that it can cause disruptions to the business if not managed effectively, and that it isn't the most efficient economically. Automation of Lifecycle Management and VxRail results in the utmost simplicity from a customer experience perspective, and offers operational freedom from maintaining infrastructure. But as shown here, our customers not only realize greater IT team efficiencies, they have also reduced downtime with fewer unplanned outages, and reduced overall cost of operations. With VxRail HCI system software, intelligent Lifecycle Management upgrades of the fully integrated hardware and software stack are automated, keeping clusters and continuously validated states while minimizing risks and operational costs. How do we ensure Continuously validated states for VxRail. VxRail labs execute an extensive, automated, repeatable process on every firmware and software upgrade and patch to ensure clusters are in continuously validated states of the customers choosing across their VxRail environment. The VxRail labs are constantly testing, analyzing, optimizing, and sequencing all of the components in the upgrade to execute in a single package for the full stack. All the while VxRail is backed by Dell EMC's world class services and support with a single point of contact for both hardware and software. IT productivity skyrockets with single click non disruptive upgrades of the fully integrated hardware and software stack without the need to do extensive research and testing. taking you to the next VxRail version of your choice, while always in a continuously validated state. You can also confidently execute automated VxRail upgrades. No matter what hardware generation or node types are in the cluster. They don't have to all be the same. And upgrades with VxRail are faster and more efficient with leapfrogging simply choose any VxRail version you desire. And be assured you will get there in a validated state while seamlessly bypassing any other release in between. Only VxRail can do that. >> All right, so Chad, the lifecycle management piece that Shannon was just talking about is, not the sexiest, it's often underappreciated. There's not only the years of experience, but the continuous work you're doing, reminds me back the early vSAN deployments versus VxRail jointly developed, jointly tested between Dell and VMware. So bring us inside why, 2020 Lifecycle Management still, a very important piece, especially in the VM family line. >> Yes, Stu, I think it's sexy, but, I'm pretty big nerd. (all laughing) Yeah, this is really always been our bread and butter. And in fact, it gets even more important, the larger the deployments come, when you start to look at data centers full of VxRails and all the different hardware software, firmware combinations that could exist out there. It's really the value that you get out of that VxRail HCI system software that Shannon was talking about and how it's optimized around the VMware use case. Very tightly integrated with each VMware component, of course, and the intelligence of being able to do all the firmware, all of the drivers, all the software all together in tremendous value to our customers. But to deliver that we really need to make a fairly large investment. So as Shannon mentioned, we run about 25,000 hours of testing across Each major release for patches, express patches, that's about 7000 hours for each of those. So, obviously, there's a lot of parallelism. And we're always developing new test scenarios for each release that we need to build in as we as we introduce new functionality. And one of the key things that we're able to do, as Shannon mentioned, is to be able to leapfrog releases and get you to that next validated state. We've got about 100 engineers just working on creating and executing those test cases on a continuous basis and obviously, a huge amount of automation. And we've talked about that investment to execute those tests. That's one worth of $60 million of investment in our lab. In fact, we've got just over 2000 VxRail units in our testbed across the US, Shanghai, China and Cork, Ireland. So a massive amount of testing of each of those components to make sure that they operate together in a validated state. >> Yeah, well, absolutely, it's super important not only for the day one, but the day two deployments. But I think this actually a great place for us to bring in that customer that Dell gave me access to. So we've got the CIO of Amarillo, Texas, he was an existing VxRail customer. And he's going to explain what happened as to how he needed to react really fast to support the work-from-home initiative, as well as we get to hear in his words the value of what Lifecycle Management means. So Andrew, if we could queue up that customer segment, please? >> It's been massive and it's been interesting to see the IT team absorb it. As we mature, I think they embrace the ability to be innovative and to work with our departments. But this instance, really justified why I was driving progress. So fervently why it was so urgent today. Three years ago, the answer would have been no. We wouldn't have been in a place where we could adapt With VxRail in place, in a week we spun up hundreds of instant balls. We spun up a 75-person call center in a day and a half, for our public health. We rolled out multiple applications for public health so they could do remote clinics. It's given us the flexibility to be able to roll out new solutions very quickly and be very adaptive. And it's not only been apparent to my team, but it's really made an impact on the business. And now what I'm seeing is those of my customers that work, a little lagging or a little conservative, or understanding the impact of modernizing the way they do business because it makes them adaptable as well. >> Alright, so great, Richard, you talked a bunch about the the efficiencies that that the IT put in place, how about that, that overall just managed, you talked about how fast you spun up these new VDI instances. need to be able to do things much simpler? So how does the overall Lifecycle Management fit into this discussion? >> It makes it so much easier. And in the old environment, one, It took a lot of man hours to make change. It was very disruptive, when we did make change, it overburdened, I guess that's the word I'm looking for. It really overburdened our staff to cause disruption to business. That wasn't cost efficient. And then simple things like, I've worked for multi billion dollar companies where we had massive QA environments that replicated production, simply can't afford that at local government. Having this sort of environment lets me do a scaled down QA environment and still get the benefit of rolling out non disruptive change. As I said earlier, it's allowed us to take all of those cycles that we were spending on Lifecycle Management because it's greatly simplified, and move those resources and rescale them in other areas where we can actually have more impact on the business. It's hard to be innovative when 100% of your cycles are just keeping the ship afloat. >> All right, well, nothing better than hearing it straight from the end user, public sector reacting very fast to the COVID-19. And, if you heard him he said, if this is his, before he had run this project, he would not have been able to respond. So I think everybody out there understands, if I didn't actually have access to the latest technology, it would be much harder. All right, I'm looking forward to doing the CrowdChat letting everybody else dig in with questions and get follow up but a little bit more, I believe one more announcement he can and got for us though. Let's roll the final video clip. >> In our latest software release VxRail 4.7.510, We continue to add new automation and self service features. New functionality enables you to schedule and run upgrade health checks in advance of upgrades, to ensure clusters are in a ready state for the next upgrade or patch. This is extremely valuable for customers that have stringent upgrade windows, as they can be assured the clusters will seamlessly upgrade within that window. Of course, running health checks on a regular basis also helps ensure that your clusters are always ready for unscheduled patches and security updates. We are also offering more flexibility and getting all nodes or clusters to a common release level with the ability to reimage nodes or clusters to a specific VxRail version, or down rev one or more nodes that may be shipped at a higher rate than the existing cluster. This enables you to easily choose your validated state when adding new nodes or repurposing nodes in a cluster. To sum up all of our announcements, whether you are accelerating data sets modernization extending HCI to harsh Edge environments, deploying an on-premises Dell Technologies Cloud platform to create a developer ready Kubernetes infrastructure. VxRail is there delivering a turn-key experience that enables you to continuously innovate, realize operational freedom and predictably evolve. VxRail provides an extensive breadth of platform configurations, automation and Lifecycle Management across the integrated hardware and software full stack and consistent hybrid Cloud operations to address the broadest range of traditional and modern applications across Core, Edge and Cloud. I now invite you to engage with us. First, the virtual passport program is an opportunity to have some fun while learning about VxRail new features and functionality and sCore some sweet digital swag while you're at it. Delivered via an augmented reality app. All you need is your device. So go to vxrail.is/passport to get started. And secondly, if you have any questions about anything I talked about or want a deeper conversation, we encourage you to join one of our exclusive VxRail Meet The Experts sessions available for a limited time. First come first served, just go to vxrail.is/expertsession to learn more. >> All right, well, obviously, with everyone being remote, there's different ways we're looking to engage. So we've got the CrowdChat right after this. But Jon, give us a little bit more as to how Dell's making sure to stay in close contact with customers and what you've got for options for them. >> Yeah, absolutely. So as Shannon said, so in lieu of not having done Tech World this year in person, where we could have those great in-person interactions and answer questions, whether it's in the booth or in meeting rooms, we are going to have these Meet The Experts sessions over the next couple weeks, and we're going to put our best and brightest from our technical community and make them accessible to everyone out there. So again, definitely encourage you. We're trying new things here in this virtual environment to ensure that we can still stay in touch, answer questions, be responsive, and really looking forward to, having these conversations over the next couple of weeks. >> All right, well, Jon and Chad, thank you so much. We definitely look forward to the conversation here and continued. If you're here live, definitely go down below and do it if you're watching this on demand. You can see the full transcript of it at crowdchat.net/vxrailrocks. For myself, Shannon on the video, Jon, Chad, Andrew, man in the booth there, thank you so much for watching, and go ahead and join the CrowdChat.
SUMMARY :
Announcer: From the Cube And the way we do things a lot of times, talk to the experts and everything. And as it relates to VDI, So, I really love to hear what both of you and the flexibility of the actual rules and the sort of the efficiencies that Shannon's going to share. the latest enhancements to really the VMware strategy and the fact that now you can build and the line I've used that delivers the full power of VxRail for bringing this to market. and operation of the "And it's going to be massive, and that it isn't the most especially in the VM family line. and all the different hardware software, And he's going to explain what happened the ability to be innovative that that the IT put in and still get the benefit it straight from the end user, for the next upgrade or patch. little bit more as to how to ensure that we can still and go ahead and join the CrowdChat.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Richard | PERSON | 0.99+ |
Jon | PERSON | 0.99+ |
Shannon | PERSON | 0.99+ |
Andrew | PERSON | 0.99+ |
Jon Siegal | PERSON | 0.99+ |
Chad Dunn | PERSON | 0.99+ |
Chad | PERSON | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
15,000 feet | QUANTITY | 0.99+ |
100% | QUANTITY | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
US | LOCATION | 0.99+ |
40 G | QUANTITY | 0.99+ |
Netherlands | LOCATION | 0.99+ |
Tom Xu | PERSON | 0.99+ |
$60 million | QUANTITY | 0.99+ |
US Navy | ORGANIZATION | 0.99+ |
131 degrees Fahrenheit | QUANTITY | 0.99+ |
Baghdad | LOCATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
113 degrees Fahrenheit | QUANTITY | 0.99+ |
vSphere 7.0 | TITLE | 0.99+ |
75-person | QUANTITY | 0.99+ |
China | LOCATION | 0.99+ |
vSphere | TITLE | 0.99+ |
45 degrees Celsius | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
VxRail | TITLE | 0.99+ |
30 days | QUANTITY | 0.99+ |
Shanghai | LOCATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
Stu | PERSON | 0.99+ |
eight | QUANTITY | 0.99+ |
VxRail 7.0 | TITLE | 0.99+ |
Amarillo | LOCATION | 0.99+ |
less than 14 days | QUANTITY | 0.99+ |
Delta Cloud | TITLE | 0.99+ |
late April | DATE | 0.99+ |
Delta | ORGANIZATION | 0.99+ |
20 inches | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
24 | QUANTITY | 0.99+ |
SAP HANA | TITLE | 0.99+ |
seven | QUANTITY | 0.99+ |
Both | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
VxRail E Series | COMMERCIAL_ITEM | 0.99+ |
each | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
a day and a half | QUANTITY | 0.98+ |
about 400 people | QUANTITY | 0.98+ |
VxRail: Taking HCI to Extremes
>> Announcer: From the Cube studios in Palo Alto in Boston, connecting with thought leaders all around the world, this is theCube Conversation. >> Hi, I'm Stu Miniman. And welcome to this special presentation. We have a launch from Dell Technologies updates from the VxRail family. We're going to do things a little bit different here. We actually have a launch video Shannon Champion, of Dell Technologies. And the way we do things a lot of times, is, analysts get a little preview or when you're watching things. You might have questions on it. So, rather than me just wanting it, or you wanting yourself I actually brought in a couple of Dell Technologies expertS two of our Cube alumni, happy to welcome you back to the program. Jon Siegal, he is the Vice President of Product Marketing, and Chad Dunn, who's the Vice President of Product Management, both of them with Dell Technologies. Gentlemen, thanks so much for joining us. >> Good to see you Stu. >> Great to be here. >> All right, and so what we're going to do is we're going to be rolling the video here. I've got a button I'm going to press, Andrew will stop it here and then we'll kind of dig in a little bit, go into some questions when we're all done. We're actually holding a crowd chat, where you will be able to ask your questions, talk to the experts and everything. And so a little bit different way to do a product announcement. Hope you enjoy it. And with that, it's VxRail. Taking HCI to the extremes is the theme. We'll see what that means and everything. But without any further ado, let's let Shannon take the video away. >> Hello, and welcome. My name is Shannon Champion, and I'm looking forward to taking you through what's new with VxRail. Let's get started. We have a lot to talk about. Our launch covers new announcements addressing use cases across the Core, Edge and Cloud and spans both new hardware platforms and options, as well as the latest in software innovations. So let's jump right in. Before we talk about our announcements, let's talk about where customers are adopting VxRail today. First of all, on behalf of the entire Dell Technologies and VxRail teams, I want to thank each of our over 8000 customers, big and small in virtually every industry, who've chosen VxRail to address a broad range of workloads, deploying nearly 100,000 nodes today. Thank you. Our promise to you is that we will add new functionality, improve serviceability, and support new use cases, so that we deliver the most value to you, whether in the Core, at the Edge or for the Cloud. In the Core, VxRail from day one has been a catalyst to accelerate IT transformation. Many of our customers started here and many will continue to leverage VxRail to simply extend and enhance your VMware environment. Now we can support even more demanding applications such as In-Memory databases, like SAP HANA, and more AI and ML applications, with support for more and more powerful GPUs. At the Edge, video surveillance, which also uses GPUs, by the way, is an example of a popular use case leveraging VxRail alongside external storage. And right now we all know the enhanced role that IT is playing. And as it relates to VDI, VxRail has always been a great option for that. In the Cloud, it's all about Kubernetes, and how Dell Technologies Cloud platform, which is VCF on VxRail can deliver consistent infrastructure for both traditional and Cloud native applications. And we're doing that together with VMware. VxRail is the only jointly engineered HCI system built with VMware for VMware environments, designed to enhance the native VMware experience. This joint engineering with VMware and investments in software innovation together deliver an optimized operational experience at reduced risk for our customers. >> Alright, so Shannon talked a bit about, the important role of IT Of course right now, with the global pandemic going on. It's really, calling in, essential things, putting, platforms to the test. So, I really love to hear what both of you are hearing from customers. Also, VDI, of course, in the early days, it was, HCI-only-does-VDI. Now, we know there are many solutions, but remote work is putting that back front and center. So, Jon, why don't we start with you as the what is (muffled speaking) >> Absolutely. So first of all, Stu, thank you, I want to do a shout out to our VxRail customers around the world. It's really been humbling, inspiring, and just amazing to see The impact of our VxRail customers around the world and what they're having on on human progress here. Just for a few examples, there are genomics companies that we have running VxRail that have rolled out testing at scale. We also have research universities out in the Netherlands, doing the antibody detection. The US Navy has stood up a floating hospital to of course care for those in need. So we are here to help that's been our message to our customers, but it's amazing to see how much they're helping society during this. So just just a pleasure there. But as you mentioned, just to hit on the VDI comments, so to your points too, HCI, VxRail, VDI, that was an initial use case years ago. And it's been great to see how many of our existing VxRail customers have been able to pivot very quickly leveraging VxRail to add and to help bring their remote workforce online and support them with their existing VxRail. Because VxRail is flexible, it is agile, to be able to support those multiple workloads. And in addition to that, we've also rolled out some new VDI bundles to make it simpler for customers more cost effective cater to everything from knowlEdge workers to multimedia workers. You name it, you know from 250, desktops up to 1000. But again, back to your point VxRail, HCI, is well beyond VDI, it crossed the chasm a couple years ago actually. And VDI now is less than a third of the typical workloads, any of our customers out there, it supports now a range of workloads that you heard from Shannon, whether it's video surveillance, whether it's general purpose, all the way to mission critical applications now with SAP HAN. So, this has changed the game for sure. But the range of work loads and the flexibility of the actual rules which really helping our existing customers during this pandemic. >> Yeah, I agree with you, Jon, we've seen customers really embrace HCI for a number of workloads in their environments, from the ones that we sure all knew and loved back in the initial days of HCI. Now, the mission critical things now to Cloud native workloads as well, and the sort of the efficiencies that customers are able to get from HCI. And specifically, VxRail gives them that ability to pivot. When these, shall we say unexpected circumstances arise? And I think that that's informing their their decisions and their opinions on what their IP strategies look like as they move forward. They want that same level of agility, and ability to react quickly with their overall infrastructure. >> Excellent. Now I want to get into the announcements. What I want my team actually, your team gave me access to the CIO from the city of Amarillo, so maybe they can dig up that footage, talk about how fast they pivoted, using VxRail to really spin up things fast. So let's hear from the announcement first and then definitely want to share that that customer story a little bit later. So let's get to the actual news that Shannon's going to share. >> Okay, now what's new? I am pleased to announce a number of exciting updates and new platforms, to further enable IT modernization across Core, Edge and Cloud. I will cover each of these announcements in more detail, demonstrating how only VxRail can offer the breadth of platform configurations, automation, orchestration and Lifecycle Management, across a fully integrated hardware and software full stack with consistent, simplified operations to address the broadest range of traditional and modern applications. I'll start with hybrid Cloud and recap what you may have seen in the Dell Technologies Cloud announcements just a few weeks ago, related to VMware Cloud foundation on VxRail. Then I'll cover two brand new VxRail hardware platforms and additional options. And finally circle back to talk about the latest enhancements to our VxRail HCI system software capabilities for Lifecycle Management. Let's get started with our new Cloud offerings based on VxRail. VxRail is the HCI foundation for Dell Technologies, Cloud Platform, bringing automation and financial models, similar to public Cloud to On-premises environments. VMware recently introduced Cloud foundation for Delta, which is based on vSphere 7.0. As you likely know by now, vSphere 7.0 was definitely an exciting and highly anticipated release. In keeping with our synchronous release commitment, we introduced VxRail 7.0 based on vSphere 7.0 in late April, which was within 30 days of VMware's release. Two key areas that VMware focused on we're embedding containers and Kubernetes into vSphere, unifying them with virtual machines. And the second is improving the work experience for vSphere administrators with vSphere Lifecycle Manager or VLCM. I'll address the second point a bit in terms of how VxRail fits in in a moment for VCF 4 with Tom Xu, based on vSphere 7.0 customers now have access to a hybrid Cloud platform that supports native Kubernetes workloads and management, as well as your traditional VM-based workloads. So containers are now first class citizens of your private Cloud alongside traditional VMs and this is now available with VCF 4.0, on VxRail 7.0. VxRail's tight integration with VMware Cloud foundation delivers a simple and direct path not only to the hybrid Cloud, but also to deliver Kubernetes at Cloud scale with one complete automated platform. The second Cloud announcement is also exciting. Recent VCF for networking advancements have made it easier than ever to get started with hybrid Cloud, because we're now able to offer a more accessible consolidated architecture. And with that Dell Technologies Cloud platform can now be deployed with a four-node configuration, lowering the cost of an entry level hybrid Cloud. This enables customers to start smaller and grow their Cloud deployment over time. VCF and VxRail can now be deployed in two different ways. For small environments, customers can utilize a consolidated architecture which starts with just four nodes. Since the management and workload domains share resources in this architecture, it's ideal for getting started with an entry level Cloud to run general purpose virtualized workloads with a smaller entry point. Both in terms of required infrastructure footprint as well as cost, but still with a Consistent Cloud operating model. For larger environments where dedicated resources and role-based access control to separate different sets of workloads is usually preferred. You can choose to deploy a standard architecture which starts at eight nodes for independent management and workload domains. A standard implementation is ideal for customers running applications that require dedicated workload domains that includes Horizon, VDI, and vSphere with Kubernetes. >> Alright, Jon, there's definitely been a lot of interest in our community around everything that VMware is doing with vSphere 7.0. understand if you wanted to use the Kubernetes piece, it's VCF as that so we've seen the announcements, Dell, partnering in there it helps us connect that story between, really the VMware strategy and how they talk about Cloud and where does VxRail fit in that overall, Delta Cloud story? >> Absolutely. So first of all Stu, the VxRail course is integral to the Delta Cloud strategy. it's been VCF on VxRail equals the Delta Cloud platform. And this is our flagship on prem Cloud offering, that we've been able to enable operational consistency across any Cloud, whether it's On-prem, in the Edge or in the public Cloud. And we've seen the Dell tech Cloud Platform embraced by customers for a couple key reasons. One is it offers the fastest hybrid Cloud deployment in the market. And this is really, thanks to a new subscription offer that we're now offering out there where in less than 14 days, it can be still up and running. And really, the Dell tech Cloud does bring a lot of flexibility in terms of consumption models, overall when it comes to VxRail. Secondly, I would say is fast and easy upgrades. This is what VxRail brings to the table for all workloads, if you will, into especially critical in the Cloud. So the full automation of Lifecycle Management across the hardware and software stack across the VMware software stack, and in the Dell software and hardware supporting that, together, this enables essentially the third thing, which is customers can just relax. They can be rest assured that their infrastructure will be continuously validated, and always be in a continuously validated state. And this is the kind of thing that those three value propositions together really fit well, with any on-prem Cloud. Now you take what Shannon just mentioned, and the fact that now you can build and run modern applications on the same VxRail infrastructure alongside traditional applications. This is a game changer. >> Yeah, I love it. I remember in the early days talking with Dunn about CI, how does that fit in with Cloud discussion and the line I've used the last couple years is, modernize the platform, then you can modernize the application. So as companies are doing their full modernization, then this plays into what you're talking about. All right, we can let Shannon continue, we can get some more before we dig into some more analysis. >> That's good. >> Let's talk about new hardware platforms and updates. that result in literally thousands of potential new configuration options. covering a wide breadth of modern and traditional application needs across a range of the actual use cases. First up, I am incredibly excited to announce a brand new Dell EMC VxRail series, the D series. This is a ruggedized durable platform that delivers the full power of VxRail for workloads at the Edge in challenging environments or for space constrained areas. VxRail D series offers the same compelling benefits as the rest of the VxRail portfolio with simplicity, agility and lifecycle management. But in a lightweight short depth at only 20 inches, it's adorable form factor that's extremely temperature-resilient, shock resistant, and easily portable. It even meets milspec standards. That means you have the full power of lifecycle automation with VxRail HCI system software and 24 by seven single point of support, enabling you to rapidly react to business needs, no matter the location or how harsh the conditions. So whether you're deploying a data center at a mobile command base, running real-time GPS mapping on the go, or implementing video surveillance in remote areas, you can ensure availability, integrity and confidence for every workload with the new VxRail ruggedized D series. >> All right, Chad we would love for you to bring us in a little bit that what customer requirement for bringing this to market. I remember seeing, Dell servers ruggedized, of course, Edge, really important growth to build on what Jon was talking about, Cloud. So, Chad, bring us inside, what was driving this piece of the offering? >> Sure Stu. Yeah, yeah, having been at the hardware platforms that can go out into some of these remote locations is really important. And that's being driven by the fact that customers are looking for compute performance and storage out at some of these Edges or some of the more exotic locations. whether that's manufacturing plants, oil rigs, submarine ships, military applications, places that we've never heard of. But it's also about extending that operational simplicity of the the sort of way that you're managing your data center that has VxRails you're managing your Edges the same way using the same set of tools. You don't need to learn anything else. So operational simplicity is absolutely key here. But in those locations, you can take a product that's designed for a data center where definitely controlling power cooling space and take it some of these places where you get sand blowing or seven to zero temperatures, could be Baghdad or it could be Ketchikan, Alaska. So we built this D series that was able to go to those extreme locations with extreme heat, extreme cold, extreme altitude, but still offer that operational simplicity. Now military is one of those applications for the rugged platform. If you look at the resistance that it has to heat, it operates at a 45 degrees Celsius or 113 degrees Fahrenheit range, but it can do an excursion up to 55 C or 131 degrees Fahrenheit for up to eight hours. It's also resistant to heat sand, dust, vibration, it's very lightweight, short depth, in fact, it's only 20 inches deep. This is a smallest form factor, obviously that we have in the VxRail family. And it's also built to be able to withstand sudden shocks certified to withstand 40 G's of shock and operation of the 15,000 feet of elevation. Pretty high. And this is sort of like wherever skydivers go to when they want the real thrill of skydiving where you actually need oxygen to, to be for that that altitude. They're milspec-certified. So, MIL-STD-810G, which I keep right beside my bed and read every night. And it comes with a VxRail stick hardening package is packaging scripts so that you can auto lock down the rail environment. And we've got a few other certifications that are on the roadmap now for naval shock requirements. EMI and radiation immunity often. >> Yeah, it's funny, I remember when we first launched it was like, "Oh, well everything's going to white boxes. "And it's going to be massive, "no differentiation between everything out there." If you look at what you're offering, if you look at how public Clouds build their things, but I called it a few years or is there's a pure optimization. So you need to scale, you need similarities but you know you need to fit some, very specific requirements, lots of places, so, interesting stuff. Yeah, certifications, always keep your teams busy. Alright, let's get back to Shannon to view on the report. >> We are also introducing three other hardware-based additions. First, a new VxRail E Series model based on where the first time AMD EPYC processors. These single socket 1U nodes, offer dual socket performance with CPU options that scale from eight to 64 Cores, up to a terabyte of memory and multiple storage options making it an ideal platform for desktop VDI analytics and computer aided design. Next, the addition of the latest Nvidia Quadro RTX GPUs brings the most significant advancement in computer graphics in over a decade to professional work flows. Designers and artists across industries can now expand the boundary of what's possible, working with the largest and most complex graphics rendering, deep learning and visual computing workloads. And Intel Optane DC persistent memory is here, and it offers high performance and significantly increased memory capacity with data persistence at an affordable price. Data persistence is a critical feature that maintains data integrity, even when power is lost, enabling quicker recovery and less downtime. With support for Intel obtain DC persistent memory customers can expand in memory intensive workloads and use cases like SAP HANA. Alright, let's finally dig into our HCI system software, which is the Core differentiation for VxRail regardless of your workload or platform choice. Our joining engineering with VMware and investments in VxRail HCI system software innovation together deliver an optimized operational experience at reduced risk for our customers. Under the covers, VxRail offers best in class hardware, married with VMware HCI software, either vSAN or VCF. But what makes us different stems from our investments to integrate the two. Dell Technologies has a dedicated VxRail team of about 400 people to build market sell and support a fully integrated hyper converged system. That team has also developed our unique VxRail HCI system software, which is a suite of integrated software elements that extend VMware native capabilities to deliver seamless, automated operational experience that customers cannot find elsewhere. The key components of VxRail HCI system software shown around the arc here that include the extra manager, full stack lifecycle management, ecosystem connectors, and support. I don't have time to get into all the details of these elements today, but if you're interested in learning more, I encourage you to meet our experts. And I will tell you how to do that in a moment. I touched on the LCM being a key feature to the vSphere 7.0 earlier and I'd like to take the opportunity to expand on that a bit in the context of VxRail Lifecycle Management. The LCM adds valuable automation to the execution of updates for customers, but it doesn't eliminate the manual work still needed to define and package the updates and validate all of the components prior to applying them. With VxRail customers have all of these areas addressed automatically on their behalf, freeing them to put their time into other important functions for their business. Customers tell us that Lifecycle management continues to be a major source of the maintenance effort they put into their infrastructure, and then it tends to lead to overburden IT staff, that it can cause disruptions to the business if not managed effectively, and that it isn't the most efficient economically. Automation of Lifecycle Management and VxRail results in the utmost simplicity from a customer experience perspective, and offers operational freedom from maintaining infrastructure. But as shown here, our customers not only realize greater IT team efficiencies, they have also reduced downtime with fewer unplanned outages, and reduced overall cost of operations. With VxRail HCI system software, intelligent Lifecycle Management upgrades of the fully integrated hardware and software stack are automated, keeping clusters and continuously validated states while minimizing risks and operational costs. How do we ensure Continuously validated states for VxRail. VxRail labs execute an extensive, automated, repeatable process on every firmware and software upgrade and patch to ensure clusters are in continuously validated states of the customers choosing across their VxRail environment. The VxRail labs are constantly testing, analyzing, optimizing, and sequencing all of the components in the upgrade to execute in a single package for the full stack. All the while VxRail is backed by Dell EMC's world class services and support with a single point of contact for both hardware and software. IT productivity skyrockets with single click non disruptive upgrades of the fully integrated hardware and software stack without the need to do extensive research and testing. taking you to the next VxRail version of your choice, while always in a continuously validated state. You can also confidently execute automated VxRail upgrades. No matter what hardware generation or node types are in the cluster. They don't have to all be the same. And upgrades with VxRail are faster and more efficient with leapfrogging simply choose any VxRail version you desire. And be assured you will get there in a validated state while seamlessly bypassing any other release in between. Only VxRail can do that. >> All right, so Chad, the lifecycle management piece that Shannon was just talking about is, not the sexiest, it's often underappreciated. There's not only the years of experience, but the continuous work you're doing, reminds me back the early vSAN deployments versus VxRail jointly developed, jointly tested between Dell and VMware. So bring us inside why, 2020 Lifecycle Management still, a very important piece, especially in the VM family line. >> Yes, Stu, I think it's sexy, but, I'm pretty big nerd. (all laughing) Yeah, this is really always been our bread and butter. And in fact, it gets even more important, the larger the deployments come, when you start to look at data centers full of VxRails and all the different hardware software, firmware combinations that could exist out there. It's really the value that you get out of that VxRail HCI system software that Shannon was talking about and how it's optimized around the VMware use case. Very tightly integrated with each VMware component, of course, and the intelligence of being able to do all the firmware, all of the drivers, all the software all together in tremendous value to our customers. But to deliver that we really need to make a fairly large investment. So as Shannon mentioned, we run about 25,000 hours of testing across Each major release for patches, express patches, that's about 7000 hours for each of those. So, obviously, there's a lot of parallelism. And we're always developing new test scenarios for each release that we need to build in as we as we introduce new functionality. And one of the key things that we're able to do, as Shannon mentioned, is to be able to leapfrog releases and get you to that next validated state. We've got about 100 engineers just working on creating and executing those test cases on a continuous basis and obviously, a huge amount of automation. And we've talked about that investment to execute those tests. That's one worth of $60 million of investment in our lab. In fact, we've got just over 2000 VxRail units in our testbed across the US, Shanghai, China and Cork, Ireland. So a massive amount of testing of each of those components to make sure that they operate together in a validated state. >> Yeah, well, absolutely, it's super important not only for the day one, but the day two deployments. But I think this actually a great place for us to bring in that customer that Dell gave me access to. So we've got the CIO of Amarillo, Texas, he was an existing VxRail customer. And he's going to explain what happened as to how he needed to react really fast to support the work-from-home initiative, as well as we get to hear in his words the value of what Lifecycle Management means. So Andrew, if we could queue up that customer segment, please? >> It's been massive and it's been interesting to see the IT team absorb it. As we mature, I think they embrace the ability to be innovative and to work with our departments. But this instance, really justified why I was driving progress. So fervently why it was so urgent today. Three years ago, the answer would have been no. We wouldn't have been in a place where we could adapt With VxRail in place, in a week we spun up hundreds of instant balls. We spun up a 75-person call center in a day and a half, for our public health. We rolled out multiple applications for public health so they could do remote clinics. It's given us the flexibility to be able to roll out new solutions very quickly and be very adaptive. And it's not only been apparent to my team, but it's really made an impact on the business. And now what I'm seeing is those of my customers that work, a little lagging or a little conservative, or understanding the impact of modernizing the way they do business because it makes them adaptable as well. >> Alright, so great, Richard, you talked a bunch about the the efficiencies that that the IT put in place, how about that, that overall just managed, you talked about how fast you spun up these new VDI instances. need to be able to do things much simpler? So how does the overall Lifecycle Management fit into this discussion? >> It makes it so much easier. And in the old environment, one, It took a lot of man hours to make change. It was very disruptive, when we did make change, it overburdened, I guess that's the word I'm looking for. It really overburdened our staff to cause disruption to business. That wasn't cost efficient. And then simple things like, I've worked for multi billion dollar companies where we had massive QA environments that replicated production, simply can't afford that at local government. Having this sort of environment lets me do a scaled down QA environment and still get the benefit of rolling out non disruptive change. As I said earlier, it's allowed us to take all of those cycles that we were spending on Lifecycle Management because it's greatly simplified, and move those resources and rescale them in other areas where we can actually have more impact on the business. It's hard to be innovative when 100% of your cycles are just keeping the ship afloat. >> All right, well, nothing better than hearing it straight from the end user, public sector reacting very fast to the COVID-19. And, if you heard him he said, if this is his, before he had run this project, he would not have been able to respond. So I think everybody out there understands, if I didn't actually have access to the latest technology, it would be much harder. All right, I'm looking forward to doing the CrowdChat letting everybody else dig in with questions and get follow up but a little bit more, I believe one more announcement he can and got for us though. Let's roll the final video clip. >> In our latest software release VxRail 4.7.510, We continue to add new automation and self service features. New functionality enables you to schedule and run upgrade health checks in advance of upgrades, to ensure clusters are in a ready state for the next upgrade or patch. This is extremely valuable for customers that have stringent upgrade windows, as they can be assured the clusters will seamlessly upgrade within that window. Of course, running health checks on a regular basis also helps ensure that your clusters are always ready for unscheduled patches and security updates. We are also offering more flexibility and getting all nodes or clusters to a common release level with the ability to reimage nodes or clusters to a specific VxRail version, or down rev one or more nodes that may be shipped at a higher rate than the existing cluster. This enables you to easily choose your validated state when adding new nodes or repurposing nodes in a cluster. To sum up all of our announcements, whether you are accelerating data sets modernization extending HCI to harsh Edge environments, deploying an on-premises Dell Technologies Cloud platform to create a developer ready Kubernetes infrastructure. VxRail is there delivering a turn-key experience that enables you to continuously innovate, realize operational freedom and predictably evolve. VxRail provides an extensive breadth of platform configurations, automation and Lifecycle Management across the integrated hardware and software full stack and consistent hybrid Cloud operations to address the broadest range of traditional and modern applications across Core, Edge and Cloud. I now invite you to engage with us. First, the virtual passport program is an opportunity to have some fun while learning about VxRail new features and functionality and sCore some sweet digital swag while you're at it. Delivered via an augmented reality app. All you need is your device. So go to vxrail.is/passport to get started. And secondly, if you have any questions about anything I talked about or want a deeper conversation, we encourage you to join one of our exclusive VxRail Meet The Experts sessions available for a limited time. First come first served, just go to vxrail.is/expertsession to learn more. >> All right, well, obviously, with everyone being remote, there's different ways we're looking to engage. So we've got the CrowdChat right after this. But Jon, give us a little bit more as to how Dell's making sure to stay in close contact with customers and what you've got for options for them. >> Yeah, absolutely. So as Shannon said, so in lieu of not having done Tech World this year in person, where we could have those great in-person interactions and answer questions, whether it's in the booth or in meeting rooms, we are going to have these Meet The Experts sessions over the next couple weeks, and we're going to put our best and brightest from our technical community and make them accessible to everyone out there. So again, definitely encourage you. We're trying new things here in this virtual environment to ensure that we can still stay in touch, answer questions, be responsive, and really looking forward to, having these conversations over the next couple of weeks. >> All right, well, Jon and Chad, thank you so much. We definitely look forward to the conversation here and continued. If you're here live, definitely go down below and do it if you're watching this on demand. You can see the full transcript of it at crowdchat.net/vxrailrocks. For myself, Shannon on the video, Jon, Chad, Andrew, man in the booth there, thank you so much for watching, and go ahead and join the CrowdChat.
SUMMARY :
Announcer: From the Cube And the way we do things a lot of times, talk to the experts and everything. And as it relates to VDI, So, I really love to hear what both of you and the flexibility of the actual rules and the sort of the efficiencies that Shannon's going to share. the latest enhancements to really the VMware strategy and the fact that now you can build and the line I've used that delivers the full power of VxRail for bringing this to market. and operation of the "And it's going to be massive, and that it isn't the most especially in the VM family line. and all the different hardware software, And he's going to explain what happened the ability to be innovative that that the IT put in and still get the benefit it straight from the end user, for the next upgrade or patch. little bit more as to how to ensure that we can still and go ahead and join the CrowdChat.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Richard | PERSON | 0.99+ |
Jon | PERSON | 0.99+ |
Shannon | PERSON | 0.99+ |
Andrew | PERSON | 0.99+ |
Jon Siegal | PERSON | 0.99+ |
Chad Dunn | PERSON | 0.99+ |
Chad | PERSON | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
15,000 feet | QUANTITY | 0.99+ |
100% | QUANTITY | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
US | LOCATION | 0.99+ |
40 G | QUANTITY | 0.99+ |
Netherlands | LOCATION | 0.99+ |
Tom Xu | PERSON | 0.99+ |
$60 million | QUANTITY | 0.99+ |
US Navy | ORGANIZATION | 0.99+ |
131 degrees Fahrenheit | QUANTITY | 0.99+ |
Baghdad | LOCATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
113 degrees Fahrenheit | QUANTITY | 0.99+ |
vSphere 7.0 | TITLE | 0.99+ |
75-person | QUANTITY | 0.99+ |
China | LOCATION | 0.99+ |
vSphere | TITLE | 0.99+ |
45 degrees Celsius | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
VxRail | TITLE | 0.99+ |
30 days | QUANTITY | 0.99+ |
Shanghai | LOCATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
Stu | PERSON | 0.99+ |
eight | QUANTITY | 0.99+ |
VxRail 7.0 | TITLE | 0.99+ |
Amarillo | LOCATION | 0.99+ |
less than 14 days | QUANTITY | 0.99+ |
Delta Cloud | TITLE | 0.99+ |
late April | DATE | 0.99+ |
Delta | ORGANIZATION | 0.99+ |
20 inches | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
24 | QUANTITY | 0.99+ |
SAP HANA | TITLE | 0.99+ |
seven | QUANTITY | 0.99+ |
Both | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
VxRail E Series | COMMERCIAL_ITEM | 0.99+ |
each | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
a day and a half | QUANTITY | 0.98+ |
about 400 people | QUANTITY | 0.98+ |
VMworld Day 1 General Session | VMworld 2018
For Las Vegas, it's the cube covering vm world 2018, brought to you by vm ware and its ecosystem partners. Ladies and gentlemen, Vm ware would like to thank it's global diamond sponsors and it's platinum sponsors for vm world 2018 with over 125,000 members globally. The vm ware User Group connects via vmware customers, partners and employees to vm ware, information resources, knowledge sharing, and networking. To learn more, visit the [inaudible] booth in the solutions exchange or the hemoglobin gene vm village become a part of the community today. This presentation includes forward looking statements that are subject to risks and uncertainties. Actual results may differ materially as a result of various risk factors including those described in the 10 k's 10 q's and k's vm ware. Files with the SEC. Ladies and Gentlemen, please welcome Pat Gelsinger. Welcome to vm world. Good morning. Let's try that again. Good morning and I'll just say it is great to be here with you today. I'm excited about the sixth year of being CEO. When it was on this stage six years ago were Paul Maritz handed me the clicker and that's the last he was seen. We have 20,000 plus here on site in Vegas and uh, you know, on behalf of everyone at Vm ware, you know, we're just thrilled that you would be with us and it's a joy and a thrill to be able to lead such a community. We have a lot to share with you today and we really think about it as a community. You know, it's my 23,000 plus employees, the souls that I'm responsible for, but it's our partners, the thousands and we kicked off our partner day yesterday, but most importantly, the vm ware community is centered on you. You know, we're very aware of this event would be nothing without you and our community and the role that we play at vm wares to build these cool breakthrough innovations that enable you to do incredible things. You're the ones who take our stuff and do amazing things. You altogether. We have truly changed the world over the last two decades and it is two decades. You know, it's our anniversary in 1998, the five people that started a vm ware, right. You know, it was, it was exactly 20 years ago and we're just thrilled and I was thinking about this over the weekend and it struck me, you know, anniversary, that's like old people, you know, we're here, we're having our birthday and it's a party, right? We can't have a drink yet, but next year. Yeah. We're 20 years old. Right. We can do that now. And I'll just say the culture of this community is something that truly is amazing and in my 38 years, 38 years in tech, that sort of sounds like I'm getting old or something, but the passion, the loyalty, almost a cult like behavior that we see in this team of people to us is simply thrilling. And you know, we put together a little video to sort of summarize the 20 years and some of that history and some of the unique and quirky aspects of our culture. Let's watch that now. We knew we had something unique and then we demonstrated that what was unique was also some reasons that we love vm ware, you know, like the community out there. So great. The technology I love it. Ware is solid and much needed. Literally. I do love Vmr. It's awesome. Super Awesome. Pardon? There's always someone that wants to listen and learn from us and we've learned so much from them as well. And we reached out to vm ware to help us start building. What's that future world look like? Since we're doing really cutting edge stuff, there's really no better people to call and Bmr has been known for continuous innovation. There's no better way to learn how to do new things in it than being with a company that's at the forefront of technology. What do you think? Don't you love that commitment? Hey Ashley, you know, but in the prep sessions for this, I thought, boy, what can I do to take my commitment to the next level? And uh, so, uh, you know, coming in a couple days early, I went to down the street to bad ass tattoo. So it's time for all of us to take our commitment up level and sometimes what happens in Vegas, you take home. Thank you. Vm Ware has had this unique role in the industry over these 20 years, you know, and for that we've seen just incredible things that have happened over this period of time and it's truly extraordinary what we've accomplished together. And you know, as we think back, you know, what vm ware has uniquely been able to do is I'll say bridge across know and we've seen time and again that we see these areas of innovation emerging and rapidly move forward. But then as they become utilized by our customers, they create this natural tension of what business wants us flexibility to use across these silos of innovation. And from the start of our history, we have collectively had this uncanny ability to bridge across these cycles of innovation. You know, an act one was clearly the server generation. You know, it may seem a little bit, uh, ancient memory now, but you remember you used to walk into your data center and it looked like the loove the museum of it passed right? You know, and you had your old p series and your z series in your sparks and your pas and your x86 cluster and Yo, it had to decide, well, which architecture or am I going to deploy and run this on? And we bridged across and that was the magic of Esx. You don't want to just changed the industry when that occurred. And I sort of called the early days of Esx and vsphere. It was like the intelligence test. If you weren't using it, you fail because Yup. Servers, 10 servers become one months, become minutes. I still have people today who come up to me and they reflect on their first experience of vsphere or be motion and it was like a holy moment in their life and in their careers. Amazing and act to the Byo d, You know, can we bridge across these devices and users wanted to be able to come in and say, I have my device and I'm productive on it. I don't want to be forced to use the corporate standard. And maybe more than anything was the power of the iphone that was introduced, the two, seven, and suddenly every employee said this is exciting and compelling. I want to use it so I can be more productive when I'm here. Bye. Jody was the rage and again it was a tough challenge and once again vm ware helped to bridge across the surmountable challenge. And clearly our workspace one community today is clearly bridging across these silos and not just about managing devices but truly enabling employee engagement and productivity. Maybe act three was the network and you know, we think about the network, you know, for 30 years we were bound to this physical view of what the network would be an in that network. We are bound to specific protocols. We had to wait months for network upgrades and firewall rules. Once every two weeks we'd upgrade them. If you had a new application that needed a firewall rule, sorry, you know, come back next month we'll put, you know, deep frustration among developers and ceos. Everyone was ready to break the chains. And that's exactly what we did. An NSX and Nice Sierra. The day we acquired it, Cisco stock drops and the industry realizes the networking has changed in a fundamental way. It will never be the same again. Maybe act for was this idea of cloud migration. And if we were here three years ago, it was student body, right to the public cloud. Everything is going there. And I remember I was meeting with a cio of federal cio and he comes up to me and he says, I tried for the last two years to replatform my 200 applications I got to done, you know, and all of a sudden that was this. How do I do cloud migration and the effective and powerful way. Once again, we bridged across, we brought these two worlds together and eliminated this, uh, you know, this gap between private and public cloud. And we'll talk a lot more about that today. You know, maybe our next act is what we'll call the multicloud era. You know, because today in a recent survey by Deloitte said that the average business today is using eight public clouds and expected to become 10 plus public clouds. And you know, as you're managing different tools, different teams, different architectures, those solution, how do you, again bridge across, and this is what we will do in the multicloud era, we will help our community to bridge across and take advantage of these powerful cycles of innovation that are going on, but be able to use them across a consistent infrastructure and operational environment. And we'll have a lot more to talk about on this topic today. You know, and maybe the last item to bridge across maybe the most important, you know, people who are profit. You know, too often we think about this as an either or question. And as a business leader, I'm are worried about the people or the And Milton Friedman probably set us up for this issue decades ago when he said, planet, right? the sole purpose of a business is to make profits. You want to create a multi-decade dilemma, right? For business leaders, could I have both people and profits? Could I do well and do good? And particularly for technology, I think we don't have a choice to think about these separately. We are permeating every aspect of business. And Society, we have the responsibility to do both and have all the things that vm ware has accomplished. I think this might be the one that I'm most proud of over, you know, w we have demonstrated by vsphere and the hypervisor alone that we have saved over 540 million tons of co two emissions. That is what you have done. Can you believe that? Five hundred 40 million tons is enough to have 68 percent of all households for a year. Wow. Thank you for what you have done. Thank you. Or another translation of that. Is that safe enough to drive a trillion miles and the average car or you could go to and from Jupiter just in case that was in your itinerary a thousand times. Right? He was just incredible. What we have done and as a result of that, and I'll say we were thrilled to accept this recognition on behalf of you and what you have done. You know, vm were recognized as number 17 in the fortune. Change the world list last week. And we really view it as accepting this honor on behalf of what you have done with our products and technology tech as a force for good. We believe that fundamentally that is our opportunity, if not our obligation, you know, fundamentally tech is neutral, you know, we together must shape it for good. You know, the printing press by Gutenberg in 1440, right? It was used to create mass education and learning materials also can be used for extremist propaganda. The technology itself is neutral. Our ecosystem has a critical role to play in shaping technology as a force for good. You know, and as we think about that tomorrow, we'll have a opportunity to have a very special guest and I really encourage you to be here, be on time tomorrow morning on the stage and you know, Sanjay's a session, we'll have Malala, Nobel Peace Prize winner and fourth will be a bit of extra security as you come in and you understand that. And I just encourage you not to be late because we see this tech being a force for good in everything that we do at vm ware. And I hope you'll enjoy, I'm quite looking forward to the session tomorrow. Now as we think about the future. I like to put it in this context, the superpowers of tech know and you know, 38 years in the industry, you know, I am so excited because I think everything that we've done over the last four decades is creating a foundation that allows us to do more and go faster together. We're unlocking game, changing opportunities that have not been available to any people in the history of humanity. And we have these opportunities now and I, and I think about these four cloud, you have unimaginable scale. You'll literally with your Amex card, you can go rent, you know, 10,000 cores for $100 per hour. Or if you have Michael's am ex card, we can rent a million cores for $10,000 an hour. Thanks Michael. But we also know that we're in many ways just getting started and we have tremendous issues to bridge across and compatible clouds, mobile unprecedented scale. Literally, your application can reach half the humans on the planet today. But we also know that five percent, the lowest five percent of humanity or the other half of humanity, they're still in the lower income brackets, less than five percent penetrated. And we know that we have customer examples that are using mobile phones to raise impoverished farmers in Africa, out of poverty just by having a smart phone with proper crop, the information field and whether a guidance that one tool alone lifting them out of poverty. Ai knows, you know, I really love the topic of ai in 1986. I'm the chief architect of the 80 46. Some of you remember what that was. Yeah, I, you know, you're, you're my folk, right? Right. And for those of you who don't, it was a real important chip at the time. And my marketing manager comes running into my office and he says, Pat, pat, we must make the 46 a great ai chip. This is 1986. What happened? Nothing an AI is today, a 30 year overnight success because the algorithms, the data have gotten so much bigger that we can produce results, that we can bring intelligence to everything. And we're seeing dramatic breakthroughs in areas like healthcare, radiology, you know, new drugs, diagnosis tools, and designer treatments. We're just scratching the surface, but ai has so many gaps, yet we don't even in many cases know why it works. Right? And we'll call that explainable ai and edge and Iot. We're connecting the physical and the digital worlds was never before possible. We're bridging technology into every dimension of human progress. And today we're largely hooking up things, right? We have so much to do yet to make them intelligent. Network secured, automated, the patch, bringing world class it to Iot, but it's not just that these are super powers. We really see that each and each one of them is a super power in and have their own right, but they're making each other more powerful as well. Cloud enables mobile conductivity. Mobile creates more data, more data makes the AI better. Ai Enables more edge use cases and more edge requires more cloud to store the data and do the computing right? They're reinforcing each other. And with that, we know that we are speeding up and these superpowers are reshaping every aspect of society from healthcare to education, the transportation, financial institutions. This is how it all comes together. Now, just a simple example, how many of you have ever worn a hardhat? Yeah, Yo. Pretty boring thing. And it has one purpose, right? You know, keep things from smacking me in the here's the modern hardhat. It's a complete heads up display with ar head. Well, vr capabilities that give the worker safety or workers or factory workers or supply people the ability to see through walls to understand what's going on inside of the equipment. I always wondered when I was a kid to have x Ray Vision, you know, some of my thoughts weren't good about why I wanted it, but you know, I wanted to. Well now you can have it, you know, but imagine in this environment, the complex application that sits behind it. You know, you're accessing maybe 50 year old building plants, right? You're accessing HVAC systems, but modern ar and vr capabilities and new containerized displays. You'll think about that application. You know, John Gage famously said the network is the computer pat today says the application is now a network and pretty typically a complicated one, you know, and this is the vm ware vision is to make that kind of environment realizable in every aspect of our business and community and we simply have been on this journey, any device, any application, any cloud with intrinsic security. And this vision has been consistent for those of you who have been joining us for a number of years. You've seen this picture, but it's been slowly evolving as we've worked in piece by piece to refine and extend this vision, you know, and for it, we're going to walk through and use this as the compass for our discussion today as we walk through our conversation. And you know, we're going to start by a focus on any cloud. And as we think about this cloud topic, you know, we see it as a multicloud world hybrid cloud, public cloud, but increasingly seeing edge and telco becoming clouds in and have their own right. And we're not gonna spend time on it today, but this area of Telco to the is an enormous opportunity for us in our community. You know, data centers and cloud today are over 80 percent virtualized. The Telco network is less than 10 percent virtualized. Wow. An industry that's almost as big as our industry entirely unvirtualized, although the technologies we've created here can be applied over here and Telco and we have an enormous buildout coming with five g and environments emerging. What an opportunity for us, a virgin market right next to us and we're getting some early mega winds in this area using the technologies that you have helped us cure rate than the So we're quite excited about this topic area as well. market. So let's look at this full view of the multicloud. Any cloud journey. And we see that businesses are on a multicloud journey, you know, and today we see this fundamentally in these two paths, a hybrid cloud and a public cloud. And these paths are complimentary and coexisting, but today, each is being driven by unique requirements and unique teams. Largely the hybrid cloud is being driven by it. And operations, the public cloud being driven more by developers and line of business requirements and as some multicloud environment. So how do we deliver upon that and for that, let's start by digging in on the hybrid cloud aspect of this and as we think about the hybrid cloud, we've been talking about this subject for a number of years and I want to give a very specific and crisp definition. You're the hybrid cloud is the public cloud and the private cloud cooperating with consistent infrastructure and consistent operations simply put seamless path to and from the cloud that my workloads don't care if it's here or there. I'm able to run them in a agile, scalable, flexible, efficient manner across those two environments, whether it's my data center or someone else's, I can bring them together to make that work is the magic of the Vm ware Cloud Foundation. The vm ware Cloud Foundation brings together computer vsphere and the core of why we are here, but combines with that networking storage delivered through a layer of management and automation. The rule of the cloud is ruthlessly automate everything. We laid out this vision of the software defined data center seven years ago and we've been steadfastly working on this vision and vm ware. Cloud Foundation provides this consistent infrastructure and operations with integrated lifecycle management automation. Patching the m ware cloud foundation is the simplest path to the hybrid cloud and the fastest way to get vm ware cloud foundation is hyperconverged infrastructure, you know, and with this we've combined integrated then validated hardware and as a building block inside of this we have validated hardware, the v Sand ready environments. We have integrated appliances and cloud delivered infrastructure, three ways that we deliver that integrate integrated hyperconverged infrastructure solution. And we have by far the broadest ecosystem of partners to do it. A broad set of the sand ready nodes from essentially everybody in the industry. Secondly, we have integrated appliances, the extract of vxrail that we have co engineered with our partners at Dell technology and today in fact Dell is releasing the power edge servers, a major step in blade servers that again are going to be powering vxrail and vxrack systems and we deliver hyperconverged infrastructure through a broader set of Vm ware cloud partners as well. At the heart of the hyperconverged infrastructure is v San and simply put, you know, be San has been the engine that's just been moving rapidly to take over the entire integration of compute and storage and expand to more and more areas. We have incredible momentum over 15,000 customers for v San Today and for those of you who joined us, we say thank you for what you have done with this product today. Really amazing you with 50 percent of the global 2000 using it know vm ware. V San Vxrail are clearly becoming the standard for how hyperconverge is done in the industry. Our cloud partner programs over 500 cloud partners are using ulv sand in their solution, you know, and finally the largest in Hci software revenue. Simply put the sand is the software defined storage technology of choice for the industry and we're seeing that customers are putting this to work in amazing ways. Vm Ware and Dell technologies believe in tech as a force for good and that it can have a major impact on the quality of life for every human on the planet and particularly for the most underdeveloped parts of the world. Those that live on less than $2 per day. In fact that this moment 5 billion people worldwide do not have access to modern affordable surgery. Mercy ships is working hard to change the global surgery crisis with greater than 400 volunteers. Mercy ships operates the largest NGO hospital ship delivering free medical care to the poorest of the poor in Africa. Let's see from them now. When the ship shows up to port, literally people line up for days to receive state of the art life, sane changing life saving surgeries, tumor site limbs, disease blindness, birth defects, but not only that, the personnel are educating and training the local healthcare providers with new skills and infrastructure so they can care for their own. After the ship has left, mercy ships runs on Vm ware, a dell technology with VX rail, Dell Isilon data protection. We are the it platform for mercy ships. Mercy ships is now building their next generation ship called global mercy, which were more than double. It's lifesaving capacity. It's the largest charity hospital ever. It will go live in 20 slash 20 serving Africa and I personally plan on being there for its launch. It is truly amazing what they are doing with our technology. Thanks. So we see this picture of the hybrid cloud. We've talked about how we do that for the private cloud. So let's look over at the public cloud and let's dig into this a little bit more deeply. You know, we're taking this incredible power of the Vm ware Cloud Foundation and making it available for the leading cloud providers in the world and with that, the partnership that we announced almost two years ago with Amazon and on the stage last year, we announced their first generation of products, no better example of the hybrid cloud. And for that it's my pleasure to bring to stage my friend, my partner, the CEO of aws. Please welcome Andy Jassy. Thank you andy. You know, you honor us with your presence, you know, and it really is a pleasure to be able to come in front of this audience and talk about what our teams have accomplished together over the last, uh, year. Yo, can you give us some perspective on that, Andy and what customers are doing with it? Well, first of all, thanks for having me. I really appreciate it. It's great to be here with all of you. Uh, you know, the offering that we have together customers because it allows them to use the same software they've been using to again, where cloud and aws is very appealing to manage their infrastructure for years to be able to deploy it an aws and we see a lot of customer momentum and a lot of customers using it. You see it in every imaginable vertical business segment in transportation. You see it with stagecoach and media and entertainment. You see it with discovery communications in education, Mit and Caltech and consulting and accenture and cognizant and dxc you see in every imaginable vertical business segment and the number of customers using the offering is doubling every quarter. So people were really excited about it and I think that probably the number one use case we see so far, although there are a lot of them, is customers who are looking to migrate on premises applications to the cloud. And a good example of that is mit. We're there right now in the process of migrating. In fact, they just did migrate 3000 vms from their data centers to Vm ware cloud native us. And this would have taken years before to do in the past, but they did it in just three months. It was really spectacular and they're just a fun company to work with and the team there. But we're also seeing other use cases as well. And you're probably the second most common example is we'll say on demand capabilities for things like disaster recovery. We have great examples of customers you that one in particular, his brakes, right? Urban in those. The brings security trucks and they all armored trucks coming by and they had a critical need to retire a secondary data center that they were using, you know, for Dr. so we quickly built to Dr Protection Environment for $600. Bdms know they migrated their mission critical workloads and Wallah stable and consistent Dr and now they're eliminating that site and looking for other migrations as well. The rate of 10 to 15 percent. It was just a great deal. One of the things I believe Andy, he'll customers should never spend capital, uh, Dr ever again with this kind of capability in place. That is just that game changing, you know, and you know, obviously we've been working on expanding our reach, you know, we promised to make the service available a year ago with the global footprint of Amazon and now we've delivered on that promise and in fact today or yesterday if you're an ozzie right down under, we announced in Sydney, uh, as well. And uh, now we're in US Europe and in APJ. Yeah. It's really, I mean it's very exciting. Of course Australia is one of the most virtualized places in the world and, and it's pretty remarkable how fast European customers have started using the offering to and just the quarter that's been out there and probably have the many requests customers has had. And you've had a, probably the number one request has been that we make the offering available in all the regions. The aws has regions and I can tell you by the end of 2019 will largely be there including with golf clubs and golf clap. You guys have been, that's been huge for you guys. Yeah. It's a government only region that we have that a lot of federal government workloads live in and we are pretty close together having the offering a fedramp authority to operate, which is a big deal on a game changer for governments because then there'll be able to use the familiar tools they use and vm ware not just to run their workloads on premises but also in the cloud as well with the data privacy requirements, security requirements they need. So it's a real game changer for government too. Yeah. And this you can see by the picture here basically before the end of next year, everywhere that you are and have an availability zone. We're going to be there running on data. Yup. Yeah. Let's get with it. Okay. We're a team go faster. Okay. You'll and you know, it's not just making it available, but this pace of innovation and you know, you guys have really taught us a few things in this respect and since we went live in the Oregon region, you know, we've been on a quarterly cadence of major releases and two was really about mission critical at scale and we added our second region. We added our hybrid cloud extension with m three. We moved the global rollout and we launched in Europe with m four. We really add a lot of these mission critical governance aspects started to attack all of the industry certifications and today we're announcing and five right. And uh, you know, with that, uh, I think we have this little cool thing you know, two of the most important priorities for that we're doing with ebs and storage. Yeah, we'll take, customers, our cost and performance. And so we have a couple of things to talk about today that we're bringing to you that I think hit both of those on a storage side. We've combined the elasticity of Amazon Elastic Block store or ebs with ware is Va v San and we've provided now a storage option that you'll be able to use that as much. It's very high capacity and much more cost effective and you'll start to see this initially on the Vm ware cloud. Native us are five instances which are compute instances, their memory optimized and so this will change the cost equation. You'll be able to use ebs by default and it'll be much more cost effective for storage or memory intensive workloads. Um, it's something that you guys have asked for. It's been very frequently requested it, it hits preview today. And then the other thing is that we've worked really hard together to integrate vm ware's Nsx along with aws direct neck to have a private even higher performance conductivity between on premises and the cloud. So very, very exciting new capabilities to show deep integration between the companies. Yeah. You know, in that aspect of the deep integration. So it's really been the thing that we committed to, you know, we have large engineering teams that are working literally every day. Right on bringing together and how do we fuse these platforms together at a deep and intimate way so that we can deliver new services just like elastic drs and the c and ebs really powerful, uh, capabilities and that pace of innovation continue. So next maybe. Um, maybe six. I don't know. We'll see. All right. You know, but we're continuing this toward pace of innovation, you know, completing all of the capabilities of Nsx. You'll full integration for all of the direct connect to capabilities. Really expanding that. You're only improving licensed capabilities on the platform. We'll be adding pks on top of for expanded developer a capabilities. So just. Oh, thank you. I, I think that was formerly known as Right, and y'all were continuing this pace of storage Chad. So anyway. innovation going forward, but I think we also have a few other things to talk about today. Andy. Yeah, I think we have some news that hopefully people here will be pretty excited about. We know we have a pretty big database business and aws and it's. It's both on the relational and on the nonrelational side and the business is billions of dollars in revenue for us and on the relational side. We have a service called Amazon relational database service or Amazon rds that we have hundreds of thousands of customers using because it makes it much easier for them to set up, operate and scale their databases and so many companies now are operating in hybrid mode and will be for a while and a lot of those customers have asked us, can you give us the ease of manageability of those databases but on premises. And so we talked about it and we thought about and we work with our partners at Vm ware and I'm excited to announce today, right now Amazon rds on Vm ware and so that will bring all the capabilities of Amazon rds to vm ware's customers for their on premises environments. And so what you'll be able to do is you'll be able to provision databases. You'll be able to scale the compute or the memory or the storage for those database instances. You'll be able to patch the operating system or database engines. You'll be able to create, read replicas to scale your database reads and you can deploy this rep because either on premises or an aws, you'll be able to deploy and high high availability configuration by replicating the data to different vm ware clusters. You'll be able to create online backups that either live on premises or an aws and then you'll be able to take all those databases and if you eventually want to move them to aws, you'll be able to do so rather easily. You have a pretty smooth path. This is going to be available in a few months. It will be available on Oracle sql server, sql postgresql and Maria DB. I think it's very exciting for our customers and I think it's also a good example of where we're continuing to deepen the partnership and listen to what customers want and then innovate on their behalf. Absolutely. Thank you andy. It is thrilling to see this and as we said, when we began the partnership, it was a deep integration of our offerings and our go to market, but also building this bi-directional hybrid highway to give customers the capabilities where they wanted cloud on premise, on premise to the cloud. It really is a unique partnership that we've built, the momentum we're feeling to our customer base and the cool innovations that we're doing. Andy, thank you so much for you Jordan Young, rural 20th. You guys appreciate it. Yeah, we really have just seen incredible momentum and as you might have heard from our earnings call that we just finished this. We finished the last quarter. We just really saw customer momentum here. Accelerating. Really exciting to see how customers are starting to really do the hybrid cloud at scale and with this we're just seeing that this vm ware cloud foundation available on Amazon available on premise. Very powerful, but it's not just the partnership with Amazon. We are thrilled to see the momentum of our Vm ware cloud provider program and this idea of the vm ware cloud providers has continued to gain momentum in the industry and go over five years. Right. This program has now accumulated more than 4,200 cloud partners in over 120 countries around the globe. It gives you choice, your local provider specialty offerings, some of your local trusted partners that you would have in giving you the greatest flexibility to choose from and cloud providers that meet your unique business requirements. And we launched last year a program called Vm ware cloud verified and this was saying you're the most complete embodiment of the Vm ware Cloud Foundation offering by our cloud partners in this program and this logo you know, allows you to that this provider has achieved the highest standard for cloud infrastructure and that you can scale and deliver your hybrid cloud and partnering with them. It know a particular. We've been thrilled to see the momentum that we've had with IBM as a huge partner and our business with them has grown extraordinarily rapidly and triple digits, but not just the customer count, which is now over 1700, but also in the depth of customers moving large portions of the workload. And as you see by the picture, we're very proud of the scope of our partnerships in a global basis. The highest standard of hybrid cloud for you, the Vm ware cloud verified partners. Now when we come back to this picture, you know we, you know, we're, we're growing in our definition of what the hybrid cloud means and through Vm Ware Cloud Foundation, we've been able to unify the private and the public cloud together as never before, but we're also seeing that many of you are interested in how do I extend that infrastructure further and farther and will simply call that the edge right? And how do we move data closer to where? How do we move data center resources and capacity closer to where the data's being generated at the operations need to be performed? Simply the edge and we'll dig into that a little bit more, but as we do that, what are the things that we offer today with what we just talked about with Amazon and our VCP p partners is that they can consume as a service this full vm ware Cloud Foundation, but today we're only offering that in the public cloud until project dimension of project dimension allows us to extend delivered as a service, private, public, and to the edge. Today we're announcing the tech preview, a project dimension Vm ware cloud foundation in a hyperconverged appliance. We're partnered deeply with Dell EMC, Lenovo for the first partners to bring this to the marketplace, built on that same proven infrastructure, a hybrid cloud control plane, so literally just like we're managing the Vm ware cloud today, we're able to do that for your on premise. You're small or remote office or your edge infrastructure through that exact same as a service management and control plane, a complete vm ware operated end to end environment. This is project dimension. Taking the vcf stack, the full vm ware cloud foundation stack, making an available in the cloud to the edge and on premise as well, a powerful solution operated by BM ware. This project dimension and project dimension allows us to have a fundamental building block in our approach to making customers even more agile, flexible, scalable, and a key component of our strategy as well. So let's click into that edge a little bit more and we think about the edge in the following layers, the compute edge, how do we get the data and operations and applications closer to where they need to be. If you remember last year I talked about this pendulum swinging of centralization and decentralization edge is a decentralization force. We're also excited that we're moving the edge of the devices as well and we're doing that in two ways. One with workspace, one for human optimized devices and the second is project pulse or Vm ware pulse. And today we're announcing pulse two point zero where you can consume it now as a service as well as with integrated security. And we've now scaled pulse to support 500 million devices. Isn't that incredible, right? I mean this is getting a scale. Billions and billions and finally networking is a key component. You all that. We're stretching the networking platform, right? And evolving how that edge operates in a more cloud and that's a service white and this is where Nsx St with Velo cloud is such a key component of delivering the edge of network services as well. Taken together the device side, the compute edge and rethinking and evolving the networking layer together is the vm ware edge strategy summary. We see businesses are on this multicloud journey, right? How do we then do that for their private of public coming together, the hybrid cloud, but they're also on a journey for how they work and operate it across the public cloud and the public cloud we have this torrid innovation, you'll want Andy's here, challenges. You know, he's announcing 1500 new services or were extraordinary innovation and you'll same for azure or Google Ibm cloud, but it also creates the same complexity as we said. Businesses are using multiple public clouds and how do I operate them? How do I make them work? You know, how do I keep track of my accounts and users that creates a set of cloud operations problems as well in the complexity of doing that. How do you make it work? Right? And your for that. We'll just see that there's this idea cloud cost compliance, analytics as these common themes that of, you know, keep coming up and we're seeing in our customers that are new role is emerging. The cloud operations role. You're the person who's figuring out how to make these multicloud environments work and keep track of who's using what and which data is landing where today I'm thrilled to tell you that the, um, where is acquiring the leader in this space? Cloudhealth technologies. Thank you. Cloudhealth technologies supports today, Amazon, azure and Google. They have some 3,500 customers, some of the largest and most respected brands in the, as a service industry. And Sasa business today rapidly span expanding feature sets. We will take cloudhealth and we're going to make it a fundamental platform and branded offering from the um, where we will add many of the other vm ware components into this platform, such as our wavefront analytics, our cloud, choreo compliance, and many of the other vm ware products will become part of the cloudhealth suite of services. We will be enabling that through our enterprise channels as well as through our MSP and BCPP partners as well know. Simply put, we will make cloudhealth the cloud operations platform of choice for the industry. I'm thrilled today to have Joe Consella, the CTO and founder. Joe, please stand up. Thank you joe to your team of a couple hundred, you know, mostly in Boston. Welcome to the Vm ware family, the Vm ware community. It is a thrill to have you part of our team. Thank you joe. Thank you. We're also announcing today, and you can think of this, much like we had v realize operations and v realize automation, the compliment to the cloudhealth operations, vm ware, cloud automation, and some of you might've heard of this in the past, this project tango. Well, today we're announcing the initial availability of Vm ware, cloud automation, assemble, manage complex applications, automate their provisioning and cloud services, and manage them through a brokerage the initial availability of cloud automation services, service. Your today, the acquisition of cloudhealth as a platform, the aware of the most complete set of multicloud management tools in the industry, and we're going to do so much more so we've seen this picture of this multicloud journey that our customers are on and you know, we're working hard to say we are going to bridge across these worlds of innovation, the multicloud world. We're doing many other things. You're gonna hear a lot at the show today about this year. We're also giving the tech preview of the Vm ware cloud marketplace for our partners and customers. Also today, Dell technologies is announcing their cloud marketplace to provide a self service, a portfolio of a Dell emc technologies. We're fundamentally in a unique position to accelerate your multicloud journey. So we've built out this any cloud piece, but right in the middle of that any cloud is the network. And when we think about the network, we're just so excited about what we have done and what we're seeing in the industry. So let's click into this a little bit further. We've gotten a lot done over the last five years. Networking. Look at these numbers. 80 million switch ports have been shipped. We are now 10 x larger than number two and software defined networking. We have over 7,500 customers running on Nsx and maybe the stat that I'm most proud of is 82 percent of the fortune 100 has now adopted nsx. You have made nsx these standard and software defined networking. Thank you very much. Thank you. When we think about this journey that we're on, we started. You're saying, Hey, we've got to break the chains inside of the data center as we said. And then Nsx became the software defined networking platform. We started to do it through our cloud provider partners. Ibm made a huge commitment to partner with us and deliver this to their customers. We then said, boy, we're going to make a fundamental to all of our cloud services including aws. We built this bridge called the hybrid cloud extension. We said we're going to build it natively into what we're doing with Telcos, with Azure and Amazon as a service. We acquired the St Wagon, right, and a Velo cloud at the hottest product of Vm ware's portfolio today. The opportunity to fundamentally transform branch and wide area networking and we're extending it to the edge. You're literally, the world has become this complex network. We have seen the world go from the old defined by rigid boundaries, simply put in a distributed world. Hardware cannot possibly work. We're empowering customers to secure their applications and the data regardless of where they sit and when we think of the virtual cloud network, we say it's these three fundamental things, a cloud centric networking fabric with intrinsic security and all of it delivered in software. The world is moving from data centers to centers of data and they need to be connected and Nsx is the way that we will do that. So you'll be aware of is well known for this idea of talking but also showing. So no vm world keynote is okay without great demonstrations of it because you shouldn't believe me only what we can actually show and to do that know I'm going to have our CTL come onstage and CTL y'all. I used to be a cto and the CTO is the certified smart guy. He's also known as the chief talking officer and today he's my demo partner. Please walk, um, Vm ware, cto ray to the stage. Right morning pat. How you doing? Oh, it's great ray, and thanks so much for joining us. Know I promised that we're going to show off some pretty cool stuff here. We've covered a lot already, but are you up to the task? We're going to try and run through a lot of demos. We're going to do it fast and you're going to have to keep me on time to ask an awkward question. Slow me down. Okay. That's my fault if you run along. Okay, I got it. I got it. Let's jump right in here. So I'm a CTO. I get to meet lots of customers that. A few weeks ago I met a cio of a large distribution company and she described her it infrastructure as consisting of a number of data centers troll to us, which he also spoke of a large number of warehouses globally, and each of these had local hyperconverged compute and storage, primarily running surveillance and warehouse management applications, and she pulls me four questions. The first question she asked me, she says, how do I migrate one of these data centers to Vm ware cloud on aws? I want to get out of one of these data centers. Okay. Sounds like something andy and I were just talking exactly, exactly what you just spoke to a few moments ago. She also wanted to simplify the management of the infrastructure in the warehouse as themselves. Okay. He's age and smaller data centers that you've had out there. Her application at the warehouses that needed to run locally, butter developers wanted to develop using cloud infrastructure. Cloud API is a little bit late. The rds we spoken with her in. Her final question was looking to the future, make all this complicated management go away. I want to be able to focus on my application, so that's what my business is about. So give me some new ways of how to automate all of this infrastructure from the edge to the cloud. Sounds pretty clear. Can we do it? Yes we can. So we're going to dive right in right now into one of these demos. And the first demo we're going to look at it is vm ware cloud on aws. This is the best solution for accelerating this public cloud journey. So can we start the demo please? So what you were looking at here is one of those data centers and you should be familiar with this product. It's a familiar vsphere client. You see it's got a bunch of virtual machines running in there. These are the virtual machines that we now want to be able to migrate and move the VMC on aws. So we're going to go through that migration right now. And to do that we use a product that you've seen already atx, however it's the x has been, has got some new cool features since the last time we download it. Probably on this stage here last year, I wanted those in particular is how do we do bulk migration and there's a new cool thing, right? Whole thing we want to move the data center en mass and his concept here is cloud motion with vsphere replication. What this does is it replicates the underlying storage of the virtual machines using vsphere replication. So if and when you want to now do the final migration, it actually becomes a vmotion. So this is what you see going on right here. The replication is in place. Now when you want to touch you move those virtual machines. What you'll do is a vmotion and the key thing to think about here is this is an actual vmotion. Those the ends as room as they're moving a hustler, migrating remained life just as you would in a v motion across one particular infrastructure. Did you feel complete application or data center migration with no dying town? It's a Standard v motion kind of appearance. Wow. That is really impressive. That's correct. Wow. You. So one of the other things we want to talk about here is as we are moving these virtual machines from the on prem infrastructure to the VMC on aws infrastructure, unfortunately when we set up the cloud on VMC and aws, we only set up for hosts, uh, that might not be, that'd be enough because she is going to move the whole infrastructure of that this was something you guys, you and Andy referred to briefly data center. Now, earlier, this concept of elastic drs. what elastic drs does, it allows the VMC on aws to react to the workloads as they're being created and pulled in onto that infrastructure and automatically pull in new hosts into the VMC infrastructure along the way. So what you're seeing here is essentially the MC growing the infrastructure to meet the needs of the workloads themselves. Very cool. So overseeing that elastic drs. we also see the ebs capabilities as well. Again, you guys spoke about this too. This is the ability to be able to take the huge amount of stories that Amazon have, an ebs and then front that by visa you get the same experience of v Sign, but you get this enormous amount of storage capabilities behind it. Wow. That's incredible. That's incredible. I'm excited about this. This is going to enable customers to migrate faster and larger than ever before. Correct. Now she had a series of little questions. Okay. The second question was around what about all those data centers and those age applications that I did not move, and this is where we introduce the project which you've heard of already tonight called project dementia. What this does, it gives you the simplicity of Vm ware cloud, but bringing that out to the age, you know what's basically going on here, vmc on aws is a service which manages your infrastructure in aws. We know stretch that service out into your infrastructure, in your data center and at the age, allowing us to be able to manage that infrastructure in the same way. Once again, let's dive down into a demo and take a look at what this looks like. So what you've got here is a familiar series of services available to you, one of them, which is project dimension. When you enter project dimension, you first get a view of all of the different infrastructure that you have available to you, your data centers, your edge locations. You can then dive deeply into one of these to get a closer look at what's going on here. We're diving into one of these The problem is there's a networking problem going on in this warehouse. warehouses and we see it as a problem here. How do we know? We know because vm ware is running this as a managed service. We are directly managing or sorry, monitoring your infrastructure or we discover there's something going wrong here. We automatically create the ASR, so somebody is dealing with this. You have visibility to what's going on, but the vm ware managed service is already chasing the problem for you. Oh, very good. So now we're seeing this dispersed infrastructure with project dementia, but what's running on it so well before we get with running out, you've got another problem and the problem is of course, if you're managing a lot of infrastructure like this, you need to keep it up to date. And so once again, this is where the vm ware managed service kicks in. We manage that infrastructure in terms of patching it and updating it for you. And as an example, when we released a security patch, here's one for the recent l, one terminal fault, the Vmr managed service is already on that and making sure that your on prem and edge infrastructure is up to date. Very good. Now, what's running? Okay. So what's running, uh, so we mentioned this case of this software running at the edge infrastructure itself, and these are workloads which are running locally in those age, uh, those edge locations. This is a surveillance application. You can see it here at the bottom it says warehouse safety monitor. So this is an application which gathers images and then stores those images He said my sql database on top there, now this is where we leverage the somewhere and it puts them in a database. technology you just learned about when Andy and pat spoke about disability to take rds and run that on your on prem infrastructure. The block of virtual machines in the moment are the rds components from Amazon running in your infrastructure or in your edge location, and this gives you the ability to allow your developers to be able to leverage and operate against those Apis, but now the actual database, the infrastructure is running on prem and you might be doing just for performance reasons because of latency, you might be doing it simply because this data center is not always connected to the cloud. When you take a look into under the hood and see what's going on here, what you actually see this is vsphere, a modified version of vsphere. You see this new concept of my custom availability zone. That is the availability zone running on your infrastructure which supports or ds. What's more interesting is you flip back to the Amazon portal. This is typically what your developers are going to do. Once again, you see an availability zone in your Amazon portal. This is the availability zone running on your equipment in your data center. So we've truly taken that already as infrastructure and moved it to the edge so the developer sees what they're comfortable with and the infrastructure sees what they're comfortable with bridging those two worlds. Fabulous. Right. So the final question of course that we got here was what's next? How do I begin to look to the future and say I am going to, I want to be able to see all of my infrastructure just handled in an automated fashion. And so when you think about that, one of the questions there is how do we leverage new technologies such as ai and ml to do that? So what you've got here is, sorry we've got a little bit later. What you've got here is how do I blend ai in a male and the power of what's in the data center itself. Okay. And we could do that. We're bringing you the AI and ml, right? And fusing them together as never before to truly change how the data center operates. Correct. And it is this introduction is this merging of these things together, which is extremely powerful in my mind. This is a little bit like a self driving vehicle, so thinking about a car driving down the street is self driving vehicle, it is consuming information from all of the environment around it, other vehicles, what's happening, everything from the wetter, but it also has a lot of built in knowledge which is built up to to self learning and training along the way in the kids collecting lots of that data for decades. Exactly. And we've got all that from all the infrastructure that we have. We can now bring that to bear. So what we're focusing on here is a project called project magna and project. Magna leverage is all of this infrastructure. What it does here is it helps connect the dots across huge datasets and again a deep insight across the stack, all the way from the application hardware, the infrastructure to the public cloud, and even the age and what it does, it leverages hundreds of control points to optimize your infrastructure on Kpis of cost performance, even user specified policies. This is the use of machine language in order to fundamentally transform. I'm sorry, machine learning. I'm going back to some. Very early was here, right? This is the use of machine learning and ai, which will automatically transform. How do you actually automate these data centers? The goal is true automation of your infrastructure, so you get to focus on the applications which really served needs of your business. Yeah, and you know, maybe you could think about that as in the past we would have described the software defined data center, but in the future we're calling it the self driving data center. Here we are taking that same acronym and redefining it, right? Because the self driving data center, the steep infusion of ai and machine learning into the management and automation into the storage, into the networking, into vsphere, redefining the self driving data center and with that we believe fundamentally is to be an enormous advance and how they can take advantage of new capabilities from bm ware. Correct. And you're already seeing some of this in pieces of projects such as some of the stuff we do in wavefront and so already this is how do we take this to a new level and that's what project magnet will do. So let's summarize what we've seen in a few demos here as we work in true each of these very quickly going through these demos. First of all, you saw the n word cloud on aws. How do I migrate an entire data center to the cloud with no downtime? Check, we saw project dementia, get the simplicity of Vm ware cloud in the data center and manage it at the age as a managed service check. Amazon rds and Vm ware. Cool Demo, seamlessly deploy a cloud service to an on premises environment. In this case already. Yes, we got that one coming in are in m five. And then finally project magna. What happens when you're looking to the future? How do we leverage ai and ml to self optimize to virtual infrastructure? Well, how did ray do as our demo guy? Thank you. Thanks. Thanks. Right. Thank you. So coming back to this picture, our gps for the day, we've covered any cloud, let's click into now any application, and as we think about any application, we really view it as this breadth of the traditional cloud native and Sas Coobernetti is quickly maybe spectacularly becoming seen as the consensus way that containers will be managed and automate as the framework for how modern APP teams are looking at their next generation environment, quickly emerging as a key to how enterprises build and deploy their applications today. And containers are efficient, lightweight, portable. They have lots of values for developers, but they need to also be run and operate and have many infrastructure challenges as well. Managing automation while patch lifecycle updates, efficient move of new application services, know can be accelerated with containers. We also have these infrastructure problems and you know, one thing we want to make clear is that the best way to run a container environment is on a virtual machine. You know, in fact, every leader in public cloud runs their containers and virtual machines. Google the creator and arguably the world leader in containers. They runs them all in containers. Both their internal it and what they run as well as G K, e for external users as well. They just announced gke on premise on vm ware for their container environments. Google and all major clouds run their containers and vms and simply put it's the best way to run containers. And we have solved through what we have done collectively the infrastructure problems and as we saw earlier, cool new container apps are also typically some ugly combination of cool new and legacy and existing environments as well. How do we bridge those two worlds? And today as people are rapidly moving forward with containers and Coobernetti's, we're seeing a certain set of problems emerge. And Dan cone, right, the director of CNCF, the Coobernetti, uh, the cloud native computing foundation, the body for Coobernetti's collaboration and that, the group that sort of stewards the standardization of this capability and he points out these four challenges. How do you secure them? How do you network and you know, how do you monitor and what do you do for the storage underneath them? Simply put, vm ware is out to be, is working to be is on our way to be the dial tone for Coobernetti's. Now, some of you who were in your twenties might not know what that means, so we know over to a gray hair or come and see me afterward. We'll explain what dial tone means to you or maybe stated differently. Enterprise grade standard for Cooper netties and for that we are working together with our partners at Google as well as pivotal to deliver Vm ware, pks, Cooper netties as an enterprise capability. It builds on Bosh. The lifecycle engine that's foundational to the pivotal have offerings today, uh, builds on and is committed to stay current with the latest Coobernetti's releases. It builds on Nsx, the SDN container, networking and additional contributions that were making like harbor the Vm ware open source contribution for the container registry. It packages those together makes them available on a hybrid cloud as well as public cloud environments with pks operators can efficiently deploy, run, upgrade their coopernetties environments on SDDC or on all public clouds. While developers have the freedom to embrace and run their applications rapidly and efficiently, simply put, pks, the standard for Coobernetti's in the enterprise and underneath that Nsx you'll is emerging as the standard for software defined networking. But when we think about and we saw that quote on the challenges of Kubernetes today, we see that networking is one of the huge challenge is underneath that and in a containerized world, things are changing even more rapidly. My network environment is moving more quickly. NSX provides the environment's easily automate networking and security for rapid deployment of containerized environments that fully supports the MRP chaos, fully supports pivotal's application service, and we're also committed to fully support all of the major kubernetes distribution such as red hat, heptio and docker as well Nsx, the only platform on the planet that can address the complexity and scale of container deployments taken together Vm Ware, pks, the production grade computer for the enterprise available on hybrid cloud, available on major public clouds. Now, let's not just talk about it again. Let's see it in action and please walk up to the stage. When di Carter with Ray, the senior director of cloud native marketing for Vm ware. Thank you. Hi everybody. So we're going to talk about pks because more and more new applications are built using kubernetes and using containers with vm ware pts. We get to simplify the deploying and the operation of Kubernetes at scale. When the. You're the experts on all of this, right? So can you take as true the scenario of how pks or vm ware pts can really help a developer operating the Kubernedes environment, developed great applications, but also from an administrator point of view, I can really handle things like networking, security and those configurations. Sounds great. I love to dive into the demo here. Okay. Our Demo is. Yeah, more pks running coubernetties vsphere. Now pks has a lot of cool functions built in, one of which is Nsx. And today what I'm going to show you is how NSX will automatically bring up network objects as quick Coobernetti's name spaces are spun up. So we're going to start with the fees per client, which has been extended to Ron pks, deployed cooper clusters. We're going to go into pks instance one, and we see that there are five clusters running. We're going to select one other clusters, call application production, and we see that it is running nsx. Now a cluster typically has multiple users and users are assigned namespaces, and these namespaces are essentially a way to provide isolation and dedicated resources to the users in that cluster. So we're going to check how many namespaces are running in this cluster and more brought up the Kubernetes Ui. We're going to click on namespace and we see that this cluster currently has four namespaces running wire. We're going to do next is bringing up a new name space and show that Nsx will automatically bring up the network objects required for that name space. So to do that, we're going to upload a Yammel file and your developer may actually use Ku Kata command to do this as well. We're going to check the namespace and there it is. We have a new name space called pks rocks. Yeah. Okay. Now why is that guy now? It's great. We have a new name space and now we want to make sure it has the network elements assigned to us, so we're going to go to the NSX manager and hit refresh and there it is. PKS rocks has a logical robber and a logical switch automatically assigned to it and it's up and running. So I want to interrupt here because you made this look so easy, right? I'm not sure people realize the power of what happened here. The developer, winton using Kubernetes, is api infrastructure to familiar with added a new namespace and behind the scenes pks and tardy took care of the networking. It combination of Nsx, a combination of what we do at pks to truly automate this function. Absolutely. So this means that if you are on the infrastructure operation, you don't need to worry about your developer springing up namespaces because Nsx will take care of bringing the networking up and then bringing them back down when the namespace is not used. So rate, but that's not it. Now, I was in operations before and I know how hard it is for enterprises to roll out a new product without visibility. Right, so pks took care of those dates, you operational needs as well, so while it's running your clusters, it's also exporting Meta data so that your developers and operators can use wavefront to gain deep visibility into the health of the cluster as well as resources consumed by the cluster. So here you see the wavefront Ui and it's showing you the number of nodes running, active parts, inactive pause, et cetera. You can also dive deeper into the analytics and take a look at information site, Georgia namespace, so you see pks rocks there and you see the number of active nodes running as well as the CPU utilization and memory consumption of that nice space. So now pks rocks is ready to run containerized applications and microservices. So you just get us a very highlight of a demo here to see a little bit what pks pks says, where can we learn more? So we'd love to show you more. Please come by the booth and we have more cool functions running on pks and we'd love to have you come by. Excellent. Thank you, Lindy. Thank you. Yeah, so when we look at these types of workloads now running on vsphere containers, Kubernedes, we also see a new type of workload beginning to appear and these are workloads which are basically machine learning and ai and in many cases they leverage a new type of infrastructure, hardware accelerators, typically gps. What we're going to talk about here is how in video and Vm ware have worked together to give you flexibility to run sophisticated Vdi workloads, but also to leverage those same gpu for deep learning inference workloads also on vsphere. So let's dive right into a demo here. Again, what you're seeing here is again, you're looking at here, you're looking at your standard view realized operations product, and you see we've got two sets of applications here, a Vdi desktop workload and machine learning, and the graph is showing what's happening with the Vdi desktops. These are office workers leveraging these desktops everyday, so of course the infrastructure is super busy during the daytime when they're in the office, but the green area shows this is not been used very heavily outside of those times. So let's take a look. What happens to the machine learning application in this case, this organization leverages those available gpu to run the machine learning operations outside the normal working hours. Let's take a little bit of a deeper dive into what the application it is before we see what we can do from an infrastructure and configuration point of view. So this machine learning application processes a vast number of images and it clarify or sorry, it categorizes these images and as it's doing so, it is moving forward and putting each of these in a database and you can see it's operating here relatively fast and it's leveraging some gps to do that. So typical image processing type of machine learning problem. Now let's take a dive in and look at the infrastructure which is making this happen. First of all, we're going to look only at the Vdi employee Dvt, a Vdi infrastructure here. So I've got a bunch of these applications running Vdi applications. What I want to do is I want to move these so that I can make this image processing out a application run a lot faster. Now normally you wouldn't do this, but pot insisted that we do this demo at 10:30 in the morning when the office workers are in there, so we're going to move older Vdi workloads over to the other cluster and that's what you're seeing is going on right now. So as they move over to this other cluster, what we are now doing is freeing up all of the infrastructure. The GPU that Vdi workload was using here. We see them moving across and now you've freed up that infrastructure. So now we want to take a look at this application itself, the machine learning application and see how we can make use of that. Now freed up infrastructure we've got here is the application is running using one gpu in a vsphere cluster, but I've got three more gpu is available now because I've moved the Vdi workloads. We simply modify the application, let it know that these are available and you suddenly see an increase in the processing capabilities because of what we've done here in terms of making the flexibility of accessing those gps. So what you see here is the same gps that youth for Vdi, which you probably have in your infrastructure today, can also be used to run sophisticated machine learning and ai type of applications on your vsphere infrastructure. So let's summarize what we've seen in the various demos here in this section. First of all, we saw how the MRPS simplifies the deployment and operating operation of Kubernetes at scale. What we've also seen is that leveraging the Nvidia Gpu, we can now run the most demanding workloads on vsphere. When we think about all of these applications and these new types of workloads that people are running. I want to take one second to speak to another workload that we're seeing beginning to appear in the data center. And this is of course blockchain. We're seeing an increasing number of organizations evaluating blockchains for smart contract and digital consensus solutions. So this tech, this technology is really becoming or potentially becoming a critical role in how businesses will interact each other, how they will work together. We'd project concord, which is an open source project that we're releasing today. You get the choice, performance and scale of verifiable trust, which you can then bring to bear and run in the enterprise, but this is not just another blockchain implementation. We have focused very squarely on making sure that this is good for enterprises. It focuses on performance, it focuses on scalability. We have seen examples where running consensus algorithms have taken over 80 days on some of the most common and widely used infrastructure in blockchain and we project conquered. You can do that in two and a half hours. So I encourage you to check out this project on get hub today. You'll also see lots of activity around the whole conference. Speaking about this. Now we're going to dive into another section which is the anti device section. And for that I need to welcome pat back up there. Thank you pat. Thanks right. So diving into any device piece of the puzzle, you and as we think about the superpowers that we have, maybe there are no more area that they are more visible than in the any device aspect of our picture. You know, and as we think about this, the superpowers, you know, think about mobility, right? You know, and how it's enabling new things like desktop as a service in the mobile area, these breadth of smartphones and devices, ai and machine learning allow us to manage them, secure them and this expanding envelope of devices in the edge that need to be connected and wearables and three d printers and so on. We've also seen increasing research that says engaged employees are at the center of business success. Engaged employees are the critical ingredient for digital transformation. And frankly this is how I run vm ware, right? You know, I have my device and my work, all my applications, every one of my 23,000 employees is running on our transformed workspace one environment. Research shows that companies that, that give employees ready anytime access are nearly three x more likely to be leaders in digital transformation. That employees spend 20 percent of their time today on manual processes that can be automated. The way team collaboration and speed of division decisions increases by 16 percent with engaged employees with modern devices. Simply put this as a critical aspect to enabling your business, but you remember this picture from the silos that we started with and each of these environments has their own tribal communities of management, security automation associated with them, and the complexity associated with these is mind boggling and we start to think about these. Remember the I'm a pc and I'm a Mac. Well now you have. I'm an Ios. I'm a droid and other bdi and I'm now a connected printer and I'm a connected watch. You remember citrix manager and good is now bad and sccm a failed model and vpns and Xanax. The chaos is now over at the center of that is vm ware, workspace one, get it out of the business of managing devices, automate them from the cloud, but still have the mentor price. Secure cloud based analytics that brings new capabilities to this critical topic. You'll focus your energy on creating employee and customer experiences. You know, new capabilities to allow like our airlift, the new capability to help customers migrate from their sccm environment to a modern management, expanding the use of workspace intelligence. Last year we announced the chromebook and a partnership with HP and today I'm happy to announce the next step in our partnerships with Dell. And uh, today we're announcing that Dell provisioning for Vm ware, workspace one as part of Dell's ready to work solutions Dallas, taking the next leap and bringing workspace one into the core of their client to offerings. And the way you can think about this as Literally a dell drop ship, lap pops showing up to new employee. day one, productivity. You give them their credential and everything else is delivered by workspace one, your image, your software, everything patched and upgraded, transforming your business, right beginning at that device experience that you give to your customer. And again, we don't want to talk about it. We want to show you how this works. Please walk to the stage with re renew the head of our desktop products marketing. Thank you. So we just heard from pat about how workspace one integrated with Dell laptops is really set up to manage windows devices. What we're broadly focused on here is how do we get a truly modern management system for these devices, but one that has an intelligence behind it to make sure that we're kept with a good understanding of how to keep these devices always up to date and secure. Can we start the demo please? So what we're seeing here is to be the the front screen that you see of workspace one and you see you've got multiple devices a little bit like that demo that patch assured. I've got Ios, android, and of course I've got windows renewal. Can you please take us through how workspace one really changes the ability of somebody an it administrator to update and manage windows into our environment? Absolutely. With windows 10, Microsoft has finally joined the modern management body and we are really excited about that. Now. The good news about modern management is the frequency of ostp updates and how quickly they come out because you can address all those security issues that are hitting our radar on a daily basis, but the bad news about modern management is the frequency of those updates because all of us in it admins, we have to test each and every one of our applications would that latest version because we don't want to roll out that update in case of causes any problems with workspace one, we saw that we simply automate and provide you with the APP compatibility information right out of the box so you can now automate that update process. Let's take a quick look. Let's drill down here further into the windows devices. What we'll see is that only a small percentage of those devices are on that latest version of operating system. Now, that's not a good thing because it might have an important security fix. Let's scroll down further and see what the issue is. We find that it's related to app compatibility. In fact, 38 percent of our devices are blocked from being upgraded and the issue is app compatibility. Now we were able to find that not by asking the admins to test each and every one of those, but we combined windows analytics data with APP intelligent out of the box and be provided that information right here inside of the console. Let's dig down further and see what those devices and apps look like. So knew this is the part that I find most interesting. If I am a system administrator at this point I'm looking at workspace one is giving me a key piece of information. It says if you proceed with this update, it's going to fail 84, 85 percent at a time. So that's an important piece of information here, but not alone. Is it telling me that? It is telling me roughly speaking why it thinks it's going to fail. We've got a number of apps which are not ready to work with this new version, particularly the Mondo card sales lead tracker APP. So what we need to do is get engineering to tackle the problems with this app and make sure that it's updated. So let's get fixing it in order to fix it. What we'll do is create an automation and we can do this right out of the box in this automation will open up a Jira ticket right from within the console to inform the engineers about the problem, not just that we can also flag and send a notification to that engineering manager so that it's top of mine and they can get working on this fixed right away. Let's go ahead and save that automation right here, ray UC. There's the automation that we just So what's happening here is essentially this update is now scheduled meeting. saved. We can go and update oldest windows devices, but workspace one is holding the process of proceeding with that update, waiting for the engineers to update the APP, which is going to cause the problem. That's going to take them some time, right? So the engineers have been working on this, they have a fixed and let's go back and see what's happened to our devices. So going back into the ios updates, what we'll find is now we've unblocked those devices from being upgraded. The 38 percent has drastically dropped down. It can rest in peace that all of the devices are compliant and on that latest version of operating system. And again, this is just a snapshot of the power of workspace one to learn more and see more. I invite you all to join our EOC showcase keynote later this evening. Okay. So we've spoken about the presence of these new devices that it needs to be able to manage and operate across everything that they do. But what we're also seeing is the emergence of a whole new class of computing device. And these are devices which are we commonly speak to have been at the age or embedded devices or Iot. And in many cases these will be in factories. They'll be in your automobiles, there'll be in the building, controlling, controlling, uh, the building itself, air conditioning, etc. Are quite often in some form of industrial environment. There's something like this where you've got A wind farm under embedded in each of these turbines. This is a new class of computing which needs to be managed, secured, or we think virtualization can do a pretty good job of that in new virtualization frontier, right at the edge for iot and iot gateways, and that's gonna. That's gonna, open up a whole new realm of innovation in that space. Let's dive down and taking the demo. This spaces. Well, let's do that. What we're seeing here is a wind turbine farm, a very different than a data center than what we're used to and all the compute infrastructure is being managed by v center and we see to edge gateway hose and they're running a very mission critical safety watchdog vm right on there. Now the safety watchdog vm is an fte mode because it's collecting a lot of the important sensor data and running the mission critical operations for the turbine, so fte mode or full tolerance mode, that's a pretty sophisticated virtualization feature allowing to applications to essentially run in lockstep. So if there's a failure, wouldn't that gets to take over immediately? So this no sophisticated virtualization feature can be brought out all the way to the edge. Exactly. So just like in the data center, we want to perform an update, so as we performed that update, the first thing we'll do is we'll suspend ft on that safety watchdog. Next, we'll put two. Oh, five into maintenance mode. Once that's done, we'll see the power of emotion that we're all familiar with. We'll start to see all the virtual machines vmotion over to the second backup host. Again, all the maintenance, all the update without skipping a heartbeat without taking down any daily operations. So what we're seeing here is the basic power of virtualization being brought out to the age v motion maintenance mode, et cetera. Great. What's the big deal? We've been doing that for years. What's the, you know, come on. What's the big deal? So what you're on the edge. So when you get to the age pack, you're dealing with a whole new class of infrastructure. You're dealing with embedded systems and new types of cpu hours and process. This whole demo has been done on an arm 64. Virtualization brought to arm 64 for embedded devices. So we're doing this on arm on the edge, correct. Specifically focused for embedded for age oems. Okay. Now that's good. Okay. Thank you ray. Actually, we've got a summary here. Pat, just a second before you disappear. A lot to rattle off what we've just seen, right? We've seen workspace one cross platform management. What we've also seen, of course esx for arm to bring the power of vfx to edge on 64, but are in platforms will go no. Okay. Okay. Thank you. Thanks. Now we've seen a look at a customer who is taking advantage of everything that we just saw and again, a story of a customer that is just changing lives in a fundamental way. Let's see. Make a wish. So when a family gets the news that a child is sick and it's a critical illness, it could be a life threatening illness. The whole family has turned upside down. Imagine somebody comes to you and they say, what's the one thing you want that's in your heart? You tell us and then we make that happen. So I was just calling to give you the good news that we're going to be able to grant jackson a wish make, which is the largest wish granting organizations in the United States. English was featured in the cbs 60 minutes episode. Interestingly, it got a lot of hits, but uh, unfortunately for the it team, the whole website crashed make a wish is going through a program right now where we're centralizing technology and putting certain security standards in place at our chapters. So what you're seeing here, we're configuring certain cloud services to make sure that they always are able to deliver on the mission whether they have a local problem or not is we continue to grow the partnership and work with vm ware. It's enabling us to become more efficient in our processes and allows us to grant more wishes. It was a little girl. She had a two year old brother. She just wanted a puppy and she was forthright and I want to name the puppy in my name so my brother would always have me to list them off a five year old. It's something we can't change their medical outcome, but we can change their spiritual outcome and we can transform their lives. Thank you. Working together with you truly making wishes come true. The last topic I want to touch on today, and maybe the most important to me personally is security. You got to fundamentally, when we think about this topic of security, I'll say it's broken today and you know, we would just say that the industry got it wrong that we're trying to bolt on or chasing bad, and when we think about our security spend, we're spending more and we're losing more, right? Every day we're investing more in this aspect of our infrastructure and we're falling more behind. We believe that we have to have much less security products and much more security. You know, fundamentally, you know, if you think about the problem, we build infrastructure, right? Generic infrastructure, we then deploy applications, all kinds of applications, and we're seeing all sorts of threats launched that as daily tens of millions. You're simple virus scanner, right? Is having tens of millions of rules running and changing many times a day. We simply believe the security model needs to change. We need to move from bolted on and chasing bad to an environment that has intrinsic security and is built to ensure good. This idea of built in security. We are taking every one of the core vm ware products and we are building security directly into it. We believe with this, we can eliminate much of the complexity. Many of the sensors and agents and boxes. Instead, they'll directly leverage the mechanisms in the infrastructure and we're using that infrastructure to lock it down to behave as we intended it to ensure good, right on the user side with workspace one on the network side with nsx and microsegmentation and storage with native encryption and on the compute with app defense, we are building in security. We're not chasing threats or adding on, but radically reducing the attack surface. When we look at our applications in the data center, you see this collection of machines running inside of it, right? You know, typically running on vsphere and those machines are increasingly connected. Through nsx and last year we introduced the breakthrough security solution called app defense and app defense. Leverages the unique insight we get into the application so that we can understand the application and map it into the infrastructure and then you can lock down, you could take that understanding, that manifest of its behavior and then lock those vms to that intended behavior and we do that without the operational and performance burden of agents and other rear looking use of attack detection. We're shrinking the attack surface, not chasing the latest attack vector, you know, and this idea of bolt on versus chasing bad. You sort of see it right in the network. Machines have lots of conductivity, lots of applications running and something bad happens. It basically has unfettered access to move horizontally through the data center and most of our security is north, south. MosT of the attacks are eastwest. We introduced this idea of microsegmentation five years ago, and by it we're enabling organizations to secure some networks and separate sensitive applications and services as never before. This idea isn't new, that just was never practical before nsx, but we're not standing still. Our teams are innovating to leap beyond 12. What's next beyond microsegmentation, and we see this in three simple words, learn, imagine a system that can look into the applications and understand their behavior and how they should operate. we're using machine learning and ai instead of chasing were to be able to ensure good where that that system can then locked down its behavior so the system consistently operates that way, but finally we know we have a world of increasing dynamic applications and as we move to more containerize the microservices, we know this world is changing, so we need to adapt. We need to have more automation to adapt to the current behavior. Today I'm very excited to have two major announcements that are delivering on this vision. The first of those vsphere platinum, our flagship vm ware vsphere product now has app defense built right in platinum will enable virtualization teams. Yeah, go ahead. Yeah, let's use it. Platinum will enable virtualization teams you to give an enormous contribution to the security profile of your enterprise. You could see whatever vm is for its purpose, its behavior until the system. That's what it's allowed to do. Dramatically reducing the attack surface without impact. On operations or performance, the capability is so powerful, so profound. We want you to be able to leverage it everywhere, and that's why we're building it directly into vsphere, vsphere platinum. I call it the burger and fries. You know, nobody leaves the restaurant without the fries who would possibly run a vm in the future without turning security on. That's how we want this to work going forward. Vsphere platinum and as powerful as microsegmentation has been as an idea. We're taking the next step with what we call adaptive microsegmentation. We are fusing Together app defense and vsphere with nsx to allow us to align the policies of the application through vsphere and the network. We can then lock down the network and the compute and enable this automation of the microsegment formation taken together adaptive microsegmentation. But again, we don't want to just tell you about it. We want to show you. Please welcome to the stage vj dante, who heads our machine learning team for app dispense. Vj a very good vj. Thanks for joining us. So, you know, I talked about this idea right, of being able to learn, lock and adapt. Uh, can you show it to us? Great. Yeah. Thank you. With vc a platinum, what we have done is we have put in everything you need to learn, lock and adapt, right with the infrastructure. The next time you bring up your wifi at line, you'll actually see a difference right in there. Let's go with that demo. There you go. And when you look at our defense there, what you see is that all your guests, virtual machines and all your host, hundreds of them and thousands of virtual machines enabling for that difference. It's in there. And what that does is immediately gets you visibility into the processes running on those virtual machines and the risk for the first time. Think about it for the first time. You're looking at the infrastructure through the lens of an application. Here, for example, the ecommerce application, you can see the components that make up that application, how they interact with each other, the specific process, a specific ip address on a specific board. That's what you get, but so we're learning the behavior. Yes. Yeah, that's very good. But how do you make sure you only learn good behavior? Exactly. How do we make sure that it's not bad? We actually verify me insured. It's all good. We ensured that everybody these reputation is verified. We ensured that the haven is verified. Let's go to svc host, for example. This process can exhibit hundreds of behaviors across numerous. Realize what we do here is we actually verify that failure saw us. It's actually a machine learning models that had been trained on millions of instances of good, bad at you said, and then automatically verify that for okay, so we said, you. We learned simply, learn now, lock. How does that work? Well, once you learned the application, locking it is as simple as clicking on that verify and protect button and then you can lock both the compute and network and it's done. So we've pushed those policies into nsx and microsegmentation has been established actually locked down the compute. What is the operating system is exactly. Let's first look at compute, protected the processes and the behaviors are locked down to exactly what is allowed for that application. And we have bacon policies and program your firewall. This is nsx being configured automatically for you, laurie, with one single click. Very good. So we said learn lock. Now, how does this adapt thing work? Well, a bad change is the only constant, but modern applications applications change on a continuous basis. What we do is actually pretty simple. We look at every change as it comes in determinant is good or bad. If it's good, we say allow it, update the policies. That's bad. We denied. Let's look at an example as asco dxc. It's exhibiting a behavior that they've not seen getting the learning period. Okay? So this machine has never behave this This hasn't been that way. But. way. But again, our machine learning models had seen thousands of instances of this process. They know this is normal. It talks on three 89 all the time. So what it's done to the few things, it's lowered the criticality of the alarm. Okay, so false positive. Exactly. The bane of security operations, false positives, and it has gone and updated. Jane does locks on compute and network to allow for that behavior. Applications continues to work on this project. Okay, so we can learn and adapt and action right through the compute and the network. What about the client? Well, we do with workplace one, intelligence protect and manage end user endpoint, but what's one intelligence? Nsx and actually work together to protect your entire data center infrastructure, but don't believe me. You can watch it for yourself tomorrow tom cornu keynote. You want to be there, at 1:00 PM, be there or be nowhere. I love you. Thank you veejay. Great job. Thank you so much. So the idea of intrinsic security and ensuring good, we believe fundamentally changing how security will be delivered in the enterprise in the future and changing the entire security industry. We've covered a lot today. I'm thrilled as I stand on stage to stand before this community that truly has been at the center of changing the world of technology over the last couple of decades. In it. We've talked about this idea of the super powers of technology and as they accelerate the huge demand for what you do, you know in the same way we together created this idea of the virtual infrastructure admin. You'll think about all the jobs that we are spawning in the discussion that we had today, the new skills, the new opportunities for each one of us in this room today, quantum program, machine learning engineer, iot and edge expert. We're on the cusp of so many new capabilities and we need you and your skills to do that. The skills that you possess, the abilities that you have to work across these silos of technology and enabled tomorrow. I'll tell you, I am now 38 years in the industry and I've never been more excited because together we have the opportunity to build on the things that collective we have done over the last four decades and truly have a positive global impact. These are hard problems, but I believe together we can successfully extend the lifespan of every human being. I believe together we can eradicate chronic diseases that have plagued mankind for centuries. I believe we can lift the remaining 10 percent of humanity out of extreme poverty. I believe that we can reschedule every worker in the age of the superpowers. I believe that we can give modern ever education to every child on the planet, even in the of slums. I believe that together we could reverse the impact of climate change. I believe that together we have the opportunity to make these a reality. I believe this possibility is only possible together with you. I asked you have a please have a wonderful vm world. Thanks for listening. Happy 20th birthday. Have a great topic.
SUMMARY :
of devices in the edge that need to be
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amazon | ORGANIZATION | 0.99+ |
Andy | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Michael | PERSON | 0.99+ |
1998 | DATE | 0.99+ |
Telco | ORGANIZATION | 0.99+ |
1986 | DATE | 0.99+ |
Telcos | ORGANIZATION | 0.99+ |
Europe | LOCATION | 0.99+ |
Paul Maritz | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
Lenovo | ORGANIZATION | 0.99+ |
10 | QUANTITY | 0.99+ |
Deloitte | ORGANIZATION | 0.99+ |
Joe | PERSON | 0.99+ |
Sydney | LOCATION | 0.99+ |
Joe Consella | PERSON | 0.99+ |
Africa | LOCATION | 0.99+ |
Pat Gelsinger | PERSON | 0.99+ |
Oregon | LOCATION | 0.99+ |
20 percent | QUANTITY | 0.99+ |
Ashley | PERSON | 0.99+ |
16 percent | QUANTITY | 0.99+ |
Vegas | LOCATION | 0.99+ |
Jupiter | LOCATION | 0.99+ |
Last year | DATE | 0.99+ |
last year | DATE | 0.99+ |
first question | QUANTITY | 0.99+ |
Lindy | PERSON | 0.99+ |
telco | ORGANIZATION | 0.99+ |
John Gage | PERSON | 0.99+ |
10 percent | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Dan cone | PERSON | 0.99+ |
68 percent | QUANTITY | 0.99+ |
200 applications | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
50 percent | QUANTITY | 0.99+ |
Vm Ware Cloud Foundation | ORGANIZATION | 0.99+ |
1440 | DATE | 0.99+ |
30 year | QUANTITY | 0.99+ |
HP | ORGANIZATION | 0.99+ |
38 percent | QUANTITY | 0.99+ |
38 years | QUANTITY | 0.99+ |
$600 | QUANTITY | 0.99+ |
20 years | QUANTITY | 0.99+ |
one months | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
windows 10 | TITLE | 0.99+ |
hundreds | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
80 million | QUANTITY | 0.99+ |
five percent | QUANTITY | 0.99+ |
second question | QUANTITY | 0.99+ |
Jody | PERSON | 0.99+ |
Today | DATE | 0.99+ |
tomorrow | DATE | 0.99+ |
Sanjay | PERSON | 0.99+ |
23,000 employees | QUANTITY | 0.99+ |
five people | QUANTITY | 0.99+ |
sixth year | QUANTITY | 0.99+ |
82 percent | QUANTITY | 0.99+ |
five instances | QUANTITY | 0.99+ |
tomorrow morning | DATE | 0.99+ |
Coobernetti | ORGANIZATION | 0.99+ |