Image Title

Search Results for two major providers:

Jack Greenfield, Walmart | A Dive into Walmart's Retail Supercloud


 

>> Welcome back to SuperCloud2. This is Dave Vellante, and we're here with Jack Greenfield. He's the Vice President of Enterprise Architecture and the Chief Architect for the global technology platform at Walmart. Jack, I want to thank you for coming on the program. Really appreciate your time. >> Glad to be here, Dave. Thanks for inviting me and appreciate the opportunity to chat with you. >> Yeah, it's our pleasure. Now we call what you've built a SuperCloud. That's our term, not yours, but how would you describe the Walmart Cloud Native Platform? >> So WCNP, as the acronym goes, is essentially an implementation of Kubernetes for the Walmart ecosystem. And what that means is that we've taken Kubernetes off the shelf as open source, and we have integrated it with a number of foundational services that provide other aspects of our computational environment. So Kubernetes off the shelf doesn't do everything. It does a lot. In particular the orchestration of containers, but it delegates through API a lot of key functions. So for example, secret management, traffic management, there's a need for telemetry and observability at a scale beyond what you get from raw Kubernetes. That is to say, harvesting the metrics that are coming out of Kubernetes and processing them, storing them in time series databases, dashboarding them, and so on. There's also an angle to Kubernetes that gets a lot of attention in the daily DevOps routine, that's not really part of the open source deliverable itself, and that is the DevOps sort of CICD pipeline-oriented lifecycle. And that is something else that we've added and integrated nicely. And then one more piece of this picture is that within a Kubernetes cluster, there's a function that is critical to allowing services to discover each other and integrate with each other securely and with proper configuration provided by the concept of a service mesh. So Istio, Linkerd, these are examples of service mesh technologies. And we have gone ahead and integrated actually those two. There's more than those two, but we've integrated those two with Kubernetes. So the net effect is that when a developer within Walmart is going to build an application, they don't have to think about all those other capabilities where they come from or how they're provided. Those are already present, and the way the CICD pipelines are set up, it's already sort of in the picture, and there are configuration points that they can take advantage of in the primary YAML and a couple of other pieces of config that we supply where they can tune it. But at the end of the day, it offloads an awful lot of work for them, having to stand up and operate those services, fail them over properly, and make them robust. All of that's provided for. >> Yeah, you know, developers often complain they spend too much time wrangling and doing things that aren't productive. So I wonder if you could talk about the high level business goals of the initiative in terms of the hardcore benefits. Was the real impetus to tap into best of breed cloud services? Were you trying to cut costs? Maybe gain negotiating leverage with the cloud guys? Resiliency, you know, I know was a major theme. Maybe you could give us a sense of kind of the anatomy of the decision making process that went in. >> Sure, and in the course of answering your question, I think I'm going to introduce the concept of our triplet architecture which we haven't yet touched on in the interview here. First off, just to sort of wrap up the motivation for WCNP itself which is kind of orthogonal to the triplet architecture. It can exist with or without it. Currently does exist with it, which is key, and I'll get to that in a moment. The key drivers, business drivers for WCNP were developer productivity by offloading the kinds of concerns that we've just discussed. Number two, improving resiliency, that is to say reducing opportunity for human error. One of the challenges you tend to run into in a large enterprise is what we call snowflakes, lots of gratuitously different workloads, projects, configurations to the extent that by developing and using WCNP and continuing to evolve it as we have, we end up with cookie cutter like consistency across our workloads which is super valuable when it comes to building tools or building services to automate operations that would otherwise be manual. When everything is pretty much done the same way, that becomes much simpler. Another key motivation for WCNP was the ability to abstract from the underlying cloud provider. And this is going to lead to a discussion of our triplet architecture. At the end of the day, when one works directly with an underlying cloud provider, one ends up taking a lot of dependencies on that particular cloud provider. Those dependencies can be valuable. For example, there are best of breed services like say Cloud Spanner offered by Google or say Cosmos DB offered by Microsoft that one wants to use and one is willing to take the dependency on the cloud provider to get that functionality because it's unique and valuable. On the other hand, one doesn't want to take dependencies on a cloud provider that don't add a lot of value. And with Kubernetes, we have the opportunity, and this is a large part of how Kubernetes was designed and why it is the way it is, we have the opportunity to sort of abstract from the underlying cloud provider for stateless workloads on compute. And so what this lets us do is build container-based applications that can run without change on different cloud provider infrastructure. So the same applications can run on WCNP over Azure, WCNP over GCP, or WCNP over the Walmart private cloud. And we have a private cloud. Our private cloud is OpenStack based and it gives us some significant cost advantages as well as control advantages. So to your point, in terms of business motivation, there's a key cost driver here, which is that we can use our own private cloud when it's advantageous and then use the public cloud provider capabilities when we need to. A key place with this comes into play is with elasticity. So while the private cloud is much more cost effective for us to run and use, it isn't as elastic as what the cloud providers offer, right? We don't have essentially unlimited scale. We have large scale, but the public cloud providers are elastic in the extreme which is a very powerful capability. So what we're able to do is burst, and we use this term bursting workloads into the public cloud from the private cloud to take advantage of the elasticity they offer and then fall back into the private cloud when the traffic load diminishes to the point where we don't need that elastic capability, elastic capacity at low cost. And this is a very important paradigm that I think is going to be very commonplace ultimately as the industry evolves. Private cloud is easier to operate and less expensive, and yet the public cloud provider capabilities are difficult to match. >> And the triplet, the tri is your on-prem private cloud and the two public clouds that you mentioned, is that right? >> That is correct. And we actually have an architecture in which we operate all three of those cloud platforms in close proximity with one another in three different major regions in the US. So we have east, west, and central. And in each of those regions, we have all three cloud providers. And the way it's configured, those data centers are within 10 milliseconds of each other, meaning that it's of negligible cost to interact between them. And this allows us to be fairly agnostic to where a particular workload is running. >> Does a human make that decision, Jack or is there some intelligence in the system that determines that? >> That's a really great question, Dave. And it's a great question because we're at the cusp of that transition. So currently humans make that decision. Humans choose to deploy workloads into a particular region and a particular provider within that region. That said, we're actively developing patterns and practices that will allow us to automate the placement of the workloads for a variety of criteria. For example, if in a particular region, a particular provider is heavily overloaded and is unable to provide the level of service that's expected through our SLAs, we could choose to fail workloads over from that cloud provider to a different one within the same region. But that's manual today. We do that, but people do it. Okay, we'd like to get to where that happens automatically. In the same way, we'd like to be able to automate the failovers, both for high availability and sort of the heavier disaster recovery model between, within a region between providers and even within a provider between the availability zones that are there, but also between regions for the sort of heavier disaster recovery or maintenance driven realignment of workload placement. Today, that's all manual. So we have people moving workloads from region A to region B or data center A to data center B. It's clean because of the abstraction. The workloads don't have to know or care, but there are latency considerations that come into play, and the humans have to be cognizant of those. And automating that can help ensure that we get the best performance and the best reliability. >> But you're developing the dataset to actually, I would imagine, be able to make those decisions in an automated fashion over time anyway. Is that a fair assumption? >> It is, and that's what we're actively developing right now. So if you were to look at us today, we have these nice abstractions and APIs in place, but people run that machine, if you will, moving toward a world where that machine is fully automated. >> What exactly are you abstracting? Is it sort of the deployment model or, you know, are you able to abstract, I'm just making this up like Azure functions and GCP functions so that you can sort of run them, you know, with a consistent experience. What exactly are you abstracting and how difficult was it to achieve that objective technically? >> that's a good question. What we're abstracting is the Kubernetes node construct. That is to say a cluster of Kubernetes nodes which are typically VMs, although they can run bare metal in certain contexts, is something that typically to stand up requires knowledge of the underlying cloud provider. So for example, with GCP, you would use GKE to set up a Kubernetes cluster, and in Azure, you'd use AKS. We are actually abstracting that aspect of things so that the developers standing up applications don't have to know what the underlying cluster management provider is. They don't have to know if it's GCP, AKS or our own Walmart private cloud. Now, in terms of functions like Azure functions that you've mentioned there, we haven't done that yet. That's another piece that we have sort of on our radar screen that, we'd like to get to is serverless approach, and the Knative work from Google and the Azure functions, those are things that we see good opportunity to use for a whole variety of use cases. But right now we're not doing much with that. We're strictly container based right now, and we do have some VMs that are running in sort of more of a traditional model. So our stateful workloads are primarily VM based, but for serverless, that's an opportunity for us to take some of these stateless workloads and turn them into cloud functions. >> Well, and that's another cost lever that you can pull down the road that's going to drop right to the bottom line. Do you see a day or maybe you're doing it today, but I'd be surprised, but where you build applications that actually span multiple clouds or is there, in your view, always going to be a direct one-to-one mapping between where an application runs and the specific cloud platform? >> That's a really great question. Well, yes and no. So today, application development teams choose a cloud provider to deploy to and a location to deploy to, and they have to get involved in moving an application like we talked about today. That said, the bursting capability that I mentioned previously is something that is a step in the direction of automatic migration. That is to say we're migrating workload to different locations automatically. Currently, the prototypes we've been developing and that we think are going to eventually make their way into production are leveraging Istio to assess the load incoming on a particular cluster and start shedding that load into a different location. Right now, the configuration of that is still manual, but there's another opportunity for automation there. And I think a key piece of this is that down the road, well, that's a, sort of a small step in the direction of an application being multi provider. We expect to see really an abstraction of the fact that there is a triplet even. So the workloads are moving around according to whatever the control plane decides is necessary based on a whole variety of inputs. And at that point, you will have true multi-cloud applications, applications that are distributed across the different providers and in a way that application developers don't have to think about. >> So Walmart's been a leader, Jack, in using data for competitive advantages for decades. It's kind of been a poster child for that. You've got a mountain of IP in the form of data, tools, applications best practices that until the cloud came out was all On Prem. But I'm really interested in this idea of building a Walmart ecosystem, which obviously you have. Do you see a day or maybe you're even doing it today where you take what we call the Walmart SuperCloud, WCNP in your words, and point or turn that toward an external world or your ecosystem, you know, supporting those partners or customers that could drive new revenue streams, you know directly from the platform? >> Great question, Steve. So there's really two things to say here. The first is that with respect to data, our data workloads are primarily VM basis. I've mentioned before some VMware, some straight open stack. But the key here is that WCNP and Kubernetes are very powerful for stateless workloads, but for stateful workloads tend to be still climbing a bit of a growth curve in the industry. So our data workloads are not primarily based on WCNP. They're VM based. Now that said, there is opportunity to make some progress there, and we are looking at ways to move things into containers that are currently running in VMs which are stateful. The other question you asked is related to how we expose data to third parties and also functionality. Right now we do have in-house, for our own use, a very robust data architecture, and we have followed the sort of domain-oriented data architecture guidance from Martin Fowler. And we have data lakes in which we collect data from all the transactional systems and which we can then use and do use to build models which are then used in our applications. But right now we're not exposing the data directly to customers as a product. That's an interesting direction that's been talked about and may happen at some point, but right now that's internal. What we are exposing to customers is applications. So we're offering our global integrated fulfillment capabilities, our order picking and curbside pickup capabilities, and our cloud powered checkout capabilities to third parties. And this means we're standing up our own internal applications as externally facing SaaS applications which can serve our partners' customers. >> Yeah, of course, Martin Fowler really first introduced to the world Zhamak Dehghani's data mesh concept and this whole idea of data products and domain oriented thinking. Zhamak Dehghani, by the way, is a speaker at our event as well. Last question I had is edge, and how you think about the edge? You know, the stores are an edge. Are you putting resources there that sort of mirror this this triplet model? Or is it better to consolidate things in the cloud? I know there are trade-offs in terms of latency. How are you thinking about that? >> All really good questions. It's a challenging area as you can imagine because edges are subject to disconnection, right? Or reduced connection. So we do place the same architecture at the edge. So WCNP runs at the edge, and an application that's designed to run at WCNP can run at the edge. That said, there are a number of very specific considerations that come up when running at the edge, such as the possibility of disconnection or degraded connectivity. And so one of the challenges we have faced and have grappled with and done a good job of I think is dealing with the fact that applications go offline and come back online and have to reconnect and resynchronize, the sort of online offline capability is something that can be quite challenging. And we have a couple of application architectures that sort of form the two core sets of patterns that we use. One is an offline/online synchronization architecture where we discover that we've come back online, and we understand the differences between the online dataset and the offline dataset and how they have to be reconciled. The other is a message-based architecture. And here in our health and wellness domain, we've developed applications that are queue based. So they're essentially business processes that consist of multiple steps where each step has its own queue. And what that allows us to do is devote whatever bandwidth we do have to those pieces of the process that are most latency sensitive and allow the queue lengths to increase in parts of the process that are not latency sensitive, knowing that they will eventually catch up when the bandwidth is restored. And to put that in a little bit of context, we have fiber lengths to all of our locations, and we have I'll just use a round number, 10-ish thousand locations. It's larger than that, but that's the ballpark, and we have fiber to all of them, but when the fiber is disconnected, and it does get disconnected on a regular basis. In fact, I forget the exact number, but some several dozen locations get disconnected daily just by virtue of the fact that there's construction going on and things are happening in the real world. When the disconnection happens, we're able to fall back to 5G and to Starlink. Starlink is preferred. It's a higher bandwidth. 5G if that fails. But in each of those cases, the bandwidth drops significantly. And so the applications have to be intelligent about throttling back the traffic that isn't essential, so that it can push the essential traffic in those lower bandwidth scenarios. >> So much technology to support this amazing business which started in the early 1960s. Jack, unfortunately, we're out of time. I would love to have you back or some members of your team and drill into how you're using open source, but really thank you so much for explaining the approach that you've taken and participating in SuperCloud2. >> You're very welcome, Dave, and we're happy to come back and talk about other aspects of what we do. For example, we could talk more about the data lakes and the data mesh that we have in place. We could talk more about the directions we might go with serverless. So please look us up again. Happy to chat. >> I'm going to take you up on that, Jack. All right. This is Dave Vellante for John Furrier and the Cube community. Keep it right there for more action from SuperCloud2. (upbeat music)

Published Date : Jan 9 2023

SUMMARY :

and the Chief Architect for and appreciate the the Walmart Cloud Native Platform? and that is the DevOps Was the real impetus to tap into Sure, and in the course And the way it's configured, and the humans have to the dataset to actually, but people run that machine, if you will, Is it sort of the deployment so that the developers and the specific cloud platform? and that we think are going in the form of data, tools, applications a bit of a growth curve in the industry. and how you think about the edge? and allow the queue lengths to increase for explaining the and the data mesh that we have in place. and the Cube community.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
StevePERSON

0.99+

Dave VellantePERSON

0.99+

Jack GreenfieldPERSON

0.99+

DavePERSON

0.99+

JackPERSON

0.99+

MicrosoftORGANIZATION

0.99+

WalmartORGANIZATION

0.99+

Martin FowlerPERSON

0.99+

USLOCATION

0.99+

Zhamak DehghaniPERSON

0.99+

TodayDATE

0.99+

eachQUANTITY

0.99+

OneQUANTITY

0.99+

twoQUANTITY

0.99+

StarlinkORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

two thingsQUANTITY

0.99+

todayDATE

0.99+

threeQUANTITY

0.99+

firstQUANTITY

0.99+

each stepQUANTITY

0.99+

FirstQUANTITY

0.99+

early 1960sDATE

0.98+

oneQUANTITY

0.98+

a dayQUANTITY

0.98+

GCPTITLE

0.97+

AzureTITLE

0.96+

WCNPTITLE

0.96+

10 millisecondsQUANTITY

0.96+

bothQUANTITY

0.96+

KubernetesTITLE

0.94+

Cloud SpannerTITLE

0.94+

LinkerdORGANIZATION

0.93+

CubeORGANIZATION

0.93+

tripletQUANTITY

0.92+

three cloud providersQUANTITY

0.91+

two core setsQUANTITY

0.88+

John FurrierPERSON

0.86+

one more pieceQUANTITY

0.86+

SuperCloud2ORGANIZATION

0.86+

two public cloudsQUANTITY

0.86+

thousand locationsQUANTITY

0.83+

Vice PresidentPERSON

0.8+

10-ishQUANTITY

0.79+

WCNPORGANIZATION

0.75+

decadesQUANTITY

0.75+

three different major regionsQUANTITY

0.74+

Luis Ceze, OctoML | Amazon re:MARS 2022


 

(upbeat music) >> Welcome back, everyone, to theCUBE's coverage here live on the floor at AWS re:MARS 2022. I'm John Furrier, host for theCUBE. Great event, machine learning, automation, robotics, space, that's MARS. It's part of the re-series of events, re:Invent's the big event at the end of the year, re:Inforce, security, re:MARS, really intersection of the future of space, industrial, automation, which is very heavily DevOps machine learning, of course, machine learning, which is AI. We have Luis Ceze here, who's the CEO co-founder of OctoML. Welcome to theCUBE. >> Thank you very much for having me in the show, John. >> So we've been following you guys. You guys are a growing startup funded by Madrona Venture Capital, one of your backers. You guys are here at the show. This is a, I would say small show relative what it's going to be, but a lot of robotics, a lot of space, a lot of industrial kind of edge, but machine learning is the centerpiece of this trend. You guys are in the middle of it. Tell us your story. >> Absolutely, yeah. So our mission is to make machine learning sustainable and accessible to everyone. So I say sustainable because it means we're going to make it faster and more efficient. You know, use less human effort, and accessible to everyone, accessible to as many developers as possible, and also accessible in any device. So, we started from an open source project that began at University of Washington, where I'm a professor there. And several of the co-founders were PhD students there. We started with this open source project called Apache TVM that had actually contributions and collaborations from Amazon and a bunch of other big tech companies. And that allows you to get a machine learning model and run on any hardware, like run on CPUs, GPUs, various GPUs, accelerators, and so on. It was the kernel of our company and the project's been around for about six years or so. Company is about three years old. And we grew from Apache TVM into a whole platform that essentially supports any model on any hardware cloud and edge. >> So is the thesis that, when it first started, that you want to be agnostic on platform? >> Agnostic on hardware, that's right. >> Hardware, hardware. >> Yeah. >> What was it like back then? What kind of hardware were you talking about back then? Cause a lot's changed, certainly on the silicon side. >> Luis: Absolutely, yeah. >> So take me through the journey, 'cause I could see the progression. I'm connecting the dots here. >> So once upon a time, yeah, no... (both chuckling) >> I walked in the snow with my bare feet. >> You have to be careful because if you wake up the professor in me, then you're going to be here for two hours, you know. >> Fast forward. >> The average version here is that, clearly machine learning has shown to actually solve real interesting, high value problems. And where machine learning runs in the end, it becomes code that runs on different hardware, right? And when we started Apache TVM, which stands for tensor virtual machine, at that time it was just beginning to start using GPUs for machine learning, we already saw that, with a bunch of machine learning models popping up and CPUs and GPU's starting to be used for machine learning, it was clear that it come opportunity to run on everywhere. >> And GPU's were coming fast. >> GPUs were coming and huge diversity of CPUs, of GPU's and accelerators now, and the ecosystem and the system software that maps models to hardware is still very fragmented today. So hardware vendors have their own specific stacks. So Nvidia has its own software stack, and so does Intel, AMD. And honestly, I mean, I hope I'm not being, you know, too controversial here to say that it kind of of looks like the mainframe era. We had tight coupling between hardware and software. You know, if you bought IBM hardware, you had to buy IBM OS and IBM database, IBM applications, it all tightly coupled. And if you want to use IBM software, you had to buy IBM hardware. So that's kind of like what machine learning systems look like today. If you buy a certain big name GPU, you've got to use their software. Even if you use their software, which is pretty good, you have to buy their GPUs, right? So, but you know, we wanted to help peel away the model and the software infrastructure from the hardware to give people choice, ability to run the models where it best suit them. Right? So that includes picking the best instance in the cloud, that's going to give you the right, you know, cost properties, performance properties, or might want to run it on the edge. You might run it on an accelerator. >> What year was that roughly, when you were going this? >> We started that project in 2015, 2016 >> Yeah. So that was pre-conventional wisdom. I think TensorFlow wasn't even around yet. >> Luis: No, it wasn't. >> It was, I'm thinking like 2017 or so. >> Luis: Right. So that was the beginning of, okay, this is opportunity. AWS, I don't think they had released some of the nitro stuff that the Hamilton was working on. So, they were already kind of going that way. It's kind of like converging. >> Luis: Yeah. >> The space was happening, exploding. >> Right. And the way that was dealt with, and to this day, you know, to a large extent as well is by backing machine learning models with a bunch of hardware specific libraries. And we were some of the first ones to say, like, know what, let's take a compilation approach, take a model and compile it to very efficient code for that specific hardware. And what underpins all of that is using machine learning for machine learning code optimization. Right? But it was way back when. We can talk about where we are today. >> No, let's fast forward. >> That's the beginning of the open source project. >> But that was a fundamental belief, worldview there. I mean, you have a world real view that was logical when you compare to the mainframe, but not obvious to the machine learning community. Okay, good call, check. Now let's fast forward, okay. Evolution, we'll go through the speed of the years. More chips are coming, you got GPUs, and seeing what's going on in AWS. Wow! Now it's booming. Now I got unlimited processors, I got silicon on chips, I got, everywhere >> Yeah. And what's interesting is that the ecosystem got even more complex, in fact. Because now you have, there's a cross product between machine learning models, frameworks like TensorFlow, PyTorch, Keras, and like that and so on, and then hardware targets. So how do you navigate that? What we want here, our vision is to say, folks should focus, people should focus on making the machine learning models do what they want to do that solves a value, like solves a problem of high value to them. Right? So another deployment should be completely automatic. Today, it's very, very manual to a large extent. So once you're serious about deploying machine learning model, you got a good understanding where you're going to deploy it, how you're going to deploy it, and then, you know, pick out the right libraries and compilers, and we automated the whole thing in our platform. This is why you see the tagline, the booth is right there, like bringing DevOps agility for machine learning, because our mission is to make that fully transparent. >> Well, I think that, first of all, I use that line here, cause I'm looking at it here on live on camera. People can't see, but it's like, I use it on a couple couple of my interviews because the word agility is very interesting because that's kind of the test on any kind of approach these days. Agility could be, and I talked to the robotics guys, just having their product be more agile. I talked to Pepsi here just before you came on, they had this large scale data environment because they built an architecture, but that fostered agility. So again, this is an architectural concept, it's a systems' view of agility being the output, and removing dependencies, which I think what you guys were trying to do. >> Only part of what we do. Right? So agility means a bunch of things. First, you know-- >> Yeah explain. >> Today it takes a couple months to get a model from, when the model's ready, to production, why not turn that in two hours. Agile, literally, physically agile, in terms of walk off time. Right? And then the other thing is give you flexibility to choose where your model should run. So, in our deployment, between the demo and the platform expansion that we announced yesterday, you know, we give the ability of getting your model and, you know, get it compiled, get it optimized for any instance in the cloud and automatically move it around. Today, that's not the case. You have to pick one instance and that's what you do. And then you might auto scale with that one instance. So we give the agility of actually running and scaling the model the way you want, and the way it gives you the right SLAs. >> Yeah, I think Swami was mentioning that, not specifically that use case for you, but that use case generally, that scale being moving things around, making them faster, not having to do that integration work. >> Scale, and run the models where they need to run. Like some day you want to have a large scale deployment in the cloud. You're going to have models in the edge for various reasons because speed of light is limited. We cannot make lights faster. So, you know, got to have some, that's a physics there you cannot change. There's privacy reasons. You want to keep data locally, not send it around to run the model locally. So anyways, and giving the flexibility. >> Let me jump in real quick. I want to ask this specific question because you made me think of something. So we're just having a data mesh conversation. And one of the comments that's come out of a few of these data as code conversations is data's the product now. So if you can move data to the edge, which everyone's talking about, you know, why move data if you don't have to, but I can move a machine learning algorithm to the edge. Cause it's costly to move data. I can move computer, everyone knows that. But now I can move machine learning to anywhere else and not worry about integrating on the fly. So the model is the code. >> It is the product. >> Yeah. And since you said, the model is the code, okay, now we're talking even more here. So machine learning models today are not treated as code, by the way. So do not have any of the typical properties of code that you can, whenever you write a piece of code, you run a code, you don't know, you don't even think what is a CPU, we don't think where it runs, what kind of CPU it runs, what kind of instance it runs. But with machine learning model, you do. So what we are doing and created this fully transparent automated way of allowing you to treat your machine learning models if you were a regular function that you call and then a function could run anywhere. >> Yeah. >> Right. >> That's why-- >> That's better. >> Bringing DevOps agility-- >> That's better. >> Yeah. And you can use existing-- >> That's better, because I can run it on the Artemis too, in space. >> You could, yeah. >> If they have the hardware. (both laugh) >> And that allows you to run your existing, continue to use your existing DevOps infrastructure and your existing people. >> So I have to ask you, cause since you're a professor, this is like a masterclass on theCube. Thank you for coming on. Professor. (Luis laughing) I'm a hardware guy. I'm building hardware for Boston Dynamics, Spot, the dog, that's the diversity in hardware, it's tends to be purpose driven. I got a spaceship, I'm going to have hardware on there. >> Luis: Right. >> It's generally viewed in the community here, that everyone I talk to and other communities, open source is going to drive all software. That's a check. But the scale and integration is super important. And they're also recognizing that hardware is really about the software. And they even said on stage, here. Hardware is not about the hardware, it's about the software. So if you believe that to be true, then your model checks all the boxes. Are people getting this? >> I think they're starting to. Here is why, right. A lot of companies that were hardware first, that thought about software too late, aren't making it. Right? There's a large number of hardware companies, AI chip companies that aren't making it. Probably some of them that won't make it, unfortunately just because they started thinking about software too late. I'm so glad to see a lot of the early, I hope I'm not just doing our own horn here, but Apache TVM, the infrastructure that we built to map models to different hardware, it's very flexible. So we see a lot of emerging chip companies like SiMa.ai's been doing fantastic work, and they use Apache TVM to map algorithms to their hardware. And there's a bunch of others that are also using Apache TVM. That's because you have, you know, an opening infrastructure that keeps it up to date with all the machine learning frameworks and models and allows you to extend to the chips that you want. So these companies pay attention that early, gives them a much higher fighting chance, I'd say. >> Well, first of all, not only are you backable by the VCs cause you have pedigree, you're a professor, you're smart, and you get good recruiting-- >> Luis: I don't know about the smart part. >> And you get good recruiting for PhDs out of University of Washington, which is not too shabby computer science department. But they want to make money. The VCs want to make money. >> Right. >> So you have to make money. So what's the pitch? What's the business model? >> Yeah. Absolutely. >> Share us what you're thinking there. >> Yeah. The value of using our solution is shorter time to value for your model from months to hours. Second, you shrink operator, op-packs, because you don't need a specialized expensive team. Talk about expensive, expensive engineers who can understand machine learning hardware and software engineering to deploy models. You don't need those teams if you use this automated solution, right? Then you reduce that. And also, in the process of actually getting a model and getting specialized to the hardware, making hardware aware, we're talking about a very significant performance improvement that leads to lower cost of deployment in the cloud. We're talking about very significant reduction in costs in cloud deployment. And also enabling new applications on the edge that weren't possible before. It creates, you know, latent value opportunities. Right? So, that's the high level value pitch. But how do we make money? Well, we charge for access to the platform. Right? >> Usage. Consumption. >> Yeah, and value based. Yeah, so it's consumption and value based. So depends on the scale of the deployment. If you're going to deploy machine learning model at a larger scale, chances are that it produces a lot of value. So then we'll capture some of that value in our pricing scale. >> So, you have direct sales force then to work those deals. >> Exactly. >> Got it. How many customers do you have? Just curious. >> So we started, the SaaS platform just launched now. So we started onboarding customers. We've been building this for a while. We have a bunch of, you know, partners that we can talk about openly, like, you know, revenue generating partners, that's fair to say. We work closely with Qualcomm to enable Snapdragon on TVM and hence our platform. We're close with AMD as well, enabling AMD hardware on the platform. We've been working closely with two hyperscaler cloud providers that-- >> I wonder who they are. >> I don't know who they are, right. >> Both start with the letter A. >> And they're both here, right. What is that? >> They both start with the letter A. >> Oh, that's right. >> I won't give it away. (laughing) >> Don't give it away. >> One has three, one has four. (both laugh) >> I'm guessing, by the way. >> Then we have customers in the, actually, early customers have been using the platform from the beginning in the consumer electronics space, in Japan, you know, self driving car technology, as well. As well as some AI first companies that actually, whose core value, the core business come from AI models. >> So, serious, serious customers. They got deep tech chops. They're integrating, they see this as a strategic part of their architecture. >> That's what I call AI native, exactly. But now there's, we have several enterprise customers in line now, we've been talking to. Of course, because now we launched the platform, now we started onboarding and exploring how we're going to serve it to these customers. But it's pretty clear that our technology can solve a lot of other pain points right now. And we're going to work with them as early customers to go and refine them. >> So, do you sell to the little guys, like us? Will we be customers if we wanted to be? >> You could, absolutely, yeah. >> What we have to do, have machine learning folks on staff? >> So, here's what you're going to have to do. Since you can see the booth, others can't. No, but they can certainly, you can try our demo. >> OctoML. >> And you should look at the transparent AI app that's compiled and optimized with our flow, and deployed and built with our flow. That allows you to get your image and do style transfer. You know, you can get you and a pineapple and see how you look like with a pineapple texture. >> We got a lot of transcript and video data. >> Right. Yeah. Right, exactly. So, you can use that. Then there's a very clear-- >> But I could use it. You're not blocking me from using it. Everyone's, it's pretty much democratized. >> You can try the demo, and then you can request access to the platform. >> But you get a lot of more serious deeper customers. But you can serve anybody, what you're saying. >> Luis: We can serve anybody, yeah. >> All right, so what's the vision going forward? Let me ask this. When did people start getting the epiphany of removing the machine learning from the hardware? Was it recently, a couple years ago? >> Well, on the research side, we helped start that trend a while ago. I don't need to repeat that. But I think the vision that's important here, I want the audience here to take away is that, there's a lot of progress being made in creating machine learning models. So, there's fantastic tools to deal with training data, and creating the models, and so on. And now there's a bunch of models that can solve real problems there. The question is, how do you very easily integrate that into your intelligent applications? Madrona Venture Group has been very vocal and investing heavily in intelligent applications both and user applications as well as enablers. So we say an enable of that because it's so easy to use our flow to get a model integrated into your application. Now, any regular software developer can integrate that. And that's just the beginning, right? Because, you know, now we have CI/CD integration to keep your models up to date, to continue to integrate, and then there's more downstream support for other features that you normally have in regular software development. >> I've been thinking about this for a long, long, time. And I think this whole code, no one thinks about code. Like, I write code, I'm deploying it. I think this idea of machine learning as code independent of other dependencies is really amazing. It's so obvious now that you say it. What's the choices now? Let's just say that, I buy it, I love it, I'm using it. Now what do I got to do if I want to deploy it? Do I have to pick processors? Are there verified platforms that you support? Is there a short list? Is there every piece of hardware? >> We actually can help you. I hope we're not saying we can do everything in the world here, but we can help you with that. So, here's how. When you have them all in the platform you can actually see how this model runs on any instance of any cloud, by the way. So we support all the three major cloud providers. And then you can make decisions. For example, if you care about latency, your model has to run on, at most 50 milliseconds, because you're going to have interactivity. And then, after that, you don't care if it's faster. All you care is that, is it going to run cheap enough. So we can help you navigate. And also going to make it automatic. >> It's like tire kicking in the dealer showroom. >> Right. >> You can test everything out, you can see the simulation. Are they simulations, or are they real tests? >> Oh, no, we run all in real hardware. So, we have, as I said, we support any instances of any of the major clouds. We actually run on the cloud. But we also support a select number of edge devices today, like ARMs and Nvidia Jetsons. And we have the OctoML cloud, which is a bunch of racks with a bunch Raspberry Pis and Nvidia Jetsons, and very soon, a bunch of mobile phones there too that can actually run the real hardware, and validate it, and test it out, so you can see that your model runs performant and economically enough in the cloud. And it can run on the edge devices-- >> You're a machine learning as a service. Would that be an accurate? >> That's part of it, because we're not doing the machine learning model itself. You come with a model and we make it deployable and make it ready to deploy. So, here's why it's important. Let me try. There's a large number of really interesting companies that do API models, as in API as a service. You have an NLP model, you have computer vision models, where you call an API and then point in the cloud. You send an image and you got a description, for example. But it is using a third party. Now, if you want to have your model on your infrastructure but having the same convenience as an API you can use our service. So, today, chances are that, if you have a model that you know that you want to do, there might not be an API for it, we actually automatically create the API for you. >> Okay, so that's why I get the DevOps agility for machine learning is a better description. Cause it's not, you're not providing the service. You're providing the service of deploying it like DevOps infrastructure as code. You're now ML as code. >> It's your model, your API, your infrastructure, but all of the convenience of having it ready to go, fully automatic, hands off. >> Cause I think what's interesting about this is that it brings the craftsmanship back to machine learning. Cause it's a craft. I mean, let's face it. >> Yeah. I want human brains, which are very precious resources, to focus on building those models, that is going to solve business problems. I don't want these very smart human brains figuring out how to scrub this into actually getting run the right way. This should be automatic. That's why we use machine learning, for machine learning to solve that. >> Here's an idea for you. We should write a book called, The Lean Machine Learning. Cause the lean startup was all about DevOps. >> Luis: We call machine leaning. No, that's not it going to work. (laughs) >> Remember when iteration was the big mantra. Oh, yeah, iterate. You know, that was from DevOps. >> Yeah, that's right. >> This code allowed for standing up stuff fast, double down, we all know the history, what it turned out. That was a good value for developers. >> I could really agree. If you don't mind me building on that point. You know, something we see as OctoML, but we also see at Madrona as well. Seeing that there's a trend towards best in breed for each one of the stages of getting a model deployed. From the data aspect of creating the data, and then to the model creation aspect, to the model deployment, and even model monitoring. Right? We develop integrations with all the major pieces of the ecosystem, such that you can integrate, say with model monitoring to go and monitor how a model is doing. Just like you monitor how code is doing in deployment in the cloud. >> It's evolution. I think it's a great step. And again, I love the analogy to the mainstream. I lived during those days. I remember the monolithic propriety, and then, you know, OSI model kind of blew it. But that OSI stack never went full stack, and it only stopped at TCP/IP. So, I think the same thing's going on here. You see some scalability around it to try to uncouple it, free it. >> Absolutely. And sustainability and accessibility to make it run faster and make it run on any deice that you want by any developer. So, that's the tagline. >> Luis Ceze, thanks for coming on. Professor. >> Thank you. >> I didn't know you were a professor. That's great to have you on. It was a masterclass in DevOps agility for machine learning. Thanks for coming on. Appreciate it. >> Thank you very much. Thank you. >> Congratulations, again. All right. OctoML here on theCube. Really important. Uncoupling the machine learning from the hardware specifically. That's only going to make space faster and safer, and more reliable. And that's where the whole theme of re:MARS is. Let's see how they fit in. I'm John for theCube. Thanks for watching. More coverage after this short break. >> Luis: Thank you. (gentle music)

Published Date : Jun 24 2022

SUMMARY :

live on the floor at AWS re:MARS 2022. for having me in the show, John. but machine learning is the And that allows you to get certainly on the silicon side. 'cause I could see the progression. So once upon a time, yeah, no... because if you wake up learning runs in the end, that's going to give you the So that was pre-conventional wisdom. the Hamilton was working on. and to this day, you know, That's the beginning of that was logical when you is that the ecosystem because that's kind of the test First, you know-- and scaling the model the way you want, not having to do that integration work. Scale, and run the models So if you can move data to the edge, So do not have any of the typical And you can use existing-- the Artemis too, in space. If they have the hardware. And that allows you So I have to ask you, So if you believe that to be true, to the chips that you want. about the smart part. And you get good recruiting for PhDs So you have to make money. And also, in the process So depends on the scale of the deployment. So, you have direct sales How many customers do you have? We have a bunch of, you know, And they're both here, right. I won't give it away. One has three, one has four. in Japan, you know, self They're integrating, they see this as it to these customers. Since you can see the booth, others can't. and see how you look like We got a lot of So, you can use that. But I could use it. and then you can request But you can serve anybody, of removing the machine for other features that you normally have It's so obvious now that you say it. So we can help you navigate. in the dealer showroom. you can see the simulation. And it can run on the edge devices-- You're a machine learning as a service. know that you want to do, I get the DevOps agility but all of the convenience it brings the craftsmanship for machine learning to solve that. Cause the lean startup No, that's not it going to work. You know, that was from DevOps. double down, we all know the such that you can integrate, and then, you know, OSI on any deice that you Professor. That's great to have you on. Thank you very much. Uncoupling the machine learning Luis: Thank you.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Luis CezePERSON

0.99+

QualcommORGANIZATION

0.99+

LuisPERSON

0.99+

2015DATE

0.99+

JohnPERSON

0.99+

John FurrierPERSON

0.99+

Boston DynamicsORGANIZATION

0.99+

two hoursQUANTITY

0.99+

NvidiaORGANIZATION

0.99+

2017DATE

0.99+

JapanLOCATION

0.99+

Madrona Venture CapitalORGANIZATION

0.99+

AMDORGANIZATION

0.99+

oneQUANTITY

0.99+

AmazonORGANIZATION

0.99+

threeQUANTITY

0.99+

IBMORGANIZATION

0.99+

OneQUANTITY

0.99+

AWSORGANIZATION

0.99+

fourQUANTITY

0.99+

2016DATE

0.99+

University of WashingtonORGANIZATION

0.99+

TodayDATE

0.99+

PepsiORGANIZATION

0.99+

BothQUANTITY

0.99+

yesterdayDATE

0.99+

FirstQUANTITY

0.99+

bothQUANTITY

0.99+

SecondQUANTITY

0.99+

todayDATE

0.99+

SiMa.aiORGANIZATION

0.99+

OctoMLTITLE

0.99+

OctoMLORGANIZATION

0.99+

IntelORGANIZATION

0.98+

one instanceQUANTITY

0.98+

DevOpsTITLE

0.98+

Madrona Venture GroupORGANIZATION

0.98+

SwamiPERSON

0.98+

MadronaORGANIZATION

0.98+

about six yearsQUANTITY

0.96+

SpotORGANIZATION

0.96+

The Lean Machine LearningTITLE

0.95+

firstQUANTITY

0.95+

theCUBEORGANIZATION

0.94+

ARMsORGANIZATION

0.94+

pineappleORGANIZATION

0.94+

Raspberry PisORGANIZATION

0.92+

TensorFlowTITLE

0.89+

SnapdragonORGANIZATION

0.89+

about three years oldQUANTITY

0.89+

a couple years agoDATE

0.88+

two hyperscaler cloud providersQUANTITY

0.88+

first onesQUANTITY

0.87+

one ofQUANTITY

0.85+

50 millisecondsQUANTITY

0.83+

Apache TVMORGANIZATION

0.82+

both laughQUANTITY

0.82+

three major cloud providersQUANTITY

0.81+

Changing the Game for Cloud Networking | Pluribus Networks


 

>>Everyone wants a cloud operating model. Since the introduction of the modern cloud. Last decade, the entire technology landscape has changed. We've learned a lot from the hyperscalers, especially from AWS. Now, one thing is certain in the technology business. It's so competitive. Then if a faster, better, cheaper idea comes along, the industry will move quickly to adopt it. They'll add their unique value and then they'll bring solutions to the market. And that's precisely what's happening throughout the technology industry because of cloud. And one of the best examples is Amazon's nitro. That's AWS has custom built hypervisor that delivers on the promise of more efficiently using resources and expanding things like processor, optionality for customers. It's a secret weapon for Amazon. As, as we, as we wrote last year, every infrastructure company needs something like nitro to compete. Why do we say this? Well, Wiki Bon our research arm estimates that nearly 30% of CPU cores in the data center are wasted. >>They're doing work that they weren't designed to do well, specifically offloading networking, storage, and security tasks. So if you can eliminate that waste, you can recapture dollars that drop right to the bottom line. That's why every company needs a nitro like solution. As a result of these developments, customers are rethinking networks and how they utilize precious compute resources. They can't, or won't put everything into the public cloud for many reasons. That's one of the tailwinds for tier two cloud service providers and why they're growing so fast. They give options to customers that don't want to keep investing in building out their own data centers, and they don't want to migrate all their workloads to the public cloud. So these providers and on-prem customers, they want to be more like hyperscalers, right? They want to be more agile and they do that. They're distributing, networking and security functions and pushing them closer to the applications. >>Now, at the same time, they're unifying their view of the network. So it can be less fragmented, manage more efficiently with more automation and better visibility. How are they doing this? Well, that's what we're going to talk about today. Welcome to changing the game for cloud networking made possible by pluribus networks. My name is Dave Vellante and today on this special cube presentation, John furrier, and I are going to explore these issues in detail. We'll dig into new solutions being created by pluribus and Nvidia to specifically address offloading, wasted resources, accelerating performance, isolating data, and making networks more secure all while unifying the network experience. We're going to start on the west coast and our Palo Alto studios, where John will talk to Mike of pluribus and AMI, but Donnie of Nvidia, then we'll bring on Alessandra Bobby airy of pluribus and Pete Lummus from Nvidia to take a deeper dive into the technology. And then we're gonna bring it back here to our east coast studio and get the independent analyst perspective from Bob Liberte of the enterprise strategy group. We hope you enjoy the program. Okay, let's do this over to John >>Okay. Let's kick things off. We're here at my cafe. One of the TMO and pluribus networks and NAMI by Dani VP of networking, marketing, and developer ecosystem at Nvidia. Great to have you welcome folks. >>Thank you. Thanks. >>So let's get into the, the problem situation with cloud unified network. What problems are out there? What challenges do cloud operators have Mike let's get into it. >>Yeah, it really, you know, the challenges we're looking at are for non hyperscalers that's enterprises, governments, um, tier two service providers, cloud service providers, and the first mandate for them is to become as agile as a hyperscaler. So they need to be able to deploy services and security policies. And second, they need to be able to abstract the complexity of the network and define things in software while it's accelerated in hardware. Um, really ultimately they need a single operating model everywhere. And then the second thing is they need to distribute networking and security services out to the edge of the host. Um, we're seeing a growth in cyber attacks. Um, it's, it's not slowing down. It's only getting worse and, you know, solving for this security problem across clouds is absolutely critical. And the way to do it is to move security out to the host. >>Okay. With that goal in mind, what's the pluribus vision. How does this tie together? >>Yeah. So, um, basically what we see is, uh, that this demands a new architecture and that new architecture has four tenants. The first tenant is unified and simplified cloud networks. If you look at cloud networks today, there's, there's sort of like discreet bespoke cloud networks, you know, per hypervisor, per private cloud edge cloud public cloud. Each of the public clouds have different networks that needs to be unified. You know, if we want these folks to be able to be agile, they need to be able to issue a single command or instantiate a security policy across all those locations with one command and not have to go to each one. The second is like I mentioned, distributed security, um, distributed security without compromise, extended out to the host is absolutely critical. So micro-segmentation and distributed firewalls, but it doesn't stop there. They also need pervasive visibility. >>You know, it's, it's, it's sort of like with security, you really can't see you can't protect what you can't see. So you need visibility everywhere. The problem is visibility to date has been very expensive. Folks have had to basically build a separate overlay network of taps, packet brokers, tap aggregation infrastructure that really needs to be built into this unified network I'm talking about. And the last thing is automation. All of this needs to be SDN enabled. So this is related to my comment about abstraction abstract, the complexity of all of these discreet networks, physic whatever's down there in the physical layer. Yeah. I don't want to see it. I want to abstract it. I wanted to find things in software, but I do want to leverage the power of hardware to accelerate that. So that's the fourth tenant is SDN automation. >>Mike, we've been talking on the cube a lot about this architectural shift and customers are looking at this. This is a big part of everyone who's looking at cloud operations next gen, how do we get there? How do customers get this vision realized? >>That's a great question. And I appreciate the tee up. I mean, we're, we're here today for that reason. We're introducing two things today. Um, the first is a unified cloud networking vision, and that is a vision of where pluribus is headed with our partners like Nvidia longterm. Um, and that is about, uh, deploying a common operating model, SDN enabled SDN, automated hardware, accelerated across all clouds. Um, and whether that's underlying overlay switch or server, um, hype, any hypervisor infrastructure containers, any workload doesn't matter. So that's ultimately where we want to get. And that's what we talked about earlier. Um, the first step in that vision is what we call the unified cloud fabric. And this is the next generation of our adaptive cloud fabric. Um, and what's nice about this is we're not starting from scratch. We have a, a, an award-winning adaptive cloud fabric product that is deployed globally. Um, and in particular, uh, we're very proud of the fact that it's deployed in over a hundred tier one mobile operators as the network fabric for their 4g and 5g virtualized cores. We know how to build carrier grade, uh, networking infrastructure, what we're doing now, um, to realize this next generation unified cloud fabric is we're extending from the switch to this Nvidia Bluefield to DPU. We know there's a, >>Hold that up real quick. That's a good, that's a good prop. That's the blue field and video. >>It's the Nvidia Bluefield two DPU data processing unit. And, um, uh, you know, what we're doing, uh, fundamentally is extending our SDN automated fabric, the unified cloud fabric out to the host, but it does take processing power. So we knew that we didn't want to do, we didn't want to implement that running on the CPU, which is what some other companies do because it consumes revenue generating CPU's from the application. So a DPU is a perfect way to implement this. And we knew that Nvidia was the leader with this blue field too. And so that is the first that's, that's the first step in the getting into realizing this vision. >>I mean, Nvidia has always been powering some great workloads of GPU. Now you've got DPU networking and then video is here. What is the relationship with clothes? How did that come together? Tell us the story. >>Yeah. So, you know, we've been working with pluribus for quite some time. I think the last several months was really when it came to fruition and, uh, what pluribus is trying to build and what Nvidia has. So we have, you know, this concept of a Bluefield data processing unit, which if you think about it, conceptually does really three things, offload, accelerate an isolate. So offload your workloads from your CPU to your data processing unit infrastructure workloads that is, uh, accelerate. So there's a bunch of acceleration engines. So you can run infrastructure workloads much faster than you would otherwise, and then isolation. So you have this nice security isolation between the data processing unit and your other CPU environment. And so you can run completely isolated workloads directly on the data processing unit. So we introduced this, you know, a couple of years ago, and with pluribus, you know, we've been talking to the pluribus team for quite some months now. >>And I think really the combination of what pluribus is trying to build and what they've developed around this unified cloud fabric, uh, is fits really nicely with the DPU and running that on the DPU and extending it really from your physical switch, all the way to your host environment, specifically on the data processing unit. So if you think about what's happening as you add data processing units to your environment. So every server we believe over time is going to have data processing units. So now you'll have to manage that complexity from the physical network layer to the host layer. And so what pluribus is really trying to do is extending the network fabric from the host, from the switch to the host, and really have that single pane of glass for network operators to be able to configure provision, manage all of the complexity of the network environment. >>So that's really how the partnership truly started. And so it started really with extending the network fabric, and now we're also working with them on security. So, you know, if you sort of take that concept of isolation and security isolation, what pluribus has within their fabric is the concept of micro-segmentation. And so now you can take that extended to the data processing unit and really have, um, isolated micro-segmentation workloads, whether it's bare metal cloud native environments, whether it's virtualized environments, whether it's public cloud, private cloud hybrid cloud. So it really is a magical partnership between the two companies with their unified cloud fabric running on, on the DPU. >>You know, what I love about this conversation is it reminds me of when you have these changing markets, the product gets pulled out of the market and, and you guys step up and create these new solutions. And I think this is a great example. So I have to ask you, how do you guys differentiate what sets this apart for customers with what's in it for the customer? >>Yeah. So I mentioned, you know, three things in terms of the value of what the Bluefield brings, right? There's offloading, accelerating, isolating, that's sort of the key core tenants of Bluefield. Um, so that, you know, if you sort of think about what, um, what Bluefields, what we've done, you know, in terms of the differentiation, we're really a robust platform for innovation. So we introduced Bluefield to, uh, last year, we're introducing Bluefield three, which is our next generation of Bluefields, you know, we'll have five X, the arm compute capacity. It will have 400 gig line rate acceleration, four X better crypto acceleration. So it will be remarkably better than the previous generation. And we'll continue to innovate and add, uh, chips to our portfolio every, every 18 months to two years. Um, so that's sort of one of the key areas of differentiation. The other is the, if you look at Nvidia and, and you know, what we're sort of known for is really known for our AI artificial intelligence and our artificial intelligence software, as well as our GPU. >>So you look at artificial intelligence and the combination of artificial intelligence plus data processing. This really creates the, you know, faster, more efficient, secure AI systems from the core of your data center, all the way out to the edge. And so with Nvidia, we really have these converged accelerators where we've combined the GPU, which does all your AI processing with your data processing with the DPU. So we have this convergence really nice convergence of that area. And I would say the third area is really around our developer environment. So, you know, one of the key, one of our key motivations at Nvidia is really to have our partner ecosystem, embrace our technology and build solutions around our technology. So if you look at what we've done with the DPU, with credit and an SDK, which is an open SDK called Doka, and it's an open SDK for our partners to really build and develop solutions using Bluefield and using all these accelerated libraries that we expose through Doka. And so part of our differentiation is really building this open ecosystem for our partners to take advantage and build solutions around our technology. >>You know, what's exciting is when I hear you talk, it's like you realize that there's no one general purpose network anymore. Everyone has their own super environment Supercloud or these new capabilities. They can really craft their own, I'd say, custom environment at scale with easy tools. Right. And it's all kind of, again, this is the new architecture Mike, you were talking about, how does customers run this effectively? Cost-effectively and how do people migrate? >>Yeah, I, I think that is the key question, right? So we've got this beautiful architecture. You, you know, Amazon nitro is a, is a good example of, of a smart NIC architecture that has been successfully deployed, but enterprises and serve tier two service providers and tier one service providers and governments are not Amazon, right? So they need to migrate there and they need this architecture to be cost-effective. And, and that's, that's super key. I mean, the reality is deep user moving fast, but they're not going to be, um, deployed everywhere on day one. Some servers will have DPS right away, some servers will have use and a year or two. And then there are devices that may never have DPS, right. IOT gateways, or legacy servers, even mainframes. Um, so that's the beauty of a solution that creates a fabric across both the switch and the DPU, right. >>Um, and by leveraging the Nvidia Bluefield DPU, what we really like about it is it's open. Um, and that drives, uh, cost efficiencies. And then, um, uh, you know, with this, with this, our architectural approach effectively, you get a unified solution across switch and DPU workload independent doesn't matter what hypervisor it is, integrated visibility, integrated security, and that can, uh, create tremendous cost efficiencies and, and really extract a lot of the expense from, from a capital perspective out of the network, as well as from an operational perspective, because now I have an SDN automated solution where I'm literally issuing a command to deploy a network service or to create or deploy our security policy and is deployed everywhere, automatically saving the oppor, the network operations team and the security operations team time. >>All right. So let me rewind that because that's super important. Get the unified cloud architecture, I'm the customer guy, but it's implemented, what's the value again, take, take me through the value to me. I have a unified environment. What's the value. >>Yeah. So I mean, the value is effectively, um, that, so there's a few pieces of value. The first piece of value is, um, I'm creating this clean D mark. I'm taking networking to the host. And like I mentioned, we're not running it on the CPU. So in implementations that run networking on the CPU, there's some conflict between the dev ops team who owned the server and the NetApps team who own the network because they're installing software on the, on the CPU stealing cycles from what should be revenue generating. Uh CPU's. So now by, by terminating the networking on the DPU, we click create this real clean DMARC. So the dev ops folks are happy because they don't necessarily have the skills to manage network and they don't necessarily want to spend the time managing networking. They've got their network counterparts who are also happy the NetApps team, because they want to control the networking. >>And now we've got this clean DMARC where the DevOps folks get the services they need and the NetApp folks get the control and agility they need. So that's a huge value. Um, the next piece of value is distributed security. This is essential. I mentioned earlier, you know, put pushing out micro-segmentation and distributed firewall, basically at the application level, right, where I create these small, small segments on an by application basis. So if a bad actor does penetrate the perimeter firewall, they're contained once they get inside. Cause the worst thing is a bad actor, penetrates a perimeter firewall and can go wherever they want and wreak havoc. Right? And so that's why this, this is so essential. Um, and the next benefit obviously is this unified networking operating model, right? Having, uh, uh, uh, an operating model across switch and server underlay and overlay, workload agnostic, making the life of the NetApps teams much easier so they can focus their time on really strategy instead of spending an afternoon, deploying a single villain, for example. >>Awesome. And I think also from my standpoint, I mean, perimeter security is pretty much, I mean, they're out there, it gets the firewall still out there exists, but pretty much they're being breached all the time, the perimeter. So you have to have this new security model. And I think the other thing that you mentioned, the separation between dev ops is cool because the infrastructure is code is about making the developers be agile and build security in from day one. So this policy aspect is, is huge. Um, new control points. I think you guys have a new architecture that enables the security to be handled more flexible. >>Right. >>That seems to be the killer feature here, >>Right? Yeah. If you look at the data processing unit, I think one of the great things about sort of this new architecture, it's really the foundation for zero trust it's. So like you talked about the perimeter is getting breached. And so now each and every compute node has to be protected. And I think that's sort of what you see with the partnership between pluribus and Nvidia is the DPU is really the foundation of zero trust. And pluribus is really building on that vision with, uh, allowing sort of micro-segmentation and being able to protect each and every compute node as well as the underlying network. >>This is super exciting. This is an illustration of how the market's evolving architectures are being reshaped and refactored for cloud scale and all this new goodness with data. So I gotta ask how you guys go into market together. Michael, start with you. What's the relationship look like in the go to market with an Nvidia? >>Sure. Um, I mean, we're, you know, we're super excited about the partnership, obviously we're here together. Um, we think we've got a really good solution for the market, so we're jointly marketing it. Um, uh, you know, obviously we appreciate that Nvidia is open. Um, that's, that's sort of in our DNA, we're about open networking. They've got other ISV who are gonna run on Bluefield too. We're probably going to run on other DPS in the, in the future, but right now, um, we're, we feel like we're partnered with the number one, uh, provider of DPS in the world and, uh, super excited about, uh, making a splash with it. >>I'm in get the hot product. >>Yeah. So Bluefield too, as I mentioned was GA last year, we're introducing, uh, well, we now also have the converged accelerator. So I talked about artificial intelligence or artificial intelligence with the Bluefield DPU, all of that put together on a converged accelerator. The nice thing there is you can either run those workloads. So if you have an artificial intelligence workload and an infrastructure workload, you can warn them separately on the same platform or you can actually use, uh, you can actually run artificial intelligence applications on the Bluefield itself. So that's what the converged accelerator really brings to the table. Uh, so that's available now. Then we have Bluefield three, which will be available late this year. And I talked about sort of, you know, uh, how much better that next generation of Bluefield is in comparison to Bluefield two. So we will see Bluefield three shipping later on this year, and then our software stack, which I talked about, which is called Doka we're on our second version are Doka one dot two. >>We're releasing Doka one dot three, uh, in about two months from now. And so that's really our open ecosystem framework. So allow you to program the Bluefields. So we have all of our acceleration libraries, um, security libraries, that's all packed into this STK called Doka. And it really gives that simplicity to our partners to be able to develop on top of Bluefield. So as we add new generations of Bluefield, you know, next, next year, we'll have, you know, another version and so on and so forth Doka is really that unified unified layer that allows, um, Bluefield to be both forwards compatible and backwards compatible. So partners only really have to think about writing to that SDK once, and then it automatically works with future generations of Bluefields. So that's sort of the nice thing around, um, around Doka. And then in terms of our go to market model, we're working with every, every major OEM. So, uh, later on this year, you'll see, you know, major server manufacturers, uh, releasing Bluefield enabled servers. So, um, more to come >>Awesome, save money, make it easier, more capabilities, more workload power. This is the future of, of cloud operations. >>Yeah. And, and, and, uh, one thing I'll add is, um, we are, um, we have a number of customers as you'll hear in the next segment, um, that are already signed up and we'll be working with us for our, uh, early field trial starting late April early may. Um, we are accepting registrations. You can go to www.pluribusnetworks.com/e F T a. If you're interested in signing up for, um, uh, being part of our field trial and providing feedback on the product, >>Awesome innovation and network. Thanks so much for sharing the news. Really appreciate it. Thanks so much. Okay. In a moment, we'll be back to look deeper in the product, the integration security zero trust use cases. You're watching the cube, the leader in enterprise tech coverage, >>Cloud networking is complex and fragmented slowing down your business. How can you simplify and unify your cloud networks to increase agility and business velocity? >>Pluribus unified cloud networking provides a unified simplify and agile network fabric across all clouds. It brings the simplicity of a public cloud operation model to private clouds, dramatically reducing complexity and improving agility, availability, and security. Now enterprises and service providers can increase their business philosophy and delight customers in the distributed multi-cloud era. We achieve this with a new approach to cloud networking, pluribus unified cloud fabric. This open vendor, independent network fabric, unifies, networking, and security across distributed clouds. The first step is extending the fabric to servers equipped with data processing units, unifying the fabric across switches and servers, and it doesn't stop there. The fabric is unified across underlay and overlay networks and across all workloads and virtualization environments. The unified cloud fabric is optimized for seamless migration to this new distributed architecture, leveraging the power of the DPU for application level micro-segmentation distributed fireball and encryption while still supporting those servers and devices that are not equipped with a DPU. Ultimately the unified cloud fabric extends seamlessly across distributed clouds, including central regional at edge private clouds and public clouds. The unified cloud fabric is a comprehensive network solution. That includes everything you need for clouds, networking built in SDN automation, distributed security without compromises, pervasive wire speed, visibility and application insight available on your choice of open networking switches and DP use all at the lowest total cost of ownership. The end result is a dramatically simplified unified cloud networking architecture that unifies your distributed clouds and frees your business to move at cloud speed, >>To learn more, visit www.pluribusnetworks.com. >>Okay. We're back I'm John ferry with the cube, and we're going to go deeper into a deep dive into unified cloud networking solution from Clovis and Nvidia. And we'll examine some of the use cases with Alessandra Burberry, VP of product management and pullovers networks and Pete Bloomberg who's director of technical marketing and video remotely guys. Thanks for coming on. Appreciate it. >>Yeah. >>So deep dive, let's get into the what and how Alexandra we heard earlier about the pluribus Nvidia partnership and the solution you're working together on what is it? >>Yeah. First let's talk about the water. What are we really integrating with the Nvidia Bluefield, the DPO technology, uh, plugable says, um, uh, there's been shipping, uh, in, uh, in volume, uh, in multiple mission critical networks. So this advisor one network operating systems, it runs today on a merchant silicone switches and effectively it's a standard open network operating system for data center. Um, and the novelty about this system that integrates a distributed control plane for, at water made effective in SDN overlay. This automation is a completely open and interoperable and extensible to other type of clouds is not enclosed them. And this is actually what we're now porting to the Nvidia DPO. >>Awesome. So how does it integrate into Nvidia hardware and specifically how has pluribus integrating its software with the Nvidia hardware? >>Yeah, I think, uh, we leverage some of the interesting properties of the Bluefield, the DPO hardware, which allows actually to integrate, uh, um, uh, our software, our network operating system in a manner which is completely isolated and independent from the guest operating system. So the first byproduct of this approach is that whatever we do at the network level on the DPU card that is completely agnostic to the hypervisor layer or OSTP layer running on, uh, on the host even more, um, uh, we can also independently manage this network, know that the switch on a Neek effectively, um, uh, managed completely independently from the host. You don't have to go through the network operating system, running on x86 to control this network node. So you throw yet the experience effectively of a top of rack for virtual machine or a top of rack for, uh, Kubernetes bots, where instead of, uh, um, if you allow me with the analogy instead of connecting a server knee directly to a switchboard, now you're connecting a VM virtual interface to a virtual interface on the switch on an ache. >>And, uh, also as part of this integration, we, uh, put a lot of effort, a lot of emphasis in, uh, accelerating the entire, uh, data plane for networking and security. So we are taking advantage of the DACA, uh, Nvidia DACA API to program the accelerators. And these accomplished two things with that. Number one, uh, you, uh, have much greater performance, much better performance. They're running the same network services on an x86 CPU. And second, this gives you the ability to free up, I would say around 20, 25% of the server capacity to be devoted either to, uh, additional workloads to run your cloud applications, or perhaps you can actually shrink the power footprint and compute footprint of your data center by 20%, if you want to run the same number of compute workloads. So great efficiencies in the overall approach, >>And this is completely independent of the server CPU, right? >>Absolutely. There is zero code from running on the x86, and this is what we think this enables a very clean demarcation between computer and network. >>So Pete, I gotta get, I gotta get you in here. We heard that, uh, the DPU is enabled cleaner separation of dev ops and net ops. Can you explain why that's important because everyone's talking DevSecOps right now, you've got net ops, net, net sec ops, this separation. Why is this clean separation important? >>Yeah, I think it's a, you know, it's a pragmatic solution in my opinion. Um, you know, we wish the world was all kind of rainbows and unicorns, but it's a little, a little messier than that. And I think a lot of the dev ops stuff and that, uh, mentality and philosophy, there's a natural fit there. Right? You have applications running on servers. So you're talking about developers with those applications integrating with the operators of those servers. Well, the network has always been this other thing and the network operators have always had a very different approach to things than compute operators. And, you know, I think that we, we in the networking industry have gotten closer together, but there's still a gap there's still some distance. And I think in that distance, isn't going to be closed. And so, you know, again, it comes down to pragmatism and I think, you know, one of my favorite phrases is look good fences, make good neighbors. And that's what this is. >>Yeah. That's a great point because dev ops has become kind of the calling card for cloud, right. But dev ops is as simply infrastructure as code and infrastructure is networking, right? So if infrastructure is code, you know, you're talking about, you know, that part of the stack under the covers under the hood, if you will, this is super important distinction. And this is where the innovation is. Can you elaborate on how you see that? Because this is really where the action is right now. >>Yeah, exactly. And I think that's where, um, one from, from the policy, the security that the zero trust aspect of this, right? If you get it wrong on that network side, all of a sudden you, you can totally open up that those capabilities. And so security is part of that. But the other part is thinking about this at scale, right? So we're taking one top of rack switch and adding, you know, up to 48 servers per rack. And so that ability to automate, orchestrate and manage at scale becomes absolutely critical. >>I'll Sandra, this is really the why we're talking about here, and this is scale. And again, getting it right. If you don't get it right, you're going to be really kind of up, you know what you know, so this is a huge deal. Networking matters, security matters, automation matters, dev ops, net ops, all coming together, clean separation, um, help us understand how this joint solution with Nvidia fits into the pluribus unified cloud networking vision, because this is what people are talking about and working on right now. >>Yeah, absolutely. So I think here with this solution, we're attacking two major problems in cloud networking. One is, uh, operation of, uh, cloud networking. And the second is a distributing security services in the cloud infrastructure. First, let me talk about the first water. We really unifying. If we're unifying something, something must be at least fragmented or this jointed and the, what is this joint that is actually the network in the cloud. If you look holistically, how networking is deployed in the cloud, you have your physical fabric infrastructure, right? Your switches and routers, you'll build your IP clause fabric leaf in spine typologies. This is actually a well understood the problem. I, I would say, um, there are multiple vendors, uh, uh, with, uh, um, uh, let's say similar technologies, um, very well standardized, whether you will understood, um, and almost a commodity, I would say building an IP fabric these days, but this is not the place where you deploy most of your services in the cloud, particularly from a security standpoint, two services are actually now moved into the compute layer where you actually were called builders, have to instrument the, a separate, uh, network virtualization layer, where they deploy segmentation and security closer to the workloads. >>And this is where the complication arise. These high value part of the cloud network is where you have a plethora of options that they don't talk to each other. And they are very dependent on the kind of hypervisor or compute solution you choose. Um, for example, the networking API to be between an GSXI environment or an hyper V or a Zen are completely disjointed. You have multiple orchestration layers. And when, and then when you throw in also Kubernetes in this, in this, in this type of architecture, uh, you're introducing yet another level of networking. And when Kubernetes runs on top of VMs, which is a prevalent approach, you actually just stacking up multiple networks on the compute layer that they eventually run on the physical fabric infrastructure. Those are all ships in the nights effectively, right? They operate as completely disjointed. And we're trying to attack this problem first with the notion of a unified fabric, which is independent from any workloads, whether it's this fabric spans on a switch, which can be con connected to a bare metal workload, or can span all the way inside the DPU, uh, where, um, you have, uh, your multi hypervisor compute environment. >>It's one API, one common network control plane, and one common set of segmentation services for the network. That's probably the number one, >>You know, it's interesting you, man, I hear you talking, I hear one network month, different operating models reminds me of the old serverless days. You know, there's still servers, but they call it serverless. Is there going to be a term network list? Because at the end of the day, it should be one network, not multiple operating models. This, this is a problem that you guys are working on. Is that right? I mean, I'm not, I'm just joking server listen network list, but the idea is it should be one thing. >>Yeah, it's effectively. What we're trying to do is we are trying to recompose this fragmentation in terms of network operation, across physical networking and server networking server networking is where the majority of the problems are because of the, uh, as much as you have standardized the ways of building, uh, physical networks and cloud fabrics with IP protocols and internet, you don't have that kind of, uh, uh, sort of, uh, um, um, uh, operational efficiency, uh, at the server layer. And, uh, this is what we're trying to attack first. The, with this technology, the second aspect we're trying to attack is are we distribute the security services throughout the infrastructure, more efficiently, whether it's micro-segmentation is a stateful firewall services, or even encryption. Those are all capabilities enabled by the blue field, uh, uh, the Butte technology and, uh, uh, we can actually integrate those capabilities directly into the nettle Fabrica, uh, limiting dramatically, at least for east-west traffic, the sprawl of, uh, security appliances, whether virtual or physical, that is typically the way the people today, uh, segment and secure the traffic in the cloud. >>Awesome. Pete, all kidding aside about network lists and serverless kind of fun, fun play on words there, the network is one thing it's basically distributed computing, right? So I love to get your thoughts about this distributed security with zero trust as the driver for this architecture you guys are doing. Can you share in more detail the depth of why DPU based approach is better than alternatives? >>Yeah, I think what's, what's beautiful and kind of what the DPU brings. That's new to this model is a completely isolated compute environment inside. So, you know, it's the, uh, yo dog, I heard you like a server, so I put a server inside your server. Uh, and so we provide, uh, you know, armed CPU's memory and network accelerators inside, and that is completely isolated from the host. So the server, the, the actual x86 host just thinks it has a regular Nick in there, but you actually have this full control plane thing. It's just like taking your top of rack switch and shoving it inside of your compute node. And so you have not only the separation, um, within the data plane, but you have this complete control plane separation. So you have this element that the network team can now control and manage, but we're taking all of the functions we used to do at the top of rack switch, and we're just shooting them now. >>And, you know, as time has gone on we've, we've struggled to put more and more and more into that network edge. And the reality is the network edge is the compute layer, not the top of rack switch layer. And so that provides this phenomenal enforcement point for security and policy. And I think outside of today's solutions around virtual firewalls, um, the other option is centralized appliances. And even if you can get one that can scale large enough, the question is, can you afford it? And so what we end up doing is we kind of hope that of aliens good enough, or we hope that if the excellent tunnel is good enough and we can actually apply more advanced techniques there because we can't physically, you know, financially afford that appliance to see all of the traffic. And now that we have a distributed model with this accelerator, we could do it. >>So what's the what's in it for the customer. I real quick, cause I think this is interesting point. You mentioned policy, everyone in networking knows policy is just a great thing and it adds, you hear it being talked about up the stack as well. When you start getting to orchestrating microservices and whatnot, all that good stuff going on there, containers and whatnot and modern applications. What's the benefit to the customers with this approach? Because what I heard was more scale, more edge deployment, flexibility, relative to security policies and application enablement. I mean, is that what what's the customer get out of this architecture? What's the enablement. >>It comes down to, uh, taking again the capabilities that were in that top of rack switch and asserting them down. So that makes simplicity smaller blast radiuses for failure, smaller failure domains, maintenance on the networks, and the systems become easier. Your ability to integrate across workloads becomes infinitely easier. Um, and again, you know, we always want to kind of separate each one of those layers. So just as in say, a VX land network, my leaf and spine don't have to be tightly coupled together. I can now do this at a different layer. And so you can run a DPU with any networking in the core there. And so you get this extreme flexibility. You can start small, you can scale large. Um, you know, to me, the, the possibilities are endless. Yes, >>It's a great security control plan. Really flexibility is key. And, and also being situationally aware of any kind of threats or new vectors or whatever's happening in the network. Alessandra, this is huge upside, right? You've already identified some successes with some customers on your early field trials. What are they doing and why are they attracted to the solution? >>Yeah, I think the response from customers has been, uh, the most, uh, encouraging and, uh, exciting, uh, for, uh, for us to, uh, to sort of continue and work and develop this product. And we have actually learned a lot in the process. Um, we talked to tier two tier three cloud providers. Uh, we talked to, uh, SP um, software Tyco type of networks, uh, as well as a large enterprise customers, um, in, uh, one particular case. Um, uh, one, uh, I think, um, let me, let me call out a couple of examples here, just to give you a flavor. Uh, there is a service provider, a cloud provider, uh, in Asia who is actually managing a cloud, uh, where they are offering services based on multiple hypervisors. They are native services based on Zen, but they also are on ramp into the cloud, uh, workloads based on, uh, ESI and, uh, uh, and KVM, depending on what the customer picks from the piece on the menu. >>And they have the problem of now orchestrating through their orchestrate or integrating with the Zen center with vSphere, uh, with, uh, open stack to coordinate these multiple environments and in the process to provide security, they actually deploy virtual appliances everywhere, which has a lot of costs, complication, and eats up into the server CPU. The problem is that they saw in this technology, they call it actually game changing is actually to remove all this complexity of in a single network and distribute the micro-segmentation service directly into the fabric. And overall, they're hoping to get out of it, uh, uh, tremendous, uh, um, opics, uh, benefit and overall, um, uh, operational simplification for the cloud infrastructure. That's one potent a use case. Uh, another, uh, large enterprise customer global enterprise customer, uh, is running, uh, both ESI and hyper V in that environment. And they don't have a solution to do micro-segmentation consistently across hypervisors. >>So again, micro-segmentation is a huge driver security looks like it's a recurring theme, uh, talking to most of these customers and in the Tyco space, um, uh, we're working with a few types of customers on the CFT program, uh, where the main goal is actually to our Monet's network operation. They typically handle all the VNF search with their own homegrown DPDK stack. This is overly complex. It is frankly also as low and inefficient, and then they have a physical network to manage the, the idea of having again, one network, uh, to coordinate the provision in our cloud services between the, the take of VNF, uh, and, uh, the rest of the infrastructure, uh, is extremely powerful on top of the offloading capability of the, by the bluefin DPOs. Those are just some examples. >>That was a great use case, a lot more potential. I see that with the unified cloud networking, great stuff, feed, shout out to you guys at Nvidia had been following your success for a long time and continuing to innovate as cloud scales and pluribus here with the unified networking, kind of bring it to the next level. Great stuff. Great to have you guys on. And again, software keeps driving the innovation again, networking is just a part of it, and it's the key solution. So I got to ask both of you to wrap this up. How can cloud operators who are interested in, in this, uh, new architecture and solution, uh, learn more because this is an architectural shift. People are working on this problem. They're trying to think about multiple clouds of trying to think about unification around the network and giving more security, more flexibility, uh, to their teams. How can people learn more? >>Yeah, so, uh, all Sandra and I have a talk at the upcoming Nvidia GTC conference. Um, so that's the week of March 21st through 24th. Um, you can go and register for free and video.com/at GTC. Um, you can also watch recorded sessions if you ended up watching us on YouTube a little bit after the fact. Um, and we're going to dive a little bit more into the specifics and the details and what we're providing in the solution. >>Alexandra, how can people learn more? >>Yeah, absolutely. People can go to the pluribus, a website, www boost networks.com/eft, and they can fill up the form and, uh, they will contact durables to either know more or to know more and actually to sign up for the actual early field trial program, which starts at the end of April. >>Okay. Well, we'll leave it there. Thanks. You both for joining. Appreciate it up next. You're going to hear an independent analyst perspective and review some of the research from the enterprise strategy group ESG. I'm John ferry with the >>Cube. Thanks for watching. >>Okay. We've heard from the folks at networks and Nvidia about their effort to transform cloud networking and unify bespoke infrastructure. Now let's get the perspective from an independent analyst and to do so. We welcome in ESG, senior analysts, Bob LA Liberte, Bob. Good to see you. Thanks for coming into our east coast studios. >>Oh, thanks for having me. It's great to be >>Here. Yeah. So this, this idea of unified cloud networking approach, how serious is it? What's what's driving it. >>Yeah, there's certainly a lot of drivers behind it, but probably the first and foremost is the fact that application environments are becoming a lot more distributed, right? So the, it pendulum tends to swing back and forth. And we're definitely on one that's swinging from consolidated to distributed. And so applications are being deployed in multiple private data centers, multiple public cloud locations, edge locations. And as a result of that, what you're seeing is a lot of complexity. So organizations are having to deal with this highly disparate environment. They have to secure it. They have to ensure connectivity to it and all that's driving up complexity. In fact, when we asked in one of our last surveys and last year about network complexity, more than half 54% came out and said, Hey, our network environment is now either more or significantly more complex than it used to be. >>And as a result of that, what you're seeing is it's really impacting agility. So everyone's moving to these modern application environments, distributing them across areas so they can improve agility yet it's creating more complexity. So a little bit counter to the fact and, you know, really counter to their overarching digital transformation initiatives. From what we've seen, you know, nine out of 10 organizations today are either beginning in process or have a mature digital transformation process or initiative, but their top goals, when you look at them, it probably shouldn't be a surprise. The number one goal is driving operational efficiency. So it makes sense. I've distributed my environment to create agility, but I've created a lot of complexity. So now I need these tools that are going to help me drive operational efficiency, drive better experience. >>I mean, I love how you bring in the data yesterday. Does a great job with that. Uh, questions is, is it about just unifying existing networks or is there sort of a need to rethink kind of a do-over network, how networks are built? >>Yeah, that's a, that's a really good point because certainly unifying networks helps right. Driving any kind of operational efficiency helps. But in this particular case, because we've made the transition to new application architectures and the impact that's having as well, it's really about changing and bringing in new frameworks and new network architectures to accommodate those new application architectures. And by that, what I'm talking about is the fact that these new modern application architectures, microservices, containers are driving a lot more east west traffic. So in the old days, it used to be easier in north south coming out of the server, one application per server, things like that. Right now you've got hundreds, if not thousands of microservices communicating with each other users communicating to them. So there's a lot more traffic and a lot of it's taking place within the servers themselves. The other issue that you starting to see as well from that security perspective, when we were all consolidated, we had those perimeter based legacy, you know, castle and moat security architectures, but that doesn't work anymore when the applications aren't in the castle, right. >>When everything's spread out that that no longer happens. So we're absolutely seeing, um, organizations trying to, trying to make a shift. And, and I think much, like if you think about the shift that we're seeing with all the remote workers and the sassy framework to enable a secure framework there, this it's almost the same thing. We're seeing this distributed services framework come up to support the applications better within the data centers, within the cloud data centers, so that you can drive that security closer to those applications and make sure they're, they're fully protected. Uh, and that's really driving a lot of the, you know, the zero trust stuff you hear, right? So never trust, always verify, making sure that everything is, is, is really secure micro-segmentation is another big area. So ensuring that these applications, when they're connected to each other, they're, they're fully segmented out. And that's again, because if someone does get a breach, if they are in your data center, you want to limit the blast radius, you want to limit the amount of damage that's done. So that by doing that, it really makes it a lot harder for them to see everything that's in there. >>You know, you mentioned zero trust. It used to be a buzzword, and now it's like become a mandate. And I love the mode analogy. You know, you build a moat to protect the queen and the castle, the Queens left the castles, it's just distributed. So how should we think about this, this pluribus and Nvidia solution. There's a spectrum, help us understand that you've got appliances, you've got pure software solutions. You've got what pluribus is doing with Nvidia, help us understand that. >>Yeah, absolutely. I think as organizations recognize the need to distribute their services to closer to the applications, they're trying different models. So from a legacy approach, you know, from a security perspective, they've got these centralized firewalls that they're deploying within their data centers. The hard part for that is if you want all this traffic to be secured, you're actually sending it out of the server up through the rack, usually to in different location in the data center and back. So with the need for agility, with the need for performance, right, that adds a lot of latency. Plus when you start needing to scale, that means adding more and more network connections, more and more appliances. So it can get very costly as well as impacting the performance. The other way that organizations are seeking to solve this problem is by taking the software itself and deploying it on the servers. Okay. So that's a, it's a great approach, right? It brings it really close to the applications, but the things you start running into there, there's a couple of things. One is that you start seeing that the DevOps team start taking on that networking and security responsibility, which they >>Don't want to >>Do, they don't want to do right. And the operations teams loses a little bit of visibility into that. Um, plus when you load the software onto the server, you're taking up precious CPU cycles. So if you're really wanting your applications to perform at an optimized state, having additional software on there, isn't going to, isn't going to do it. So, you know, when we think about all those types of things, right, and certainly the other side effects of that is the impact of the performance, but there's also a cost. So if you have to buy more servers because your CPU's are being utilized, right, and you have hundreds or thousands of servers, right, those costs are going to add up. So what, what Nvidia and pluribus have done by working together is to be able to take some of those services and be able to deploy them onto a smart Nick, right? >>To be able to deploy the DPU based smart SMARTNICK into the servers themselves. And then pluribus has come in and said, we're going to unify create that unified fabric across the networking space, into those networking services all the way down to the server. So the benefits of having that are pretty clear in that you're offloading that capability from the server. So your CPU's are optimized. You're saving a lot of money. You're not having to go outside of the server and go to a different rack somewhere else in the data center. So your performance is going to be optimized as well. You're not going to incur any latency hit for every trip round trip to the, to the firewall and back. So I think all those things are really important. Plus the fact that you're going to see from a, an organizational aspect, we talked about the dev ops and net ops teams. The network operations teams now can work with the security teams to establish the security policies and the networking policies. So that they've dev ops teams. Don't have to worry about that. So essentially they just create the guardrails and let the dev op team run. Cause that's what they want. They want that agility and speed. >>Yeah. Your point about CPU cycles is key. I mean, it's estimated that 25 to 30% of CPU cycles in the data center are wasted. The cores are wasted doing storage offload or, or networking or security offload. And, you know, I've said many times everybody needs a nitro like Amazon nugget, but you can't go, you can only buy Amazon nitro if you go into AWS. Right. Everybody needs a nitro. So is that how we should think about this? >>Yeah. That's a great analogy to think about this. Um, and I think I would take it a step further because it's, it's almost the opposite end of the spectrum because pluribus and video are doing this in a very open way. And so pluribus has always been a proponent of open networking. And so what they're trying to do is extend that now to these distributed services. So leverage working with Nvidia, who's also open as well, being able to bring that to bear so that organizations can not only take advantage of these distributed services, but also that unified networking fabric, that unified cloud fabric across that environment from the server across the switches, the other key piece of what pluribus is doing, because they've been doing this for a while now, and they've been doing it with the older application environments and the older server environments, they're able to provide that unified networking experience across a host of different types of servers and platforms. So you can have not only the modern application supported, but also the legacy environments, um, you know, bare metal. You could go any type of virtualization, you can run containers, et cetera. So a wide gambit of different technologies hosting those applications supported by a unified cloud fabric from pluribus. >>So what does that mean for the customer? I don't have to rip and replace my whole infrastructure, right? >>Yeah. Well, think what it does for, again, from that operational efficiency, when you're going from a legacy environment to that modern environment, it helps with the migration helps you accelerate that migration because you're not switching different management systems to accomplish that. You've got the same unified networking fabric that you've been working with to enable you to run your legacy as well as transfer over to those modern applications. Okay. >>So your people are comfortable with the skillsets, et cetera. All right. I'll give you the last word. Give us the bottom line here. >>So yeah, I think obviously with all the modern applications that are coming out, the distributed application environments, it's really posing a lot of risk on these organizations to be able to get not only security, but also visibility into those environments. And so organizations have to find solutions. As I said, at the beginning, they're looking to drive operational efficiency. So getting operational efficiency from a unified cloud networking solution, that it goes from the server across the servers to multiple different environments, right in different cloud environments is certainly going to help organizations drive that operational efficiency. It's going to help them save money for visibility, for security and even open networking. So a great opportunity for organizations, especially large enterprises, cloud providers who are trying to build that hyperscaler like environment. You mentioned the nitro card, right? This is a great way to do it with an open solution. >>Bob, thanks so much for, for coming in and sharing your insights. Appreciate it. >>You're welcome. Thanks. >>Thanks for watching the program today. Remember all these videos are available on demand@thekey.net. You can check out all the news from today@siliconangle.com and of course, pluribus networks.com many thanks diplomas for making this program possible and sponsoring the cube. This is Dave Volante. Thanks for watching. Be well, we'll see you next time.

Published Date : Mar 16 2022

SUMMARY :

And one of the best examples is Amazon's nitro. So if you can eliminate that waste, and Pete Lummus from Nvidia to take a deeper dive into the technology. Great to have you welcome folks. Thank you. So let's get into the, the problem situation with cloud unified network. and the first mandate for them is to become as agile as a hyperscaler. How does this tie together? Each of the public clouds have different networks that needs to be unified. So that's the fourth tenant How do customers get this vision realized? And I appreciate the tee up. That's the blue field and video. And so that is the first that's, that's the first step in the getting into realizing What is the relationship with clothes? So we have, you know, this concept of a Bluefield data processing unit, which if you think about it, the host, from the switch to the host, and really have that single pane of glass for So it really is a magical partnership between the two companies with pulled out of the market and, and you guys step up and create these new solutions. Um, so that, you know, if you sort of think about what, So if you look at what we've done with the DPU, with credit and an SDK, which is an open SDK called And it's all kind of, again, this is the new architecture Mike, you were talking about, how does customers So they need to migrate there and they need this architecture to be cost-effective. And then, um, uh, you know, with this, with this, our architectural approach effectively, Get the unified cloud architecture, I'm the customer guy, So now by, by terminating the networking on the DPU, Um, and the next benefit obviously So you have to have this new security model. And I think that's sort of what you see with the partnership between pluribus and Nvidia is the DPU is really the the go to market with an Nvidia? in the future, but right now, um, we're, we feel like we're partnered with the number one, And I talked about sort of, you know, uh, how much better that next generation of Bluefield So as we add new generations of Bluefield, you know, next, This is the future of, of cloud operations. You can go to www.pluribusnetworks.com/e Thanks so much for sharing the news. How can you simplify and unify your cloud networks to increase agility and business velocity? Ultimately the unified cloud fabric extends seamlessly across And we'll examine some of the use cases with Alessandra Burberry, Um, and the novelty about this system that integrates a distributed control So how does it integrate into Nvidia hardware and specifically So the first byproduct of this approach is that whatever And second, this gives you the ability to free up, I would say around 20, and this is what we think this enables a very clean demarcation between computer and So Pete, I gotta get, I gotta get you in here. And so, you know, again, it comes down to pragmatism and I think, So if infrastructure is code, you know, you're talking about, you know, that part of the stack And so that ability to automate, into the pluribus unified cloud networking vision, because this is what people are talking but this is not the place where you deploy most of your services in the cloud, particularly from a security standpoint, on the kind of hypervisor or compute solution you choose. That's probably the number one, I mean, I'm not, I'm just joking server listen network list, but the idea is it should the Butte technology and, uh, uh, we can actually integrate those capabilities directly So I love to get your thoughts about Uh, and so we provide, uh, you know, armed CPU's memory scale large enough, the question is, can you afford it? What's the benefit to the customers with this approach? And so you can run a DPU You've already identified some successes with some customers on your early field trials. couple of examples here, just to give you a flavor. And overall, they're hoping to get out of it, uh, uh, tremendous, and then they have a physical network to manage the, the idea of having again, one network, So I got to ask both of you to wrap this up. Um, so that's the week of March 21st through 24th. more or to know more and actually to sign up for the actual early field trial program, You're going to hear an independent analyst perspective and review some of the research from the enterprise strategy group ESG. Now let's get the perspective It's great to be What's what's driving it. So organizations are having to deal with this highly So a little bit counter to the fact and, you know, really counter to their overarching digital transformation I mean, I love how you bring in the data yesterday. So in the old days, it used to be easier in north south coming out of the server, So that by doing that, it really makes it a lot harder for them to see And I love the mode analogy. but the things you start running into there, there's a couple of things. So if you have to buy more servers because your CPU's are being utilized, the server and go to a different rack somewhere else in the data center. So is that how we should think about this? environments and the older server environments, they're able to provide that unified networking experience across environment, it helps with the migration helps you accelerate that migration because you're not switching different management I'll give you the last word. that it goes from the server across the servers to multiple different environments, right in different cloud environments Bob, thanks so much for, for coming in and sharing your insights. You're welcome. You can check out all the news from today@siliconangle.com and of course,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DonniePERSON

0.99+

Bob LibertePERSON

0.99+

Dave VellantePERSON

0.99+

AmazonORGANIZATION

0.99+

Alessandra BurberryPERSON

0.99+

SandraPERSON

0.99+

Dave VolantePERSON

0.99+

NvidiaORGANIZATION

0.99+

Pete BloombergPERSON

0.99+

MichaelPERSON

0.99+

AsiaLOCATION

0.99+

AlexandraPERSON

0.99+

hundredsQUANTITY

0.99+

Pete LummusPERSON

0.99+

AWSORGANIZATION

0.99+

Bob LA LibertePERSON

0.99+

MikePERSON

0.99+

JohnPERSON

0.99+

ESGORGANIZATION

0.99+

BobPERSON

0.99+

two companiesQUANTITY

0.99+

25QUANTITY

0.99+

Alessandra BobbyPERSON

0.99+

two yearsQUANTITY

0.99+

Palo AltoLOCATION

0.99+

thousandsQUANTITY

0.99+

BluefieldORGANIZATION

0.99+

NetAppsORGANIZATION

0.99+

demand@thekey.netOTHER

0.99+

20%QUANTITY

0.99+

last yearDATE

0.99+

a yearQUANTITY

0.99+

March 21stDATE

0.99+

FirstQUANTITY

0.99+

www.pluribusnetworks.com/eOTHER

0.99+

TycoORGANIZATION

0.99+

late AprilDATE

0.99+

DokaTITLE

0.99+

400 gigQUANTITY

0.99+

yesterdayDATE

0.99+

second versionQUANTITY

0.99+

two servicesQUANTITY

0.99+

first stepQUANTITY

0.99+

third areaQUANTITY

0.99+

oneQUANTITY

0.99+

second aspectQUANTITY

0.99+

OneQUANTITY

0.99+

EachQUANTITY

0.99+

www.pluribusnetworks.comOTHER

0.99+

PetePERSON

0.99+

last yearDATE

0.99+

one applicationQUANTITY

0.99+

two thingsQUANTITY

0.99+

Vasanth Kumar, MongoDB Principal Solutions Architect | Io-Tahoe Episode 7


 

>> Okay. We're here with Vasanth Kumar who's the Principal Solutions Architect for MongoDB. Vasanth, welcome to "theCube." >> Thanks Dave. >> Hey, listen, I feel like you were born to be an architect in technology. I mean, you've worked for big SIs, you've worked with many customers, you have experience in financial services and banking. Tell us, the audience, a little bit more about yourself, and what you're up to these days. >> Yeah. Hi, thanks for the for inviting me for this discussion. I'm based out of Bangalore, India, having around 18 years experience in IT industry, building enterprise products for different domains, verticals, finance built and enterprise banking applications, IOT platforms, digital experience solutions. Now being with MongoDB nearly two years, been working in a partner team as a principal solutions architect, especially working with ISBs to build the best practices of handling the data and embed the right database as part of their product. I also worked with technology partners to integrate the compatible technology compliance with MongoDB. And also worked with the private cloud providers to provide a database as a service. >> Got it. So, you know, I have to Vasanth, I think Mongo, you kind of nailed it. They were early on with the trends of managing unstructured data, making it really simple. There was always a developer appeal, which has lasted and then doing so with an architecture that scales out, and back in the early days when Mongo was founded, I remember those days, I mean, digital transformation, wasn't a thing, it wasn't a buzz word, but it just so happens that Mongo's approach, it dovetails very nicely with a digital business. So I wonder if you could talk about that, talk about the fit and how MongoDB thinks about accelerating digital transformation and why you're different from like a traditional RDBMS. >> Sure, exactly, yeah. You had a right understanding, let me elaborate it. So we all know that the customer expectation changes day by day, because of the business agility functionality changes, how they want to experience the applications, or in apps that changes okay. And obviously this yields to the agility of the information which transforms between the multiple systems or layers. And to achieve this, obviously the way of architecting or developing the product as completely a different shift, might be moving from the monolith to microservices or event-based architecture and so on. And obviously the database has to be opt for these environment to adopt these changes, to adopt the scale of load and the other thing. Okay. And also like we see that the common, the protocol for the information exchange is JSON, and something like you, you adopt it. The database adopts it natively to that is a perfect fit. Okay. So that's where the MongoDB fits perfectly for billing or transforming the modern applications, because it's a general purpose database which accepts the JSON as a payload and stores it in a BSON format. You don't need to be, suppose like to develop any particular application or to transfer an existing application, typically they see the what is the effort required and how much, what is the cost involved in it, and how quickly I can do that. That's main important thing without disturbing the functionality here where, since it is a multimodal database in a JSON format, you don't easily build an application. Okay? Don't need a lot of transformation in case of an RDBMS, you get the JSON payload, you transform into a tabular structure or a different format, and then probably you build an ORM layer and then map it and save it. There are lot of work involved in it. There are a lot of components need to be written in between. But in case of MongoDB, what they can do is you get the information from the multiple sources. And as is, you can put it in a DB based on where, or you can transform it based on the access patterns. And then you can store it quickly. >> Dave: Got it. And I tell Dave, because today you haven't context data, which has a selected set of information. Probably tomorrow the particular customer has more information to put it. So how do you capture that? In case of an RDBMS, you need to change the schema. Once you scheme change the schema, your application breaks down. But here it magically adopts it. Like you pass the extra information, it's open for extension. It adopts it easily. You don't need to redeploy or change the schema or do something like that. >> Right. That's the genius of Mongo. And then of course, you know, in the early days people say, oh, you know, Mongo, it won't scale. And then of course we, through the cloud. And I follow very closely Atlas. I look at the numbers every quarter. I mean, overall cloud adoption is increasing like crazy, you know, our Wiki Bon analyst team. We got the big four cloud vendors just in IAS growing beyond a 115 billion this year. That's 35% on top of, you know, 80-90 billion last year. So talk more about how MongoDB fits with the cloud and how it helps with the whole migration story. 'Cause you're killing it in that space. >> Yeah. Sure. Just to add one more point on the previous question. So for continuously, for past four to five years, we have been the number one in the wanted database. >> Dave: Right Okay. That that's how like the popularity is getting done. That's how the adoption has happened. >> Dave: Right. >> I'm coming back to your question- >> Yeah let's talk about the cloud and database as a service, you guys actually have packaged that very nicely I have to say. >> Yeah. So we have spent lot of effort and time in developing Atlas, our managed database as a service, which typically gives the customer the way of just concentrating on their application rather than maintaining and managing the whole set of database or how to scale infrastructure. All those things on work is taken care. You don't need to be an expert of DB, like when you are using an Atlas. So we provide the managed database in three major cloud providers, AWS, GCP, and Azure, and also it's a purely a multicloud, you know, like you can have a primary in AWS and you have the replicated nodes in GCP or Azure. It's a purely multicloud. So that like, you don't have a cloud blocking. You feel that, okay, your business is, I mean, if this is the right for your business you are choosing the model, you think that I need to move to GCP. You don't need to bother, you easily migrate this to GCP. Okay. No vendor lock in, no cloud lock in this particular- >> So Vasanth, maybe you could talk a little bit more about Atlas and some of the differentiable features and things that you can do with Atlas that maybe people don't know about. >> Yeah, sure Dave like, Atlas is not just a manage database as a service, you know, like it's a complete data platform and it provides many features. Like for example, you build an application and probably down the line of three years, the data which you captured three years back might be an old data. Like how do you do it? Like there's no need for you to manually purge or do thing. Like we do have an online archival where you configure the data. So that like the data, which is older than two years, just purge it. So automatically this is taken care. So that like you have hot data kept in Atlas cluster and the cold data moved up to an ARKit. And also like we have a data lake where you can run a federated queries . For example, you've done an archival, but what if people want to access the data? So with data lake, what it can do is, on a single connection, you can fire a- you can run a federated queries both on the active and the archival data. That's the beauty, like you archive the data, but still you can able to query it. And we do also have a charts where like, you can build in visualization on top of the data, what you have captured. You can build in graphs or you can build in graphs and also embed these graphs as part of your application, or you can collaborate to the customers, to the CXOs and other theme. >> Dave: Got it. >> It's a complete data platform. >> Okay. Well, speaking of data platform, let's talk about Io-Tahoe's data RPA platform, and coupling that with Mongo DB. So maybe you could help us understand how you're helping with process automation, which is a very hot topic and just this whole notion of a modern application development. >> Sure. See, the process automation is more with respect to the data and how you manage this data and what to derive and build a business process on top of it. I see there are two parts into it. Like one is the source of data. How do you identify, how do you discover the data? How do you enrich the context or transform it, give a business context to it. And then you build a business rules or act on it, and then you store the data or you derive the insights or enrich it and store it into DB. The first part is completely taken by Io-Tahoe, where you can tag the data for the multiple data sources. For example, if we take an customer 360 view, you can grab the data from multiple data sources using Io-Tahoe and you discover this data, you can tag it, you can label it and you build a view of the complete customer context, and use a realm web book and then the data is ingested back to Mongo. So that's all like more sort of like server-less fashion. You can build this particular customer 360 view for example. And just to talk about the realm I spoke, right? The realm web book, realm is a backend APA that you can create on top of the data on Mongo cluster, which is available in addclass. Okay. Then once you run, the APS are ready. Data as a service, you build it as a data as a service, and you fully secure APIs, which are available. These APS can be integrated within a mobile app or an web application to build in a built in modern application. But what left out is like, just build a UI artifacts and integrate these APIs. >> Yeah, I mean we live in this API economy companies. People throw that out as sort of a buzz phrase, but Mongo lives that. I mean, that's why developers really like the Mongo. So what's your take on DevOps? Maybe you could talk a little bit about, you know, your perspective there, how you help Devs and data engineers build faster pipelines. >> Yeah, sure. Like, okay, this is the most favorite topic. Like, no, and it's a buzzword along, like all the DevOps moving out from the traditional deployment, what I learned online. So like we do support like the deployment automation in multiple ways okay, and also provide the diagnostic under the hood. We have two options in Mongo DB. One is an enterprise option, which is more on the on-prem's version. And Atlas is more with respect to the cloud one manage database service. Okay. In case of an enterprise advanced, like we do have an Ops manager and the Kubernetes operator, like a Ops manager will manage all sort of deployment automation. Upgrades, provides your diagnostics, both with respect to the hardwares, and also with respect to the MongoDB gives you a profiling, slow running queries and what you can get a context of what's working on the data using that. I'm using an enterprise operator. You can integrate with existing Kubernetes cluster, either in a different namespace on an existing namespace. And orchestrate the deployment. And in case of Atlas, we do have an Atlas-Kubernetes operator, which helps you to integrate your Kubernetes operator. And you don't need to leave your Kubernetes. And also we have worked with the cloud providers. For example, we have we haven't cloud formation templates where you can just in one click, you can just roll out an Atlas cluster with a complete platform. So that's one, like we are continuously working, evolving on the DevOps site to roll out the might be a helm chart, or we do have an operator, which has a standard (indistinct) for different types of deployments. >> You know, some really important themes here. Obviously, anytime you talk about Mongo, simplicity comes in, automation, you know, that big, big push that Io-Tahoe was making. What you said about data context was interesting because a lot of data systems, organizations, they lack context and context is very important. So auto classification and things like that. And the other thing you said about federated queries I think fits very well into the trend toward decentralized data architecture. So very important there. And of course, hybridisity. I call it hybridisity. On-prem, cloud, abstracting that complexity away and allowing people to really focus on their digital transformations. I tell ya, Vasanth, it's great stuff. It's always a pleasure chatting with Io-Tahoe partners, and really getting into the tech with folks like yourself. So thanks so much for coming on theCube. >> Thanks. Thanks, Dave. Thanks for having a nice discussion with you. >> Okay. Stay right there. We've got one more quick session that you don't want to miss.

Published Date : Aug 10 2021

SUMMARY :

Okay. We're here with Vasanth Kumar you have experience in of handling the data and and back in the early days And then you can store it quickly. So how do you capture that? And then of course, you know, on the previous question. That's how the adoption has happened. you guys actually have So that like, you don't So Vasanth, maybe you could talk the data which you So maybe you could help us and then you store the data little bit about, you know, and what you can get a context And the other thing you discussion with you. that you don't want to miss.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Vasanth KumarPERSON

0.99+

MongoORGANIZATION

0.99+

DavePERSON

0.99+

two partsQUANTITY

0.99+

VasanthPERSON

0.99+

35%QUANTITY

0.99+

tomorrowDATE

0.99+

last yearDATE

0.99+

115 billionQUANTITY

0.99+

first partQUANTITY

0.99+

Bangalore, IndiaLOCATION

0.99+

three yearsQUANTITY

0.99+

JSONTITLE

0.99+

AWSORGANIZATION

0.99+

Io-TahoeORGANIZATION

0.99+

80-90 billionQUANTITY

0.99+

MongoDBORGANIZATION

0.99+

ARKitTITLE

0.98+

two optionsQUANTITY

0.98+

one clickQUANTITY

0.98+

todayDATE

0.98+

AtlasTITLE

0.98+

this yearDATE

0.98+

bothQUANTITY

0.97+

older than two yearsQUANTITY

0.97+

around 18 yearsQUANTITY

0.96+

OneQUANTITY

0.96+

nearly two yearsQUANTITY

0.96+

AzureTITLE

0.96+

Wiki BonORGANIZATION

0.96+

MongoDBTITLE

0.95+

three years backDATE

0.94+

VasanthORGANIZATION

0.93+

Io-TahoeTITLE

0.92+

DevOpsTITLE

0.91+

AtlasORGANIZATION

0.91+

KubernetesTITLE

0.91+

360 viewQUANTITY

0.89+

oneQUANTITY

0.89+

single connectionQUANTITY

0.88+

five yearsQUANTITY

0.85+

one more quick sessionQUANTITY

0.83+

GCPORGANIZATION

0.83+

four cloud vendorsQUANTITY

0.82+

GCPTITLE

0.79+

three major cloud providersQUANTITY

0.76+

one more pointQUANTITY

0.73+

IoTITLE

0.72+

AzureORGANIZATION

0.72+

-TahoeORGANIZATION

0.68+

fourQUANTITY

0.67+

Mongo DBTITLE

0.65+

APATITLE

0.58+

ISBsORGANIZATION

0.54+

Christian Keynote with Disclaimer


 

(upbeat music) >> Hi everyone, thank you for joining us at the Data Cloud Summit. The last couple of months have been an exciting time at Snowflake. And yet, what's even more compelling to all of us at Snowflake is what's ahead. Today I have the opportunity to share new product developments that will extend the reach and impact of our Data Cloud and improve the experience of Snowflake users. Our product strategy is focused on four major areas. First, Data Cloud content. In the Data Cloud silos are eliminated and our vision is to bring the world's data within reach of every organization. You'll hear about new data sets and data services available in our data marketplace and see how previous barriers to sourcing and unifying data are eliminated. Second, extensible data pipelines. As you gain frictionless access to a broader set of data through the Data Cloud, Snowflakes platform brings additional capabilities and extensibility to your data pipelines, simplifying data ingestion, and transformation. Third, data governance. The Data Cloud eliminates silos and breaks down barriers and in a world where data collaboration is the norm, the importance of data governance is ratified and elevated. We'll share new advancements to support how the world's most demanding organizations mobilize your data while maintaining high standards of compliance and governance. Finally, our fourth area focuses on platform performance and capabilities. We remain laser focused on continuing to lead with the most performant and capable data platform. We have some exciting news to share about the core engine of Snowflake. As always, we love showing you Snowflake in action, and we prepared some demos for you. Also, we'll keep coming back to the fact that one of the characteristics of Snowflake that we're proud as staff is that we offer a single platform from which you can operate all of your data workloads, across clouds and across regions, which workloads you may ask, specifically, data warehousing, data lake, data science, data engineering, data applications, and data sharing. Snowflake makes it possible to mobilize all your data in service of your business without the cost, complexity and overhead of managing multiple systems, tools and vendors. Let's dive in. As you heard from Frank, the Data Cloud offers a unique capability to connect organizations and create collaboration and innovation across industries fueled by data. The Snowflake data marketplace is the gateway to the Data Cloud, providing visibility for organizations to browse and discover data that can help them make better decisions. For data providers on the marketplace, there is a new opportunity to reach new customers, create new revenue streams, and radically decrease the effort and time to data delivery. Our marketplace dramatically reduces the friction of sharing and collaborating with data opening up new possibilities to all participants in the Data Cloud. We introduced the Snowflake data marketplace in 2019. And it is now home to over 100 data providers, with half of them having joined the marketplace in the last four months. Since our most recent product announcements in June, we have continued broadening the availability of the data marketplace, across regions and across clouds. Our data marketplace provides the opportunity for data providers to reach consumers across cloud and regional boundaries. A critical aspect of the Data Cloud is that we envisioned organizations collaborating not just in terms of data, but also data powered applications and services. Think of instances where a provider doesn't want to open access to the entirety of a data set, but wants to provide access to business logic that has access and leverages such data set. That is what we call data services. And we want Snowflake to be the platform of choice for developing discovering and consuming such rich building blocks. To see How the data marketplace comes to live, and in particular one of these data services, let's jump into a demo. For all of our demos today, we're going to put ourselves in the shoes of a fictional global insurance company. We've called it Insureco. Insurance is a data intensive and highly regulated industry. Having the right access control and insight from data is core to every insurance company's success. I'm going to turn it over to Prasanna to show how the Snowflake data marketplace can solve a data discoverability and access problem. >> Let's look at how Insureco can leverage data and data services from the Snowflake data marketplace and use it in conjunction with its own data in the Data Cloud to do three things, better detect fraudulent claims, arm its agents with the right information, and benchmark business health against competition. Let's start with detecting fraudulent claims. I'm an analyst in the Claims Department. I have auto claims data in my account. I can see there are 2000 auto claims, many of these submitted by auto body shops. I need to determine if they are valid and legitimate. In particular, could some of these be insurance fraud? By going to the Snowflake data marketplace where numerous data providers and data service providers can list their offerings, I find the quantifying data service. It uses a combination of external data sources and predictive risk typology models to inform the risk level of an organization. Quantifying external sources include sanctions and blacklists, negative news, social media, and real time search engine results. That's a wealth of data and models built on that data which we don't have internally. So I'd like to use Quantifind to determine a fraud risk score for each auto body shop that has submitted a claim. First, the Snowflake data marketplace made it really easy for me to discover a data service like this. Without the data marketplace, finding such a service would be a lengthy ad hoc process of doing web searches and asking around. Second, once I find Quantifind, I can use Quantifind service against my own data in three simple steps using data sharing. I create a table with the names and addresses of auto body shops that have submitted claims. I then share the table with Quantifind to start the risk assessment. Quantifind does the risk scoring and shares the data back with me. Quantifind uses external functions which we introduced in June to get results from their risk prediction models. Without Snowflake data sharing, we would have had to contact Quantifind to understand what format they wanted the data in, then extract this data into a file, FTP the file to Quantifind, wait for the results, then ingest the results back into our systems for them to be usable. Or I would have had to write code to call Quantifinds API. All of that would have taken days. In contrast, with data sharing, I can set this up in minutes. What's more, now that I have set this up, as new claims are added in the future, they will automatically leverage Quantifind's data service. I view the scores returned by Quantifind and see the two entities in my claims data have a high score for insurance fraud risk. I open up the link returned by Quantifind to read more, and find that this organization has been involved in an insurance crime ring. Looks like that is a claim that we won't be approving. Using the Quantifind data service through the Snowflake data marketplace gives me access to a risk scoring capability that we don't have in house without having to call custom APIs. For a provider like Quantifind this drives new leads and monetization opportunities. Now that I have identified potentially fraudulent claims, let's move on to the second part. I would like to share this fraud risk information with the agents who sold the corresponding policies. To do this, I need two things. First, I need to find the agents who sold these policies. Then I need to share with these agents the fraud risk information that we got from Quantifind. But I want to share it such that each agent only sees the fraud risk information corresponding to claims for policies that they wrote. To find agents who sold these policies, I need to look up our Salesforce data. I can find this easily within Insureco's internal data exchange. I see there's a listing with Salesforce data. Our sales Ops team has published this listing so I know it's our officially blessed data set, and I can immediately access it from my Snowflake account without copying any data or having to set up ETL. I can now join Salesforce data with my claims to identify the agents for the policies that were flagged to have fraudulent claims. I also have the Snowflake account information for each agent. Next, I create a secure view that joins on an entitlements table, such that each agent can only see the rows corresponding to policies that they have sold. I then share this directly with the agents. This share contains the secure view that I created with the names of the auto body shops, and the fraud risk identified by Quantifind. Finally, let's move on to the third and last part. Now that I have detected potentially fraudulent claims, I'm going to move on to building a dashboard that our executives have been asking for. They want to see how Insureco compares against other auto insurance companies on key metrics, like total claims paid out for the auto insurance line of business nationwide. I go to the Snowflake data marketplace and find SNL U.S. Insurance Statutory Data from SNP. This data is included with Insureco's existing subscription with SMP so when I request access to it, SMP can immediately share this data with me through Snowflake data sharing. I create a virtual database from the share, and I'm ready to query this data, no ETL needed. And since this is a virtual database, pointing to the original data in SNP Snowflake account, I have access to the latest data as it arrives in SNPs account. I see that the SNL U.S. Insurance Statutory Data from SNP has data on assets, premiums earned and claims paid out by each us insurance company in 2019. This data is broken up by line of business and geography and in many cases goes beyond the data that would be available from public financial filings. This is exactly the data I need. I identify a subset of comparable insurance companies whose net total assets are within 20% of Insureco's, and whose lines of business are similar to ours. I can now create a Snow site dashboard that compares Insureco against similar insurance companies on key metrics, like net earned premiums, and net claims paid out in 2019 for auto insurance. I can see that while we are below median our net earned premiums, we are doing better than our competition on total claims paid out in 2019, which could be a reflection of our improved claims handling and fraud detection. That's a good insight that I can share with our executives. In summary, the Data Cloud enabled me to do three key things. First, seamlessly fine data and data services that I need to do my job, be it an external data service like Quantifind and external data set from SNP or internal data from Insureco's data exchange. Second, get immediate live access to this data. And third, control and manage collaboration around this data. With Snowflake, I can mobilize data and data services across my business ecosystem in just minutes. >> Thank you Prasanna. Now I want to turn our focus to extensible data pipelines. We believe there are two different and important ways of making Snowflakes platform highly extensible. First, by enabling teams to leverage services or business logic that live outside of Snowflake interacting with data within Snowflake. We do this through a feature called external functions, a mechanism to conveniently bring data to where the computation is. We announced this feature for calling regional endpoints via AWS gateway in June, and it's currently available in public preview. We are also now in public preview supporting Azure API management and will soon support Google API gateway and AWS private endpoints. The second extensibility mechanism does the converse. It brings the computation to Snowflake to run closer to the data. We will do this by enabling the creation of functions and procedures in SQL, Java, Scala or Python ultimately providing choice based on the programming language preference for you or your organization. You will see Java, Scala and Python available through private and public previews in the future. The possibilities enabled by these extensibility features are broad and powerful. However, our commitment to being a great platform for data engineers, data scientists and developers goes far beyond programming language. Today, I am delighted to announce Snowpark a family of libraries that will bring a new experience to programming data in Snowflake. Snowpark enables you to write code directly against Snowflake in a way that is deeply integrated into the languages I mentioned earlier, using familiar concepts like DataFrames. But the most important aspect of Snowpark is that it has been designed and optimized to leverage the Snowflake engine with its main characteristics and benefits, performance, reliability, and scalability with near zero maintenance. Think of the power of a declarative SQL statements available through a well known API in Scala, Java or Python, all these against data governed in your core data platform. We believe Snowpark will be transformative for data programmability. I'd like to introduce Sri to showcase how our fictitious insurance company Insureco will be able to take advantage of the Snowpark API for data science workloads. >> Thanks Christian, hi, everyone? I'm Sri Chintala, a product manager at Snowflake focused on extensible data pipelines. And today, I'm very excited to show you a preview of Snowpark. In our first demo, we saw how Insureco could identify potentially fraudulent claims. Now, for all the valid claims InsureCo wants to ensure they're providing excellent customer service. To do that, they put in place a system to transcribe all of their customer calls, so they can look for patterns. A simple thing they'd like to do is detect the sentiment of each call so they can tell which calls were good and which were problematic. They can then better train their claim agents for challenging calls. Let's take a quick look at the work they've done so far. InsureCo's data science team use Snowflakes external functions to quickly and easily train a machine learning model in H2O AI. Snowflake has direct integrations with H2O and many other data science providers giving Insureco the flexibility to use a wide variety of data science libraries frameworks or tools to train their model. Now that the team has a custom trained sentiment model tailored to their specific claims data, let's see how a data engineer at Insureco can use Snowpark to build a data pipeline that scores customer call logs using the model hosted right inside of Snowflake. As you can see, we have the transcribed call logs stored in the customer call logs table inside Snowflake. Now, as a data engineer trained in Scala, and used to working with systems like Spark and Pandas, I want to use familiar programming concepts to build my pipeline. Snowpark solves for this by letting me use popular programming languages like Java or Scala. It also provides familiar concepts in APIs, such as the DataFrame abstraction, optimized to leverage and run natively on the Snowflake engine. So here I am in my ID, where I've written a simple scalar program using the Snowpark libraries. The first step in using the Snowpark API is establishing a session with Snowflake. I use the session builder object and specify the required details to connect. Now, I can create a DataFrame for the data in the transcripts column of the customer call logs table. As you can see, the Snowpark API provides native language constructs for data manipulation. Here, I use the Select method provided by the API to specify the column names to return rather than writing select transcripts as a string. By using the native language constructs provided by the API, I benefit from features like IntelliSense and type checking. Here you can see some of the other common methods that the DataFrame class offers like filters like join and others. Next, I define a get sentiment user defined function that will return a sentiment score for an input string by using our pre trained H2O model. From the UDF, we call the score method that initializes and runs the sentiment model. I've built this helper into a Java file, which along with the model object and license are added as dependencies that Snowpark will send to Snowflake for execution. As a developer, this is all programming that I'm familiar with. We can now call our get sentiment function on the transcripts column of the DataFrame and right back the results of the score transcripts to a new target table. Let's run this code and switch over to Snowflake to see the score data and also all the work that Snowpark has done for us on the back end. If I do a select star from scored logs, we can see the sentiment score of each call right alongside the transcript. With Snowpark all the logic in my program is pushed down into Snowflake. I can see in the query history that Snowpark has created a temporary Java function to host the pre trained H20 model, and that the model is running right in my Snowflake warehouse. Snowpark has allowed us to do something completely new in Snowflake. Let's recap what we saw. With Snowpark, Insureco was able to use their preferred programming language, Scala and use the familiar DataFrame constructs to score data using a machine learning model. With support for Java UDFs, they were able to run a train model natively within Snowflake. And finally, we saw how Snowpark executed computationally intensive data science workloads right within Snowflake. This simplifies Insureco's data pipeline architecture, as it reduces the number of additional systems they have to manage. We hope that extensibility with Scala, Java and Snowpark will enable our users to work with Snowflake in their preferred way while keeping the architecture simple. We are very excited to see how you use Snowpark to extend your data pipelines. Thank you for watching and with that back to you, Christian. >> Thank you Sri. You saw how Sri could utilize Snowpark to efficiently perform advanced sentiment analysis. But of course, if this use case was important to your business, you don't want to fully automate this pipeline and analysis. Imagine being able to do all of the following in Snowflake, your pipeline could start far upstream of what you saw in the demo. By storing your actual customer care call recordings in Snowflake, you may notice that this is new for Snowflake. We'll come back to the idea of storing unstructured data in Snowflake at the end of my talk today. Once you have the data in Snowflake, you can use our streams and past capabilities to call an external function to transcribe these files. To simplify this flow even further, we plan to introduce a serverless execution model for tasks where Snowflake can automatically size and manage resources for you. After this step, you can use the same serverless task to execute sentiment scoring of your transcript as shown in the demo with incremental processing as each transcript is created. Finally, you can surface the sentiment score either via snow side, or through any tool you use to share insights throughout your organization. In this example, you see data being transformed from a raw asset into a higher level of information that can drive business action, all fully automated all in Snowflake. Turning back to Insureco, you know how important data governance is for any major enterprise but particularly for one in this industry. Insurance companies manage highly sensitive data about their customers, and have some of the strictest requirements for storing and tracking such data, as well as managing and governing it. At Snowflake, we think about governance as the ability to know your data, manage your data and collaborate with confidence. As you saw in our first demo, the Data Cloud enables seamless collaboration, control and access to data via the Snowflake data marketplace. And companies may set up their own data exchanges to create similar collaboration and control across their ecosystems. In future releases, we expect to deliver enhancements that create more visibility into who has access to what data and provide usage information of that data. Today, we are announcing a new capability to help Snowflake users better know and organize your data. This is our new tagging framework. Tagging in Snowflake will allow user defined metadata to be attached to a variety of objects. We built a broad and robust framework with powerful implications. Think of the ability to annotate warehouses with cost center information for tracking or think of annotating tables and columns with sensitivity classifications. Our tagging capability will enable the creation of companies specific business annotations for objects in Snowflakes platform. Another key aspect of data governance in Snowflake is our policy based framework where you specify what you want to be true about your data, and Snowflake enforces those policies. We announced one such policy earlier this year, our dynamic data masking capability, which is now available in public preview. Today, we are announcing a great complimentary a policy to achieve row level security to see how role level security can enhance InsureCo's ability to govern and secure data. I'll hand it over to Artin for a demo. >> Hello, I'm Martin Avanes, Director of Product Management for Snowflake. As Christian has already mentioned, the rise of the Data Cloud greatly accelerates the ability to access and share diverse data leading to greater data collaboration across teams and organizations. Controlling data access with ease and ensuring compliance at the same time is top of mind for users. Today, I'm thrilled to announce our new row access policies that will allow users to define various rules for accessing data in the Data Cloud. Let's check back in with Insureco to see some of these in action and highlight how those work with other existing policies one can define in Snowflake. Because Insureco is a multinational company, it has to take extra measures to ensure data across geographic boundaries is protected to meet a wide range of compliance requirements. The Insureco team has been asked to segment what data sales team members have access to based on where they are regionally. In order to make this possible, they will use Snowflakes row access policies to implement row level security. We are going to apply policies for three Insureco's sales team members with different roles. Alice, an executive must be able to view sales data from both North America and Europe. Alex in North America sales manager will be limited to access sales data from North America only. And Jordan, a Europe sales manager will be limited to access sales data from Europe only. As a first step, the security administrator needs to create a lookup table that will be used to determine which data is accessible based on each role. As you can see, the lookup table has the row and their associated region, both of which will be used to apply policies that we will now create. Row access policies are implemented using standard SQL syntax to make it easy for administrators to create policies like the one our administrators looking to implement. And similar to masking policies, row access policies are leveraging our flexible and expressive policy language. In this demo, our admin users to create a row access policy that uses the row and region of a user to determine what row level data they have access to when queries are executed. When users queries are executed against the table protected by such a row access policy, Snowflakes query engine will dynamically generate and apply the corresponding predicate to filter out rows the user is not supposed to see. With the policy now created, let's log in as our Sales Users and see if it worked. Recall that as a sales executive, Alice should have the ability to see all rows from North America and Europe. Sure enough, when she runs her query, she can see all rows so we know the policy is working for her. You may also have noticed that some columns are showing masked data. That's because our administrator's also using our previously announced data masking capabilities to protect these data attributes for everyone in sales. When we look at our other users, we should notice that the same columns are also masked for them. As you see, you can easily combine masking and row access policies on the same data sets. Now let's look at Alex, our North American sales manager. Alex runs to st Korea's Alice, row access policies leverage the lookup table to dynamically generate the corresponding predicates for this query. The result is we see that only the data for North America is visible. Notice too that the same columns are still masked. Finally, let's try Jordan, our European sales manager. Jordan runs the query and the result is only the data for Europe with the same columns also masked. And you reintroduced masking policies, today you saw row access policies in action. And similar to our masking policies, row access policies in Snowflake will be accepted Hands of capability integrated seamlessly across all of Snowflake everywhere you expect it to work it does. If you're accessing data stored in external tables, semi structured JSON data, or building data pipelines via streams or plan to leverage Snowflakes data sharing functionality, you will be able to implement complex row access policies for all these diverse use cases and workloads within Snowflake. And with Snowflakes unique replication feature, you can instantly apply these new policies consistently to all of your Snowflake accounts, ensuring governance across regions and even across different clouds. In the future, we plan to demonstrate how to combine our new tagging capabilities with Snowflakes policies, allowing advanced audit and enforcing those policies with ease. And with that, let's pass it back over to Christian. >> Thank you Artin. We look forward to making this new tagging and row level security capabilities available in private preview in the coming months. One last note on the broad area of data governance. A big aspect of the Data Cloud is the mobilization of data to be used across organizations. At the same time, privacy is an important consideration to ensure the protection of sensitive, personal or potentially identifying information. We're working on a set of product capabilities to simplify compliance with privacy related regulatory requirements, and simplify the process of collaborating with data while preserving privacy. Earlier this year, Snowflake acquired a company called Crypto Numerix to accelerate our efforts on this front, including the identification and anonymization of sensitive data. We look forward to sharing more details in the future. We've just shown you three demos of new and exciting ways to use Snowflake. However, I want to also remind you that our commitment to the core platform has never been greater. As you move workloads on to Snowflake, we know you expect exceptional price performance and continued delivery of new capabilities that benefit every workload. On price performance, we continue to drive performance improvements throughout the platform. Let me give you an example comparing an identical set of customers submitted queries that ran both in August of 2019, and August of 2020. If I look at the set of queries that took more than one second to compile 72% of those improved by at least 50%. When we make these improvements, execution time goes down. And by implication, the required compute time is also reduced. Based on our pricing model to charge for what you use, performance improvements not only deliver faster insights, but also translate into cost savings for you. In addition, we have two new major announcements on performance to share today. First, we announced our search optimization service during our June event. This service currently in public preview can be enabled on a table by table basis, and is able to dramatically accelerate lookup queries on any column, particularly those not used as clustering columns. We initially support equality comparisons only, and today we're announcing expanded support for searches in values, such as pattern matching within strings. This will unlock a number of additional use cases such as analytics on logs data for performance or security purposes. This expanded support is currently being validated by a few customers in private preview, and will be broadly available in the future. Second, I'd like to introduce a new service that will be in private preview in a future release. The query acceleration service. This new feature will automatically identify and scale out parts of a query that could benefit from additional resources and parallelization. This means that you will be able to realize dramatic improvements in performance. This is especially impactful for data science and other scan intensive workloads. Using this feature is pretty simple. You define a maximum amount of additional resources that can be recruited by a warehouse for acceleration, and the service decides when it would be beneficial to use them. Given enough resources, a query over a massive data set can see orders of magnitude performance improvement compared to the same query without acceleration enabled. In our own usage of Snowflake, we saw a common query go 15 times faster without changing the warehouse size. All of these performance enhancements are extremely exciting, and you will see continued improvements in the future. We love to innovate and continuously raise the bar on what's possible. More important, we love seeing our customers adopt and benefit from our new capabilities. In June, we announced a number of previews, and we continue to roll those features out and see tremendous adoption, even before reaching general availability. Two have those announcements were the introduction of our geospatial support and policies for dynamic data masking. Both of these features are currently in use by hundreds of customers. The number of tables using our new geography data type recently crossed the hundred thousand mark, and the number of columns with masking policies also recently crossed the same hundred thousand mark. This momentum and level of adoption since our announcements in June is phenomenal. I have one last announcement to highlight today. In 2014, Snowflake transformed the world of data management and analytics by providing a single platform with first class support for both structured and semi structured data. Today, we are announcing that Snowflake will be adding support for unstructured data on that same platform. Think of the abilities of Snowflake used to store access and share files. As an example, would you like to leverage the power of SQL to reason through a set of image files. We have a few customers as early adopters and we'll provide additional details in the future. With this, you will be able to leverage Snowflake to mobilize all your data in the Data Cloud. Our customers rely on Snowflake as the data platform for every part of their business. However, the vision and potential of Snowflake is actually much bigger than the four walls of any organization. Snowflake has created a Data Cloud a data connected network with a vision where any Snowflake customer can leverage and mobilize the world's data. Whether it's data sets, or data services from traditional data providers for SaaS vendors, our marketplace creates opportunities for you and raises the bar in terms of what is possible. As examples, you can unify data across your supply chain to accelerate your time and quality to market. You can build entirely new revenue streams, or collaborate with a consortium on data for good. The possibilities are endless. Every company has the opportunity to gain richer insights, build greater products and deliver better services by reaching beyond the data that he owns. Our vision is to enable every company to leverage the world's data through seamless and governing access. Snowflake is your window into this data network into this broader opportunity. Welcome to the Data Cloud. (upbeat music)

Published Date : Nov 19 2020

SUMMARY :

is the gateway to the Data Cloud, FTP the file to Quantifind, It brings the computation to Snowflake and that the model is running as the ability to know your data, the ability to access is the mobilization of data to

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
InsurecoORGANIZATION

0.99+

ChristianPERSON

0.99+

AlicePERSON

0.99+

August of 2020DATE

0.99+

August of 2019DATE

0.99+

JuneDATE

0.99+

InsureCoORGANIZATION

0.99+

Martin AvanesPERSON

0.99+

EuropeLOCATION

0.99+

QuantifindORGANIZATION

0.99+

PrasannaPERSON

0.99+

15 timesQUANTITY

0.99+

2019DATE

0.99+

AlexPERSON

0.99+

SNPORGANIZATION

0.99+

2014DATE

0.99+

JordanPERSON

0.99+

AWSORGANIZATION

0.99+

ScalaTITLE

0.99+

JavaTITLE

0.99+

72%QUANTITY

0.99+

SQLTITLE

0.99+

TodayDATE

0.99+

North AmericaLOCATION

0.99+

each agentQUANTITY

0.99+

SMPORGANIZATION

0.99+

second partQUANTITY

0.99+

FirstQUANTITY

0.99+

SecondQUANTITY

0.99+

SnowflakeORGANIZATION

0.99+

SnowflakeTITLE

0.99+

PythonTITLE

0.99+

each callQUANTITY

0.99+

Sri ChintalaPERSON

0.99+

each roleQUANTITY

0.99+

todayDATE

0.99+

TwoQUANTITY

0.99+

twoQUANTITY

0.99+

BothQUANTITY

0.99+

Crypto NumerixORGANIZATION

0.99+

two entitiesQUANTITY

0.99+

Krish Prasad and Manuvir Das | VMworld 2020


 

>> Narrator: From around the globe, it's theCube. With digital coverage of VMworld 2020. Brought to you by VMware and its ecosystem partners. >> Hello, and welcome back to theCube virtual coverage of VMworld 2020. I'm John Furrier, host of theCube. VMworld's not in person this year, it's on the virtual internet. A lot of content, check it out, vmworld.com, a lot of great stuff, online demos, and a lot of great keynotes. Here we got a great conversation to unpack, the NVIDIA, the AI and all things Cloud Native. With Krish Prasad, who's the SVP and GM of Cloud Platform, Business Unit, and Manuvir Das head of enterprise computing at NVIDIA. Gentlemen, great to see you virtually. Thanks for joining me on the virtual Cube, for the virtual VMworld 2020. >> Thank you John. >> Pleasure to be here. >> Quite a world. And I think one of the things that obviously we've been talking about all year since COVID is the acceleration of this virtualized environment with media and everyone working at home remote. Really puts the pressure on digital transformation Has been well discussed and documented. You guys have some big news, obviously on the main stage NVIDIA CEO, Jensen there legend. And of course, you know, big momentum with with AI and GPUs and all things, you know, computing. Krish, what are your announcements today? You got some big news. Could you take a minute to explain the big announcements today? >> Yeah, John. So today we want to make two major announcements regarding our partnership with NVIDIA. So let's take the first one, and talk through it and then we can get to the second announcement later. In the first one, as you well know, NVIDIA is the leader in AI and VMware as the leader in virtualization and cloud. This announcement is about us teaming up, deliver a jointly engineered solution to the market to bring AI to every enterprise. So as you well know, VMware has more than 300,000 customers worldwide. And we believe that this solution would enable our customers to transform their data centers or AI applications running on top of their virtualized VMware infrastructure that they already have. And we think that this is going to vastly accelerate the adoption of AI and essentially democratize AI in the enterprise. >> Why AI? Why now Manuvir? Obviously we know the GPUs have set the table for many cool things, from mining Bitcoin to really providing a great user experience. But AI has been a big driver. Why now? Why VMware now? >> Yes. Yeah. And I think it's important to understand this is about AI more than even about GPUs, you know. This is a great moment in time where AI has finally come to life, because the hardware and software has come together to make it possible. And if you just look at industries and different parts of life, how is AI impacting? So for example, if you're a company on the internet doing business, everything you do revolves around making recommendations to your customers about what they should do next. This is based on AI. Think about the world we live in today, with the importance of healthcare, drug discovery, finding vaccines for something like COVID. That work is dramatically accelerated if you use AI. And what we've been doing in NVIDIA over the years is, we started with the hardware technology with the GPU, the Parallel Processor, if you will, that could really make these algorithms real. And then we worked very hard on building up the ecosystem. You know, we have 2 million developers today who work with NVIDIA AI. That's thousands of companies that are using AI today. But then if you think about what Krish said, you know about the number of customers that VMware has, which is in the hundreds of thousands, the opportunity before us really now is, how do we democratize this? How do we take this power of AI, that makes every customer and every person better and put it in the hands of every enterprise customer? And we need a great vehicle for that, and that vehicle is VMware. >> Guys, before we get to the next question, I would just want to get your personal take on this, because again, we've talked many times, both of you've been on theCube on this topic. But now I want to highlight, you mentioned the GPU that's hardware. This is software. VMware had hardware partners and then still software's driving it. Software's driving everything. Whether it's something in space, it's an IOT device or anything at the edge of the network. Software, is the value. This has become so obvious. Just share your personal take on this for folks who are now seeing this for the first time. >> Yeah. I mean, I'll give you my take first. I'm a software guy by background, I learned a few years ago for the first time that an array is a storage device and not a data structure in programming. And that was a shock to my system. Definitely the world is based on algorithms. Algorithms are implemented in software. Great hardware enables those algorithms. >> Krish, your thoughts. we live we're living in the future right now. >> Yeah, yeah. I would say that, I mean, the developers are becoming the center. They are actually driving the transformation in this industry, right? It's all about the application development, it's all about software, the infrastructure itself is becoming software defined. And the reason for that is you want the developers to be able to craft the infrastructure the way they need for the applications to run on top of. So it's all about software like I said. >> Software defined. Yeah, just want to get that quick self-congratulatory high five amongst ourselves virtually. (laughs) Congratulations. >> Exactly. >> Krish, last time we spoke at VMworld, we were obviously in person, but we talked about Tanzu and vSphere. Okay, you had Project Pacific. Does this expand? Does this announcement expand on that offering? >> Absolutely. As you know John, for the past several years, VMware has been on this journey to define the Hybrid Cloud Infrastructure, right? Essentially is the software stack that we have, which will enable our customers to provide a cloud operating model to their developers, irrespective of where they want to land their workloads. Whether they want to land their workloads On-Premise, or if they want it to be on top of AWS, Google, Azure, VMware stack is already running across all of them as you well know. And in addition to that, we have around, you know, 4,000, 5,000 service providers who are also running our Platform to deliver cloud services to their customers. So as part of that journey, last year, we took the Platform and we added one further element to it. Traditionally, our platform has been used by customers for running via VMs. Last year, we natively integrated Kubernetes into our platform. This was the big re architecture of vSphere, as we talked about. That was delivered to the market. And essentially now customers can use the same platform to run Kubernetes, Containers and VM workloads. The exact same platform, it is operationally the same. So the same skillsets, tools and processes can be used to run Kubernetes as well as VM applications. And the same platform runs, whether you want to run it On-Premise or in any of the clouds, as we talked about before. So that vastly simplifies the operational complexity that our customers have to deal with. And this is the next chapter in that journey, by doing the same thing for AI workload. >> You guys had great success with these Co-Engineering joined efforts. VMware and now with NVIDIA is interesting. It's very relevant and is very cool. So it's cool and relevant, so check, check. Manuvir, talk about this, because how do you bring that vision to the enterprises? >> Yeah, John, I think, you know, it's important to understand there is some real deep Computer Science here between the Engineers at VMware and NVIDIA. Just to lay that out, you can think of this as a three layer stack, right? The first thing that you need is, clearly you need the hardware that is capable of running these algorithms, that's what the GPU enable. Then you need a great software stack for AI, all the right Algorithmics that take advantage of that hardware. This is actually where NVIDIA spends most of its effort today. People may sometimes think of NVIDIA as a GPU company, but we have much more a software company now, where we have over the years created a body of work of all of the software that it actually takes to do good AI. But then how do you marry the software stack with the hardware? You need a platform in the middle that supports the applications and consumes the hardware and exposes it properly. And that's where vSphere, you know, as Krish described with either VMs or Containers comes into the picture. So the Computer Science here is, to wire all these things up together with the right algorithmics so that you get real acceleration. So as examples of early work that the two teams have done together, we have workloads in healthcare, for example. In cancer detection, where the acceleration we get with this new stack is 30X, right? The workload is running 30 times faster than it was running before this integration just on CPUs. >> Great performance increase again. You guys are hiring a lot of software developers. I can attest to knowing folks in Silicon Valley and around the world. So I know you guys are bringing the software jobs to the table on a great product by the way, so congratulations. Krish, Democratization of AI for the enterprise. This is a liberating opportunity, because one of the things we've heard from your customers and also from VMware, but mostly from the customer's successes, is that there's two types of extremes. There's the, I'm going to modernize my business, certainly COVID forcing companies, whether they're airlines or whatever, not a lot going on, they have an opportunity to modernize, to essentially modern apps that are getting a tailwind from these new digital transformation accelerated. How does AI democratize this? Cause you got people and you've got technology. (laughs) Right? So share your thoughts on how you see this democratizing. >> That's a very good question. I think if you look at how people are running AI applications today, like you go to an enterprise, you would see that there is a silo of bare metal sun works on the side, where the AI stack is run. And you have people with specialized skills and different tools and utilities that manage that environment. And that is what is standing in the way of AI taking off in the enterprise, right? It is not the use case. There are all these use cases which are mission critical that all companies want to do, right? Worldwide, that has been the case. It is about the complexity of life that is standing in the way. So what we are doing with this is we are saying, "hey, that whole solution stack that Manuvir talked about, is integrated into the VMware Virtualized Infrastructure." Whether it's On-Prem or in the cloud. And you can manage that environment with the exact same tools and processes and skills that you traditionally had for running any other application on VMware infrastructure. So, you don't need to have anything special to run this. And that's what is going to give us the acceleration that we talked about and essentially hive the Democratization of AI. >> That's a great point. I just want to highlight that and call that out, because AI's every use case. You could almost say theCube could have AI and we do actually have a little bit of AI and some of our transcriptions and work. But it's not so much just use cases, it's actually not just saying you got to do it. So taking down that blocker, the complexity, certainly is the key. And that's a great point. We're going to call that out after. Alright, let's move on to the second part of the announcement. Krish Project Monterey. This is a big deal. And it looks like a, you know, kind of this elusive, it's architectural thing, but it's directionally really strategic for VMware. Could you take a minute to explain this announcement? Frame this for us. >> Absolutely. I think John, you remember Pat got on stage last year at Vmworld and said, you know, "we are undertaking the biggest re architecture of the vSphere platform in the last 10 years." And he was talking about natively embedding Kubernetes, in vSphere, right? Remember Tanzu and Project Pacific. This year we are announcing Project Monterrey. It's a project that is significant with several partners in the industry, along with NVIDIA was one of the key partners. And what we are doing is we are reimagination of the data center for the next generation applications. And at the center of it, what we are going to do is rearchitect vSphere and ESX. So that the ESX can normally run on the CPU, but it'll also run on the Smart Mix. And what this gives us is the whole, let's say data center, infrastructure type services to be offloaded from running on the CPU onto the Smart Mix. So what does this provide the applications? The applications then will perform better. And secondly, it provides an extra layer of security for the next generation applications. Now we are not going to stop there. We are going to use this architecture and extended it so that we can finally eliminate one of the big silos that exist in the enterprise, which is the bare metal silo. Right? Today we have virtualized environments and bare metal, and what this architecture will do is bring those bare metal environments also under ESX management. So you ESX will manage environments which are virtualized and environments which are running bare metal OS. And so that's one big breakthrough and simplification for the elimination of silo or the elimination of, you know, specialized skills to keep it running. And lastly, but most importantly, where we are going with this. That just on the question you asked us earlier about software defined and developers being in control. Where we want to go with this is give developers, the application developers, the ability to really define and create their run time on the Fly, dynamically. So think about it. If dynamically they're able to describe how the application should run. And the infrastructure essentially kind of attaches computer resources on the Fly, whether they are sitting in the same server or somewhere in the network as pools of resources. Bring it all together and compose the runtime environment for them. That's going to be huge. And they won't be constrained anymore by the resources that are tied to the physical server that they are running on. And that's the vision of where we are taking it. It is going to be the next big change in the industry in terms of enterprise computing. >> Sounds like an Operating System to me. Yeah. Run time, assembly orchestration, all these things coming together, exciting stuff. Looking forward to digging in more after Vmworld. Manuvir, how does this connect to NVIDIA and AI? Tie that together for us. >> Yeah, It's an interesting question, because you would think, you know, okay, so NVIDIA this GPU company or this AI company. But you have to remember that INVIDIA is also a networking company. Because friends at Mellanox joined us not that long ago. And the interesting thing is that there's a Yin and Yang here, because, Krish described the software vision, which is brilliant. And what this does is it imposes a lot on the host CPU of the server to do. And so what we've be doing in parallel is developing hardware. A new kind of "Nick", if you will, we call it a DPU or a Data Processing Unit or a Smart Nick that is capable of hosting all this stuff. So, amusingly when Krish and I started talking, we exchanged slides and we basically had the same diagram for our vision of where things go with that software, the infrastructure software being offloaded, data center infrastructure on a chip, if you will. Right? And so it's a very natural confluence. We are very excited to be part of this, >> Yeah. >> Monterey program with Krish and his team. And we think our DPU, which is called the NVIDIA BlueField-2, is a pretty good device to empower the work that Krish's team is doing. >> Guys it's awesome stuff. And I got to say, you know, I've been covering Vmworld now 11 years with theCube, and I've known VMware since its founding, just the evolution. And just recently before VMworld, you know, you saw the biggest IPO in the history of Wall Street, Snowflake an Enterprise Data Cloud Company. The number one IPO ever. Enterprise tech is so exciting. This is really awesome. And NVIDIA obviously well known, great brand. You own some chip company as well, and get processors and data and software. Guys, customers are going to be very interested in this, so what should customers do to find out more? Obviously you've got Project Monterey, strategic direction, right? Framed perfectly. You got this announcement. If I'm a customer, how do I get involved? How do I learn more? And what's in it for me. >> Yeah, John, I would say, sorry, go ahead, Krish. >> No, I was just going to say sorry Manuvir. I was just going to say like a lot of these discussions are going to be happening, there are going to be panel discussions there are going to be presentations at Vmworld. So I would encourage customers to really look at these topics around Project Monterey and also about the AI work we are doing with NVIDIA and attend those sessions and be active and we will have a ways for them to connect with us in terms of our early access programs and whatnot. And then as Manuvir was about to say, I think Manuvir, I will give it to you about GTC. >> Yeah, I think right after that, we have the NVIDIA conference, which is GTC, where we'll also go over this. And I think some of this work is a lot closer to hand than people might imagine. So I would encourage watching all the sessions and learning more about how to get started. >> Yeah, great stuff. And just for the folks @vmworld.com watching, Cloud City's got 60 solution demos, go look for the sessions. You got the EX, the expert sessions, Raghu, Joe Beda amongst other people from VMware are going to be there. And of course, a lot of action on the content. Guys, thanks so much for coming on. Congratulations on the news, big news. NVIDIA on the Bay in Virtual stage here at VMworld. And of course you're in theCube. Thanks for coming. Appreciate it. >> Thank you for having us. Okay. >> Thank you very much. >> This is Cube's coverage of VMworld 2020 virtual. I'm John Furrier, host of theCube virtual, here in Palo Alto, California for VMworld 2020. Thanks for watching. (upbeat music)

Published Date : Sep 18 2020

SUMMARY :

Brought to you by VMware Thanks for joining me on the virtual Cube, is the acceleration of this and VMware as the leader GPUs have set the table the Parallel Processor, if you will, Software, is the value. the first time that an array the future right now. for the applications to run on top of. Yeah, just want to get that quick Okay, you had Project Pacific. And the same platform runs, because how do you bring that the acceleration we get and around the world. that is standing in the way. certainly is the key. the ability to really define Sounds like an Operating System to me. of the server to do. And we think our DPU, And I got to say, you know, Yeah, John, I would say, and also about the AI work And I think some of this And just for the folks Thank you for having us. This is Cube's coverage

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NVIDIAORGANIZATION

0.99+

JohnPERSON

0.99+

KrishPERSON

0.99+

30 timesQUANTITY

0.99+

John FurrierPERSON

0.99+

Krish PrasadPERSON

0.99+

VMwareORGANIZATION

0.99+

Silicon ValleyLOCATION

0.99+

RaghuPERSON

0.99+

Joe BedaPERSON

0.99+

Last yearDATE

0.99+

two teamsQUANTITY

0.99+

last yearDATE

0.99+

MellanoxORGANIZATION

0.99+

Manuvir DasPERSON

0.99+

todayDATE

0.99+

more than 300,000 customersQUANTITY

0.99+

Project PacificORGANIZATION

0.99+

PatPERSON

0.99+

11 yearsQUANTITY

0.99+

30XQUANTITY

0.99+

first oneQUANTITY

0.99+

ESXTITLE

0.99+

VmworldORGANIZATION

0.99+

hundreds of thousandsQUANTITY

0.99+

two typesQUANTITY

0.99+

AWSORGANIZATION

0.99+

Palo Alto, CaliforniaLOCATION

0.99+

VMworldORGANIZATION

0.99+

first timeQUANTITY

0.99+

vSphereTITLE

0.99+

INVIDIAORGANIZATION

0.99+

second partQUANTITY

0.99+

TodayDATE

0.98+

VMworld 2020EVENT

0.98+

SnowflakeORGANIZATION

0.98+

first thingQUANTITY

0.98+

oneQUANTITY

0.98+

bothQUANTITY

0.98+

60 solution demosQUANTITY

0.98+

first oneQUANTITY

0.98+

GoogleORGANIZATION

0.97+

This yearDATE

0.97+

firstQUANTITY

0.97+

vmworld.comOTHER

0.97+

Thomas LaRock, SolarWinds | Microsoft Ignite 2019


 

>>Live from Orlando, Florida. It's the cube covering Microsoft ignite brought to you by Cohesity. >>Hello cube nation and welcome to back to the cubes live coverage of Microsoft ignite here in Orlando, Florida. I'm your host, Rebecca Knight. Along with my cohost Stu Miniman. We are closing down the second day of the three days of coverage. This is day two >>wall to wall to wall coverage. Joining us is Thomas LA rock, best job title ever, head geek at solar winds or speaker data expert and SQL rockstar and Microsoft MVP and Microsoft MVP and yes importantly and you saved me. You didn't have me on yesterday. You waited to the second day, the end of the second day. Thomas, we wanted to make sure that by the time you came on that you had got some time to really absorb some of those announcements and be ready to give us a different perspective on some of the items. All right. Precisely. So this is your 10th Microsoft ignite. It is my first go back to tech ed 2010 so yeah, my 10th consecutive between tech ed and McKnight. Thoughts, impressions of 2019, how is it different? How is the show evolving? What does the show all about? So your perspective, you know, I do a lot of events and shows and what my impression right now just over these two days is that this is one of the only shows this year I've been to where I feel the from year over year, the expo hall is say bigger. >>I mean I know it's the same size the last year. I think there's actually more vendors here this year. There are, and there's more people here. This year in the expo hall. Our traffic at the booth yesterday was amazing and continued through today. Uh, other events I've been to, I feel it's kind of shrinking a little bit. So to me the partners and the ecosystem for Microsoft in general is grow or I should just say Azure because that's what I think this show really is. Now I think the old tech ed you had mentioned was more like a windows type a show. But now this is th these shows between AWS and this, you're talking about the two biggest providers of infrastructure. This is an Azure show. Yeah. Well and Thomas, if you come follow us along, I'll be at CubeCon in two weeks and I'll be at AWS re invent. >>Yup. Right after Thanksgiving. Both of those shows are growing. The ecosystem are growing there too. So the cloud is definitely one of those. The raising tide is moving all boats. I want to poke you say Azure. Azure is definitely one of the main pieces, but you know, the applications that data are so important to your last year. AI front and center. Um, it was, it was more, you know, they didn't use the term AI as much here. You know, Satya, I was talking about, you know, tech intensity and all of the things we can do with data. So this, while a cloud is a major piece, I wouldn't call this just a cloud show because I think that would limit what we're actually talking about here. Cause there's so many of the apps and so many of the things. When I talked to some of the ecosystem providers, you know, they're looking for that solution that fits it and therefore they're go into the ecosystem and talking about all of those pieces. >>So for an infrastructure guy like me, cloud's a big piece of it, but it's way more than that. And that's one of the challenges is there's, you know, everything from, you know, the latest Azure arc all the way through big edge and mobile devices and, uh, you know, heck, there's even, you know, in the store they've got people playing Xbox. Uh, so it's, there's a lot in your Microsoft community here. So. Absolutely. So I, I didn't say cloud though. I said it's an Azure show. And then as your show is to me is almost synonymous with Microsoft and all that stuff. You see, uh, over there, that entire hall, you're right. They have all those other things. They have the, all the power apps, they have those applications, they have everything for developers that you need. But still to me, uh, so what was that stat you just gave me? >>We were debating, it's roughly eight upwards of 80% of workloads are still earth on premises, right? It's still there. So with Azure Ark now they have the ability to take an Azure surface and put it in your data center wherever you want it. So when I say it's an Azure show, it's not even that. It's just cloud. The cloud is coming to you and we see it with VMware, we see it with AWS and outposts that they have decided that 80% is a huge market and they're coming for it. Right? So, so Thomas, if you'd asked me two years ago, uh, which of the hyperscale providers as best as hybrid, my answer would have been Microsoft because they're in both places. The hybrid discussion at this show is way different. There was a lot of retooling. We talked about what was going on. Azure stacks has been there, but arc kind of is a new big push and everybody is trying to look at that and say, wait, is this a management tool? >>Is this just the latest Kubernetes flavor? In your viewpoint, how does arc fit in the Microsoft story? And you know, what should we be comparing it to from the other Amazon, VMware, you know, red hat type of pliers out there? Well Brian, >> I think it's the same thing is that, I was just saying is that arc to me, we can talk about the plumbing. So yeah, they put a fancy name on whether it's Kubernetes, Coobernetti's and all that stuff, but no arc to me is a way for Microsoft to get their hands on as many data estates as possible. Right? I know data state, right? I have a data state and it's next to my data Lake and I work at the data factory and everything's stored in the data warehouse and I shop at the data Mark. We can go on forever with this stuff, but that is the reality of the world. >>And the thing is all those things exist and they're, as your arc is, it's the ability to extend into there because what is Azure and AWS, they're nothing more than an electric company. Their utility and the utility, you're going to offer similar services and that's what they have. And of course VM Ware's in the mix as well. And it's just the ability for all those companies to have their hands on your data, wherever it is, whether it's in your data center or with them, they don't care. They just want the ability to have a piece of that data as it's in transit or at rest. >>And so what's the end there? I mean, you're making that sound like there's some sort of nefarious, uh, end game here. >>It's, I wouldn't say so. Farias I would just say it's market share. What's the end is to survive, to have the market share, to continue to build new cool things. Right. Um, I, I think the end is some consolidation. I don't think the end is, I don't know. Let's say there's five major players. I don't think those five will always exist. I think the are gonna see it shrink over time, but it really, that depends on how well they partner with each other too. Um, I think there's room for everybody, but it's just depends on where they want to say, um, if they want the co-exist or not. Right. So for some of them like VMware, that's really just kind of software, right? They're partnering with clouds. But the clouds are the infrastructure hose. And so how long does VMware really have? Now they've done the nice pivot and I think they're going to last a little bit longer. >>But had they not taken that pivot in the last year or two? I think their timeline with a much shorter, yeah, it's interesting cause we've been looking at, you talk about that cloud adoption, some of the traditional vendors out there, um, many of which are, you know, ecosystem providers that have show here it has to react and deal with the cloud. You know, everybody's jumped on the Kubernetes fly and bandwagon. Everybody's partnering especially with Azure but also AWS and the like. Um, you know, Thomas, you and your company deal with a lot of end users out there. What are they looking for when it comes to being a trusted provider? You know, who, what, what, what's there and how does Microsoft stack up? When we talk about that Satya talked about trust a lot and you know, just curious to how you see them being perceived out there and you know, when customer want to lead partner, what do they want? >>Well, uh, for us, we have, uh, I believe over 300,000 customers at this point and, uh, I think roughly 53% of them are Azure base and that's a higher percentage than what we have for AWS, for our customer base. So we have taken steps to be that trusted partner. So when these companies are going to take that 80% workload that isn't there yet, uh, just in the booth discussions this week where they come to us and they say, Hey, we're going to owe three 65, how can you help us? We're going here at small steps at the time, so that workload that will chip away at it, but we're a company that can help with that transition as people move their workloads and their systems into a place like Azure. Uh, I think what you're gonna also see is our ability to, um, help people understand wherever they want their for structure. >>So for example, last week we announced how we have 15 of our products are now, um, deployed to the Azure marketplace. So you're talking two clicks and everything's deployed for you and you're up and running. And then if you want, if you want to, you know, manage the nodes that are still in your data center, you can just point everything to go up to Azure and Azure, handle a lot of those infrastructure needs for you. So that to me is the trust where you partner with a company like Microsoft and you say, what will it take for us to get in the marketplace? What will it take for us to help help us help, help us help you get that data into your data, into your cloud, right? I think our customers really want to know that when it comes to, Hey, I got to go to Azure. Are you somebody who could help us get there and stay there and manage and monitor the stuff for us? >>I want to talk productivity because I think you have a pretty different take from Satya Nadella. So he had a, he on the, on the main stage yesterday, he said the human act, human attention to inattention is at the root of all productivity. He's, he laid out a stat when you multitask it takes 25 minutes. I'm sorry I got distracted. So it was a 25 minutes. Yes, 25 minutes and you lose 40% of your productivity with that 25 minute lapse. So I w I felt that compelling and that rang true to me. But absolutely >>it's true. So right after he got done with that, Microsoft told us the answer was they were going to take Yammer and shove it inside teams on a shoving inside outlook. I don't think we need more productivity tools. I don't think we need more ways of distracting us. They say they say, Hey, it's great. We'll put tasks from outlook right inside teams. I'm like maybe I'm in teams cause I shut down outlook because I'm distracted by email and other things right now maybe I don't need that. Is it a nice to have and it's a possible thing I guess, but at the end of the day, I don't need you shoving all these extra things into all the things. You're just making the problem worse. We need fewer productivity tools. At what point do we hit peak productivity? I guess? I think we're there. I think I have all the tools that enable me to do my job already. I don't need them all tightly integrated. I need to shut more things off. Right. In order to get stuff done. >>That's a, that's an excellent point because when I want to get work done, I go to a place where I can't get online. Right. Because that's, that's the biggest, >>that's why, uh, I work remote from home one that one of my advantages is I don't have people just walking by my desk and, and distracting me with all sorts of things. That's a huge advantage. I try to take advantage of what, cause I work remote, but for people in an office, bells, whistles, lists that and the other, you know, uh, I just, I get a cup of coffee. You know, it's, it's difficult and I'm not sure that these companies, not just Microsoft, I just don't think companies are really thinking through if they're making things better or not. Every one of them Slack, all of them, they all think that they're the one that's all you need. It's not true and it's not making things better. Yeah, it's a true, we've had good feedback about teams overall here. Especially you've talked to a number of people that are remote workers and they feel that that does help them get connected with teams and, uh, you know, in the remote areas and by itself, but, you know, create point, uh, on the productivity stuff too. >>Do you use teams to use teams? Uh, kind of reluctant at first, like, do I need the another tool? But now that, uh, we've all kind of started switching to it and my company went O three 65 as well. Some teams comes with it and, uh, I do find that very useful, um, uh, much more so than I have any of the other tools in the past. I think teams took a lot of good things from a lot of different tools and they rolled out of them to the one they, and it works for me. It doesn't work for everybody though. Right? >>Exactly. Exactly. So what, so what else are you taking away from the, from your 10th ever ignite, you go back to the office, but is your home on Monday? What kinds of conversations are you going to have most stayed with you, have most resonated you? Okay. >>For me, uh, I, I focus a lot on the data platform and uh, I think the thing that's going to resonate the most with me, it really is Azure arc and what that, what the, what that really means and getting a little more involved with, uh, understanding where they're headed with it. Like just the idea they're going to give me that one management console that can control everything. Earth and cloud. Uh, that's an interesting thing. I see. Come at me. I work for a tools vendor, so as a tools vendor, I'm sitting there going, so Microsoft's building something that gives visibility into both. Now, what does that mean for me and where we might, we want to think about pivoting to make sure that we stay ahead and keep offering value where Microsoft might have a gap. Um, so I think those are the things I'll probably be thinking about. >>My role as head geek is to, you know, help our users and the people who write the code and, you know, connect, share and learn and figure out where things are going. And also involves partnering and having conversations with folks at Microsoft, uh, to help our company, you know, continue to have that edge. So I think that's all I'll be thinking about on Monday, probably now on the plane ride home on Friday, but who knows, right. Uh, Thomas, any other final words about the community here? Uh, you know, you're a Microsoft MVP is we set up in front, uh, you know, Microsoft should get great kudos for, they put the unity in community and they talk about diversity and inclusion, something they highlight something that, at least from the viewpoint we've had, uh, they seem to be doing a good job in moving the needle here. >>But, uh, you know, as an insider to the Microsoft community, uh, anything particular that you'd call out? Well, certainly the changes and the emphasis I've seen on diversity inclusion over the years. You're absolutely right. I think, I know this, you were having some interviews earlier to have those specific discussions and, uh, it's an important conversation to have, uh, uh, as somebody who organizes events, it becomes, you know, what's the diversity, how diverse should the event be? At what point are we diverse enough? Right? And what does that really mean? And so I look at it and I say, if I'm going to run an event that caters to say an it community, well, what's the makeup of the it community? Then the speakers should represent the community that they're trying to speak to. So what I've seen over these 11 years is a lot more focus for events, especially like ones I help organize where it's like, no, what I'm going to go out and recruit the speakers that I need to represent the people that I want them to be presenting to. >>Uh, I don't think I will recall that I'm old. I don't recall a lot of things, but you know, 11 years ago when I was, when I joined to became an MVP, I, I don't think that the diversity was there and I don't think the efforts were being done. I think those efforts have come just in the past few years, four or five years maybe society as a whole, but specifically inside Microsoft and, and their programs. And I think it's fabulous. Uh, I, I think you could never be diverse enough. I guess. I don't know how to say that. I think he could always do more to, uh, include, I always say inclusion is better than the exclusion any day. You can never do enough for that. And I think Microsoft's made great efforts. I'm, I'm really proud to call myself a Microsoft MVP. Uh, I, I think it's a great program. I'm glad that I questioned, you know, their selection method maybe because they keep inviting me back, but they do and, but I love it. I, it's been a great ride, >>a great note to end on. Thomas law crock head geek. Great. Great to have you on the show. Great. Great. Thanks for having me back. I really appreciate it. I'm Rebecca Knight for Stu minimums. Come back tomorrow for more of the cubes live coverage of Microsoft ignite.

Published Date : Nov 5 2019

SUMMARY :

Microsoft ignite brought to you by Cohesity. We are closing down the second day of the three days of coverage. the time you came on that you had got some time to really absorb Now I think the old tech ed you had mentioned was more like a windows type a Azure is definitely one of the main pieces, but you know, And that's one of the challenges is there's, you know, everything from, you know, The cloud is coming to you and we see it with VMware, I think it's the same thing is that, I was just saying is that arc to me, we can talk about the the ability for all those companies to have their hands on your data, wherever it is, I mean, you're making that sound like there's some sort of nefarious, I don't think those five will always exist. you know, ecosystem providers that have show here it has to react and deal with the cloud. owe three 65, how can you help us? So that to me is the trust where you partner with a company like Microsoft and I want to talk productivity because I think you have a pretty different take from Satya Nadella. but at the end of the day, I don't need you shoving all these extra things into Because that's, that's the biggest, they feel that that does help them get connected with teams and, uh, you know, in the remote areas and I think teams took a lot of good things from a lot of different tools and they rolled out of them to the one they, So what, so what else are you taking away from the, from your 10th ever ignite, I think the thing that's going to resonate the most with me, it really is Azure arc and what that, conversations with folks at Microsoft, uh, to help our company, you know, But, uh, you know, as an insider to the Microsoft community, uh, anything particular that you'd call out? Uh, I, I think you could never be diverse enough. Great to have you on the show.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Rebecca KnightPERSON

0.99+

ThomasPERSON

0.99+

MicrosoftORGANIZATION

0.99+

AWSORGANIZATION

0.99+

FridayDATE

0.99+

40%QUANTITY

0.99+

BrianPERSON

0.99+

AmazonORGANIZATION

0.99+

SatyaPERSON

0.99+

Satya NadellaPERSON

0.99+

25 minutesQUANTITY

0.99+

80%QUANTITY

0.99+

MondayDATE

0.99+

fiveQUANTITY

0.99+

2019DATE

0.99+

Stu MinimanPERSON

0.99+

25 minuteQUANTITY

0.99+

last weekDATE

0.99+

Orlando, FloridaLOCATION

0.99+

todayDATE

0.99+

yesterdayDATE

0.99+

two yearsQUANTITY

0.99+

three daysQUANTITY

0.99+

second dayQUANTITY

0.99+

firstQUANTITY

0.99+

last yearDATE

0.99+

tomorrowDATE

0.99+

this yearDATE

0.99+

10thQUANTITY

0.99+

This yearDATE

0.99+

bothQUANTITY

0.99+

BothQUANTITY

0.99+

two clicksQUANTITY

0.99+

five major playersQUANTITY

0.99+

11 years agoDATE

0.98+

two daysQUANTITY

0.98+

ThanksgivingEVENT

0.98+

AzureTITLE

0.98+

over 300,000 customersQUANTITY

0.98+

fourQUANTITY

0.98+

VMwareORGANIZATION

0.98+

11 yearsQUANTITY

0.98+

XboxCOMMERCIAL_ITEM

0.98+

53%QUANTITY

0.97+

Thomas LaRockPERSON

0.97+

two biggest providersQUANTITY

0.97+

oneQUANTITY

0.96+

SolarWindsORGANIZATION

0.96+

five yearsQUANTITY

0.95+

this weekDATE

0.95+

FariasPERSON

0.94+

both placesQUANTITY

0.94+

YammerPERSON

0.94+

outlookTITLE

0.93+

Maheswaran Surendra, IBM GTS & Dave Link, ScienceLogic | ScienceLogic Symposium 2019


 

>> From Washington D.C. it's theCUBE covering ScienceLogic Symposium 2019. Brought to you by ScienceLogic. >> Hi, I'm Stu Miniman and this is theCUBE's coverage of ScienceLogic Symposium 2019 here at The Ritz-Carlton in Washington D.C. About 460 people here, the events' grown about 50%, been digging in with a lot of the practitioners, the technical people as well as some of the partners. And for this session I'm happy to welcome to the program for the first time guest, Surendra who is the vice president and CTO for automation in IBM's global technology services. And joining us also is Dave Link who is the co-founder and CEO of ScienceLogic. Gentlemen, thank you so much for joining us. >> Thank you for having us. >> Thanks for having us. >> Alright, so Surendra let's start with you. Anybody that knows IBM services at the core of your business, primary driver, large number of the presented to the employees at IBM are there. You've got automation in your title so, let's flush out a little bit for us, your part of the organization and your role there. >> Alright, so as you pointed out, IBM, a big part of IBM is services; it's a large component. And that two major parts of that and though we come together as one in terms of IBM services, one is much more focused on infrastructure services and the other one on business services. So, the automation I'm dealing with primarily is in the infrastructure services area which means all the way from resources you have in a persons data center going into now much more of course in a hybrid environment, hybrid multi-cloud, with different clouds out there including our own and providing the automation around that. And when we mean automation we mean the things that we have to do to keep our clients' environments healthy from a availability and performance standpoint; making sure that our environment then we respond to the changes that they need to the environment because it obviously evolves over time, we do that effectively and correctly and certainly another very important part is to make sure that they're secure and compliant. So, if you think of Maslow's hierarchy of the things that IT operations has to do that in a nutshell sums it up. That's what we do for our clients. >> Yeah, so Dave luckily we've got a one on one with you today to dig out lots of nuggets from the kino and talk a bit about the company but, you talk about IT operations and one of the pieces I've got infrastructure, I've got applications, ScienceLogic sits at an interesting place in this heterogeneous ever-changing world that we live in today. >> It does and the world's changing quickly because the clouds transforming the way people build applications. And that is causing a lot of applications to be refactored to take advantage of some of these technologies. The especially focused global scale we've seen them, we've used them, applications that we use on our phone. They require a different footprint and that requires then a different set of tools to manage an application that lives in the cloud and it also might live in a multi-cloud environment with some data coming from private clouds that populate information on public clouds. What we found is the tools industry is at a bit of a crossroads because the applications now need to be infrastructure aware, but the infrastructure could be served from a lot of different places, meaning they've got lots of data sources to sort together and contextualize to understand how they relate to one another real time. And that's the challenge that we've been focused on solving for our customers. >> Alright, Surendra I want to know if we can get a little bit more to automation and we talk automation, >> There's also IBM use for a number of years, the cognitive and there was the analyst that spoke in the kino this morning. He put cognitive as this overarching umbrella and underneath that you had the AI and underneath that you had that machine learning and deep learning pieces there. Can you help tease out a little bit for IBM global services in your customers? How do they think of the relationship between the MLAI cognitive piece in automation? >> So I think the way you laid it out, the way it was talked about this morning absolutely makes sense, so cognitive is a broad definition and then within that of course AI and the different techniques within AI, machine learning being one, natural language processing, national languages understanding which not as much statistically driven as being another type of AI. And we use all of these techniques to make our automation smarter. So, often times when we're trying to automate something, there can be very prescriptive type of automation, say a particular event comes in and then you take a response to it. But then often times you have situations where you have events especially what Dave was talking about; when an application is distributed not just a classic of distributed application, but now distributed of infrastructure you may have. Some of it may be running on the main frame, some of it actually running in different clouds. And all of this comes together, you have events and signals coming from all of this and trying to reason over where a problem may be originating from because now you have a slow performance. What's the reason for the slow performance? Trying to do some degree of root cause determination, problem determination; that's where some of the smarts comes in in terms of how we actually want to be able to diagnose a problem and then actually kick off maybe more diagnostics and eventually kick off actions to automatically fix that or give the practitioner the ability to fix that in a effective fashion. So that's one place. The other areas that one type of machine learning I shouldn't say one type, but deadly machine learning techniques lend themselves to that. There's another arena of causes a lot of knowledge and information buried in tickets and knowledge documents and things like that. And to be able to extract from that, the things that are most meaningful and that's where the natural language understanding comes in and now you marry that with the information that's coming from machines, which is far more contextualized. And to be able to reason over these two together and be able to make decisions, so that's where the automation. >> Wonder if we can actually, let's some of those terms I want to up level a little bit. I hear knowledge I hear information; the core of everything that people are doing these today, it's data. And what I heard, and was really illuminated to me listening to what I've seen of ScienceLogic is that data collection and leveraging and unlocking value of data is such an important piece of what they're doing. From an IBM standpoint and your customers, where does data fit into that whole discussion? How do things like ScienceLogic fit in the overall portfolio of solutions that you're helping customers through either manager, deploying and services? >> So definitely in the IT Ops arena, a big part of IT Ops, at the heart of it really is monitoring and keeping track of systems. So, all sets of systems throw off a lot of data whether it's log data, real time performance data, events that are happening, monitoring of the performance of the application and that's tons and tons of data. And that's where a platform like ScienceLogic comes in, as a monitoring system with capabilities to do what we call also event management. And in the old days, actually probably would have thought about monitoring event management and logs as somewhat different things; these worlds are collapsing together a bit more. And so this is where ScienceLogic has a platform that lends itself to a marriage of these faces in that sense. And then that would feed a downstream automation system of informing it what actions to take. Dave, thoughts on that? >> Dave, if you want to comment on that I've got some follow ups too, but. >> Yeah, there's many areas of automation. There's layers of automation and I think Surendra's worked with customers over a story career to help them through the different layered cakes of automation. You have automation related to provisioning systems, the provision and in some case provision based on capacity analytics. There's automation based on analysis of a root cause and then once you know it, conducting other layers of automation to augment the root cause with other insights so that when you send up a case or a ticket, it's not just the event but other information that somebody would have to go and do, after they get the event to figure out what's going on. So you do that at time of event that's another automation layer and then the final automation layer, is if you know predictively about how to solve the problem just going ahead if you have 99% confidence that you can solve it based on these use case conditions just solve it. So when you look at the different layers of automation, ScienceLogic is in some cases a data engine, to get accurate clean data to make the right decisions. In other cases, we'll kick off automations in other tools. In some cases we'll automate into ecosystem platforms whether it's a ticketing system, a service desk system, a notifications systems, that augment our platform. So, all those layers really have to work together real time to create service assurance that IBM's customers expect. They expect perfection they expect that excellence the brand that IBM presents means it just works. And so you got to have the right tooling in place and the right automation layers to deliver that kind of service quality. >> Yeah, Dave I actually been, one of the things that really impressed me is that the balance between on the one hand, we've talked to customers that take many many tools and replace it with ScienceLogic. But, we understand that there is no one single pane of glass or one tool to rule them all, the theme of the shows; you get the superheros together because it takes a team. You give a little bit of a history lesson which resonated me. I remember SNMP was going to solve everything for us, right? But, the lot of focus on all the integrations that works, so if you've got your APM tools, your ITSM tools or things you're doing in the cloud. It's the API economy today, so balancing that you want to provide the solutions for your customers, but you're going to work with many of the things that they have; it's been an interesting balance to watch. >> Yeah, I think that's the one thing we've realized over the years; you can't rip and replace years and years of work that's been done for good reason. I did hear today that one of our new customers is replacing a record 51 tools with our product. But a lot of these might be shadow IT tools that they've built on top of special instrumentation they might have for a specific use cases or applications or a reason that a subject matter expert would apply another tool, another automation. So, the thing that we've realized is that you've got to pull data from so many sources today to get machine learning, artificial intelligence is only as good as the data that it's making those decisions upon. >> Absolutely. >> So you've got to pull data from many different sources, understand how they relate to one another and then make the right recommendations so that you get that smooth service assurance that everybody's shooting for. And in a time where systems are ephemeral where they're coming and going and moving around a lot, that's compounding the challenge that operations has not just in all the different technologies that make up the service; where those technologies are being delivered from, but the data sources that need to be mashed together in a common format to make intelligent decisions and that's really the problem we've been tackling. >> Alright, Surendra I wonder if you can bring us inside your, you talked to a lot of enterprise customers and it helped share their voices to in this space, not sure if they're probably not calling it AI ops there, but some of the big challenges that they're facing where you're helping them to meet those challenges and where ScienceLogic fits in. >> So certainly the, yes, they probably don't want to talk about it that. They want to make sure that their applications are always up and performing the way they expect them to be and at the same time, being responsive to changes because they need to respond to their business demands where the applications and what they have out there continually has to evolve, but at the same time be very available. So, all the way from even if you think about something that is traditional and is batch jobs which they have large processing of batch jobs; sometimes those things slow down and because now they're running through multiple systems and trying to understand the precedence and actions you take when a batch job is not running properly; as just one example, right? Then what actions we want, first diagnosing why it's not working well. Is it because some upstream system is not providing it the data it needs? Is it clogged up because it's waiting on instructions from some downstream system? And then how do you recover from this? Do you stop the thing? Just kill it or do you have to then understand what downstream further subsequent batch jobs needs to or other jobs will be impacted because you killed this one? And all of that planning needs to be done in some fashion and the actions taken such that if we have to take an action because something has failed, we take the right kind of action. So that's one type of thing where it matters for clients. Certainly, performance is one that matters a lot and even on the most modern of applications because it may be an application that's entirely sitting on the cloud, but it's using five or 10 different SAS providers. Understanding which of those interactions may be causing a performance issue is a challenge because you need to be able to diagnose that and take some actions against that. Maybe it's a log in or the IDN management service that you getting from somewhere else and understanding if they have any issues and whether that provider is providing the right kind of monitoring or information about their system such that you can reason over it and understand; okay my service which is dependent on this other service is actually being impacted. And all these kind of things, it's a lot of data and these need to come together. That's where the platform something like ScienceLogic would come into play. And then taking actions on top of that is now where a platform also starts to matter because you start to develop different types of what we call content. So we distinguish the space between an automation platform or a framework plus and the content you need to have that. And ScienceLogic they talk about power packs and these things you need to have that essentially call out the work flows of the kind of actions you need to take when you have the falling signature of a certain bundle of events that have come together. Can you reason over it to say okay, this is what I need to do? And that's where a lot of our focus is to make sure that we have the right content to make sure that our clients applications stay healthy. Did that get to, I think build on what you were talking about a bit? >> Absolutely. Yes, you've got, it's this confluence of a know how an intelligence from working with customers, solving problems for them and being proactive against the applications that really run their business; and that means you're constantly adjusting. These networks I think Surendra's said it before, they're like living organisms. Based on load, based on so many factors; they're not stagnant, they're changing all the time, unless you need the right tools to understand not just anomaly's what's different, but the new technologies that come in to augmenting solutions and enhancing them and how that effects the whole service delivery cadence. >> Mr. Surendra, I want to give you the final word. One of the things I found heartening when I look at this big wave of AI that's been coming is, there's been good focus on what kind of business outcomes customers are having. >> Okay. >> Because back in the big data wave I remember we did the survey's and it was like what was the most common use case? And it was custom. And what you don't want to have is a science project, right? >> Right. >> Yes. >> You actually want to get things done. So any kinds you can give as to, I know you understand we're still early in a lot of these deployments and rollouts but what do you seeing out there? What are some of the lighthouse use cases? >> So, certainly for us, right? We've been at using data for a while now to improve the service assurance for our clients and I'll be talking about this tomorrow, but one of the things we have done is we found that now in terms of the events and incidents that we deal with, we can automatically respond with essentially no human interference or involvement I should say about 55% of them. And a lot of this is because we have an engine behind it where we get data from multiple different sources. So, monitoring event data, configuration data of the systems that matter, tickets; not just incident tickets but change tickets and all of these things and a lot of that's unstructured information and you essentially make decisions over this and say okay, I know I have seen this kind of event before in these other situations and I can identify an automation whether it's a power pack, an automotor, an Ansible module, playbook. that has worked in the situation before in another client and these two situations are similar enough such that I can now say with these kind of events coming in, or group events I can respond to it in this particular fashion; that's how we keep pushing the envelope in terms of driving more and more automation and automated response such that the I would say certainly the easy or the trivial kinds of I shouldn't say trivial, but the easy kinds of events and monitoring things we see in monitoring are being taken care of even the more somewhat moderate ones where file systems are filling out for some unknown reasons we know how to act on them. Some services are going down in some strange ways we know how to act on them to getting to even more complex things like the batch job type of thing. Example I gave you because those can be some really pernicious things can be happening in a broad network and we have to be able to diagnose that problem, hopefully with smarts to be able to fix that. And into this we bring in lots of different techniques. When you have the incident tickets, change tickets and all of that, that's unstructured information; we need to reason over that using natural language understanding to pick out the right I'm getting a bit technical here, verp no pas that matter that say okay this probably led to these kind of incidents downstream from typical changes. In another client in a similar environment. Can we see that? And can we then do something proactively in this case. So those are all the different places that we're bringing in AI, call it whatever you want, AIML into a very practical environment of improving certainly how we respond to the incidents that we have in our clients environments. Understanding when I talked about the next level changes when people are making changes to systems, understanding the risks associated with that change; based on all the learning that we have because we are very large service provider with essentially, approximately 1,000 clients. We get learning over a very diverse and heterogeneous experience and we reason over that to understand okay, how risky is this change? And all the way into the compliance arena, understanding how much risk there is in the environment that our clients facing because they're not keeping up with patches or configurations for security parameters that are not as optimal as they could be. >> Alright, well Surendra we really appreciate you sharing a glimpse into some of your customers and the opportunities that they're facing. >> Thank you. >> Thanks so much for joining us. Alright and Dave, we'll be talking to you a little bit more later. >> Great, thanks for having me. >> All right. >> Thank you. >> And thank you as always for watching. I'm Stu Miniman and thanks for watching theCUBE. >> Thank you Dave. >> Thank you. (upbeat techno music)

Published Date : Apr 25 2019

SUMMARY :

Brought to you by ScienceLogic. And for this session I'm happy to welcome to the program of the presented to the employees at IBM are there. And that two major parts of that and though we come together Yeah, so Dave luckily we've got a one on one with you And that's the challenge that we've been focused on solving that you had the AI and underneath that you had that machine give the practitioner the ability to fix that in a effective the core of everything And in the old days, actually probably would have thought Dave, if you want to comment on that I've got some And so you got to have the right tooling in place and the It's the API economy today, so balancing that you want to the years; you can't rip and replace but the data sources that need to be mashed together in but some of the big challenges that they're facing where flows of the kind of actions you need to take when you have different, but the new technologies that come in to One of the things I found heartening when I look at this big Because back in the big data wave I remember we did the but what do you seeing out there? found that now in terms of the events and incidents that we Alright, well Surendra we really appreciate you sharing to you a little bit more later. And thank you as always for watching. Thank you.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SurendraPERSON

0.99+

DavePERSON

0.99+

Dave LinkPERSON

0.99+

fiveQUANTITY

0.99+

IBMORGANIZATION

0.99+

99%QUANTITY

0.99+

Stu MinimanPERSON

0.99+

Washington D.C.LOCATION

0.99+

51 toolsQUANTITY

0.99+

two situationsQUANTITY

0.99+

todayDATE

0.99+

Maheswaran SurendraPERSON

0.99+

ScienceLogic Symposium 2019EVENT

0.99+

ScienceLogicORGANIZATION

0.98+

first timeQUANTITY

0.98+

one toolQUANTITY

0.98+

approximately 1,000 clientsQUANTITY

0.98+

oneQUANTITY

0.98+

one exampleQUANTITY

0.98+

tomorrowDATE

0.98+

theCUBEORGANIZATION

0.97+

about 50%QUANTITY

0.97+

twoQUANTITY

0.97+

one typeQUANTITY

0.95+

firstQUANTITY

0.95+

OneQUANTITY

0.94+

about 55%QUANTITY

0.94+

this morningDATE

0.93+

two major partsQUANTITY

0.92+

About 460 peopleQUANTITY

0.89+

SNMPORGANIZATION

0.86+

MaslowORGANIZATION

0.85+

tons and tons of dataQUANTITY

0.84+

one placeQUANTITY

0.83+

IBM GTSORGANIZATION

0.79+

single pane ofQUANTITY

0.77+

ScienceLogic SymposiumEVENT

0.76+

big wave ofEVENT

0.75+

10 different SAS providersQUANTITY

0.72+

ScienceLogicTITLE

0.72+

Ritz-CarltonLOCATION

0.67+

AnsibleORGANIZATION

0.65+

kinoORGANIZATION

0.6+

Eric Herzog, IBM | IBM Think 2019


 

>> Live from San Francisco, it's theCUBE. Covering IBM Think 2019, brought to you by IBM. >> Hello everyone welcome back to theCUBE's live coverage here at IBM Think 2019 in San Francisco, our exclusive coverage, day four, four days of coverage events winding down, I'm John Furrier with Stu Miniman, our next guest, Eric Herzog, CUBE alumni, CMO of IBM storage and VP of storage channels, Eric great to see you wearing the Hawaiian shirt as usual. >> Great, I can't come to theCUBE and not wear the Hawaiian shirt. You guys give me too much of a heart attack. >> Love getting you on to get down and dirty on storage and the impact of Cloud and infrastructure. First, you gave a great talk yesterday to a packed house, I saw that on social media, great response, what's going on for you at the show, tell us. >> So the big focuses for us are around four key initiatives. One is multi-cloud particularly from a hybrid perspective and in fact, I had three presenters with me, panelists and users, all of them were using multiple public cloud providers and all of them had a private cloud. One of them also was a software as a service vendor, so clearly they're really monetizing it. So that's one, the second one is around AI, both AI that we use inside of our storage to make it more efficient and more cost effective for the end user, but also as the platform for AI work loads and applications. Cyber resiliency is our other big theme, we've got all kinds of security, yes everyone is used to of course the Great Wall of China protecting you and then of course chasing the bad guy down when they breach you, but when they breach you it'd sure be nice if everything had data at rest encryption, or when you tiered out to the cloud you knew that it was being backed up or tiered out fully encrypted or how about something that can help you with ransomware and malware. So we have that, and that's a storage product not a regular, you know what you think of from a security vendor. So those are the big things that we've been harking on at the show. >> One of the things that I've observed, you've been very active out in the field, we've seen you at a lot of different events, Cisco Live, others, you guys have had an interesting storage product portfolio, very broad and specific leadership categories, but you also have the ability to work with other partners. This has been a big part of your strategy, you get the channels. What is, how would you summarize the current story around IBM storage and systems, because it's now an ingredient part of other people's infrastructure with cloud storage then becomes a key equation, how would you describe the IBM storage posture, product portfolio, what are the key things? >> So I think the key thing from a portfolio perspective, while it looks broad it's really four things. Software defined storage which we also happened to have bet on on array so theoretically that's one product line, same exact software. Other vendors don't do that, they have an array pack and you buy the array but if you buy their software defined storage it's actually different software, for us it's the same software. Then we have modern data protection and then we have management playing. That's kind of it. I do think one of the big differentiator for us, is even though we're part of IBM, we have already been working with everyone any way. So as we talked about at Cisco Live, for Spectrum Protect alone, our modern data protection platform, we have 400 small and medium cloud service providers all over the world that their back up service is based on it, so even though IBM Cloud has their own cloud division theoretically, we're enabling the competition but we've had that story at IBM storage now for four years. >> So storage anywhere basically is the theme here, AI anywhere storage anywhere, I mean it's not the official tagline but that's the philosophy with software. >> And that's yeah, so even if you think look at AI. We have an AI reference architecture with the power product line, we also have an AI reference architecture with the Nvidia product line, and we're working on a third one right now with another major server vendor because we want our storage to be anywhere there's AI and anywhere there's a cloud, big medium or small. >> Alright, Eric let's tease that out a little bit because I had a great conversation with an IBM fellow yesterday and we think back ten years ago, when you talked about hybrid and multi cloud, when you talked about an application it's "Am I spanning between environments? "Am I bursting between environments?" And architectures just didn't work that way. Today microservices architecture, there's pieces of the solution that can live in lots of environments, Compute I can spin up almost anywhere at any time, data doesn't move and I need to worry about my data, I need to worry about security so there's certain things that multi cloud like data protection, cyber resiliency, those kind of ones need to live everywhere, but when I talk about storage, I'm not moving my storage and my persistent database all over the place. So help us kind of tease out as to what is the multi everywhere and what is the you know the data that the Compute's going to actually move to that data, help us squint through that a little bit. >> So let's do the storage part first. So most applications, workloads, and use cases that are either business critical or mission critical are going to stay on prem, doesn't mean you can't use a public cloud provider for overflow whether that be IBM or Amazon or Microsoft or like I said the 400 cloud providers that we sell to that are not IBM, so but you're still going to have this hybridness where the data is partially on prem and off prem, in that case you're going to be using the public cloud provider, and by the way we did a survey, IBM did, and when you're looking enterprise, so let's say companies that are three or four billion US and up, anywhere in the world, you're seeing that most of them are using five or six different public clouds, whether that be salesforce.com which really is sales enablement software as a service. We have a startup that we work with who uses IBM's flash system and they do cyber security as a service, that's their whole business. So all of this software vendors that now deliver not on prem but you know over the cloud. Then you've got regular public cloud providers for file, block, and object for example we not only support IBM Cloud object storage protocol, but S3. So we have customers that put data out in S3, we have customers that put it out on other clouds because as you know S3's become the de facto standard so all the mid to small cloud providers use it. So I think what you've got is hybrid cloud is a sort of a subset of multi cloud and then multi cloud what you're seeing is because of software as service could even be geographic issues, we have a lot of data centers at IBM Cloud so do the three major cloud providers, but we are not in all 212 countries so if you have the law like in Canada where the data has to physically stay within the premises of Canada, now we all happen to have data centers that are big enough, but that doesn't mean we have data centers in every country, so you have legal issues, you have applications what applications are good, that make sense, what about pricing, and as you know some big companies still buy regionally. >> Eric, one of the things I'd love to get your perspective on is the SAS providers because if we look at the storage market in many ways, you know there was like the threat of public cloud, but really you got to follow the application, follow the data and as SAS proliferation happens, your data is going to go with that, you know you have them as customers in a lot of environments, what are you seeing from the SAS providers, how do they choose what offerings they have and how do they look at their data center versus public cloud mix? >> So when you look at a SAS provider, they've got a couple of different parameters that they look at which is why we've been very successful. One is performance, they already know their subject to the vicissitudes of the cloud so you can't have any bottle neck in your core data center because you're serving that app up, and if it's too slow or it doesn't work right, then of course the end user will go buy a different piece of software from another SAS provider. Second one is availability, because you have no idea when wiki bomb theCUBE is going to turn on that service, it could be the middle of the night right? If you guys expand to Asia, you guys will be asleep but your guy in Australia will be using that software, so it can't ever go down, so availability. Resiliency, can it handle pounding. If CUBE wiki bomb becomes ginormous, and you buy all these other analyst firms and the next thing you know the biggest analyst firm in the world, if you have thousands of people guess what now you're hammering on that software, so it's got to be able to take that workload abuse, right? And that's the kind of thing, so they look for that. >> That's scale basically, scale is critical. >> Right, they cannot have any issues of resiliency or availability and performance so A: they're usually going all flash, some of them will buy like a tape or the older all hard drive arrays as a backup store, ideal for IBM cloud object storage but again the main thing they focus on is flash because they're serving up that software. >> Let me ask you a question, so I know you've been in this business for a long time, storage you know everything about the speeds and fees but also you've been a historian too, you're on the front edge. IBM has got a killer strategy with cloud private, doing very well with Openshift and Redhat acquisition, you're now poised to essentially bring cloud scale across multiple clouds and with AI, it really puts storage at the center of the action. How is storage now positioned and how should customers think about storage, because scale is table stakes, enabling developers to program infrastructure as code, how does storage and how has it changed and how are you guys positioned to take advantage of that? How would you kind of explain that to a customer? >> Yeah so I think there's a couple of changes, first of all you're looking for a storage vendor which should be us, but you're looking for a storage vendor that is always making sure, for example when micro services first came out and containers, okay great except when containers came out and it's still a problem, you don't have storage consistency whereas in a VM ware or a hyper V or you know KVM environment, you do. So when you move things around, you don't lose the dataset, well we have persistency storage. So the key thing that you want to look for is a storage vendor that will stay on that leading edge as you move. Our copy data manager has an API so the developers can spit up their own environments but use real data, so as you guys know well from your pasts that the last thing you want to do is have the dev ops guy be developing things on faux datasets, try to put it in production, and then the real dataset doesn't work, at the same time if they put it out to a public cloud provider you could have a legal or security breach, right? So by being able to take modern data protection, as an example, and not just to have grandfather, father son back up, we all remember that I remember it better than you guys since I'm older, but that's back up right? It's not back up any more, it's modern data protection. You need to be able to take the snapshot, the replica or the back up dataset and use it for development, so you want a storage vendor that's going to be on the leading edge of that. We've done that at IBM on the Kenner side, the modern data protection side, and we'll continue to the do that. The whole multi cloud thing, IBM as you know is now all about multi cloud, what Redhat's been in, the storage division of IBM has been working with Redhat for 15 years. Going to the Redhat summit every year, I know you guys do theCUBE from there sometimes. >> You're on, but this is software defined so at the end of the day a software defined bet with arrays have paid off. >> Yes. >> You'd say that would be kind of a key linchpin. >> I would argue that, while there's some hardware aspects to it, so for example our flash core modules give us a big differentiator from a flash perspective, in general the number one differentiator for a strong, powerful array vendor is actually the underlying software code. The RAID stack, what you can wrap around it, file block and object support, what could you enhance, our Spectrum discover, allowing you to use metadata about unstructured data whether that be in the file space of the object store. That allows the data scientist to dramatically reduce the time it takes to prep the data when they're doing either AI or an analytic workload, so we just saved them money but we're really a storage company that came up with something that a data scientist could use because we understand how storage is at the central foundation and how you could literally use the metadata for something actually valuable, not to a storage person because a data scientist is not the storage guy of course. >> Yeah and Eric I would love to get your feedback, what are some of those key discussions you're having with customers here at the show? We've been talking a lot this week digital transformation, AI into everything there, are those some of the themes? What are the struggles that really the enterprises of today are facing and how your group's helping them? >> So one of the big things is understanding that it's going to be multi cloud and so because we've already been the Switzerland of the storage industry and working with every cloud provider, all the big ones, including ones that compete with our own sister division, but all the little small ones too, right? And all the software as service vendors we work with that we're the safe bet, you don't have to worry about it. Because whoever you pick, or for a big enterprise, in fact I had Aetna on stage with me and he said he's using seven different clouds, one of which is their private cloud and then six different cloud providers they use, and he said not counting salesforce.com and I forgot the other name, so really if you count the softwares there, she really got like nine clouds. She said I use IBM cause I know it's going to work with whoever, and you're not going to say oh I don't work with this one or that one. So that's been obviously making sure everyone realizes that, the whole company is embracing it as you saw and what we're going to do obviously with Redhat and continue for them to participate with all of their existing customer base that they've been doing for years. >> So you see multi cloud and sweet spot, that highlights your value proposition, would you say that to be true? >> I would say that and then the second one is around AI. All the storage vendors including us have had AI sort of inside, what I'll call inside of the box, inside of the array and use that to make the array better, but now with AI being ubiquitous from a work load perspective, you have to have the right foundation underneath that, again performance resiliency availability, if you're going to use AI in a giant car factory, and it's going to run all of those machines, you better make sure the thing never fails because then the assembly line goes down and those things are hundreds of millions of dollars of build every day. So that's the kind of thing you got to look for, so AI's got to have the right platform underneath it as well. >> Eric you have some reporting from the field as you're out in the, doing a lot of talks a lot of customers, give it a couple of anecdotal examples of where the leading edge is in storage and where are use cases that would be a good tell sign of where this kind of multi cloud is going. Can you just give some examples of the use cases, situation, and kind of why is that relevant for where everyone will be going? Where is the puck going to be, so I can skate to where the puck is, as they say. >> So from a multi cloud perspective, A: you've got to deal with how your company is structured, if you have a divisionalized company or one that really lets the regions make their own buy decisions, then you may have NTT Cloud in Japan, you may have Ali Baba in China, you may have IBM Cloud Australia, and then you might have Amazon in Latin America. And as IT guys you got to make sure you're dealing with that, and embrace it. One of the things I think from an IT perspective is why I'm wearing the Hawaiian shirt, you don't fight the wave, you ride the wave. And that's what everyone's got to realize so, they're going to use multi cloud, and remember the cloud was the web was the internet, it's actually all the same stuff from a long time ago, the mid 90's, which also means now procurement's involved and when procurement's involved, what are they going to say to you? Did you get a bid from IBM Cloud, did you see that bid from Amazon and Microsoft? So it's changed the whole thing of, I can just go to any cloud I want to, now procurement's involved that even mid-size companies procurement says you did get another bid right, did you not? Which for server, storage, and network vendors that's been the way it's been for 35, 40 years. >> The bids are changing too, so what are the requirements now? Amazon has a cloud, they have storage, you have storage, but people have on premise they have multiple environments. If the world is one big data center, with multiple regions and locations, this is the resilience you spoke of, what's the new requirements as procurement gets involved because procurement isn't dictating the requirements, they're getting the requirements from the application work loads and the infrastructure, so what are the new requirements that you see? >> So I think the thing you're seeing is if you take cloud just a couple years ago, I'm going to put my storage out there, okay great, I need this kind of availability, ooh that's extra money, sorry Mr. Wikibomb, Mr. CUBE we got to charge you a little extra for that. Oh we need a certain amount of performance, oh that's a little extra. And then for heavy transactional work loads the data's constantly moving back and forth, oh we forgot to tell you that we're charging you every time you move the data in and every time you move the data out. So as you're putting together these RFPs you needs to be aware of that. >> Those are hidden costs. >> Those are hidden costs that are, I think the reason you're seeing such the ride of the hybrid is people went to public cloud and then someone in finance, or maybe even in the IT group sat down with a spread sheet and said "Oh my god, we could've just bought an IBM array "or someone else's array" and actually had less money even counting support, because all every time we're moving the data, but for archive, for back up we don't move the data around a lot, it's a great solution for anything. Then you have the whole factoring of software as a service, so part of that is the software itself, if you're going to go up against salesforce.com then whoever does, they better make sure the software's good, then on top of that again you negotiate with the software vendor, I need it globally, okay what's the fee for that? So I think the IT guys need to understand that with the ubiquity of the cloud, you've got to ask way more questions, in the storage array business, everyone's got five nines and almost everybody's got six nines, well way back when it was four nines then it was five and now it's six, so you don't ask anymore because you know it just changes right? And the cloud is still new enough and the whole software as a service is a different angle, and a lot of people don't even realize software as a service is cloud, but when you say that they go, what are you talking about, it's just I'm getting it over a service. Where do you think it comes from? A cloud data center. >> Well the trend is software defined, you guys are on that early. Congratulations, and don't forget the hardware, the high performance hardware as well, arrays and what not. So great job. Eric thanks for coming on, appreciate it. >> Great, thank you very much. >> CUBE coverage here, I'm John Furrier, Stu Miniman. Day four of our live coverage here in Moscone North, in San Francisco for IBM Think 2019. Great packed house here at IBM Think, back for more coverage after this short break. (electronic outro music)

Published Date : Feb 14 2019

SUMMARY :

Covering IBM Think 2019, brought to you by IBM. Eric great to see you wearing the Hawaiian shirt as usual. Great, I can't come to theCUBE and the impact of Cloud and infrastructure. to the cloud you knew that it was being backed up leadership categories, but you also have the ability and you buy the array but if you buy their software So storage anywhere basically is the theme here, And that's yeah, so even if you think look at AI. the you know the data that the Compute's going to actually move and as you know some big companies still buy regionally. and the next thing you know the biggest analyst firm the main thing they focus on is flash and how are you guys positioned to take advantage of that? So the key thing that you want to look for so at the end of the day a software defined bet is at the central foundation and how you could literally use and I forgot the other name, so really if you count So that's the kind of thing you got to look for, Eric you have some reporting from the field And as IT guys you got to make sure you're dealing so what are the new requirements that you see? oh we forgot to tell you that we're charging you as a service, so part of that is the software itself, Congratulations, and don't forget the hardware, Day four of our live coverage here in Moscone North,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MicrosoftORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

IBMORGANIZATION

0.99+

Eric HerzogPERSON

0.99+

Stu MinimanPERSON

0.99+

AustraliaLOCATION

0.99+

CanadaLOCATION

0.99+

threeQUANTITY

0.99+

AsiaLOCATION

0.99+

EricPERSON

0.99+

fiveQUANTITY

0.99+

John FurrierPERSON

0.99+

Moscone NorthLOCATION

0.99+

sixQUANTITY

0.99+

NvidiaORGANIZATION

0.99+

JapanLOCATION

0.99+

San FranciscoLOCATION

0.99+

oneQUANTITY

0.99+

15 yearsQUANTITY

0.99+

ChinaLOCATION

0.99+

CUBEORGANIZATION

0.99+

yesterdayDATE

0.99+

OneQUANTITY

0.99+

Latin AmericaLOCATION

0.99+

four yearsQUANTITY

0.99+

FirstQUANTITY

0.99+

mid 90'sDATE

0.99+

nine cloudsQUANTITY

0.99+

WikibombPERSON

0.99+

ten years agoDATE

0.99+

RedhatORGANIZATION

0.99+

AetnaORGANIZATION

0.98+

S3TITLE

0.98+

hundreds of millions of dollarsQUANTITY

0.98+

RedhatEVENT

0.98+

three presentersQUANTITY

0.98+

Day fourQUANTITY

0.98+

TodayDATE

0.97+

thousands of peopleQUANTITY

0.97+

212 countriesQUANTITY

0.97+

four billionQUANTITY

0.97+

four daysQUANTITY

0.97+

Great Wall of ChinaLOCATION

0.97+

six different cloud providersQUANTITY

0.96+

bothQUANTITY

0.96+

second oneQUANTITY

0.96+

Second oneQUANTITY

0.96+

Cisco LiveORGANIZATION

0.95+

todayDATE

0.94+

SwitzerlandLOCATION

0.94+

USLOCATION

0.94+

Lenovo Transform 2.0 Keynote | Lenovo Transform 2018


 

(electronic dance music) (Intel Jingle) (ethereal electronic dance music) ♪ Okay ♪ (upbeat techno dance music) ♪ Oh oh oh oh ♪ ♪ Oh oh oh oh ♪ ♪ Oh oh oh oh oh ♪ ♪ Oh oh oh oh ♪ ♪ Oh oh oh oh oh ♪ ♪ Take it back take it back ♪ ♪ Take it back ♪ ♪ Take it back take it back ♪ ♪ Take it back ♪ ♪ Take it back take it back ♪ ♪ Yeah everybody get loose yeah ♪ ♪ Yeah ♪ ♪ Ye-yeah yeah ♪ ♪ Yeah yeah ♪ ♪ Everybody everybody yeah ♪ ♪ Whoo whoo ♪ ♪ Whoo whoo ♪ ♪ Whoo yeah ♪ ♪ Everybody get loose whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ >> As a courtesy to the presenters and those around you, please silence all mobile devices, thank you. (electronic dance music) ♪ Everybody get loose ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ ♪ Whoo ♪ (upbeat salsa music) ♪ Ha ha ha ♪ ♪ Ah ♪ ♪ Ha ha ha ♪ ♪ So happy ♪ ♪ Whoo whoo ♪ (female singer scatting) >> Ladies and gentlemen, please take your seats. Our program will begin momentarily. ♪ Hey ♪ (female singer scatting) (male singer scatting) ♪ Hey ♪ ♪ Whoo ♪ (female singer scatting) (electronic dance music) ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ Red red red red ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ Red red red red ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ Red red red red ♪ ♪ Red don't go ♪ ♪ All hands are in don't go ♪ ♪ In don't go ♪ ♪ Oh red go ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ Red red red red ♪ ♪ All hands are red don't go ♪ ♪ All hands are in red red red red ♪ ♪ All hands are in don't go ♪ ♪ All hands are in red go ♪ >> Ladies and gentlemen, there are available seats. Towards house left, house left there are available seats. If you are please standing, we ask that you please take an available seat. We will begin momentarily, thank you. ♪ Let go ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ ♪ All hands are in don't go ♪ ♪ Red all hands are in don't go ♪ (upbeat electronic dance music) ♪ Just make me ♪ ♪ Just make me ♪ ♪ Just make me ♪ ♪ Just make me ♪ ♪ Just make me ♪ ♪ I live ♪ ♪ Just make me ♪ ♪ Just make me ♪ ♪ Hey ♪ ♪ Yeah ♪ ♪ Oh ♪ ♪ Ah ♪ ♪ Ah ah ah ah ah ah ♪ ♪ Just make me ♪ ♪ Just make me ♪ (bouncy techno music) >> Ladies and gentlemen, once again we ask that you please take the available seats to your left, house left, there are many available seats. If you are standing, please make your way there. The program will begin momentarily, thank you. Good morning! This is Lenovo Transform 2.0! (keyboard clicks) >> Progress. Why do we always talk about it in the future? When will it finally get here? We don't progress when it's ready for us. We need it when we're ready, and we're ready now. Our hospitals and their patients need it now, our businesses and their customers need it now, our cities and their citizens need it now. To deliver intelligent transformation, we need to build it into the products and solutions we make every day. At Lenovo, we're designing the systems to fight disease, power businesses, and help you reach more customers, end-to-end security solutions to protect your data and your companies reputation. We're making IT departments more agile and cost efficient. We're revolutionizing how kids learn with VR. We're designing smart devices and software that transform the way you collaborate, because technology shouldn't just power industries, it should power people. While everybody else is talking about tomorrow, we'll keep building today, because the progress we need can't wait for the future. >> Please welcome to the stage Lenovo's Rod Lappen! (electronic dance music) (audience applauding) >> Alright. Good morning everyone! >> Good morning. >> Ooh, that was pretty good actually, I'll give it one more shot. Good morning everyone! >> Good morning! >> Oh, that's much better! Hope everyone's had a great morning. Welcome very much to the second Lenovo Transform event here in New York. I think when I got up just now on the steps I realized there's probably one thing in common all of us have in this room including myself which is, absolutely no one has a clue what I'm going to say today. So, I'm hoping very much that we get through this thing very quickly and crisply. I love this town, love New York, and you're going to hear us talk a little bit about New York as we get through here, but just before we get started I'm going to ask anyone who's standing up the back, there are plenty of seats down here, and down here on the right hand side, I think he called it house left is the professional way of calling it, but these steps to my right, your left, get up here, let's get you all seated down so that you can actually sit down during the keynote session for us. Last year we had our very first Lenovo Transform. We had about 400 people. It was here in New York, fantastic event, today, over 1,000 people. We have over 62 different technology demonstrations and about 15 breakout sessions, which I'll talk you through a little bit later on as well, so it's a much bigger event. Next year we're definitely going to be shooting for over 2,000 people as Lenovo really transforms and starts to address a lot of the technology that our commercial customers are really looking for. We were however hampered last year by a storm, I don't know if those of you who were with us last year will remember, we had a storm on the evening before Transform last year in New York, and obviously the day that it actually occurred, and we had lots of logistics. Our media people from AMIA were coming in. They took the, the plane was circling around New York for a long time, and Kamran Amini, our General Manager of our Data Center Infrastructure Group, probably one of our largest groups in the Lenovo DCG business, took 17 hours to get from Raleigh, North Carolina to New York, 17 hours, I think it takes seven or eight hours to drive. Took him 17 hours by plane to get here. And then of course this year, we have Florence. And so, obviously the hurricane Florence down there in the Carolinas right now, we tried to help, but still Kamran has made it today. Unfortunately, very tragically, we were hoping he wouldn't, but he's here today to do a big presentation a little bit later on as well. However, I do want to say, obviously, Florence is a very serious tragedy and we have to take it very serious. We got, our headquarters is in Raleigh, North Carolina. While it looks like the hurricane is just missing it's heading a little bit southeast, all of our thoughts and prayers and well wishes are obviously with everyone in the Carolinas on behalf of Lenovo, everyone at our headquarters, everyone throughout the Carolinas, we want to make sure everyone stays safe and out of harm's way. We have a great mixture today in the crowd of all customers, partners, industry analysts, media, as well as our financial analysts from all around the world. There's over 30 countries represented here and people who are here to listen to both YY, Kirk, and Christian Teismann speak today. And so, it's going to be a really really exciting day, and I really appreciate everyone coming in from all around the world. So, a big round of applause for everyone whose come in. (audience applauding) We have a great agenda for you today, and it starts obviously a very consistent format which worked very successful for us last year, and that's obviously our keynote. You'll hear from YY, our CEO, talk a little bit about the vision he has in the industry and how he sees Lenovo's turned the corner and really driving some great strategy to address our customer's needs. Kirk Skaugen, our Executive Vice President of DCG, will be up talking about how we've transformed the DCG business and once again are hitting record growth ratios for our DCG business. And then you'll hear from Christian Teismann, our SVP and General Manager for our commercial business, get up and talk about everything that's going on in our IDG business. There's really exciting stuff going on there and obviously ThinkPad being the cornerstone of that I'm sure he's going to talk to us about a couple surprises in that space as well. Then we've got some great breakout sessions, I mentioned before, 15 breakout sessions, so while this keynote section goes until about 11:30, once we get through that, please go over and explore, and have a look at all of the breakout sessions. We have all of our subject matter experts from both our PC, NBG, and our DCG businesses out to showcase what we're doing as an organization to better address your needs. And then obviously we have the technology pieces that I've also spoken about, 62 different technology displays there arranged from everything IoT, 5G, NFV, everything that's really cool and hot in the industry right now is going to be on display up there, and I really encourage all of you to get up there. So, I'm going to have a quick video to show you from some of the setup yesterday on a couple of the 62 technology displays we've got on up on stage. Okay let's go, so we've got a demonstrations to show you today, one of the greats one here is the one we've done with NC State, a high-performance computing artificial intelligence demonstration of fresh produce. It's about modeling the population growth of the planet, and how we're going to supply water and food as we go forward. Whoo. Oh, that is not an apple. Okay. (woman laughs) Second one over here is really, hey Jonas, how are you? Is really around virtual reality, and how we look at one of the most amazing sites we've got, as an install on our high-performance computing practice here globally. And you can see, obviously, that this is the Barcelona supercomputer, and, where else in New York can you get access to being able to see something like that so easily? Only here at Lenovo Transform. Whoo, okay. (audience applauding) So there's two examples of some of the technology. We're really encouraging everyone in the room after the keynote to flow into that space and really get engaged, and interact with a lot of the technology we've got up there. It seems I need to also do something about my fashion, I've just realized I've worn a vest two days in a row, so I've got to work on that as well. Alright so listen, the last thing on the agenda, we've gone through the breakout sessions and the demo, tonight at four o'clock, there's about 400 of you registered to be on the cruise boat with us, the doors will open behind me. the boat is literally at the pier right behind us. You need to make sure you're on the boat for 4:00 p.m. this evening. Outside of that, I want everyone to have a great time today, really enjoy the experience, make it as experiential as you possibly can, get out there and really get in and touch the technology. There's some really cool AI displays up there for us all to get involved in as well. So ladies and gentlemen, without further adieu, it gives me great pleasure to introduce to you a lover of tennis, as some of you would've heard last year at Lenovo Transform, as well as a lover of technology, Lenovo, and of course, New York City. I am obviously very pleasured to introduce to you Yang Yuanqing, our CEO, as we like to call him, YY. (audience applauding) (upbeat funky music) >> Good morning, everyone. >> Good morning. >> Thank you Rod for that introduction. Welcome to New York City. So, this is the second year in a row we host our Transform event here, because New York is indeed one of the most transformative cities in the world. Last year on this stage, I spoke about the Fourth Industrial Revolution, and our vision around the intelligent transformation, how it would fundamentally change the nature of business and the customer relationships. And why preparing for this transformation is the key for the future of our company. And in the last year I can assure you, we were being very busy doing just that, from searching and bringing global talents around the world to the way we think about every product and every investment we make. I was here in New York just a month ago to announce our fiscal year Q1 earnings, which was a good day for us. I think now the world believes it when we say Lenovo has truly turned the corner to a new phase of growth and a new phase of acceleration in executing the transformation strategy. That's clear to me is that the last few years of a purposeful disruption at Lenovo have led us to a point where we can now claim leadership of the coming intelligent transformation. People often asked me, what is the intelligent transformation? I was saying this way. This is the unlimited potential of the Fourth Industrial Revolution driven by artificial intelligence being realized, ordering a pizza through our speaker, and locking the door with a look, letting your car drive itself back to your home. This indeed reflect the power of AI, but it just the surface of it. The true impact of AI will not only make our homes smarter and offices more efficient, but we are also completely transformed every value chip in every industry. However, to realize these amazing possibilities, we will need a structure built around the key components, and one that touches every part of all our lives. First of all, explosions in new technology always lead to new structures. This has happened many times before. In the early 20th century, thousands of companies provided a telephone service. City streets across the US looked like this, and now bundles of a microscopic fiber running from city to city bring the world closer together. Here's what a driving was like in the US, up until 1950s. Good luck finding your way. (audience laughs) And today, millions of vehicles are organized and routed daily, making the world more efficient. Structure is vital, from fiber cables and the interstate highways, to our cells bounded together to create humans. Thankfully the structure for intelligent transformation has emerged, and it is just as revolutionary. What does this new structure look like? We believe there are three key building blocks, data, computing power, and algorithms. Ever wondered what is it behind intelligent transformation? What is fueling this miracle of human possibility? Data. As the Internet becomes ubiquitous, not only PCs, mobile phones, have come online and been generating data. Today it is the cameras in this room, the climate controls in our offices, or the smart displays in our kitchens at home. The number of smart devices worldwide will reach over 20 billion in 2020, more than double the number in 2017. These devices and the sensors are connected and generating massive amount of data. By 2020, the amount of data generated will be 57 times more than all the grains of sand on Earth. This data will not only make devices smarter, but will also fuel the intelligence of our homes, offices, and entire industries. Then we need engines to turn the fuel into power, and the engine is actually the computing power. Last but not least the advanced algorithms combined with Big Data technology and industry know how will form vertical industrial intelligence and produce valuable insights for every value chain in every industry. When these three building blocks all come together, it will change the world. At Lenovo, we have each of these elements of intelligent transformations in a single place. We have built our business around the new structure of intelligent transformation, especially with mobile and the data center now firmly part of our business. I'm often asked why did you acquire these businesses? Why has a Lenovo gone into so many fields? People ask the same questions of the companies that become the leaders of the information technology revolution, or the third industrial transformation. They were the companies that saw the future and what the future required, and I believe Lenovo is the company today. From largest portfolio of devices in the world, leadership in the data center field, to the algorithm-powered intelligent vertical solutions, and not to mention the strong partnership Lenovo has built over decades. We are the only company that can unify all these essential assets and deliver end to end solutions. Let's look at each part. We now understand the important importance data plays as fuel in intelligent transformation. Hundreds of billions of devices and smart IoTs in the world are generating better and powering the intelligence. Who makes these devices in large volume and variety? Who puts these devices into people's home, offices, manufacturing lines, and in their hands? Lenovo definitely has the front row seats here. We are number one in PCs and tablets. We also produces smart phones, smart speakers, smart displays. AR/VR headsets, as well as commercial IoTs. All of these smart devices, or smart IoTs are linked to each other and to the cloud. In fact, we have more than 20 manufacturing facilities in China, US, Brazil, Japan, India, Mexico, Germany, and more, producing various devices around the clock. We actually make four devices every second, and 37 motherboards every minute. So, this factory located in my hometown, Hu-fi, China, is actually the largest laptop factory in the world, with more than three million square feet. So, this is as big as 42 soccer fields. Our scale and the larger portfolio of devices gives us access to massive amount of data, which very few companies can say. So, why is the ability to scale so critical? Let's look again at our example from before. The early days of telephone, dozens of service providers but only a few companies could survive consolidation and become the leader. The same was true for the third Industrial Revolution. Only a few companies could scale, only a few could survive to lead. Now the building blocks of the next revolution are locking into place. The (mumbles) will go to those who can operate at the scale. So, who could foresee the total integration of cloud, network, and the device, need to deliver intelligent transformation. Lenovo is that company. We are ready to scale. Next, our computing power. Computing power is provided in two ways. On one hand, the modern supercomputers are providing the brute force to quickly analyze the massive data like never before. On the other hand the cloud computing data centers with the server storage networking capabilities, and any computing IoT's, gateways, and miniservers are making computing available everywhere. Did you know, Lenovo is number one provider of super computers worldwide? 170 of the top 500 supercomputers, run on Lenovo. We hold 89 World Records in key workloads. We are number one in x86 server reliability for five years running, according to ITIC. a respected provider of industry research. We are also the fastest growing provider of hyperscale public cloud, hyper-converged and aggressively growing in edge computing. cur-ges target, we are expand on this point soon. And finally to run these individual nodes into our symphony, we must transform the data and utilize the computing power with advanced algorithms. Manufactured, industry maintenance, healthcare, education, retail, and more, so many industries are on the edge of intelligent transformation to improve efficiency and provide the better products and services. We are creating advanced algorithms and the big data tools combined with industry know-how to provide intelligent vertical solutions for several industries. In fact, we studied at Lenovo first. Our IT and research teams partnered with our global supply chain to develop an AI that improved our demand forecasting accuracy. Beyond managing our own supply chain we have offered our deep learning supply focused solution to other manufacturing companies to improve their efficiency. In the best case, we have improved the demand, focused the accuracy by 30 points to nearly 90 percent, for Baosteel, the largest of steel manufacturer in China, covering the world as well. Led by Lenovo research, we launched the industry-leading commercial ready AR headset, DaystAR, partnering with companies like the ones in this room. This technology is being used to revolutionize the way companies service utility, and even our jet engines. Using our workstations, servers, and award-winning imaging processing algorithms, we have partnered with hospitals to process complex CT scan data in minutes. So, this enable the doctors to more successfully detect the tumors, and it increases the success rate of cancer diagnosis all around the world. We are also piloting our smart IoT driven warehouse solution with one of the world's largest retail companies to greatly improve the efficiency. So, the opportunities are endless. This is where Lenovo will truly shine. When we combine the industry know-how of our customers with our end-to-end technology offerings, our intelligent vertical solutions like this are growing, which Kirk and Christian will share more. Now, what will drive this transformation even faster? The speed at which our networks operate, specifically 5G. You may know that Lenovo just launched the first-ever 5G smartphone, our Moto Z3, with the new 5G Moto model. We are partnering with multiple major network providers like Verizon, China Mobile. With the 5G model scheduled to ship early next year, we will be the first company to provide a 5G mobile experience to any users, customers. This is amazing innovation. You don't have to buy a new phone, just the 5G clip on. What can I say, except wow. (audience laughs) 5G is 10 times the fast faster than 4G. Its download speed will transform how people engage with the world, driverless car, new types of smart wearables, gaming, home security, industrial intelligence, all will be transformed. Finally, accelerating with partners, as ready as we are at Lenovo, we need partners to unlock our full potential, partners here to create with us the edge of the intelligent transformation. The opportunities of intelligent transformation are too profound, the scale is too vast. No company can drive it alone fully. We are eager to collaborate with all partners that can help bring our vision to life. We are dedicated to open partnerships, dedicated to cross-border collaboration, unify the standards, share the advantage, and market the synergies. We partner with the biggest names in the industry, Intel, Microsoft, AMD, Qualcomm, Google, Amazon, and Disney. We also find and partner with the smaller innovators as well. We're building the ultimate partner experience, open, shared, collaborative, diverse. So, everything is in place for intelligent transformation on a global scale. Smart devices are everywhere, the infrastructure is in place, networks are accelerating, and the industries demand to be more intelligent, and Lenovo is at the center of it all. We are helping to drive change with the hundreds of companies, companies just like yours, every day. We are your partner for intelligent transformation. Transformation never stops. This is what you will hear from Kirk, including details about Lenovo NetApp global partnership we just announced this morning. We've made the investments in every single aspect of the technology. We have the end-to-end resources to meet your end-to-end needs. As you attend the breakout session this afternoon, I hope you see for yourself how much Lenovo has transformed as a company this past year, and how we truly are delivering a future of intelligent transformation. Now, let me invite to the stage Kirk Skaugen, our president of Data Center growth to tell you about the exciting transformation happening in the global Data C enter market. Thank you. (audience applauding) (upbeat music) >> Well, good morning. >> Good morning. >> Good morning! >> Good morning! >> Excellent, well, I'm pleased to be here this morning to talk about how we're transforming the Data Center and taking you as our customers through your own intelligent transformation journey. Last year I stood up here at Transform 1.0, and we were proud to announce the largest Data Center portfolio in Lenovo's history, so I thought I'd start today and talk about the portfolio and the progress that we've made over the last year, and the strategies that we have going forward in phase 2.0 of Lenovo's transformation to be one of the largest data center companies in the world. We had an audacious vision that we talked about last year, and that is to be the most trusted data center provider in the world, empowering customers through the new IT, intelligent transformation. And now as the world's largest supercomputer provider, giving something back to humanity, is very important this week with the hurricanes now hitting North Carolina's coast, but we take this most trusted aspect very seriously, whether it's delivering the highest quality products on time to you as customers with the highest levels of security, or whether it's how we partner with our channel partners and our suppliers each and every day. You know we're in a unique world where we're going from hundreds of millions of PCs, and then over the next 25 years to hundred billions of connected devices, so each and every one of you is going through this intelligent transformation journey, and in many aspects were very early in that cycle. And we're going to talk today about our role as the largest supercomputer provider, and how we're solving humanity's greatest challenges. Last year we talked about two special milestones, the 25th anniversary of ThinkPad, but also the 25th anniversary of Lenovo with our IBM heritage in x86 computing. I joined the workforce in 1992 out of college, and the IBM first personal server was launching at the same time with an OS2 operating system and a free mouse when you bought the server as a marketing campaign. (audience laughing) But what I want to be very clear today, is that the innovation engine is alive and well at Lenovo, and it's really built on the culture that we're building as a company. All of these awards at the bottom are things that we earned over the last year at Lenovo. As a Fortune now 240 company, larger than companies like Nike, or AMEX, or Coca-Cola. The one I'm probably most proud of is Forbes first list of the top 2,000 globally regarded companies. This was something where 15,000 respondents in 60 countries voted based on ethics, trustworthiness, social conduct, company as an employer, and the overall company performance, and Lenovo was ranked number 27 of 2000 companies by our peer group, but we also now one of-- (audience applauding) But we also got a perfect score in the LGBTQ Equality Index, exemplifying the diversity internally. We're number 82 in the top working companies for mothers, top working companies for fathers, top 100 companies for sustainability. If you saw that factory, it's filled with solar panels on the top of that. And now again, one of the top global brands in the world. So, innovation is built on a customer foundation of trust. We also said last year that we'd be crossing an amazing milestone. So we did, over the last 12 months ship our 20 millionth x86 server. So, thank you very much to our customers for this milestone. (audience applauding) So, let me recap some of the transformation elements that have happened over the last year. Last year I talked about a lot of brand confusion, because we had the ThinkServer brand from the legacy Lenovo, the System x, from IBM, we had acquired a number of networking companies, like BLADE Network Technologies, et cetera, et cetera. Over the last year we've been ramping based on two brand structures, ThinkAgile for next generation IT, and all of our software-defined infrastructure products and ThinkSystem as the world's highest performance, highest reliable x86 server brand, but for servers, for storage, and for networking. We have transformed every single aspect of the customer experience. A year and a half ago, we had four different global channel programs around the world. Typically we're about twice the mix to our channel partners of any of our competitors, so this was really important to fix. We now have a single global Channel program, and have technically certified over 11,000 partners to be technical experts on our product line to deliver better solutions to our customer base. Gardner recently recognized Lenovo as the 26th ranked supply chain in the world. And, that's a pretty big honor, when you're up there with Amazon and Walmart and others, but in tech, we now are in the top five supply chains. You saw the factory network from YY, and today we'll be talking about product shipping in more than 160 countries, and I know there's people here that I've met already this morning, from India, from South Africa, from Brazil and China. We announced new Premier Support services, enabling you to go directly to local language support in nine languages in 49 countries in the world, going directly to a native speaker level three support engineer. And today we have more than 10,000 support specialists supporting our products in over 160 countries. We've delivered three times the number of engineered solutions to deliver a solutions orientation, whether it's on HANA, or SQL Server, or Oracle, et cetera, and we've completely reengaged our system integrator channel. Last year we had the CIO of DXE on stage, and here we're talking about more than 175 percent growth through our system integrator channel in the last year alone as we've brought that back and really built strong relationships there. So, thank you very much for amazing work here on the customer experience. (audience applauding) We also transformed our leadership. We thought it was extremely important with a focus on diversity, to have diverse talent from the legacy IBM, the legacy Lenovo, but also outside the industry. We made about 19 executive changes in the DCG group. This is the most senior leadership team within DCG, all which are newly on board, either from our outside competitors mainly over the last year. About 50 percent of our executives were now hired internally, 50 percent externally, and 31 percent of those new executives are diverse, representing the diversity of our global customer base and gender. So welcome, and most of them you're going to be able to meet over here in the breakout sessions later today. (audience applauding) But some things haven't changed, they're just keeping getting better within Lenovo. So, last year I got up and said we were committed with the new ThinkSystem brand to be a world performance leader. You're going to see that we're sponsoring Ducati for MotoGP. You saw the Ferrari out there with Formula One. That's not a surprise. We want the Lenovo ThinkSystem and ThinkAgile brands to be synonymous with world record performance. So in the last year we've gone from 39 to 89 world records, and partners like Intel would tell you, we now have four times the number of world record workloads on Lenovo hardware than any other server company on the planet today, with more than 89 world records across HPC, Java, database, transaction processing, et cetera. And we're proud to have just brought on Doug Fisher from Intel Corporation who had about 10-17,000 people on any given year working for him in workload optimizations across all of our software. It's just another testament to the leadership team we're bringing in to keep focusing on world-class performance software and solutions. We also per ITIC, are the number one now in x86 server reliability five years running. So, this is a survey where CIOs are in a blind survey asked to submit their reliability of their uptime on their x86 server equipment over the last 365 days. And you can see from 2016 to 2017 the downtime, there was over four hours as noted by the 750 CXOs in more than 20 countries is about one percent for the Lenovo products, and is getting worse generation from generation as we went from Broadwell to Pearlie. So we're taking our reliability, which was really paramount in the IBM System X heritage, and ensuring that we don't just recognize high performance but we recognize the highest level of reliability for mission-critical workloads. And what that translates into is that we at once again have been ranked number one in customer satisfaction from you our customers in 19 of 22 attributes, in North America in 18 of 22. This is a survey by TVR across hundreds of customers of us and our top competitors. This is the ninth consecutive study that we've been ranked number one in customer satisfaction, so we're taking this extremely seriously, and in fact YY now has increased the compensation of every single Lenovo employee. Up to 40 percent of their compensation bonus this year is going to be based on customer metrics like quality, order to ship, and things of this nature. So, we're really putting every employee focused on customer centricity this year. So, the summary on Transform 1.0 is that every aspect of what you knew about Lenovo's data center group has transformed, from the culture to the branding to dedicated sales and marketing, supply chain and quality groups, to a worldwide channel program and certifications, to new system integrator relationships, and to the new leadership team. So, rather than me just talk about it, I thought I'd share a quick video about what we've done over the last year, if you could run the video please. Turn around for a second. (epic music) (audience applauds) Okay. So, thank you to all our customers that allowed us to publicly display their logos in that video. So, what that means for you as investors, and for the investor community out there is, that our customers have responded, that this year Gardner just published that we are the fastest growing server company in the top 10, with 39 percent growth quarter-on-quarter, and 49 percent growth year-on-year. If you look at the progress we've made since the transformation the last three quarters publicly, we've grown 17 percent, then 44 percent, then 68 percent year on year in revenue, and I can tell you this quarter I'm as confident as ever in the financials around the DCG group, and it hasn't been in one area. You're going to see breakout sessions from hyperscale, software-defined, and flash, which are all growing more than a 100 percent year-on-year, supercomputing which we'll talk about shortly, now number one, and then ultimately from profitability, delivering five consecutive quarters of pre-tax profit increase, so I think, thank you very much to the customer base who's been working with us through this transformation journey. So, you're here to really hear what's next on 2.0, and that's what I'm excited to talk about today. Last year I came up with an audacious goal that we would become the largest supercomputer company on the planet by 2020, and this graph represents since the acquisition of the IBM System x business how far we were behind being the number one supercomputer. When we started we were 182 positions behind, even with the acquisition for example of SGI from HP, we've now accomplished our goal actually two years ahead of time. We're now the largest supercomputer company in the world. About one in every four supercomputers, 117 on the list, are now Lenovo computers, and you saw in the video where the universities are said, but I think what I'm most proud of is when your customers rank you as the best. So the awards at the bottom here, are actually Readers Choice from the last International Supercomputing Show where the scientific researchers on these computers ranked their vendors, and we were actually rated the number one server technology in supercomputing with our ThinkSystem SD530, and the number one storage technology with our ThinkSystem DSS-G, but more importantly what we're doing with the technology. You're going to see we won best in life sciences, best in data analytics, and best in collaboration as well, so you're going to see all of that in our breakout sessions. As you saw in the video now, 17 of the top 25 research institutions in the world are now running Lenovo supercomputers. And again coming from Raleigh and watching that hurricane come across the Atlantic, there are eight supercomputers crunching all of those models you see from Germany to Malaysia to Canada, and we're happy to have a SciNet from University of Toronto here with us in our breakout session to talk about what they're doing on climate modeling as well. But we're not stopping there. We just announced our new Neptune warm water cooling technology, which won the International Supercomputing Vendor Showdown, the first time we've won that best of show in 25 years, and we've now installed this. We're building out LRZ in Germany, the first ever warm water cooling in Peking University, at the India Space Propulsion Laboratory, at the Malaysian Weather and Meteorological Society, at Uninett, at the largest supercomputer in Norway, T-Systems, University of Birmingham. This is truly amazing technology where we're actually using water to cool the machine to deliver a significantly more energy-efficient computer. Super important, when we're looking at global warming and some of the electric bills can be millions of dollars just for one computer, and could actually power a small city just with the technology from the computer. We've built AI centers now in Morrisville, Stuttgart, Taipei, and Beijing, where customers can bring their AI workloads in with experts from Intel, from Nvidia, from our FPGA partners, to work on their workloads, and how they can best implement artificial intelligence. And we also this year launched LICO which is Lenovo Intelligent Compute Orchestrator software, and it's a software solution that simplifies the management and use of distributed clusters in both HPC and AI model development. So, what it enables you to do is take a single cluster, and run both HPC and AI workloads on it simultaneously, delivering better TCO for your environment, so check out LICO as well. A lot of the customers here and Wall Street are very excited and using it already. And we talked about solving humanity's greatest challenges. In the breakout session, you're going to have a virtual reality experience where you're going to be able to walk through what as was just ranked the world's most beautiful data center, the Barcelona Supercomputer. So, you can actually walk through one of the largest supercomputers in the world from Barcelona. You can see the work we're doing with NC State where we're going to have to grow the food supply of the world by 50 percent, and there's not enough fresh water in the world in the right places to actually make all those crops grow between now and 2055, so you're going to see the progression of how they're mapping the entire globe and the water around the world, how to build out the crop population over time using AI. You're going to see our work with Vestas is this largest supercomputer provider in the wind turbine areas, how they're working on wind energy, and then with University College London, how they're working on some of the toughest particle physics calculations in the world. So again, lots of opportunity here. Take advantage of it in the breakout sessions. Okay, let me transition to hyperscale. So in hyperscale now, we have completely transformed our business model. We are now powering six of the top 10 hyperscalers in the world, which is a significant difference from where we were two years ago. And the reason we're doing that, is we've coined a term called ODM+. We believe that hyperscalers want more procurement power than an ODM, and Lenovo is doing about $18 billion of procurement a year. They want a broader global supply chain that they can get from a local system integrator. We're more than 160 countries around the world, but they want the same world-class quality and reliability like they get from an MNC. So, what we're doing now is instead of just taking off the shelf motherboards from somewhere, we're starting with a blank sheet of paper, we're working with the customer base on customized SKUs and you can see we already are developing 33 custom solutions for the largest hyperscalers in the world. And then we're not just running notebooks through this factory where YY said, we're running 37 notebook boards a minute, we're now putting in tens and tens and tens of thousands of server board capacity per month into this same factory, so absolutely we can compete with the most aggressive ODM's in the world, but it's not just putting these things in in the motherboard side, we're also building out these systems all around the world, India, Brazil, Hungary, Mexico, China. This is an example of a new hyperscale customer we've had this last year, 34,000 servers we delivered in the first six months. The next 34,000 servers we delivered in 68 days. The next 34,000 servers we delivered in 35 days, with more than 99 percent on-time delivery to 35 data centers in 14 countries as diverse as South Africa, India, China, Brazil, et cetera. And I'm really ashamed to say it was 99.3, because we did have a forklift driver who rammed their forklift right through the middle of the one of the server racks. (audience laughing) At JFK Airport that we had to respond to, but I think this gives you a perspective of what it is to be a top five global supply chain and technology. So last year, I said we would invest significantly in IP, in joint ventures, and M and A to compete in software defined, in networking, and in storage, so I wanted to give you an update on that as well. Our newest software-defined partnership is with Cloudistics, enabling a fully composable cloud infrastructure. It's an exclusive agreement, you can see them here. I think Nag, our founder, is going to be here today, with a significant Lenovo investment in the company. So, this new ThinkAgile CP series delivers the simplicity of the public cloud, on-premise with exceptional support and a marketplace of essential enterprise applications all with a single click deployment. So simply put, we're delivering a private cloud with a premium experience. It's simple in that you need no specialists to deploy it. An IT generalist can set it up and manage it. It's agile in that you can provision dozens of workloads in minutes, and it's transformative in that you get all of the goodness of public cloud on-prem in a private cloud to unlock opportunity for use. So, we're extremely excited about the ThinkAgile CP series that's now shipping into the marketplace. Beyond that we're aggressively ramping, and we're either doubling, tripling, or quadrupling our market share as customers move from traditional server technology to software-defined technology. With Nutanix we've been public, growing about more than 150 percent year-on-year, with Nutanix as their fastest growing Nutanix partner, but today I want to set another audacious goal. I believe we cannot just be Nutanix's fastest growing partner but we can become their largest partner within two years. On Microsoft, we are already four times our market share on Azure stack of our traditional business. We were the first to launch our ThinkAgile on Broadwell and on Skylake with the Azure Stack Infrastructure. And on VMware we're about twice our market segment share. We were the first to deliver an Intel-optimized Optane-certified VSAN node. And with Optane technology, we're delivering 50 percent more VM density than any competitive SSD system in the marketplace, about 10 times lower latency, four times the performance of any SSD system out there, and Lenovo's first to market on that. And at VMworld you saw CEO Pat Gelsinger of VMware talked about project dimension, which is Edge as a service, and we're the only OEM beyond the Dell family that is participating today in project dimension. Beyond that you're going to see a number of other partnerships we have. I'm excited that we have the city of Bogota Columbia here, an eight million person city, where we announced a 3,000 camera video surveillance solution last month. With pivot three you're going to see city of Bogota in our breakout sessions. You're going to see a new partnership with Veeam around backup that's launching today. You're going to see partnerships with scale computing in IoT and hyper-converged infrastructure working on some of the largest retailers in the world. So again, everything out in the breakout session. Transitioning to storage and data management, it's been a great year for Lenovo, more than a 100 percent growth year-on-year, 2X market growth in flash arrays. IDC just reported 30 percent growth in storage, number one in price performance in the world and the best HPC storage product in the top 500 with our ThinkSystem DSS G, so strong coverage, but I'm excited today to announce for Transform 2.0 that Lenovo is launching the largest data management and storage portfolio in our 25-year data center history. (audience applauding) So a year ago, the largest server portfolio, becoming the largest fastest growing server OEM, today the largest storage portfolio, but as you saw this morning we're not doing it alone. Today Lenovo and NetApp, two global powerhouses are joining forces to deliver a multi-billion dollar global alliance in data management and storage to help customers through their intelligent transformation. As the fastest growing worldwide server leader and one of the fastest growing flash array and data management companies in the world, we're going to deliver more choice to customers than ever before, global scale that's never been seen, supply chain efficiencies, and rapidly accelerating innovation and solutions. So, let me unwrap this a little bit for you and talk about what we're announcing today. First, it's the largest portfolio in our history. You're going to see not just storage solutions launching today but a set of solution recipes from NetApp that are going to make Lenovo server and NetApp or Lenovo storage work better together. The announcement enables Lenovo to go from covering 15 percent of the global storage market to more than 90 percent of the global storage market and distribute these products in more than 160 countries around the world. So we're launching today, 10 new storage platforms, the ThinkSystem DE and ThinkSystem DM platforms. They're going to be centrally managed, so the same XClarity management that you've been using for server, you can now use across all of your storage platforms as well, and it'll be supported by the same 10,000 plus service personnel that are giving outstanding customer support to you today on the server side. And we didn't come up with this in the last month or the last quarter. We're announcing availability in ordering today and shipments tomorrow of the first products in this portfolio, so we're excited today that it's not just a future announcement but something you as customers can take advantage of immediately. (audience applauding) The second part of the announcement is we are announcing a joint venture in China. Not only will this be a multi-billion dollar global partnership, but Lenovo will be a 51 percent owner, NetApp a 49 percent owner of a new joint venture in China with the goal of becoming in the top three storage companies in the largest data and storage market in the world. We will deliver our R and D in China for China, pooling our IP and resources together, and delivering a single route to market through a complementary channel, not just in China but worldwide. And in the future I just want to tell everyone this is phase one. There is so much exciting stuff. We're going to be on the stage over the next year talking to you about around integrated solutions, next-generation technologies, and further synergies and collaborations. So, rather than just have me talk about it, I'd like to welcome to the stage our new partner NetApp and Brad Anderson who's the senior vice president and general manager of NetApp Cloud Infrastructure. (upbeat music) (audience applauding) >> Thank You Kirk. >> So Brad, we've known each other a long time. It's an exciting day. I'm going to give you the stage and allow you to say NetApp's perspective on this announcement. >> Very good, thank you very much, Kirk. Kirk and I go back to I think 1994, so hey good morning and welcome. My name is Brad Anderson. I manage the Cloud Infrastructure Group at NetApp, and I am honored and privileged to be here at Lenovo Transform, particularly today on today's announcement. Now, you've heard a lot about digital transformation about how companies have to transform their IT to compete in today's global environment. And today's announcement with the partnership between NetApp and Lenovo is what that's all about. This is the joining of two global leaders bringing innovative technology in a simplified solution to help customers modernize their IT and accelerate their global digital transformations. Drawing on the strengths of both companies, Lenovo's high performance compute world-class supply chain, and NetApp's hybrid cloud data management, hybrid flash and all flash storage solutions and products. And both companies providing our customers with the global scale for them to be able to meet their transformation goals. At NetApp, we're very excited. This is a quote from George Kurian our CEO. George spent all day yesterday with YY and Kirk, and would have been here today if it hadn't been also our shareholders meeting in California, but I want to just convey how excited we are for all across NetApp with this partnership. This is a partnership between two companies with tremendous market momentum. Kirk took you through all the amazing results that Lenovo has accomplished, number one in supercomputing, number one in performance, number one in x86 reliability, number one in x86 customers sat, number five in supply chain, really impressive and congratulations. Like Lenovo, NetApp is also on a transformation journey, from a storage company to the data authority in hybrid cloud, and we've seen some pretty impressive momentum as well. Just last week we became number one in all flash arrays worldwide, catching EMC and Dell, and we plan to keep on going by them, as we help customers modernize their their data centers with cloud connected flash. We have strategic partnerships with the largest hyperscalers to provide cloud native data services around the globe and we are having success helping our customers build their own private clouds with just, with a new disruptive hyper-converged technology that allows them to operate just like hyperscalers. These three initiatives has fueled NetApp's transformation, and has enabled our customers to change the world with data. And oh by the way, it has also fueled us to have meet or have beaten Wall Street's expectations for nine quarters in a row. These are two companies with tremendous market momentum. We are also building this partnership for long term success. We think about this as phase one and there are two important components to phase one. Kirk took you through them but let me just review them. Part one, the establishment of a multi-year commitment and a collaboration agreement to offer Lenovo branded flash products globally, and as Kurt said in 160 countries. Part two, the formation of a joint venture in PRC, People's Republic of China, that will provide long term commitment, joint product development, and increase go-to-market investment to meet the unique needs to China. Both companies will put in storage technologies and storage expertise to form an independent JV that establishes a data management company in China for China. And while we can dream about what phase two looks like, our entire focus is on making phase one incredibly successful and I'm pleased to repeat what Kirk, is that the first products are orderable and shippable this week in 160 different countries, and you will see our two companies focusing on the here and now. On our joint go to market strategy, you'll see us working together to drive strategic alignment, focused execution, strong governance, and realistic expectations and milestones. And it starts with the success of our customers and our channel partners is job one. Enabling customers to modernize their legacy IT with complete data center solutions, ensuring that our customers get the best from both companies, new offerings the fuel business success, efficiencies to reinvest in game-changing initiatives, and new solutions for new mission-critical applications like data analytics, IoT, artificial intelligence, and machine learning. Channel partners are also top of mind for both our two companies. We are committed to the success of our existing and our future channel partners. For NetApp channel partners, it is new pathways to new segments and to new customers. For Lenovo's channel partners, it is the competitive weapons that now allows you to compete and more importantly win against Dell, EMC, and HP. And the good news for both companies is that our channel partner ecosystem is highly complementary with minimal overlap. Today is the first day of a very exciting partnership, of a partnership that will better serve our customers today and will provide new opportunities to both our companies and to our partners, new products to our customers globally and in China. I am personally very excited. I will be on the board of the JV. And so, I look forward to working with you, partnering with you and serving you as we go forward, and with that, I'd like to invite Kirk back up. (audience applauding) >> Thank you. >> Thank you. >> Well, thank you, Brad. I think it's an exciting overview, and these products will be manufactured in China, in Mexico, in Hungary, and around the world, enabling this amazing supply chain we talked about to deliver in over 160 countries. So thank you Brad, thank you George, for the amazing partnership. So again, that's not all. In Transform 2.0, last year, we talked about the joint ventures that were coming. I want to give you a sneak peek at what you should expect at future Lenovo events around the world. We have this Transform in Beijing in a couple weeks. We'll then be repeating this in 20 different locations roughly around the world over the next year, and I'm excited probably more than ever about what else is coming. Let's talk about Telco 5G and network function virtualization. Today, Motorola phones are certified on 46 global networks. We launched the world's first 5G upgradable phone here in the United States with Verizon. Lenovo DCG sells to 58 telecommunication providers around the world. At Mobile World Congress in Barcelona and Shanghai, you saw China Telecom and China Mobile in the Lenovo booth, China Telecom showing a video broadband remote access server, a VBRAS, with video streaming demonstrations with 2x less jitter than they had seen before. You saw China Mobile with a virtual remote access network, a VRAN, with greater than 10 times the throughput and 10x lower latency running on Lenovo. And this year, we'll be launching a new NFV company, a software company in China for China to drive the entire NFV stack, delivering not just hardware solutions, but software solutions, and we've recently hired a new CEO. You're going to hear more about that over the next several quarters. Very exciting as we try to drive new economics into the networks to deliver these 20 billion devices. We're going to need new economics that I think Lenovo can uniquely deliver. The second on IoT and edge, we've integrated on the device side into our intelligent devices group. With everything that's going to consume electricity computes and communicates, Lenovo is in a unique position on the device side to take advantage of the communications from Motorola and being one of the largest device companies in the world. But this year, we're also going to roll out a comprehensive set of edge gateways and ruggedized industrial servers and edge servers and ISP appliances for the edge and for IoT. So look for that as well. And then lastly, as a service, you're going to see Lenovo delivering hardware as a service, device as a service, infrastructure as a service, software as a service, and hardware as a service, not just as a glorified leasing contract, but with IP, we've developed true flexible metering capability that enables you to scale up and scale down freely and paying strictly based on usage, and we'll be having those announcements within this fiscal year. So Transform 2.0, lots to talk about, NetApp the big news of the day, but a lot more to come over the next year from the Data Center group. So in summary, I'm excited that we have a lot of customers that are going to be on stage with us that you saw in the video. Lots of testimonials so that you can talk to colleagues of yourself. Alamos Gold from Canada, a Canadian gold producer, Caligo for data optimization and privacy, SciNet, the largest supercomputer we've ever put into North America, and the largest in Canada at the University of Toronto will be here talking about climate change. City of Bogota again with our hyper-converged solutions around smart city putting in 3,000 cameras for criminal detection, license plate detection, et cetera, and then more from a channel mid market perspective, Jerry's Foods, which is from my home state of Wisconsin, and Minnesota which has about 57 stores in the specialty foods market, and how they're leveraging our IoT solutions as well. So again, about five times the number of demos that we had last year. So in summary, first and foremost to the customers, thank you for your business. It's been a great journey and I think we're on a tremendous role. You saw from last year, we're trying to build credibility with you. After the largest server portfolio, we're now the fastest-growing server OEM per Gardner, number one in performance, number one in reliability, number one in customer satisfaction, number one in supercomputing. Today, the largest storage portfolio in our history, with the goal of becoming the fastest growing storage company in the world, top three in China, multibillion-dollar collaboration with NetApp. And the transformation is going to continue with new edge gateways, edge servers, NFV solutions, telecommunications infrastructure, and hardware as a service with dynamic metering. So thank you for your time. I've looked forward to meeting many of you over the next day. We appreciate your business, and with that, I'd like to bring up Rod Lappen to introduce our next speaker. Rod? (audience applauding) >> Thanks, boss, well done. Alright ladies and gentlemen. No real secret there. I think we've heard why I might talk about the fourth Industrial Revolution in data and exactly what's going on with that. You've heard Kirk with some amazing announcements, obviously now with our NetApp partnership, talk about 5G, NFV, cloud, artificial intelligence, I think we've hit just about all the key hot topics. It's with great pleasure that I now bring up on stage Mr. Christian Teismann, our senior vice president and general manager of commercial business for both our PCs and our IoT business, so Christian Teismann. (techno music) Here, take that. >> Thank you. I think I'll need that. >> Okay, Christian, so obviously just before we get down, you and I last year, we had a bit of a chat about being in New York. >> Exports. >> You were an expat in New York for a long time. >> That's true. >> And now, you've moved from New York. You're in Munich? >> Yep. >> How does that feel? >> Well Munich is a wonderful city, and it's a great place to live and raise kids, but you know there's no place in the world like New York. >> Right. >> And I miss it a lot, quite frankly. >> So what exactly do you miss in New York? >> Well there's a lot of things in New York that are unique, but I know you spent some time in Japan, but I still believe the best sushi in the world is still in New York City. (all laughing) >> I will beg to differ. I will beg to differ. I think Mr. Guchi-san from Softbank is here somewhere. He will get up an argue very quickly that Japan definitely has better sushi than New York. But obviously you know, it's a very very special place, and I have had sushi here, it's been fantastic. What about Munich? Anything else that you like in Munich? >> Well I mean in Munich, we have pork knuckles. >> Pork knuckles. (Christian laughing) Very similar sushi. >> What is also very fantastic, but we have the real, the real Oktoberfest in Munich, and it starts next week, mid-September, and I think it's unique in the world. So it's very special as well. >> Oktoberfest. >> Yes. >> Unfortunately, I'm not going this year, 'cause you didn't invite me, but-- (audience chuckling) How about, I think you've got a bit of a secret in relation to Oktoberfest, probably not in Munich, however. >> It's a secret, yes, but-- >> Are you going to share? >> Well I mean-- >> See how I'm putting you on the spot? >> In the 10 years, while living here in New York, I was a regular visitor of the Oktoberfest at the Lower East Side in Avenue C at Zum Schneider, where I actually met my wife, and she's German. >> Very good. So, how about a big round of applause? (audience applauding) Not so much for Christian, but more I think, obviously for his wife, who obviously had been drinking and consequently ended up with you. (all laughing) See you later, mate. >> That's the beauty about Oktoberfest, but yes. So first of all, good morning to everybody, and great to be back here in New York for a second Transform event. New York clearly is the melting pot of the world in terms of culture, nations, but also business professionals from all kind of different industries, and having this event here in New York City I believe is manifesting what we are trying to do here at Lenovo, is transform every aspect of our business and helping our customers on the journey of intelligent transformation. Last year, in our transformation on the device business, I talked about how the PC is transforming to personalized computing, and we've made a lot of progress in that journey over the last 12 months. One major change that we have made is we combined all our device business under one roof. So basically PCs, smart devices, and smart phones are now under the roof and under the intelligent device group. But from my perspective makes a lot of sense, because at the end of the day, all devices connect in the modern world into the cloud and are operating in a seamless way. But we are also moving from a device business what is mainly a hardware focus historically, more and more also into a solutions business, and I will give you during my speech a little bit of a sense of what we are trying to do, as we are trying to bring all these components closer together, and specifically also with our strengths on the data center side really build end-to-end customer solution. Ultimately, what we want to do is make our business, our customer's businesses faster, safer, and ultimately smarter as well. So I want to look a little bit back, because I really believe it's important to understand what's going on today on the device side. Many of us have still grown up with phones with terminals, ultimately getting their first desktop, their first laptop, their first mobile phone, and ultimately smartphone. Emails and internet improved our speed, how we could operate together, but still we were defined by linear technology advances. Today, the world has changed completely. Technology itself is not a limiting factor anymore. It is how we use technology going forward. The Internet is pervasive, and we are not yet there that we are always connected, but we are nearly always connected, and we are moving to the stage, that everything is getting connected all the time. Sharing experiences is the most driving force in our behavior. In our private life, sharing pictures, videos constantly, real-time around the world, with our friends and with our family, and you see the same behavior actually happening in the business life as well. Collaboration is the number-one topic if it comes down to workplace, and video and instant messaging, things that are coming from the consumer side are dominating the way we are operating in the commercial business as well. Most important beside technology, that a new generation of workforce has completely changed the way we are working. As the famous workforce the first generation of Millennials that have now fully entered in the global workforce, and the next generation, it's called Generation Z, is already starting to enter the global workforce. By 2025, 75 percent of the world's workforce will be composed out of two of these generations. Why is this so important? These two generations have been growing up using state-of-the-art IT technology during their private life, during their education, school and study, and are taking these learnings and taking these behaviors in the commercial workspace. And this is the number one force of change that we are seeing in the moment. Diverse workforces are driving this change in the IT spectrum, and for years in many of our customers' focus was their customer focus. Customer experience also in Lenovo is the most important thing, but we've realized that our own human capital is equally valuable in our customer relationships, and employee experience is becoming a very important thing for many of our customers, and equally for Lenovo as well. As you have heard YY, as we heard from YY, Lenovo is focused on intelligent transformation. What that means for us in the intelligent device business is ultimately starting with putting intelligence in all of our devices, smartify every single one of our devices, adding value to our customers, traditionally IT departments, but also focusing on their end users and building products that make their end users more productive. And as a world leader in commercial devices with more than 33 percent market share, we can solve problems been even better than any other company in the world. So, let's talk about transformation of productivity first. We are in a device-led world. Everything we do is connected. There's more interaction with devices than ever, but also with spaces who are increasingly becoming smart and intelligent. YY said it, by 2020 we have more than 20 billion connected devices in the world, and it will grow exponentially from there on. And users have unique personal choices for technology, and that's very important to recognize, and we call this concept a digital wardrobe. And it means that every single end-user in the commercial business is composing his personal wardrobe on an ongoing basis and is reconfiguring it based on the work he's doing and based where he's going and based what task he is doing. I would ask all of you to put out all the devices you're carrying in your pockets and in your bags. You will see a lot of you are using phones, tablets, laptops, but also cameras and even smartwatches. They're all different, but they have one underlying technology that is bringing it all together. Recognizing digital wardrobe dynamics is a core factor for us to put all the devices under one roof in IDG, one business group that is dedicated to end-user solutions across mobile, PC, but also software services and imaging, to emerging technologies like AR, VR, IoT, and ultimately a AI as well. A couple of years back there was a big debate around bring-your-own-device, what was called consumerization. Today consumerization does not exist anymore, because consumerization has happened into every single device we build in our commercial business. End users and commercial customers today do expect superior display performance, superior audio, microphone, voice, and touch quality, and have it all connected and working seamlessly together in an ease of use space. We are already deep in the journey of personalized computing today. But the center point of it has been for the last 25 years, the mobile PC, that we have perfected over the last 25 years, and has been the undisputed leader in mobility computing. We believe in the commercial business, the ThinkPad is still the core device of a digital wardrobe, and we continue to drive the success of the ThinkPad in the marketplace. We've sold more than 140 million over the last 26 years, and even last year we exceeded nearly 11 million units. That is about 21 ThinkPads per minute, or one Thinkpad every three seconds that we are shipping out in the market. It's the number one commercial PC in the world. It has gotten countless awards but we felt last year after Transform we need to build a step further, in really tailoring the ThinkPad towards the need of the future. So, we announced a new line of X1 Carbon and Yoga at CES the Consumer Electronics Show. And the reason is not we want to sell to consumer, but that we do recognize that a lot of CIOs and IT decision makers need to understand what consumers are really doing in terms of technology to make them successful. So, let's take a look at the video. (suspenseful music) >> When you're the number one business laptop of all time, your only competition is yourself. (wall shattering) And, that's different. Different, like resisting heat, ice, dust, and spills. Different, like sharper, brighter OLA display. The trackpoint that reinvented controls, and a carbon fiber roll cage to protect what's inside, built by an engineering and design team, doing the impossible for the last 25 years. This is the number one business laptop of all time, but it's not a laptop. It's a ThinkPad. (audience applauding) >> Thank you very much. And we are very proud that Lenovo ThinkPad has been selected as the best laptop in the world in the second year in a row. I think it's a wonderful tribute to what our engineers have been done on this one. And users do want awesome displays. They want the best possible audio, voice, and touch control, but some users they want more. What they want is super power, and I'm really proud to announce our newest member of the X1 family, and that's the X1 extreme. It's exceptionally featured. It has six core I9 intel chipset, the highest performance you get in the commercial space. It has Nvidia XTX graphic, it is a 4K UHD display with HDR with Dolby vision and Dolby Atmos Audio, two terabyte in SSD, so it is really the absolute Ferrari in terms of building high performance commercial computer. Of course it has touch and voice, but it is one thing. It has so much performance that it serves also a purpose that is not typical for commercial, and I know there's a lot of secret gamers also here in this room. So you see, by really bringing technology together in the commercial space, you're creating productivity solutions of one of a kind. But there's another category of products from a productivity perspective that is incredibly important in our commercial business, and that is the workstation business . Clearly workstations are very specifically designed computers for very advanced high-performance workloads, serving designers, architects, researchers, developers, or data analysts. And power and performance is not just about the performance itself. It has to be tailored towards the specific use case, and traditionally these products have a similar size, like a server. They are running on Intel Xeon technology, and they are equally complex to manufacture. We have now created a new category as the ultra mobile workstation, and I'm very proud that we can announce here the lightest mobile workstation in the industry. It is so powerful that it really can run AI and big data analysis. And with this performance you can go really close where you need this power, to the sensors, into the cars, or into the manufacturing places where you not only wannna read the sensors but get real-time analytics out of these sensors. To build a machine like this one you need customers who are really challenging you to the limit. and we're very happy that we had a customer who went on this journey with us, and ultimately jointly with us created this product. So, let's take a look at the video. (suspenseful music) >> My world involves pathfinding both the hardware needs to the various work sites throughout the company, and then finding an appropriate model of desktop, laptop, or workstation to match those needs. My first impressions when I first seen the ThinkPad P1 was I didn't actually believe that we could get everything that I was asked for inside something as small and light in comparison to other mobile workstations. That was one of the I can't believe this is real sort of moments for me. (engine roars) >> Well, it's better than general when you're going around in the wind tunnel, which isn't alway easy, and going on a track is not necessarily the best bet, so having a lightweight very powerful laptop is extremely useful. It can take a Xeon processor, which can support ECC from when we try to load a full car, and when we're analyzing live simulation results. through and RCFT post processor or example. It needs a pretty powerful machine. >> It's come a long way to be able to deliver this. I hate to use the word game changer, but it is that for us. >> Aston Martin has got a lot of different projects going. There's some pretty exciting projects and a pretty versatile range coming out. Having Lenovo as a partner is certainly going to ensure that future. (engine roars) (audience applauds) >> So, don't you think the Aston Martin design and the ThinkPad design fit very well together? (audience laughs) So if Q, would get a new laptop, I think you would get a ThinkPad X P1. So, I want to switch gears a little bit, and go into something in terms of productivity that is not necessarily on top of the mind or every end user but I believe it's on top of the mind of every C-level executive and of every CEO. Security is the number one threat in terms of potential risk in your business and the cost of cybersecurity is estimated by 2020 around six trillion dollars. That's more than the GDP of Japan and we've seen a significant amount of data breach incidents already this years. Now, they're threatening to take companies out of business and that are threatening companies to lose a huge amount of sensitive customer data or internal data. At Lenovo, we are taking security very, very seriously, and we run a very deep analysis, around our own security capabilities in the products that we are building. And we are announcing today a new brand under the Think umbrella that is called ThinkShield. Our goal is to build the world's most secure PC, and ultimately the most secure devices in the industry. And when we looked at this end-to-end, there is no silver bullet around security. You have to go through every aspect where security breaches can potentially happen. That is why we have changed the whole organization, how we look at security in our device business, and really have it grouped under one complete ecosystem of solutions, Security is always something where you constantly are getting challenged with the next potential breach the next potential technology flaw. As we keep innovating and as we keep integrating, a lot of our partners' software and hardware components into our products. So for us, it's really very important that we partner with companies like Intel, Microsoft, Coronet, Absolute, and many others to really as an example to drive full encryption on all the data seamlessly, to have multi-factor authentication to protect your users' identity, to protect you in unsecured Wi-Fi locations, or even simple things like innovation on the device itself, to and an example protect the camera, against usage with a little thing like a thinkShutter that you can shut off the camera. SO what I want to show you here, is this is the full portfolio of ThinkShield that we are announcing today. This is clearly not something I can even read to you today, but I believe it shows you the breadth of security management that we are announcing today. There are four key pillars in managing security end-to-end. The first one is your data, and this has a lot of aspects around the hardware and the software itself. The second is identity. The third is the security around online, and ultimately the device itself. So, there is a breakout on security and ThinkShield today, available in the afternoon, and encourage you to really take a deeper look at this one. The first pillar around productivity was the device, and around the device. The second major pillar that we are seeing in terms of intelligent transformation is the workspace itself. Employees of a new generation have a very different habit how they work. They split their time between travel, working remotely but if they do come in the office, they expect a very different office environment than what they've seen in the past in cubicles or small offices. They come into the office to collaborate, and they want to create ideas, and they really work in cross-functional teams, and they want to do it instantly. And what we've seen is there is a huge amount of investment that companies are doing today in reconfiguring real estate reconfiguring offices. And most of these kind of things are moving to a digital platform. And what we are doing, is we want to build an entire set of solutions that are just focused on making the workspace more productive for remote workforce, and to create technology that allow people to work anywhere and connect instantly. And the core of this is that we need to be, the productivity of the employee as high as possible, and make it for him as easy as possible to use these kind of technologies. Last year in Transform, I announced that we will enter the smart office space. By the end of last year, we brought the first product into the market. It's called the Hub 500. It's already deployed in thousands of our customers, and it's uniquely focused on Microsoft Skype for Business, and making meeting instantly happen. And the product is very successful in the market. What we are announcing today is the next generation of this product, what is the Hub 700, what has a fantastic audio quality. It has far few microphones, and it is usable in small office environment, as well as in major conference rooms, but the most important part of this new announcement is that we are also announcing a software platform, and this software platform allows you to run multiple video conferencing software solutions on the same platform. Many of you may have standardized for one software solution or for another one, but as you are moving in a world of collaborating instantly with partners, customers, suppliers, you always will face multiple software standards in your company, and Lenovo is uniquely positioned but providing a middleware platform for the device to really enable multiple of these UX interfaces. And there's more to come and we will add additional UX interfaces on an ongoing base, based on our customer requirements. But this software does not only help to create a better experience and a higher productivity in the conference room or the huddle room itself. It really will allow you ultimately to manage all your conference rooms in the company in one instance. And you can run AI technologies around how to increase productivity utilization of your entire conference room ecosystem in your company. You will see a lot more devices coming from the node in this space, around intelligent screens, cameras, and so on, and so on. The idea is really that Lenovo will become a core provider in the whole movement into the smart office space. But it's great if you have hardware and software that is really supporting the approach of modern IT, but one component that Kirk also mentioned is absolutely critical, that we are providing this to you in an as a service approach. Get it what you want, when you need it, and pay it in the amount that you're really using it. And within UIT there is also I think a new philosophy around IT management, where you're much more focused on the value that you are consuming instead of investing into technology. We are launched as a service two years back and we already have a significant number of customers running PC as a service, but we believe as a service will stretch far more than just the PC device. It will go into categories like smart office. It might go even into categories like phone, and it will definitely go also in categories like storage and server in terms of capacity management. I want to highlight three offerings that we are also displaying today that are sort of building blocks in terms of how we really run as a service. The first one is that we collaborated intensively over the last year with Microsoft to be the launch pilot for their Autopilot offering, basically deploying images easily in the same approach like you would deploy a new phone on the network. The purpose really is to make new imaging and enabling new PC as seamless as it's used to be in the phone industry, and we have a complete set of offerings, and already a significant number customers have deployed Autopilot with Lenovo. The second major offering is Premier Support, like in the in the server business, where Premier Support is absolutely critical to run critical infrastructure, we see a lot of our customers do want to have Premier Support for their end users, so they can be back into work basically instantly, and that you have the highest possible instant repair on every single device. And then finally we have a significant amount of time invested into understanding how the software as a service really can get into one philosophy. And many of you already are consuming software as a service in many different contracts from many different vendors, but what we've created is one platform that really can manage this all together. All these things are the foundation for a device as a service offering that really can manage this end-to-end. So, implementing an intelligent workplace can be really a daunting prospect depending on where you're starting from, and how big your company ultimately is. But how do you manage the transformation of technology workspace if you're present in 50 or more countries and you run an infrastructure for more than 100,000 people? Michelin, famous for their tires, infamous for their Michelin star restaurant rating, especially in New York, and instantly recognizable by the Michelin Man, has just doing that. Please welcome with me Damon McIntyre from Michelin to talk to us about the challenges and transforming collaboration and productivity. (audience applauding) (electronic dance music) Thank you, David. >> Thank you, thank you very much. >> We on? >> So, how do you feel here? >> Well good, I want to thank you first of all for your partnership and the devices you create that helped us design, manufacture, and distribute the best tire in the world, okay? I just had to say it and put out there, alright. And I was wondering, were those Michelin tires on that Aston Martin? >> I'm pretty sure there is no other tire that would fit to that. >> Yeah, no, thank you, thank you again, and thank you for the introduction. >> So, when we talk about the transformation happening really in the workplace, the most tangible transformation that you actually see is the drastic change that companies are doing physically. They're breaking down walls. They're removing cubes, and they're moving to flexible layouts, new desks, new huddle rooms, open spaces, but the underlying technology for that is clearly not so visible very often. So, tell us about Michelin's strategy, and the technology you are deploying to really enable this corporation. >> So we, so let me give a little bit a history about the company to understand the daunting tasks that we had before us. So we have over 114,000 people in the company under 170 nationalities, okay? If you go to the corporate office in France, it's Clermont. It's about 3,000 executives and directors, and what have you in the marketing, sales, all the way up to the chain of the global CIO, right? Inside of the Americas, we merged in Americas about three years ago. Now we have the Americas zone. There's about 28,000 employees across the Americas, so it's really, it's really hard in a lot of cases. You start looking at the different areas that you lose time, and you lose you know, your productivity and what have you, so there, it's when we looked at different aspects of how we were going to manage the meeting rooms, right? because we have opened up our areas of workspace, our CIO, CEOs in our zones will no longer have an office. They'll sit out in front of everybody else and mingle with the crowd. So, how do you take those spaces that were originally used by an individual but now turn them into like meeting rooms? So, we went through a large process, and looked at the Hub 500, and that really met our needs, because at the end of the day what we noticed was, it was it was just it just worked, okay? We've just added it to the catalog, so we're going to be deploying it very soon, and I just want to again point that I know everybody struggles with this, and if you look at all the minutes that you lose in starting up a meeting, and we know you know what I'm talking about when I say this, it equates to many many many dollars, okay? And so at the end the day, this product helps us to be more efficient in starting up the meeting, and more productive during the meeting. >> Okay, it's very good to hear. Another major trend we are seeing in IT departments is taking a more hands-off approach to hardware. We're seeing new technologies enable IT to create a more efficient model, how IT gets hardware in the hands of end-users, and how they are ultimately supporting themselves. So what's your strategy around the lifecycle management of the devices? >> So yeah you mentioned, again, we'll go back to the 114,000 employees in the company, right? You imagine looking at all the devices we use. I'm not going to get into the number of devices we have, but we have a set number that we use, and we have to go through a process of deploying these devices, which we right now service our own image. We build our images, we service them through our help desk and all that process, and we go through it. If you imagine deploying 25,000 PCs in a year, okay? The time and the daunting task that's behind all that, you can probably add up to 20 or 30 people just full-time doing that, okay? So, with partnering with Lenovo and their excellent technology, their technical teams, and putting together the whole process of how we do imaging, it now lifts that burden off of our folks, and it shifts it into a more automated process through the cloud, okay? And, it's with the Autopilot on the end of the project, we'll have Autopilot fully engaged, but what I really appreciate is how Lenovo really, really kind of got with us, and partnered with us for the whole process. I mean it wasn't just a partner between Michelin and Lenovo. Microsoft was also partnered during that whole process, and it really was a good project that we put together, and we hope to have something in a full production mode next year for sure. >> So, David thank you very, very much to be here with us on stage. What I really want to say, customers like you, who are always challenging us on every single aspect of our capabilities really do make the big difference for us to get better every single day and we really appreciate the partnership. >> Yeah, and I would like to say this is that I am, I'm doing what he's exactly said he just said. I am challenging Lenovo to show us how we can innovate in our work space with your devices, right? That's a challenge, and it's going to be starting up next year for sure. We've done some in the past, but I'm really going to challenge you, and my whole aspect about how to do that is bring you into our workspace. Show you how we make how we go through the process of making tires and all that process, and how we distribute those tires, so you can brainstorm, come back to the table and say, here's a device that can do exactly what you're doing right now, better, more efficient, and save money, so thank you. >> Thank you very much, David. (audience applauding) Well it's sometimes really refreshing to get a very challenging customers feedback. And you know, we will continue to grow this business together, and I'm very confident that your challenge will ultimately help to make our products even more seamless together. So, as we now covered productivity and how we are really improving our devices itself, and the transformation around the workplace, there is one pillar left I want to talk about, and that's really, how do we make businesses smarter than ever? What that really means is, that we are on a journey on trying to understand our customer's business, deeper than ever, understanding our customer's processes even better than ever, and trying to understand how we can help our customers to become more competitive by injecting state-of-the-art technology in this intelligent transformation process, into core processes. But this cannot be done without talking about a fundamental and that is the journey towards 5G. I really believe that 5G is changing everything the way we are operating devices today, because they will be connected in a way like it has never done before. YY talked about you know, 20 times 10 times the amount of performance. There are other studies that talk about even 200 times the performance, how you can use these devices. What it will lead to ultimately is that we will build devices that will be always connected to the cloud. And, we are preparing for this, and Kirk already talked about, and how many operators in the world we already present with our Moto phones, with how many Telcos we are working already on the backend, and we are working on the device side on integrating 5G basically into every single one of our product in the future. One of the areas that will benefit hugely from always connected is the world of virtual reality and augmented reality. And I'm going to pick here one example, and that is that we have created a commercial VR solution for classrooms and education, and basically using consumer type of product like our Mirage Solo with Daydream and put a solution around this one that enables teachers and schools to use these products in the classroom experience. So, students now can have immersive learning. They can studying sciences. They can look at environmental issues. They can exploring their careers, or they can even taking a tour in the next college they're going to go after this one. And no matter what grade level, this is how people will continue to learn in the future. It's quite a departure from the old world of textbooks. In our area that we are looking is IoT, And as YY already elaborated, we are clearly learning from our own processes around how we improve our supply chain and manufacturing and how we improve also retail experience and warehousing, and we are working with some of the largest companies in the world on pilots, on deploying IoT solutions to make their businesses, their processes, and their businesses, you know, more competitive, and some of them you can see in the demo environment. Lenovo itself already is managing 55 million devices in an IoT fashion connecting to our own cloud, and constantly improving the experience by learning from the behavior of these devices in an IoT way, and we are collecting significant amount of data to really improve the performance of these systems and our future generations of products on a ongoing base. We have a very strong partnership with a company called ADLINK from Taiwan that is one of the leading manufacturers of manufacturing PC and hardened devices to create solutions on the IoT platform. The next area that we are very actively investing in is commercial augmented reality. I believe augmented reality has by far more opportunity in commercial than virtual reality, because it has the potential to ultimately improve every single business process of commercial customers. Imagine in the future how complex surgeries can be simplified by basically having real-time augmented reality information about the surgery, by having people connecting into a virtual surgery, and supporting the surgery around the world. Visit a furniture store in the future and see how this furniture looks in your home instantly. Doing some maintenance on some devices yourself by just calling the company and getting an online manual into an augmented reality device. Lenovo is exploring all kinds of possibilities, and you will see a solution very soon from Lenovo. Early when we talked about smart office, I talked about the importance of creating a software platform that really run all these use cases for a smart office. We are creating a similar platform for augmented reality where companies can develop and run all their argumented reality use cases. So you will see that early in 2019 we will announce an augmented reality device, as well as an augmented reality platform. So, I know you're very interested on what exactly we are rolling out, so we will have a first prototype view available there. It's still a codename project on the horizon, and we will announce it ultimately in 2019, but I think it's good for you to take a look what we are doing here. So, I just wanted to give you a peek on what we are working beyond smart office and the device productivity in terms of really how we make businesses smarter. It's really about increasing productivity, providing you the most secure solutions, increase workplace collaboration, increase IT efficiency, using new computing devices and software and services to make business smarter in the future. There's no other company that will enable to offer what we do in commercial. No company has the breadth of commercial devices, software solutions, and the same data center capabilities, and no other company can do more for your intelligent transformation than Lenovo. Thank you very much. (audience applauding) >> Thanks mate, give me that. I need that. Alright, ladies and gentlemen, we are done. So firstly, I've got a couple of little housekeeping pieces at the end of this and then we can go straight into going and experiencing some of the technology we've got on the left-hand side of the room here. So, I want to thank Christian obviously. Christian, awesome as always, some great announcements there. I love the P1. I actually like the Aston Martin a little bit better, but I'll take either if you want to give me one for free. I'll take it. We heard from YY obviously about the industry and how the the fourth Industrial Revolution is impacting us all from a digital transformation perspective, and obviously Kirk on DCG, the great NetApp announcement, which is going to be really exciting, actually that Twitter and some of the social media panels are absolutely going crazy, so it's good to see that the industry is really taking some impact. Some of the publications are really great, so thank you for the media who are obviously in the room publishing right no. But now, I really want to say it's all of your turn. So, all of you up the back there who are having coffee, it's your turn now. I want everyone who's sitting down here after this event move into there, and really take advantage of the 15 breakouts that we've got set there. There are four breakout sessions from a time perspective. I want to try and get you all out there at least to use up three of them and use your fourth one to get out and actually experience some of the technology. So, you've got four breakout sessions. A lot of the breakout sessions are actually done twice. If you have not downloaded the app, please download the app so you can actually see what time things are going on and make sure you're registering correctly. There's a lot of great experience of stuff out there for you to go do. I've got one quick video to show you on some of the technology we've got and then we're about to close. Alright, here we are acting crazy. Now, you can see obviously, artificial intelligence machine learning in the browser. God, I hate that dance, I'm not a Millenial at all. It's effectively going to be implemented by healthcare. I want you to come around and test that out. Look at these two guys. This looks like a Lenovo management meeting to be honest with you. These two guys are actually concentrating, using their brain power to race each others in cars. You got to come past and give that a try. Give that a try obviously. Fantastic event here, lots of technology for you to experience, and great partners that have been involved as well. And so, from a Lenovo perspective, we've had some great alliance partners contribute, including obviously our number one partner, Intel, who's been a really big loyal contributor to us, and been a real part of our success here at Transform. Excellent, so please, you've just seen a little bit of tech out there that you can go and play with. I really want you, I mean go put on those black things, like Scott Hawkins our chief marketing officer from Lenovo's DCG business was doing and racing around this little car with his concentration not using his hands. He said it's really good actually, but as soon as someone comes up to speak to him, his car stops, so you got to try and do better. You got to try and prove if you can multitask or not. Get up there and concentrate and talk at the same time. 62 different breakouts up there. I'm not going to go into too much detai, but you can see we've got a very, very unusual numbering system, 18 to 18.8. I think over here we've got a 4849. There's a 4114. And then up here we've got a 46.1 and a 46.2. So, you need the decoder ring to be able to understand it. Get over there have a lot of fun. Remember the boat leaves today at 4:00 o'clock, right behind us at the pier right behind us here. There's 400 of us registered. Go onto the app and let us know if there's more people coming. It's going to be a great event out there on the Hudson River. Ladies and gentlemen that is the end of your keynote. I want to thank you all for being patient and thank all of our speakers today. Have a great have a great day, thank you very much. (audience applauding) (upbeat music) ♪ Ba da bop bop bop ♪ ♪ Ba da bop bop bop ♪ ♪ Ba da bop bop bop ♪ ♪ Ba da bop bop bop ♪ ♪ Ba da bop bop bop ♪ ♪ Ba da bop bop bop ♪ ♪ Ba da bop bop bop ba do ♪

Published Date : Sep 13 2018

SUMMARY :

and those around you, Ladies and gentlemen, we ask that you please take an available seat. Ladies and gentlemen, once again we ask and software that transform the way you collaborate, Good morning everyone! Ooh, that was pretty good actually, and have a look at all of the breakout sessions. and the industries demand to be more intelligent, and the strategies that we have going forward I'm going to give you the stage and allow you to say is that the first products are orderable and being one of the largest device companies in the world. and exactly what's going on with that. I think I'll need that. Okay, Christian, so obviously just before we get down, You're in Munich? and it's a great place to live and raise kids, And I miss it a lot, but I still believe the best sushi in the world and I have had sushi here, it's been fantastic. (Christian laughing) the real Oktoberfest in Munich, in relation to Oktoberfest, at the Lower East Side in Avenue C at Zum Schneider, and consequently ended up with you. and is reconfiguring it based on the work he's doing and a carbon fiber roll cage to protect what's inside, and that is the workstation business . and then finding an appropriate model of desktop, in the wind tunnel, which isn't alway easy, I hate to use the word game changer, is certainly going to ensure that future. And the core of this is that we need to be, and distribute the best tire in the world, okay? that would fit to that. and thank you for the introduction. and the technology you are deploying and more productive during the meeting. how IT gets hardware in the hands of end-users, You imagine looking at all the devices we use. and we really appreciate the partnership. and it's going to be starting up next year for sure. and how many operators in the world Ladies and gentlemen that is the end of your keynote.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

GeorgePERSON

0.99+

DellORGANIZATION

0.99+

KirkPERSON

0.99+

LenovoORGANIZATION

0.99+

BradPERSON

0.99+

AmazonORGANIZATION

0.99+

EMCORGANIZATION

0.99+

George KurianPERSON

0.99+

MichelinORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

NikeORGANIZATION

0.99+

WalmartORGANIZATION

0.99+

QualcommORGANIZATION

0.99+

DisneyORGANIZATION

0.99+

CaliforniaLOCATION

0.99+

IBMORGANIZATION

0.99+

HPORGANIZATION

0.99+

FranceLOCATION

0.99+

JapanLOCATION

0.99+

CanadaLOCATION

0.99+

ChinaLOCATION

0.99+

NutanixORGANIZATION

0.99+

AmericasLOCATION

0.99+

Christian TeismannPERSON

0.99+

New YorkLOCATION

0.99+

Kirk SkaugenPERSON

0.99+

MalaysiaLOCATION

0.99+

AMEXORGANIZATION

0.99+

NvidiaORGANIZATION

0.99+

Rod LappenPERSON

0.99+

University College LondonORGANIZATION

0.99+

BrazilLOCATION

0.99+

KurtPERSON

0.99+

2016DATE

0.99+

GermanyLOCATION

0.99+

17QUANTITY

0.99+

2019DATE

0.99+

AMDORGANIZATION

0.99+

VerizonORGANIZATION

0.99+

IndiaLOCATION

0.99+

sevenQUANTITY

0.99+

Hudson RiverLOCATION

0.99+

twoQUANTITY

0.99+

10xQUANTITY

0.99+

NetAppORGANIZATION

0.99+

MotorolaORGANIZATION

0.99+

USLOCATION

0.99+

South AfricaLOCATION

0.99+

Fireside Chat with Andy Jassy, AWS CEO, at the AWS Summit SF 2017


 

>> Announcer: Please welcome Vice President of Worldwide Marketing, Amazon Web Services, Ariel Kelman. (applause) (techno music) >> Good afternoon, everyone. Thank you for coming. I hope you guys are having a great day here. It is my pleasure to introduce to come up on stage here, the CEO of Amazon Web Services, Andy Jassy. (applause) (techno music) >> Okay. Let's get started. I have a bunch of questions here for you, Andy. >> Just like one of our meetings, Ariel. >> Just like one of our meetings. So, I thought I'd start with a little bit of a state of the state on AWS. Can you give us your quick take? >> Yeah, well, first of all, thank you, everyone, for being here. We really appreciate it. We know how busy you guys are. So, hope you're having a good day. You know, the business is growing really quickly. In the last financials, we released, in Q four of '16, AWS is a 14 billion dollar revenue run rate business, growing 47% year over year. We have millions of active customers, and we consider an active customer as a non-Amazon entity that's used the platform in the last 30 days. And it's really a very broad, diverse customer set, in every imaginable size of customer and every imaginable vertical business segment. And I won't repeat all the customers that I know Werner went through earlier in the keynote, but here are just some of the more recent ones that you've seen, you know NELL is moving their their digital and their connected devices, meters, real estate to AWS. McDonalds is re-inventing their digital platform on top of AWS. FINRA is moving all in to AWS, yeah. You see at Reinvent, Workday announced AWS was its preferred cloud provider, and to start building on top of AWS further. Today, in press releases, you saw both Dunkin Donuts and Here, the geo-spatial map company announced they'd chosen AWS as their provider. You know and then I think if you look at our business, we have a really large non-US or global customer base and business that continues to expand very dramatically. And we're also aggressively increasing the number of geographic regions in which we have infrastructure. So last year in 2016, on top of the broad footprint we had, we added Korea, India, and Canada, and the UK. We've announced that we have regions coming, another one in China, in Ningxia, as well as in France, as well as in Sweden. So we're not close to being done expanding geographically. And then of course, we continue to iterate and innovate really quickly on behalf of all of you, of our customers. I mean, just last year alone, we launched what we considered over 1,000 significant services and features. So on average, our customers wake up every day and have three new capabilities they can choose to use or not use, but at their disposal. You've seen it already this year, if you look at Chime, which is our new unified communication service. It makes meetings much easier to conduct, be productive with. You saw Connect, which is our new global call center routing service. If you look even today, you look at Redshift Spectrum, which makes it easy to query all your data, not just locally on disk in your data warehouse but across all of S3, or DAX, which puts a cash in front of DynamoDB, we use the same interface, or all the new features in our machine learning services. We're not close to being done delivering and iterating on your behalf. And I think if you look at that collection of things, it's part of why, as Gartner looks out at the infrastructure space, they estimate the AWS is several times the size business of the next 14 providers combined. It's a pretty significant market segment leadership position. >> You talked a lot about adopts in there, a lot of customers moving to AWS, migrating large numbers of workloads, some going all in on AWS. And with that as kind of backdrop, do you still see a role for hybrid as being something that's important for customers? >> Yeah, it's funny. The quick answer is yes. I think the, you know, if you think about a few years ago, a lot of the rage was this debate about private cloud versus what people call public cloud. And we don't really see that debate very often anymore. I think relatively few companies have had success with private clouds, and most are pretty substantially moving in the direction of building on top of clouds like AWS. But, while you increasingly see more and more companies every month announcing that they're going all in to the cloud, we will see most enterprises operate in some form of hybrid mode for the next number of years. And I think in the early days of AWS and the cloud, I think people got confused about this, where they thought that they had to make this binary decision to either be all in on the public cloud and AWS or not at all. And of course that's not the case. It's not a binary decision. And what we know many of our enterprise customers want is they want to be able to run the data centers that they're not ready to retire yet as seamlessly as they can alongside of AWS. And it's why we've built a lot of the capabilities we've built the last several years. These are things like PPC, which is our virtual private cloud, which allows you to cordon off a portion of our network, deploy resources into it and connect to it through VPN or Direct Connect, which is a private connection between your data centers and our regions or our storage gateway, which is a virtual storage appliance, or Identity Federation, or a whole bunch of capabilities like that. But what we've seen, even though the vast majority of the big hybrid implementations today are built on top of AWS, as more and more of the mainstream enterprises are now at the point where they're really building substantial cloud adoption plans, they've come back to us and they've said, well, you know, actually you guys have made us make kind of a binary decision. And that's because the vast majority of the world is virtualized on top of VMWare. And because VMWare and AWS, prior to a few months ago, had really done nothing to try and make it easy to use the VMWare tools that people have been using for many years seamlessly with AWS, customers were having to make a binary choice. Either they stick with the VMWare tools they've used for a while but have a really tough time integrating with AWS, or they move to AWS and they have to leave behind the VMWare tools they've been using. And it really was the impetus for VMWare and AWS to have a number of deep conversations about it, which led to the announcement we made late last fall of VMWare and AWS, which is going to allow customers who have been using the VMWare tools to manage their infrastructure for a long time to seamlessly be able to run those on top of AWS. And they get to do so as they move workloads back and forth and they evolve their hybrid implementation without having to buy any new hardware, which is a big deal for companies. Very few companies are looking to find ways to buy more hardware these days. And customers have been very excited about this prospect. We've announced that it's going to be ready in the middle of this year. You see companies like Amadeus and Merck and Western Digital and the state of Louisiana, a number of others, we've a very large, private beta and preview happening right now. And people are pretty excited about that prospect. So we will allow customers to run in the mode that they want to run, and I think you'll see a huge transition over the next five to 10 years. >> So in addition to hybrid, another question we get a lot from enterprises around the concept of lock-in and how they should think about their relationship with the vendor and how they should think about whether to spread the workloads across multiple infrastructure providers. How do you think about that? >> Well, it's a question we get a lot. And Oracle has sure made people care about that issue. You know, I think people are very sensitive about being locked in, given the experience that they've had over the last 10 to 15 years. And I think the reality is when you look at the cloud, it really is nothing like being locked into something like Oracle. The APIs look pretty similar between the various providers. We build an open standard, it's like Linux and MySQL and Postgres. All the migration tools that we build allow you to migrate in or out of AWS. It's up to customers based on how they want to run their workload. So it is much easier to move away from something like the cloud than it is from some of the old software services that has created some of this phobia. But I think when you look at most CIOs, enterprise CIOs particularly, as they think about moving to the cloud, many of them started off thinking that they, you know, very well might split their workloads across multiple cloud providers. And I think when push comes to shove, very few decide to do so. Most predominately pick an infrastructure provider to run their workloads. And the reason that they don't split it across, you know, pretty evenly across clouds is a few reasons. Number one, if you do so, you have to standardize in the lowest common denominator. And these platforms are in radically different stages at this point. And if you look at something like AWS, it has a lot more functionality than anybody else by a large margin. And we're also iterating more quickly than you'll find from the other providers. And most folks don't want to tie the hands of their developers behind their backs in the name of having the ability of splitting it across multiple clouds, cause they actually are, in most of their spaces, competitive, and they have a lot of ideas that they want to actually build and invent on behalf of their customers. So, you know, they don't want to actually limit their functionality. It turns out the second reason is that they don't want to force their development teams to have to learn multiple platforms. And most development teams, if any of you have managed multiple stacks across different technologies, and many of us have had that experience, it's a pain in the butt. And trying to make a shift from what you've been doing for the last 30 years on premises to the cloud is hard enough. But then forcing teams to have to get good at running across two or three platforms is something most teams don't relish, and it's wasteful of people's time, it's wasteful of natural resources. That's the second thing. And then the third reason is that you effectively diminish your buying power because all of these cloud providers have volume discounts, and then you're splitting what you buy across multiple providers, which gives you a lower amount you buy from everybody at a worse price. So when most CIOs and enterprises look at this carefully, they don't actually end up splitting it relatively evenly. They predominately pick a cloud provider. Some will just pick one. Others will pick one and then do a little bit with a second, just so they know they can run with a second provider, in case that relationship with the one they choose to predominately run with goes sideways in some fashion. But when you really look at it, CIOs are not making that decision to split it up relatively evenly because it makes their development teams much less capable and much less agile. >> Okay, let's shift gears a little bit, talk about a subject that's on the minds of not just enterprises but startups and government organizations and pretty much every organization we talk to. And that's AI and machine learning. Reinvent, we introduced our Amazon AI services and just this morning Werner announced the general availability of Amazon Lex. So where are we overall on machine learning? >> Well it's a hugely exciting opportunity for customers, and I think, we believe it's exciting for us as well. And it's still in the relatively early stages, if you look at how people are using it, but it's something that we passionately believe is going to make a huge difference in the world and a huge difference with customers, and that we're investing a pretty gigantic amount of resource and capability for our customers. And I think the way that we think about, at a high level, the machine learning and deep learning spaces are, you know, there's kind of three macro layers of the stack. I think at that bottom layer, it's generally for the expert machine learning practitioners, of which there are relatively few in the world. It's a scarce resource relative to what I think will be the case in five, 10 years from now. And these are folks who are comfortable working with deep learning engines, know how to build models, know how to tune those models, know how to do inference, know how to get that data from the models into production apps. And for that group of people, if you look at the vast majority of machine learning and deep learning that's being done in the cloud today, it's being done on top of AWS, are P2 instances, which are optimized for deep learning and our deep learning AMIs, that package, effectively the deep learning engines and libraries inside those AMIs. And you see companies like Netflix, Nvidia, and Pinterest and Stanford and a whole bunch of others that are doing significant amounts of machine learning on top of those optimized instances for machine learning and the deep learning AMIs. And I think that you can expect, over time, that we'll continue to build additional capabilities and tools for those expert practitioners. I think we will support and do support every single one of the deep learning engines on top of AWS, and we have a significant amount of those workloads with all those engines running on top of AWS today. We also are making, I would say, a disproportionate investment of our own resources and the MXNet community just because if you look at running deep learning models once you get beyond a few GPUs, it's pretty difficult to have those scale as you get into the hundreds of GPUs. And most of the deep learning engines don't scale very well horizontally. And so what we've found through a lot of extensive testing, cause remember, Amazon has thousands of deep learning experts inside the company that have built very sophisticated deep learning capabilities, like the ones you see in Alexa, we have found that MXNet scales the best and almost linearly, as we continue to add nodes, as we continue to horizontally scale. So we have a lot of investment at that bottom layer of the stack. Now, if you think about most companies with developers, it's still largely inaccessible to them to do the type of machine learning and deep learning that they'd really like to do. And that's because the tools, I think, are still too primitive. And there's a number of services out there, we built one ourselves in Amazon Machine Learning that we have a lot of customers use, and yet I would argue that all of those services, including our own, are still more difficult than they should be for everyday developers to be able to build machine learning and access machine learning and deep learning. And if you look at the history of what AWS has done, in every part of our business, and a lot of what's driven us, is trying to democratize technologies that were really only available and accessible before to a select, small number of companies. And so we're doing a lot of work at what I would call that middle layer of the stack to get rid of a lot of the muck associated with having to do, you know, building the models, tuning the models, doing the inference, figuring how to get the data into production apps, a lot of those capabilities at that middle layer that we think are really essential to allow deep learning and machine learning to reach its full potential. And then at the top layer of the stack, we think of those as solutions. And those are things like, pass me an image and I'll tell you what that image is, or show me this face, does it match faces in this group of faces, or pass me a string of text and I'll give you an mpg file, or give me some words and what your intent is and then I'll be able to return answers that allow people to build conversational apps like the Lex technology. And we have a whole bunch of other services coming in that area, atop of Lex and Polly and Recognition, and you can imagine some of those that we've had to use in Amazon over the years that we'll continue to make available for you, our customers. So very significant level of investment at all three layers of that stack. We think it's relatively early days in the space but have a lot of passion and excitement for that. >> Okay, now for ML and AI, we're seeing customers wanting to load in tons of data, both to train the models and to actually process data once they've built their models. And then outside of ML and AI, we're seeing just as much demand to move in data for analytics and traditional workloads. So as people are looking to move more and more data to the cloud, how are we thinking about making it easier to get data in? >> It's a great question. And I think it's actually an often overlooked question because a lot of what gets attention with customers is all the really interesting services that allow you to do everything from compute and storage and database and messaging and analytics and machine learning and AI. But at the end of the day, if you have a significant amount of data already somewhere else, you have to get it into the cloud to be able to take advantage of all these capabilities that you don't have on premises. And so we have spent a disproportionate amount of focus over the last few years trying to build capabilities for our customers to make this easier. And we have a set of capabilities that really is not close to matched anywhere else, in part because we have so many customers who are asking for help in this area that it's, you know, that's really what drives what we build. So of course, you could use the good old-fashioned wire to send data over the internet. Increasingly, we find customers that are trying to move large amounts of data into S3, is using our S3 transfer acceleration service, which basically uses our points of presence, or POPs, all over the world to expedite delivery into S3. You know, a few years ago, we were talking to a number of companies that were looking to make big shifts to the cloud, and they said, well, I need to move lots of data that just isn't viable for me to move it over the wire, given the connection we can assign to it. It's why we built Snowball. And so we launched Snowball a couple years ago, which is really, it's a 50 terabyte appliance that is encrypted, the data's encrypted three different ways, and you ingest the data from your data center into Snowball, it has a Kindle connected to it, it allows you to, you know, that makes sure that you send it to the right place, and you can also track the progress of your high-speed ingestion into our data centers. And when we first launched Snowball, we launched it at Reinvent a couple years ago, I could not believe that we were going to order as many Snowballs to start with as the team wanted to order. And in fact, I reproached the team and I said, this is way too much, why don't we first see if people actually use any of these Snowballs. And so the team thankfully didn't listen very carefully to that, and they really only pared back a little bit. And then it turned out that we, almost from the get-go, had ordered 10X too few. And so this has been something that people have used in a very broad, pervasive way all over the world. And last year, at the beginning of the year, as we were asking people what else they would like us to build in Snowball, customers told us a few things that were pretty interesting to us. First, one that wasn't that surprising was they said, well, it would be great if they were bigger, you know, if instead of 50 terabytes it was more data I could store on each device. Then they said, you know, one of the problems is when I load the data onto a Snowball and send it to you, I have to still keep my local copy on premises until it's ingested, cause I can't risk losing that data. So they said it would be great if you could find a way to provide clustering, so that I don't have to keep that copy on premises. That was pretty interesting. And then they said, you know, there's some of that data that I'd actually like to be loading synchronously to S3, and then, or some things back from S3 to that data that I may want to compare against. That was interesting, having that endpoint. And then they said, well, we'd really love it if there was some compute on those Snowballs so I can do analytics on some relatively short-term signals that I want to take action on right away. Those were really the pieces of feedback that informed Snowball Edge, which is the next version of Snowball that we launched, announced at Reinvent this past November. So it has, it's a hundred-terabyte appliance, still the same level of encryption, and it has clustering so that you don't have to keep that copy of the data local. It allows you to have an endpoint to S3 to synchronously load data back and forth, and then it has a compute inside of it. And so it allows customers to use these on premises. I'll give you a good example. GE is using these for their wind turbines. And they collect all kinds of data from those turbines, but there's certain short-term signals they want to do analytics on in as close to real time as they can, and take action on those. And so they use that compute to do the analytics and then when they fill up that Snowball Edge, they detach it and send it back to AWS to do broad-scale analytics in the cloud and then just start using an additional Snowball Edge to capture that short-term data and be able to do those analytics. So Snowball Edge is, you know, we just launched it a couple months ago, again, amazed at the type of response, how many customers are starting to deploy those all over the place. I think if you have exabytes of data that you need to move, it's not so easy. An exabyte of data, if you wanted to move from on premises to AWS, would require 10,000 Snowball Edges. Those customers don't want to really manage a fleet of 10,000 Snowball Edges if they don't have to. And so, we tried to figure out how to solve that problem, and it's why we launched Snowmobile back at Reinvent in November, which effectively, it's a hundred-petabyte container on a 45-foot trailer that we will take a truck and bring out to your facility. It comes with its own power and its own network fiber that we plug in to your data center. And if you want to move an exabyte of data over a 10 gigabit per second connection, it would take you 26 years. But using 10 Snowmobiles, it would take you six months. So really different level of scale. And you'd be surprised how many companies have exabytes of data at this point that they want to move to the cloud to get all those analytics and machine learning capabilities running on top of them. Then for streaming data, as we have more and more companies that are doing real-time analytics of streaming data, we have Kinesis, where we built something called the Kinesis Firehose that makes it really simple to stream all your real-time data. We have a storage gateway for companies that want to keep certain data hot, locally, and then asynchronously be loading the rest of their data to AWS to be able to use in different formats, should they need it as backup or should they choose to make a transition. So it's a very broad set of storage capabilities. And then of course, if you've moved a lot of data into the cloud or into anything, you realize that one of the hardest parts that people often leave to the end is ETL. And so we have announced an ETL service called Glue, which we announced at Reinvent, which is going to make it much easier to move your data, be able to find your data and map your data to different locations and do ETL, which of course is hugely important as you're moving large amounts. >> So we've talked a lot about moving things to the cloud, moving applications, moving data. But let's shift gears a little bit and talk about something not on the cloud, connected devices. >> Yeah. >> Where do they fit in and how do you think about edge? >> Well, you know, I've been working on AWS since the start of AWS, and we've been in the market for a little over 11 years at this point. And we have encountered, as I'm sure all of you have, many buzzwords. And of all the buzzwords that everybody has talked about, I think I can make a pretty strong argument that the one that has delivered fastest on its promise has been IOT and connected devices. Just amazing to me how much is happening at the edge today and how fast that's changing with device manufacturers. And I think that if you look out 10 years from now, when you talk about hybrid, I think most companies, majority on premise piece of hybrid will not be servers, it will be connected devices. There are going to be billions of devices all over the place, in your home, in your office, in factories, in oil fields, in agricultural fields, on ships, in cars, in planes, everywhere. You're going to have these assets that sit at the edge that companies are going to want to be able to collect data on, do analytics on, and then take action. And if you think about it, most of these devices, by their very nature, have relatively little CPU and have relatively little disk, which makes the cloud disproportionately important for them to supplement them. It's why you see most of the big, successful IOT applications today are using AWS to supplement them. Illumina has hooked up their genome sequencing to AWS to do analytics, or you can look at Major League Baseball Statcast is an IOT application built on top of AWS, or John Deer has over 200,000 telematically enabled tractors that are collecting real-time planting conditions and information that they're doing analytics on and sending it back to farmers so they can figure out where and how to optimally plant. Tata Motors manages their truck fleet this way. Phillips has their smart lighting project. I mean, there're innumerable amounts of these IOT applications built on top of AWS where the cloud is supplementing the device's capability. But when you think about these becoming more mission-critical applications for companies, there are going to be certain functions and certain conditions by which they're not going to want to connect back to the cloud. They're not going to want to take the time for that round trip. They're not going to have connectivity in some cases to be able to make a round trip to the cloud. And what they really want is customers really want the same capabilities they have on AWS, with AWS IOT, but on the devices themselves. And if you've ever tried to develop on these embedded devices, it's not for mere mortals. It's pretty delicate and it's pretty scary and there's a lot of archaic protocols associated with it, pretty tough to do it all and to do it without taking down your application. And so what we did was we built something called Greengrass, and we announced it at Reinvent. And Greengrass is really like a software module that you can effectively have inside your device. And it allows developers to write lambda functions, it's got lambda inside of it, and it allows customers to write lambda functions, some of which they want to run in the cloud, some of which they want to run on the device itself through Greengrass. So they have a common programming model to build those functions, to take the signals they see and take the actions they want to take against that, which is really going to help, I think, across all these IOT devices to be able to be much more flexible and allow the devices and the analytics and the actions you take to be much smarter, more intelligent. It's also why we built Snowball Edge. Snowball Edge, if you think about it, is really a purpose-built Greengrass device. We have Greengrass, it's inside of the Snowball Edge, and you know, the GE wind turbine example is a good example of that. And so it's to us, I think it's the future of what the on-premises piece of hybrid's going to be. I think there're going to be billions of devices all over the place and people are going to want to interact with them with a common programming model like they use in AWS and the cloud, and we're continuing to invest very significantly to make that easier and easier for companies. >> We've talked about several feature directions. We talked about AI, machine learning, the edge. What are some of the other areas of investment that this group should care about? >> Well there's a lot. (laughs) That's not a suit question, Ariel. But there's a lot. I think, I'll name a few. I think first of all, as I alluded to earlier, we are not close to being done expanding geographically. I think virtually every tier-one country will have an AWS region over time. I think many of the emerging countries will as well. I think the database space is an area that is radically changing. It's happening at a faster pace than I think people sometimes realize. And I think it's good news for all of you. I think the database space over the last few decades has been a lonely place for customers. I think that they have felt particularly locked into companies that are expensive and proprietary and have high degrees of lock-in and aren't so customer-friendly. And I think customers are sick of it. And we have a relational database service that we launched many years ago and has many flavors that you can run. You can run MySQL, you can run Postgres, you can run MariaDB, you can run SQLServer, you can run Oracle. And what a lot of our customers kept saying to us was, could you please figure out a way to have a database capability that has the performance characteristics of the commercial-grade databases but the customer-friendly and pricing model of the more open engines like the MySQL and Postgres and MariaDB. What you do on your own, we do a lot of it at Amazon, but it's hard, I mean, it takes a lot of work and a lot of tuning. And our customers really wanted us to solve that problem for them. And it's why we spent several years building Aurora, which is our own database engine that we built, but that's fully compatible with MySQL and with Postgres. It's at least as fault tolerant and durable and performant as the commercial-grade databases, but it's a tenth of the cost of those. And it's also nice because if it turns out that you use Aurora and you decide for whatever reason you don't want to use Aurora anymore, because it's fully compatible with MySQL and Postgres, you just dump it to the community versions of those, and off you are. So there's really hardly any transition there. So that is the fastest-growing service in the history of AWS. I'm amazed at how quickly it's grown. I think you may have heard earlier, we've had 23,000 database migrations just in the last year or so. There's a lot of pent-up demand to have database freedom. And we're here to help you have it. You know, I think on the analytic side, it's just never been easier and less expensive to collect, store, analyze, and share data than it is today. Part of that has to do with the economics of the cloud. But a lot of it has to do with the really broad analytics capability that we provide you. And it's a much broader capability than you'll find elsewhere. And you know, you can manage Hadoop and Spark and Presto and Hive and Pig and Yarn on top of AWS, or we have a managed elastic search service, and you know, of course we have a very high scale, very high performing data warehouse in Redshift, that just got even more performant with Spectrum, which now can query across all of your S3 data, and of course you have Athena, where you can query S3 directly. We have a service that allows you to do real-time analytics of streaming data in Kinesis. We have a business intelligence service in QuickSight. We have a number of machine learning capabilities I talked about earlier. It's a very broad array. And what we find is that it's a new day in analytics for companies. A lot of the data that companies felt like they had to throw away before, either because it was too expensive to hold or they didn't really have the tools accessible to them to get the learning from that data, it's a totally different day today. And so we have a pretty big investment in that space, I mentioned Glue earlier to do ETL on all that data. We have a lot more coming in that space. I think compute, super interesting, you know, I think you will find, I think we will find that companies will use full instances for many, many years and we have, you know, more than double the number of instances than you'll find elsewhere in every imaginable shape and size. But I would also say that the trend we see is that more and more companies are using smaller units of compute, and it's why you see containers becoming so popular. We have a really big business in ECS. And we will continue to build out the capability there. We have companies really running virtually every type of container and orchestration and management service on top of AWS at this point. And then of course, a couple years ago, we pioneered the event-driven serverless capability in compute that we call Lambda, which I'm just again, blown away by how many customers are using that for everything, in every way. So I think the basic unit of compute is continuing to get smaller. I think that's really good for customers. I think the ability to be serverless is a very exciting proposition that we're continuing to to fulfill that vision that we laid out a couple years ago. And then, probably, the last thing I'd point out right now is, I think it's really interesting to see how the basic procurement of software is changing. In significant part driven by what we've been doing with our Marketplace. If you think about it, in the old world, if you were a company that was buying software, you'd have to go find bunch of the companies that you should consider, you'd have to have a lot of conversations, you'd have to talk to a lot of salespeople. Those companies, by the way, have to have a big sales team, an expensive marketing budget to go find those companies and then go sell those companies and then both companies engage in this long tap-dance around doing an agreement and the legal terms and the legal teams and it's just, the process is very arduous. Then after you buy it, you have to figure out how you're going to actually package it, how you're deploy to infrastructure and get it done, and it's just, I think in general, both consumers of software and sellers of software really don't like the process that's existed over the last few decades. And then you look at AWS Marketplace, and we have 35 hundred product listings in there from 12 hundred technology providers. If you look at the number of hours, that software that's been running EC2 just in the last month alone, it's several hundred million hours, EC2 hours, of that software being run on top of our Marketplace. And it's just completely changing how software is bought and procured. I think that if you talk to a lot of the big sellers of software, like Splunk or Trend Micro, there's a whole number of them, they'll tell you it totally changes their ability to be able to sell. You know, one of the things that really helped AWS in the early days and still continues to help us, is that we have a self-service model where we don't actually have to have a lot of people talk to every customer to get started. I think if you're a seller of software, that's very appealing, to allow people to find your software and be able to buy it. And if you're a consumer, to be able to buy it quickly, again, without the hassle of all those conversations and the overhead associated with that, very appealing. And I think it's why the marketplace has just exploded and taken off like it has. It's also really good, by the way, for systems integrators, who are often packaging things on top of that software to their clients. This makes it much easier to build kind of smaller catalogs of software products for their customers. I think when you layer on top of that the capabilities that we've announced to make it easier for SASS providers to meter and to do billing and to do identity is just, it's a very different world. And so I think that also is very exciting, both for companies and customers as well as software providers. >> We certainly touched on a lot here. And we have a lot going on, and you know, while we have customers asking us a lot about how they can use all these new services and new features, we also tend to get a lot of questions from customers on how we innovate so quickly, and they can think about applying some of those lessons learned to their own businesses. >> So you're asking how we're able to innovate quickly? >> Mmm hmm. >> I think there's a few things that have helped us, and it's different for every company. But some of these might be helpful. I'll point to a few. I think the first thing is, I think we disproportionately index on hiring builders. And we think of builders as people who are inventors, people who look at different customer experiences really critically, are honest about what's flawed about them, and then seek to reinvent them. And then people who understand that launch is the starting line and not the finish line. There's very little that any of us ever built that's a home run right out of the gate. And so most things that succeed take a lot of listening to customers and a lot of experimentation and a lot of iterating before you get to an equation that really works. So the first thing is who we hire. I think the second thing is how we organize. And we have, at Amazon, long tried to organize into as small and separable and autonomous teams as we can, that have all the resources in those teams to own their own destiny. And so for instance, the technologists and the product managers are part of the same team. And a lot of that is because we don't want the finger pointing that goes back and forth between the teams, and if they're on the same team, they focus all their energy on owning it together and understanding what customers need from them, spending a disproportionate amount of time with customers, and then they get to own their own roadmaps. One of the reasons we don't publish a 12 to 18 month roadmap is we want those teams to have the freedom, in talking to customers and listening to what you tell us matters, to re-prioritize if there are certain things that we assumed mattered more than it turns out it does. So, you know I think that the way that we organize is the second piece. I think a third piece is all of our teams get to use the same AWS building blocks that all of you get to use, which allow you to move much more quickly. And I think one of the least told stories about Amazon over the last five years, in part because people have gotten interested in AWS, is people have missed how fast our consumer business at Amazon has iterated. Look at the amount of invention in Amazon's consumer business. And they'll tell you that a big piece of that is their ability to use the AWS building blocks like they do. I think a fourth thing is many big companies, as they get larger, what starts to happen is what people call the institutional no, which is that leaders walk into meetings on new ideas looking to find ways to say no, and not because they're ill intended but just because they get more conservative or they have a lot on their plate or things are really managed very centrally, so it's hard to imagine adding more to what you're already doing. At Amazon, it's really the opposite, and in part because of the way we're organized in such a decoupled, decentralized fashion, and in part because it's just part of our DNA. When the leaders walk into a meeting, they are looking for ways to say yes. And we don't say yes to everything, we have a lot of proposals. But we say yes to a lot more than I think virtually any other company on the planet. And when we're having conversations with builders who are proposing new ideas, we're in a mode where we're trying to problem-solve with them to get to yes, which I think is really different. And then I think the last thing is that we have mechanisms inside the company that allow us to make fast decisions. And if you want a little bit more detail, you should read our founder and CEO Jeff Bezos's shareholder letter, which just was released. He talks about the fast decision-making that happens inside the company. It's really true. We make fast decisions and we're willing to fail. And you know, we sometimes talk about how we're working on several of our next biggest failures, and we hope that most of the things we're doing aren't going to fail, but we know, if you're going to push the envelope and if you're going to experiment at the rate that we're trying to experiment, to find more pillars that allow us to do more for customers and allow us to be more relevant, you are going to fail sometimes. And you have to accept that, and you have to have a way of evaluating people that recognizes the inputs, meaning the things that they actually delivered as opposed to the outputs, cause on new ventures, you don't know what the outputs are going to be, you don't know consumers or customers are going to respond to the new thing you're trying to build. So you have to be able to reward employees on the inputs, you have to have a way for them to continue to progress and grow in their career even if they work on something didn't work. And you have to have a way of thinking about, when things don't work, how do I take the technology that I built as part of that, that really actually does work, but I didn't get it right in the form factor, and use it for other things. And I think that when you think about a culture like Amazon, that disproportionately hires builders, organizes into these separable, autonomous teams, and allows them to use building blocks to move fast, and has a leadership team that's looking to say yes to ideas and is willing to fail, you end up finding not only do you do more inventing but you get the people at every level of the organization spending their free cycles thinking about new ideas because it actually pays to think of new ideas cause you get a shot to try it. And so that has really helped us and I think most of our customers who have made significant shifts to AWS and the cloud would argue that that's one of the big transformational things they've seen in their companies as well. >> Okay. I want to go a little bit deeper on the subject of culture. What are some of the things that are most unique about the AWS culture that companies should know about when they're looking to partner with us? >> Well, I think if you're making a decision on a predominant infrastructure provider, it's really important that you decide that the culture of the company you're going to partner with is a fit for yours. And you know, it's a super important decision that you don't want to have to redo multiple times cause it's wasted effort. And I think that, look, I've been at Amazon for almost 20 years at this point, so I have obviously drank the Kool Aid. But there are a few things that I think are truly unique about Amazon's culture. I'll talk about three of them. The first is I think that we are unusually customer-oriented. And I think a lot of companies talk about being customer-oriented, but few actually are. I think most of the big technology companies truthfully are competitor-focused. They kind of look at what competitors are doing and then they try to one-up one another. You have one or two of them that I would say are product-focused, where they say, hey, it's great, you Mr. and Mrs. Customer have ideas on a product, but leave that to the experts, and you know, you'll like the products we're going to build. And those strategies can be good ones and successful ones, they're just not ours. We are driven by what customers tell us matters to them. We don't build technology for technology's sake, we don't become, you know, smitten by any one technology. We're trying to solve real problems for our customers. 90% of what we build is driven by what you tell us matters. And the other 10% is listening to you, and even if you can't articulate exactly what you want, trying to read between the lines and invent on your behalf. So that's the first thing. Second thing is that we are pioneers. We really like to invent, as I was talking about earlier. And I think most big technology companies at this point have either lost their will or their DNA to invent. Most of them acquire it or fast follow. And again, that can be a successful strategy. It's just not ours. I think in this day and age, where we're going through as big a shift as we are in the cloud, which is the biggest technology shift in our lifetime, as dynamic as it is, being able to partner with a company that has the most functionality, it's iterating the fastest, has the most customers, has the largest ecosystem of partners, has SIs and ISPs, that has had a vision for how all these pieces fit together from the start, instead of trying to patch them together in a following act, you have a big advantage. I think that the third thing is that we're unusually long-term oriented. And I think that you won't ever see us show up at your door the last day of a quarter, the last day of a year, trying to harass you into doing some kind of deal with us, not to be heard from again for a couple years when we either audit you or try to re-up you for a deal. That's just not the way that we will ever operate. We are trying to build a business, a set of relationships, that will outlast all of us here. And I think something that always ties it together well is this trusted advisor capability that we have inside our support function, which is, you know, we look at dozens of programmatic ways that our customers are using the platform and reach out to you if you're doing something we think's suboptimal. And one of the things we do is if you're not fully utilizing resources, or hardly, or not using them at all, we'll reach out and say, hey, you should stop paying for this. And over the last couple of years, we've sent out a couple million of these notifications that have led to actual annualized savings for customers of 350 million dollars. So I ask you, how many of your technology partners reach out to you and say stop spending money with us? To the tune of 350 million dollars lost revenue per year. Not too many. And I think when we first started doing it, people though it was gimmicky, but if you understand what I just talked about with regard to our culture, it makes perfect sense. We don't want to make money from customers unless you're getting value. We want to reinvent an experience that we think has been broken for the prior few decades. And then we're trying to build a relationship with you that outlasts all of us, and we think the best way to do that is to provide value and do right by customers over a long period of time. >> Okay, keeping going on the culture subject, what about some of the quirky things about Amazon's culture that people might find interesting or useful? >> Well there are a lot of quirky parts to our culture. And I think any, you know lots of companies who have strong culture will argue they have quirky pieces but I think there's a few I might point to. You know, I think the first would be the first several years I was with the company, I guess the first six years or so I was at the company, like most companies, all the information that was presented was via PowerPoint. And we would find that it was a very inefficient way to consume information. You know, you were often shaded by the charisma of the presenter, sometimes you would overweight what the presenters said based on whether they were a good presenter. And vice versa. You would very rarely have a deep conversation, cause you have no room on PowerPoint slides to have any depth. You would interrupt the presenter constantly with questions that they hadn't really thought through cause they didn't think they were going to have to present that level of depth. You constantly have the, you know, you'd ask the question, oh, I'm going to get to that in five slides, you want to do that now or you want to do that in five slides, you know, it was just maddening. And we would often find that most of the meetings required multiple meetings. And so we made a decision as a company to effectively ban PowerPoints as a communication vehicle inside the company. Really the only time I do PowerPoints is at Reinvent. And maybe that shows. And what we found is that it's a much more substantive and effective and time-efficient way to have conversations because there is no way to fake depth in a six-page narrative. So what we went to from PowerPoint was six-page narrative. You can write, have as much as you want in the appendix, but you have to assume nobody will read the appendices. Everything you have to communicate has to be done in six pages. You can't fake depth in a six-page narrative. And so what we do is we all get to the room, we spend 20 minutes or so reading the document so it's fresh in everybody's head. And then where we start the conversation is a radically different spot than when you're hearing a presentation one kind of shallow slide at a time. We all start the conversation with a fair bit of depth on the topic, and we can really hone in on the three or four issues that typically matter in each of these conversations. So we get to the heart of the matter and we can have one meeting on the topic instead of three or four. So that has been really, I mean it's unusual and it takes some time getting used to but it is a much more effective way to pay attention to the detail and have a substantive conversation. You know, I think a second thing, if you look at our working backwards process, we don't write a lot of code for any of our services until we write and refine and decide we have crisp press release and frequently asked question, or FAQ, for that product. And in the press release, what we're trying to do is make sure that we're building a product that has benefits that will really matter. How many times have we all gotten to the end of products and by the time we get there, we kind of think about what we're launching and think, this is not that interesting. Like, people are not going to find this that compelling. And it's because you just haven't thought through and argued and debated and made sure that you drew the line in the right spot on a set of benefits that will really matter to customers. So that's why we use the press release. The FAQ is to really have the arguments up front about how you're building the product. So what technology are you using? What's the architecture? What's the customer experience? What's the UI look like? What's the pricing dimensions? Are you going to charge for it or not? All of those decisions, what are people going to be most excited about, what are people going to be most disappointed by. All those conversations, if you have them up front, even if it takes you a few times to go through it, you can just let the teams build, and you don't have to check in with them except on the dates. And so we find that if we take the time up front we not only get the products right more often but the teams also deliver much more quickly and with much less churn. And then the third thing I'd say that's kind of quirky is it is an unusually truth-seeking culture at Amazon. I think we have a leadership principle that we say have backbone, disagree, and commit. And what it means is that we really expect people to speak up if they believe that we're headed down a path that's wrong for customers, no matter who is advancing it, what level in the company, everybody is empowered and expected to speak up. And then once we have the debate, then we all have to pull the same way, even if it's a different way than you were advocating. And I think, you always hear the old adage of where, two people look at a ceiling and one person says it's 14 feet and the other person says, it's 10 feet, and they say, okay let's compromise, it's 12 feet. And of course, it's not 12 feet, there is an answer. And not all things that we all consider has that black and white answer, but most things have an answer that really is more right if you actually assess it and debate it. And so we have an environment that really empowers people to challenge one another and I think it's part of why we end up getting to better answers, cause we have that level of openness and rigor. >> Okay, well Andy, we have time for one more question. >> Okay. >> So other than some of the things you've talked about, like customer focus, innovation, and long-term orientation, what is the single most important lesson that you've learned that is really relevant to this audience and this time we're living in? >> There's a lot. But I'll pick one. I would say I'll tell a short story that I think captures it. In the early days at Amazon, our sole business was what we called an owned inventory retail business, which meant we bought the inventory from distributors or publishers or manufacturers, stored it in our own fulfillment centers and shipped it to customers. And around the year 1999 or 2000, this third party seller model started becoming very popular. You know, these were companies like Half.com and eBay and folks like that. And we had a really animated debate inside the company about whether we should allow third party sellers to sell on the Amazon site. And the concerns internally were, first of all, we just had this fundamental belief that other sellers weren't going to care as much about the customer experience as we did cause it was such a central part of everything we did DNA-wise. And then also we had this entire business and all this machinery that was built around owned inventory business, with all these relationships with publishers and distributors and manufacturers, who we didn't think would necessarily like third party sellers selling right alongside us having bought their products. And so we really debated this, and we ultimately decided that we were going to allow third party sellers to sell in our marketplace. And we made that decision in part because it was better for customers, it allowed them to have lower prices, so more price variety and better selection. But also in significant part because we realized you can't fight gravity. If something is going to happen, whether you want it to happen or not, it is going to happen. And you are much better off cannibalizing yourself or being ahead of whatever direction the world is headed than you are at howling at the wind or wishing it away or trying to put up blockers and find a way to delay moving to the model that is really most successful and has the most amount of benefits for the customers in question. And that turned out to be a really important lesson for Amazon as a company and for me, personally, as well. You know, in the early days of doing Marketplace, we had all kinds of folks, even after we made the decision, that despite the have backbone, disagree and commit weren't really sure that they believed that it was going to be a successful decision. And it took several months, but thankfully we really were vigilant about it, and today in roughly half of the units we sell in our retail business are third party seller units. Been really good for our customers. And really good for our business as well. And I think the same thing is really applicable to the space we're talking about today, to the cloud, as you think about this gigantic shift that's going on right now, moving to the cloud, which is, you know, I think in the early days of the cloud, the first, I'll call it six, seven, eight years, I think collectively we consumed so much energy with all these arguments about are people going to move to the cloud, what are they going to move to the cloud, will they move mission-critical applications to the cloud, will the enterprise adopt it, will public sector adopt it, what about private cloud, you know, we just consumed a huge amount of energy and it was, you can see both in the results in what's happening in businesses like ours, it was a form of fighting gravity. And today we don't really have if conversations anymore with our customers. They're all when and how and what order conversations. And I would say that this going to be a much better world for all of us, because we will be able to build in a much more cost effective fashion, we will be able to build much more quickly, we'll be able to take our scarce resource of engineers and not spend their resource on the undifferentiated heavy lifting of infrastructure and instead on what truly differentiates your business. And you'll have a global presence, so that you have lower latency and a better end user customer experience being deployed with your applications and infrastructure all over the world. And you'll be able to meet the data sovereignty requirements of various locales. So I think it's a great world that we're entering right now, I think we're at a time where there's a lot less confusion about where the world is headed, and I think it's an unprecedented opportunity for you to reinvent your businesses, reinvent your applications, and build capabilities for your customers and for your business that weren't easily possible before. And I hope you take advantage of it, and we'll be right here every step of the way to help you. Thank you very much. I appreciate it. (applause) >> Thank you, Andy. And thank you, everyone. I appreciate your time today. >> Thank you. (applause) (upbeat music)

Published Date : May 3 2017

SUMMARY :

of Worldwide Marketing, Amazon Web Services, Ariel Kelman. It is my pleasure to introduce to come up on stage here, I have a bunch of questions here for you, Andy. of a state of the state on AWS. And I think if you look at that collection of things, a lot of customers moving to AWS, And of course that's not the case. and how they should think about their relationship And I think the reality is when you look at the cloud, talk about a subject that's on the minds And I think that you can expect, over time, So as people are looking to move and it has clustering so that you don't and talk about something not on the cloud, And I think that if you look out 10 years from now, What are some of the other areas of investment and we have, you know, more than double and you know, while we have customers and listening to what you tell us matters, What are some of the things that are most unique And the other 10% is listening to you, And I think any, you know lots of companies moving to the cloud, which is, you know, And thank you, everyone. Thank you.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmadeusORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Western DigitalORGANIZATION

0.99+

AndyPERSON

0.99+

NvidiaORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

FranceLOCATION

0.99+

SwedenLOCATION

0.99+

NingxiaLOCATION

0.99+

ChinaLOCATION

0.99+

Andy JassyPERSON

0.99+

StanfordORGANIZATION

0.99+

six monthsQUANTITY

0.99+

Ariel KelmanPERSON

0.99+

Jeff BezosPERSON

0.99+

twoQUANTITY

0.99+

threeQUANTITY

0.99+

2000DATE

0.99+

OracleORGANIZATION

0.99+

12QUANTITY

0.99+

26 yearsQUANTITY

0.99+

20 minutesQUANTITY

0.99+

ArielPERSON

0.99+

two peopleQUANTITY

0.99+

10 feetQUANTITY

0.99+

six pagesQUANTITY

0.99+

90%QUANTITY

0.99+

GEORGANIZATION

0.99+

six-pageQUANTITY

0.99+

second pieceQUANTITY

0.99+

last yearDATE

0.99+

14 feetQUANTITY

0.99+

sixQUANTITY

0.99+

PowerPointTITLE

0.99+

47%QUANTITY

0.99+

50 terabytesQUANTITY

0.99+

Amazon Web ServicesORGANIZATION

0.99+

12 feetQUANTITY

0.99+

sevenQUANTITY

0.99+

five slidesQUANTITY

0.99+

TodayDATE

0.99+

fourQUANTITY

0.99+

oneQUANTITY

0.99+

10%QUANTITY

0.99+

2016DATE

0.99+

350 million dollarsQUANTITY

0.99+

10XQUANTITY

0.99+

NetflixORGANIZATION

0.99+

NovemberDATE

0.99+

USLOCATION

0.99+

second reasonQUANTITY

0.99+

McDonaldsORGANIZATION

0.99+

Jerry Chen, Greylock | AWS Re:Invent 2013


 

okay welcome back day two of the cube here and Las Vegas for live this is looking angles exclusive coverage of Amazon Web Services reinvent I'm John furrier with Dave vellante co-host of the cube Dave we got our first segment here we're pleased to have Jerry chin new venture capitalist cloud guru was at VMware it's been in the enterprise for a while guys welcome welcome to the cube Jay to kick off here at amazon reinvent Jerry welcome back decided Amy thanks for having guys cube alumni how was Hong Kong you just back from I'm stack I think Hong Kong was great my my body and time clocks someplace our Pacific though so I don't know them jet lag but thank God in Vegas I never need to leave the building so I don't need to know what time is on my mom actually in so it's good to be here so Amazon's pushing the cloud hard obviously they are the cloud huge market share on infrastructure as a service check the boxes there they got like thirty six percent by are not I think it's much higher than that actually her but jesse was saying today well I mean by vechs the next 14 it's got to be higher than thirty six percent I think it's closer to seven but ok that's infrastructure service but the actions platform as a service and SAS yeah if you can I got to get your take on guys we're following OpenStack you were just in Hong Kong you got amazon public cloud you get OpenStack coming up you know as that horse those a two-horse race right now clouds Dax out there but really it's OpenStack is like the enterprise hope it's the great hope for the enterprise with Amazon kind of rolling rolling out massive services what's your take on the two and and and is it a two-horse race and what's what's what's the what's the difference between the two you know I don't think it's a it's a two horse race yet but Amazon is quickly becoming the marker soph monopoly of the public cloud at the rate they're going and and it there have the size and scale that pretty soon to be really hard to compete and I think only google and maybe Marcus off and the public cloud space can really compete but if you take a step back and look at you know to your question OpenStack versus amazon I was in Hong Kong last week the OpenStack design summit and openstax philosophies one be all things to all people right it's open source multiple projects Amazon's philosophy is they want to be one cloud all people so you saw their announcements today around enterprise use cases desktop use cases startup use cases me to use cases there won't be one cloud to all people so it is not the race isn't over yet but very different philosophies right now between the two different cams was there much to talk about incorporating amazon api's into the whole OpenStack framework you know six months ago you heard a lot about that we had a crowd chatter on that run what was the the buzz there you know I I'll be honest into to the point that you guys brought up early around the Amazon ap is almost are becoming a lingua franca for infrastructure of a service but quite frankly debating whatnot they're the right api's or not isn't I think where the actions and the actions add to the point you made around pass and other developer services so the actual API so you do the api's right should be pretty easy for developers to adopt you just create really great developer service around it database services storage services security services those are what developers really care about so I feel like we have you know sometimes called cloud plus there are infrastructure service plus and you got sass minus you know it's like what you have with Salesforce do you feel like we really need that pass layer does that just sort of bifurcate into one of those two there's there's a there's a school of thought that says the world goes into two worlds a long telus a sax so there's an app for everything in which case you have SAS or SATA minus and then you know infrastructure private cloud for a budget likes the apps there's no middle ground for pass you know I'm more towards the middle ground because in a world where we have multiple SAS providers in multiple clouds I believe you're going to have multiple SAS multiple clouds you're going to need to integrate and stitch together a mash-up of applications right you have work day for HCM Salesforce for crm applications your own custom website running on amazon there are three different kinds now servers now how are you connect the data are going to move data around there's going to be at least some kind of past layer integration layer or cloud layer that needs to help stitch together this multi-cloud world so you like the pivotal play a pill I think the concept Indian concept right I think Paul is is a pulse of visionary and bus my friends to work there their announcement yes sir was was I think a step in the right direction that they're planning a flag saying that there has to be something beyond amazon there has to be a relevant private cloud initiative be it VMware or OpenStack of someplace else and let's create some services around it and the angle are taking around data and data services i think is proud of the right the right bed because all these new applications will need these data services to be relevant we were talking about pivotal yesterday one of the things that we were critical on and but also hopeful as you pointed out it's early right so true pivotal a mulligan or a pass if you will is this early and it's really a new company if you think about a 1,600 employees but new but it's window dressing announcement it really wasn't really i mean so the same logos i mean come on that they're trying to overhype and that's that was that's what people are talking about saying hey guys just be honest and say we're working as fast as i can because amazon is not going to break the enterprise right away I mean they also have a longer road going hard at the enterprise so they are going after IBM we must saw in the keynote that called out IBM specifically around some of the advertising there on the show yeah so Amazon is clearly trying to knock on the door or the enterprise so the question we are asking and talking about is how much time is it till they proliferate the enterprise I mean they're in there now toe in the water little beachhead still not enterprise-ready in the ends of the SLA s and the demands or does it matter so what's your take how much time is really on the radar for Amazon when will the clock be expiring for the IBM's HP pivotal's in terms of retooling so I think the evolution around enterprise public cloud like Amazon would take three potential paths so path one around amazon amazon invests enough engineering and product talent to make their cloud enterprise friendly privacy security reliability and they're they're hiring a bunch of folks a bunch of folks my old place vmware try to do that that's path one path to is you see a category of startups out there trying to meet amazon more cloud and enterprise friendly security privacy reliability right so that's path to and as a Greylock a venture capitalist we're investing a bunch of companies trying to you make that happen or past three is developers out there I'm engineer around the weaknesses amazon so the new Amazon is an enterprise friendly they know and about Amazon's got a bunch of weakness around security and privacy and he's just right there application around those weaknesses so I think those are the three evolutionary path paths I think it's a race to see who wins right one two or three yeah there's no doubt that Amazon is forcing the hand of the big guys he's seeing that clearly we have a question on our crowd check go to crowd chatting at / reinvent we've got a live live crowd-sourced thought leader chat there all those to Twitter and LinkedIn pendulum will you sign in but the question Jerry to you is how our cloud providers catering to provide low latency access to developing markets like India Indonesia Philippines etc you know given that the Hurricanes just destroyed all the infrastructure considering there's huge potential explosive internet growth so given that those new emerging markets are essentially refreshing their infrastructure what is the the cloud providers take on the end you do you work in that area what you're giving the opinion on what's going on in those areas sure I mean I think that the world is looking at two or three different clouds you say there's a u.s. dominated cloud maybe a China dominate cloud and rest of the world right generally a lot of analyst kind of segment the world in three major pockets when you think about developing markets or other geographies like Asia South Asia or South America huge markets lot of developers all applications it's the reason why I think there's only a handful of providers that can have the scoop in the reeds to reach globally I think Equinix Rackspace on Google Marcus off or all global footprint players everyone else I think you're going to look at a Federation of multiple players so every region has a local telco cloud provider it could be like an entity or rakuten in Japan it could be a sink tell in Singapore South East Asia so I think you're going to see a global brand around like Amazon or or VMware and VMware trying to franchise our own cloud or Microsoft and then I would see partnerships working between the different geographies and maybe OpenStack is that partnership maybe amazon API is the way different class communicate its remains to be seen what that interface between the different gos look like in the future what do you see as IBM's role I mean first of all do they have the global scale are you sort of purposefully leaving them out or just forget about them and just don't feel like they can compete on that global scale what do you see is their role in OpenStack so um bunch of questions there IBM didn't mean to leave them out there are definitely relevant especially for the large enterprises so I think you're seeing enterprise adoption come from large startups or small starts growing up in the cloud as well as large enterprises that are looking to modernize your applications and I think IBM has a great role to play from kind of that top-down approach I think IBM between a combination of a soft layers which is their their acquired cloud provider combined with their global services and their consulting business will be really relevant to large enterprises my mind so talk about the Amazon enterprise marchi obviously they're talking about cloud trails which is kind of like a monitoring service compliance oriented and I'll see vbi so you you've been close to the vdi movement so that's those are I started VDI hearted the beady eye movement so you know being there what is your take on that because that's very enterprising and that's rude good for business I'm what sir what's their chances there well I think so first on the vdi market we started that at VMware at 05 06 we coined the term VDI and I think it's a great service for large enterprises than need secure mass desktops I think I would love to see in a VDI service from VMware in Amazon five six seven years ago because now video i think is part of a larger solution it's it's it's significant but not enough right he's now enterprise to care about their madness desktops like VDI but my ipad devices iOS devices Android devices they really want kind of a holistically managed desktop or workspace environment so if i were amazon i would expand beyond windows and two other you know operating systems to manage like android and iOS but that's other serious about you know managing enterprise workspaces do they have do they have advantage and you're in your opinion despite the fact that they're so late to market do they have an advantage in that and I mean in essence they are starting around mobile developers aren't they whereas when you started that was especially a consideration Wright and Citrix sort of found its way there right but I think between um amazon I think Google's in a great position because they own so much of the Android stack right if they want to create an enterprise friendly manage um Android environment for Chromebooks Android devices they can start creating a bunch of great developer services like magic google drive but secured on on kind of a google cloud or something like that that could be pretty compelling I don't know if they're going there i think dropbox has a great opportunity kind of be that back and platform obviously Greylock investment but dropbox has a huge opportunity to be that kind of manage secure servers across mobile devices and desktop devices it's all a sudden the one overarching fact you have between Windows iOS and Android is your data and drop boxes on all three platforms chair we got to get rolling and we got in our next guest but I want to ask you actually talk about what you're investing in at greylock rate locked here 1dc you guys have done amazing deals I mean just recently in the past decade Greylock has emerged from just a tier 1 BC to a mega success good investments and if you're on the enterprise team they're actually the consumer side kick ass what's going on for you guys what are you investing in what are you looking at and if price is not an easy game to invest in obviously it's hard but what are you guys doing what are you investing what are you looking for I'm thinking about looking at across the categories most relevant for this audience is I'm really interested looking at startups that can either a make amazon a more enterprise funding cloud or be startups that will pose alternative or challenge to amazon in the enterprise cloud space and you do that either by you know focus on enterprise requirements or focus on enterprise services like data storage security that matter enterprises focus on doing that really really well better than vmware better than Microsoft there in the Amazon I think in the build a really big enterprise cloud business around those technology services you're essentially betting on that transformation from the way the world is the cloud is post of the world known as buying servers they're all trying to find a lab partner that's the direction and and are you bullish on this integrated stack offering obviously DevOps has been a big success you see Facebook you see Google you see Amazon building their own gear they were kind of saying we're not playing an open compute but sure that aside DevOps is a software model absolutely and so the integrated stack which are common on integrated stack and how that's going to involve for both the mainstream of DevOps absolutely so you see this DevOps culture permeating first development of applications now how you manage your infrastructure so you look at what's happened with open compute and open source switches which I think open compute project announced a couple days ago you're seeing that kind of DevOps culture and how they manage and update their applications / minate storage compute and now networking that's going to be kind of a common adoption curve throughout the cloud so the way DevOps technologies are getting adopted from languages to frameworks of databases is the same way we're seeing storage compute and networking technologies get adopted in this next cloud wave what's your take on the iphone for the enterprise amazon cloud kind of metaphor and OpenStack being more the Android we were talking earlier right just get your thoughts there an OpenStack also has a lot of legs right now but it's very open iPhone model or Amazon is kind of closed or some say lock in alright but it still apps are not closed right so the metaphor the metaphor was you know iphone is to Amazon as Android is to OpenStack and I think at a high level that kind of makes sense but not really because there's no Google behind OpenStack like there's a google behind Android so I think Rackspace is was an early leader and still as a leader in the OpenStack space but there's also red hat there's a bunch of the players there so as a result there's no single entity kind of driving OpenStack like Google's driving Android so that analogy can breaks down and then as far as Apple analogy to Amazon I I think Amazon is a lot more open than the iOS ecosystem is because just the fact that there's no governing board to prove her apps to launch on amazon right I can go stand up on an ec2 instance lost my application use it I don't need wait for this there's not a 20-page approval process so knowingly directionally that's more correct than not but it's analogy breaks down when you really get into it and OpenStack your prospects roman sec what's your what's your outlook on OpenStack real quick I think OpenStack so holistically i think is great a more bullets than sort of sub projects that i am overall I think they keep launching new projects some are better than others the core processing around compute and storage and this um API management I'm bullish on I'm supposed to be bullish on what they're doing around containers like docker and core OS and kind of adopting this next generation of cloud platforms well we got to go we got some fans out there want to hear what your take on VDI so go tweet to at jerry chen j ER are wide CH en we got a break here we'd love to have you on a little longer we got our next guest coming on it's the cube live in Las Vegas day two of Amazon's reinvent changing the cloud game and the enterprise and we get all the detailed coverage here on the key we'll be right back after this short break the cute

Published Date : Nov 13 2013

SUMMARY :

the question Jerry to you is how our

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SingaporeLOCATION

0.99+

JapanLOCATION

0.99+

twoQUANTITY

0.99+

AmazonORGANIZATION

0.99+

amazonORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

androidTITLE

0.99+

20-pageQUANTITY

0.99+

iOSTITLE

0.99+

VegasLOCATION

0.99+

Las VegasLOCATION

0.99+

IBMORGANIZATION

0.99+

Hong KongLOCATION

0.99+

thirty six percentQUANTITY

0.99+

threeQUANTITY

0.99+

iPhoneCOMMERCIAL_ITEM

0.99+

first segmentQUANTITY

0.99+

GreylockORGANIZATION

0.99+

Jerry ChenPERSON

0.99+

AndroidTITLE

0.99+

South AmericaLOCATION

0.99+

VMwareORGANIZATION

0.99+

AmyPERSON

0.99+

Dave vellantePERSON

0.99+

yesterdayDATE

0.99+

two-horseQUANTITY

0.99+

WindowsTITLE

0.99+

jessePERSON

0.99+

last weekDATE

0.99+

GoogleORGANIZATION

0.99+

FacebookORGANIZATION

0.99+

six months agoDATE

0.99+

Hong KongLOCATION

0.99+

iphoneCOMMERCIAL_ITEM

0.99+

ipadCOMMERCIAL_ITEM

0.99+

two horseQUANTITY

0.99+

todayDATE

0.98+

LinkedInORGANIZATION

0.98+

three major pocketsQUANTITY

0.98+

googleORGANIZATION

0.98+

DavePERSON

0.98+

JohnPERSON

0.98+

windowsTITLE

0.98+

two different camsQUANTITY

0.98+

Amazon Web ServicesORGANIZATION

0.98+

AWSORGANIZATION

0.98+

IndiaLOCATION

0.98+