SiliconANGLE News | Red Hat Collaborates with Nvidia, Samsung and Arm on Efficient, Open Networks
(upbeat music) >> Hello, everyone; I'm John Furrier with SiliconANGLE NEWS and host of theCUBE, and welcome to our SiliconANGLE NEWS MWC NEWS UPDATE in Barcelona where MWC is the premier event for the cloud telecommunication industry, and in the news here is Red Hat, Red Hat announcing a collaboration with NVIDIA, Samsung and Arm on Efficient Open Networks. Red Hat announced updates across various fields including advanced 5G telecommunications cloud, industrial edge, artificial intelligence, and radio access networks, RAN, and Efficiency. Red Hat's enterprise Kubernetes platform, OpenShift, has added support for NVIDIA's converged accelerators and aerial SDK facilitating RAND deployments on industry standard service across hybrid and multicloud platforms. This composable infrastructure enables telecom firms to support heavier compute demands for edge computing, AI, private 5G, and more, and just also helps network operators adopt open architectures, allowing them to choose non-proprietary components from multiple suppliers. In addition to the NVIDIA collaboration, Red Hat is working with Samsung to offer a new vRAN solution for service providers to better manage their open RAN networks. They're also working with UK chip designer, Arm, to create new networking solutions for energy efficient Red Hat Open Source Kubernetes-based Efficient Power Level Exporter project, or Kepler, has been donated to the open Cloud Native Compute Foundation, allowing enterprise to better understand their cloud native workloads and power consumptions. Kepler can also help in the development of sustainable software by creating less power hungry applications. Again, Red Hat continuing to provide OpenSource, OpenRAN, and contributing an open source project to the CNCF, continuing to create innovation for developers, and, of course, Red Hat knows what, a lot about operating systems and the telco could be the next frontier. That's SiliconANGLE NEWS. I'm John Furrier; thanks for watching. (monotone music)
SUMMARY :
and in the news here is Red Hat,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
NVIDIA | ORGANIZATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Samsung | ORGANIZATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Barcelona | LOCATION | 0.99+ |
Cloud Native Compute Foundation | ORGANIZATION | 0.99+ |
CNCF | ORGANIZATION | 0.98+ |
UK | LOCATION | 0.95+ |
OpenRAN | TITLE | 0.93+ |
telco | ORGANIZATION | 0.93+ |
Kubernetes | TITLE | 0.92+ |
Kepler | ORGANIZATION | 0.9+ |
SiliconANGLE NEWS | ORGANIZATION | 0.88+ |
vRAN | TITLE | 0.88+ |
SiliconANGLE | ORGANIZATION | 0.87+ |
Arm | ORGANIZATION | 0.87+ |
MWC | EVENT | 0.86+ |
Arm on Efficient Open Networks | ORGANIZATION | 0.86+ |
theCUBE | ORGANIZATION | 0.84+ |
OpenShift | TITLE | 0.78+ |
Hat | TITLE | 0.73+ |
SiliconANGLE News | ORGANIZATION | 0.65+ |
OpenSource | TITLE | 0.61+ |
NEWS | ORGANIZATION | 0.51+ |
Red | ORGANIZATION | 0.5+ |
SiliconANGLE | TITLE | 0.43+ |
Anthony Dina, Dell Technologies and Bob Crovella, NVIDIA | SuperComputing 22
>>How do y'all, and welcome back to Supercomputing 2022. We're the Cube, and we are live from Dallas, Texas. I'm joined by my co-host, David Nicholson. David, hello. Hello. We are gonna be talking about data and enterprise AI at scale during this segment. And we have the pleasure of being joined by both Dell and Navidia. Anthony and Bob, welcome to the show. How you both doing? Doing good. >>Great. Great show so far. >>Love that. Enthusiasm, especially in the afternoon on day two. I think we all, what, what's in that cup? Is there something exciting in there that maybe we should all be sharing with you? >>Just say it's just still Yeah, water. >>Yeah. Yeah. I love that. So I wanna make sure that, cause we haven't talked about this at all during the show yet, on the cube, I wanna make sure that everyone's on the same page when we're talking about data unstructured versus structured data. I, it's in your title, Anthony, tell me what, what's the difference? >>Well, look, the world has been based in analytics around rows and columns, spreadsheets, data warehouses, and we've made predictions around the forecast of sales maintenance issues. But when we take computers and we give them eyes, ears, and fingers, cameras, microphones, and temperature and vibration sensors, we now translate that into more human experience. But that kind of data, the sensor data, that video camera is unstructured or semi-structured, that's what that >>Means. We live in a world of unstructured data structure is something we add to later after the fact. But the world that we see and the world that we experience is unstructured data. And one of the promises of AI is to be able to take advantage of everything that's going on around us and augment that, improve that, solve problems based on that. And so if we're gonna do that job effectively, we can't just depend on structured data to get the problem done. We have to be able to incorporate everything that we can see here, taste, smell, touch, and use >>That as, >>As part of the problem >>Solving. We want the chaos, bring it. >>Chaos has been a little bit of a theme of our >>Show. It has been, yeah. And chaos is in the eye of the beholder. You, you think about, you think about the reason for structuring data to a degree. We had limited processing horsepower back when everything was being structured as a way to allow us to be able to, to to reason over it and gain insights. So it made sense to put things into rows and tables. How does, I'm curious, diving right into where Nvidia fits into this, into this puzzle, how does NVIDIA accelerate or enhance our ability to glean insight from or reason over unstructured data in particular? >>Yeah, great question. It's really all about, I would say it's all about ai and Invidia is a leader in the AI space. We've been investing and focusing on AI since at least 2012, if not before, accelerated computing that we do it. Invidia is an important part of it, really. We believe that AI is gonna revolutionize nearly every aspect of computing. Really nearly every aspect of problem solving, even nearly every aspect of programming. And one of the reasons is for what we're talking about now is it's a little impact. Being able to incorporate unstructured data into problem solving is really critical to being able to solve the next generation of problems. AI unlocks, tools and methodologies that we can realistically do that with. It's not realistic to write procedural code that's gonna look at a picture and solve all the problems that we need to solve if we're talking about a complex problem like autonomous driving. But with AI and its ability to naturally absorb unstructured data and make intelligent reason decisions based on it, it's really a breakthrough. And that's what NVIDIA's been focusing on for at least a decade or more. >>And how does NVIDIA fit into Dell's strategy? >>Well, I mean, look, we've been partners for many, many years delivering beautiful experiences on workstations and laptops. But as we see the transition away from taking something that was designed to make something pretty on screen to being useful in solving problems in life sciences, manufacturing in other places, we work together to provide integrated solutions. So take for example, the dgx a 100 platform, brilliant design, revolutionary bus technologies, but the rocket ship can't go to Mars without the fuel. And so you need a tank that can scale in performance at the same rate as you throw GPUs at it. And so that's where the relationship really comes alive. We enable people to curate the data, organize it, and then feed those algorithms that get the answers that Bob's been talking about. >>So, so as a gamer, I must say you're a little shot at making things pretty on a screen. Come on. That was a low blow. That >>Was a low blow >>Sassy. What I, >>I Now what's in your cup? That's what I wanna know, Dave, >>I apparently have the most boring cup of anyone on you today. I don't know what happened. We're gonna have to talk to the production team. I'm looking at all of you. We're gonna have to make that better. One of the themes that's been on this show, and I love that you all embrace the chaos, we're, we're seeing a lot of trend in the experimentation phase or stage rather. And it's, we're in an academic zone of it with ai, companies are excited to adopt, but most companies haven't really rolled out their strategy. What is necessary for us to move from this kind of science experiment, science fiction in our heads to practical application at scale? Well, >>Let me take this, Bob. So I've noticed there's a pattern of three levels of maturity. The first level is just what you described. It's about having an experience, proof of value, getting stakeholders on board, and then just picking out what technology, what algorithm do I need? What's my data source? That's all fun, but it is chaos over time. People start actually making decisions based on it. This moves us into production. And what's important there is normality, predictability, commonality across, but hidden and embedded in that is a center of excellence. The community of data scientists and business intelligence professionals sharing a common platform in the last stage, we get hungry to replicate those results to other use cases, throwing even more information at it to get better accuracy and precision. But to do this in a budget you can afford. And so how do you figure out all the knobs and dials to turn in order to make, take billions of parameters and process that, that's where casual, what's >>That casual decision matrix there with billions of parameters? >>Yeah. Oh, I mean, >>But you're right that >>That's, that's exactly what we're, we're on this continuum, and this is where I think the partnership does really well, is to marry high performant enterprise grade scalability that provides the consistency, the audit trail, all of the things you need to make sure you don't get in trouble, plus all of the horsepower to get to the results. Bob, what would you >>Add there? I think the thing that we've been talking about here is complexity. And there's complexity in the AI problem solving space. There's complexity everywhere you look. And we talked about the idea that NVIDIA can help with some of that complexity from the architecture and the software development side of it. And Dell helps with that in a whole range of ways, not the least of which is the infrastructure and the server design and everything that goes into unlocking the performance of the technology that we have available to us today. So even the center of excellence is an example of how do I take this incredibly complex problem and simplify it down so that the real world can absorb and use this? And that's really what Dell and Vidia are partnering together to do. And that's really what the center of excellence is. It's an idea to help us say, let's take this extremely complex problem and extract some good value out of >>It. So what is Invidia's superpower in this realm? I mean, look, we're we are in, we, we are in the era of Yeah, yeah, yeah. We're, we're in a season of microprocessor manufacturers, one uping, one another with their latest announcements. There's been an ebb and a flow in our industry between doing everything via the CPU versus offloading processes. Invidia comes up and says, Hey, hold on a second, gpu, which again, was focused on graphics processing originally doing something very, very specific. How does that translate today? What's the Nvidia again? What's, what's, what's the superpower? Because people will say, well, hey, I've got a, I've got a cpu, why do I need you? >>I think our superpower is accelerated computing, and that's really a hardware and software thing. I think your question is slanted towards the hardware side, which is, yes, it is very typical and we do make great processors, but the processor, the graphics processor that you talked about from 10 or 20 years ago was designed to solve a very complex task. And it was exquisitely designed to solve that task with the resources that we had available at that time. Time. Now, fast forward 10 or 15 years, we're talking about a new class of problems called ai. And it requires both exquisite, soft, exquisite processor design as well as very complex and exquisite software design sitting on top of it as well. And the systems and infrastructure knowledge, high performance storage and everything that we're talking about in the solution today. So Nvidia superpower is really about that accelerated computing stack at the bottom. You've got hardware above that, you've got systems above that, you have middleware and libraries and above that you have what we call application SDKs that enable the simplification of this really complex problem to this domain or that domain or that domain, while still allowing you to take advantage of that processing horsepower that we put in that exquisitely designed thing called the gpu >>Decreasing complexity and increasing speed to very key themes of the show. Shocking, no one, you all wanna do more faster. Speaking of that, and I'm curious because you both serve a lot of different unique customers, verticals and use cases, is there a specific project that you're allowed to talk about? Or, I mean, you know, you wanna give us the scoop, that's totally cool too. We're here for the scoop on the cube, but is there a specific project or use case that has you personally excited Anthony? We'll start with that. >>Look, I'm, I've always been a big fan of natural language processing. I don't know why, but to derive intent based on the word choices is very interesting to me. I think what compliments that is natural language generation. So now we're having AI programs actually discover and describe what's inside of a package. It wouldn't surprise me that over time we move from doing the typical summary on the economic, the economics of the day or what happened in football. And we start moving that towards more of the creative advertising and marketing arts where you are no longer needed because the AI is gonna spit out the result. I don't think we're gonna get there, but I really love this idea of human language and computational linguistics. >>What a, what a marriage. I agree. Think it's fascinating. What about you, Bob? It's got you >>Pumped. The thing that really excites me is the problem solving, sort of the tip of the spear in problem solving. The stuff that you've never seen before, the stuff that you know, in a geeky way kind of takes your breath away. And I'm gonna jump or pivot off of what Anthony said. Large language models are really one of those areas that are just, I think they're amazing and they're just kind of surprising everyone with what they can do here on the show floor. I was looking at a demonstration from a large language model startup, basically, and they were showing that you could ask a question about some obscure news piece that was reported only in a German newspaper. It was about a little shipwreck that happened in a hardware. And I could type in a query to this system and it would immediately know where to find that information as if it read the article, summarized it for you, and it even could answer questions that you could only only answer by looking pic, looking at pictures in that article. Just amazing stuff that's going on. Just phenomenal >>Stuff. That's a huge accessibility. >>That's right. And I geek out when I see stuff like that. And that's where I feel like all this work that Dell and Invidia and many others are putting into this space is really starting to show potential in ways that we wouldn't have dreamed of really five years ago. Just really amazing. And >>We see this in media and entertainment. So in broadcasting, you have a sudden event, someone leaves this planet where they discover something new where they get a divorce and they're a major quarterback. You wanna go back somewhere in all of your archives to find that footage. That's a very laborist project. But if you can use AI technology to categorize that and provide the metadata tag so you can, it's searchable, then we're off to better productions, more interesting content and a much richer viewer experience >>And a much more dynamic picture of what's really going on. Factoring all of that in, I love that. I mean, David and I are both nerds and I know we've had take our breath away moments, so I appreciate that you just brought that up. Don't worry, you're in good company. In terms of the Geek Squad over >>Here, I think actually maybe this entire show for Yes, exactly. >>I mean, we were talking about how steampunk some of the liquid cooling stuff is, and you know, this is the only place on earth really, or the only show where you would come and see it at this level in scale and, and just, yeah, it's, it's, it's very, it's very exciting. How important for the future of innovation in HPC are partnerships like the one that Navia and Dell have? >>You wanna start? >>Sure, I would, I would just, I mean, I'm gonna be bold and brash and arrogant and say they're essential. Yeah, you don't not, you do not want to try and roll this on your own. This is, even if we just zoomed in to one little beat, little piece of the technology, the software stack that do modern, accelerated deep learning is incredibly complicated. There can be easily 20 or 30 components that all have to be the right version with the right buttons pushed, built the right way, assembled the right way, and we've got lots of technologies to help with that. But you do not want to be trying to pull that off on your own. That's just one little piece of the complexity that we talked about. And we really need, as technology providers in this space, we really need to do as much as we do to try to unlock the potential. We have to do a lot to make it usable and capable as well. >>I got a question for Anthony. All >>Right, >>So in your role, and I, and I'm, I'm sort of, I'm sort of projecting here, but I think, I think, I think your superpower personally is likely in the realm of being able to connect the dots between technology and the value that that technology holds in a variety of contexts. That's right. Whether it's business or, or whatever, say sentences. Okay. Now it's critical to have people like you to connect those dots. Today in the era of pervasive ai, how important will it be to have AI have to explain its answer? In other words, words, should I trust the information the AI is giving me? If I am a decision maker, should I just trust it on face value? Or am I going to want a demand of the AI kind of what you deliver today, which is No, no, no, no, no, no. You need to explain this to me. How did you arrive at that conclusion, right? How important will that be for people to move forward and trust the results? We can all say, oh hey, just trust us. Hey, it's ai, it's great, it's got Invidia, you know, Invidia acceleration and it's Dell. You can trust us, but come on. So many variables in the background. It's >>An interesting one. And explainability is a big function of ai. People want to know how the black box works, right? Because I don't know if you have an AI engine that's looking for potential maladies in an X-ray, but it misses it. Do you sue the hospital, the doctor or the software company, right? And so that accountability element is huge. I think as we progress and we trust it to be part of our everyday decision making, it's as simply as a recommendation engine. It isn't actually doing all of the decisions. It's supporting us. We still have, after decades of advanced technology algorithms that have been proven, we can't predict what the market price of any object is gonna be tomorrow. And you know why? You know why human beings, we are so unpredictable. How we feel in the moment is radically different. And whereas we can extrapolate for a population to an individual choice, we can't do that. So humans and computers will not be separated. It's a, it's a joint partnership. But I wanna get back to your point, and I think this is very fundamental to the philosophy of both companies. Yeah, it's about a community. It's always about the people sharing ideas, getting the best. And anytime you have a center of excellence and algorithm that works for sales forecasting may actually be really interesting for churn analysis to make sure the employees or students don't leave the institution. So it's that community of interest that I think is unparalleled at other conferences. This is the place where a lot of that happens. >>I totally agree with that. We felt that on the show. I think that's a beautiful note to close on. Anthony, Bob, thank you so much for being here. I'm sure everyone feels more educated and perhaps more at peace with the chaos. David, thanks for sitting next to me asking the best questions of any host on the cube. And thank you all for being a part of our community. Speaking of community here on the cube, we're alive from Dallas, Texas. It's super computing all week. My name is Savannah Peterson and I'm grateful you're here. >>So I.
SUMMARY :
And we have the pleasure of being joined by both Dell and Navidia. Great show so far. I think we all, cause we haven't talked about this at all during the show yet, on the cube, I wanna make sure that everyone's on the same page when we're talking about But that kind of data, the sensor data, that video camera is unstructured or semi-structured, And one of the promises of AI is to be able to take advantage of everything that's going on We want the chaos, bring it. And chaos is in the eye of the beholder. And one of the reasons is for what we're talking about now is it's a little impact. scale in performance at the same rate as you throw GPUs at it. So, so as a gamer, I must say you're a little shot at making things pretty on a I apparently have the most boring cup of anyone on you today. But to do this in a budget you can afford. the horsepower to get to the results. and simplify it down so that the real world can absorb and use this? What's the Nvidia again? So Nvidia superpower is really about that accelerated computing stack at the bottom. We're here for the scoop on the cube, but is there a specific project or use case that has you personally excited And we start moving that towards more of the creative advertising and marketing It's got you And I'm gonna jump or pivot off of what That's a huge accessibility. And I geek out when I see stuff like that. and provide the metadata tag so you can, it's searchable, then we're off to better productions, so I appreciate that you just brought that up. I mean, we were talking about how steampunk some of the liquid cooling stuff is, and you know, this is the only place on earth really, There can be easily 20 or 30 components that all have to be the right version with the I got a question for Anthony. to have people like you to connect those dots. And anytime you have a center We felt that on the show.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
David Nicholson | PERSON | 0.99+ |
Bob | PERSON | 0.99+ |
Anthony | PERSON | 0.99+ |
Bob Crovella | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
20 | QUANTITY | 0.99+ |
Invidia | ORGANIZATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Mars | LOCATION | 0.99+ |
Vidia | ORGANIZATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
10 | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
Dallas, Texas | LOCATION | 0.99+ |
Dell Technologies | ORGANIZATION | 0.99+ |
15 years | QUANTITY | 0.99+ |
Dallas, Texas | LOCATION | 0.99+ |
Navidia | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
first level | QUANTITY | 0.99+ |
both companies | QUANTITY | 0.98+ |
Today | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
2012 | DATE | 0.98+ |
today | DATE | 0.98+ |
billions | QUANTITY | 0.98+ |
earth | LOCATION | 0.97+ |
10 | DATE | 0.96+ |
Anthony Dina | PERSON | 0.96+ |
five years ago | DATE | 0.96+ |
30 components | QUANTITY | 0.95+ |
Navia | ORGANIZATION | 0.95+ |
day two | QUANTITY | 0.94+ |
one little piece | QUANTITY | 0.91+ |
tomorrow | DATE | 0.87+ |
three levels | QUANTITY | 0.87+ |
HPC | ORGANIZATION | 0.86+ |
20 years ago | DATE | 0.83+ |
one little | QUANTITY | 0.77+ |
billions of parameters | QUANTITY | 0.75+ |
a decade | QUANTITY | 0.74+ |
decades | QUANTITY | 0.68+ |
German | OTHER | 0.68+ |
dgx a 100 platform | COMMERCIAL_ITEM | 0.67+ |
themes | QUANTITY | 0.63+ |
second | QUANTITY | 0.57+ |
22 | QUANTITY | 0.48+ |
Squad | ORGANIZATION | 0.4+ |
Supercomputing 2022 | ORGANIZATION | 0.36+ |
Ami Badani, NVIDIA & Mike Capuano, Pluribus Networks
(upbeat music) >> Let's kick things off. We're here at Mike Capuano the CMO of Pluribus Networks, and Ami Badani VP of Networking, Marketing, and Developer of Ecosystem at NVIDIA. Great to have you welcome folks. >> Thank you. >> Thanks. >> So let's get into the the problem situation with cloud unified networking. What problems are out there? What challenges do cloud operators have Mike? Let's get into it. >> The challenges that we're looking at are for non hyperscalers that's enterprises, governments Tier 2 service providers, cloud service providers. And the first mandate for them is to become as agile as a hyperscaler. So they need to be able to deploy services and security policies in seconds. They need to be able to abstract the complexity of the network and define things in software while it's accelerated in hardware. Really ultimately they need a single operating model everywhere. And then the second thing is they need to distribute networking and security services out to the edge of the host. We're seeing a growth cyber attacks. It's not slowing down. It's only getting worse and solving for this security problem across clouds is absolutely critical. And the way to do it is to move security out to the host. >> With that goal in mind, what's the Pluribus vision how does this tie together? >> So basically what we see is that this demands a new architecture and that new architecture has four tenets. The first tenet is unified and simplified cloud networks. If you look at cloud networks today, there's sort of like discreet bespoke cloud networks per hypervisor, per private cloud, edge cloud, public cloud. Each of the public clouds have different networks, that needs to be unified. If we want these folks to be able to be agile they need to be able to issue a single command or instantiate a security policy across all of those locations with one command and not have to go to each one. The second is, like I mentioned distributed security. Distributed security without compromise, extended out to the host is absolutely critical. So micro segmentation and distributed firewalls. But it doesn't stop there. They also need pervasive visibility. It's sort of like with security you really can't see you can't protect you can't see. So you need visibility everywhere. The problem is visibility to date has been very expensive. Folks have had to basically build a separate overlay network of taps, packet brokers, tap aggregation infrastructure, that really needs to be built in to this unified network I'm talking about. And the last thing is automation. All of this needs to be SDN enabled. So this is related to my comment about abstraction. Abstract the complexity of all these discreet networks whatever's down there in the physical layer. I don't want to see it. I want to abstract it. I want to define things in software but I do want to leverage the power of hardware to accelerate that. So that's the fourth tenet is SDN automation. >> Mike, we've been talking on theCUBE a lot about this architectural shift and customers are looking at this. This is a big part of everyone who's looking at cloud operations, NextGen. How do we get there? How do customer customers get this vision realized? >> That's a great question. And I appreciate the tee up. We're here today for that reason. We're introducing two things today. The first is a unified cloud networking vision. And that is a vision of where Pluribus is headed with our partners like NVIDIA long term. And that is about deploying a common operating model SDN enabled, SDN automated, hardware accelerated across all clouds. And whether that's underlay and overlay switch or server, any hypervisor infrastructure containers, any workload doesn't matter. So that's ultimately where we want to get. And that's what we talked about earlier. The first step in that vision is what we call the unified cloud fabric. And this is the next generation of our adaptive cloud fabric. And what's nice about this is we're not starting from scratch. We have an award-winning adaptive cloud fabric product that is deployed globally. And in particular, we're very proud of the fact that it's deployed in over 100 Tier 1 mobile operators as the network fabric for their 4G and 5G virtualized cores. We know how to build carrier grade networking infrastructure. What we're doing now to realize this next generation unified cloud fabric is we're extending from the switch to this NVIDIA BlueField-2 DPU. We know there's. >> Hold that up real quick. That's a good prop. That's the BlueField NVIDIA card. >> It's the NVIDIA BlueField-2 DPU, data processing unit. What we're doing fundamentally is extending our SDN automated fabric, the unified cloud fabric, out to the host. But it does take processing power. So we knew that we didn't want to do we didn't want to implement that running on the CPUs which is what some other companies do. Because it consumes revenue generating CPUs from the application. So a DPU is a perfect way to implement this. And we knew that NVIDIA was the leader with this BlueField-2. And so that is the first, that's the first step into getting, into realizing this vision. >> NVIDIA has always been powering some great workloads of GPUs, now you got DPUs. Networking and NVIDIA as here. What is the relationship with Pluribus? How did that come together? Tell us the story. >> We've been working with Pluribus for quite some time. I think the last several months was really when it came to fruition. And what Pluribus is trying to build and what NVIDIA has. So we have, this concept of a blue field data processing unit, which, if you think about it, conceptually does really three things, offload, accelerate, and isolate. So offload your workloads from your CPU to your data processing unit, infrastructure workloads that is. Accelerate, so there's a bunch of acceleration engines. You can run infrastructure workloads much faster than you would otherwise. And then isolation, So you have this nice security isolation between the data processing unit and your other CPU environment. And so you can run completely isolated workloads directly on the data processing unit. So we introduced this, a couple years ago. And with Pluribus we've been talking to the Pluribus team for quite some months now. And I think really the combination of what Pluribus is trying to build, and what they've developed around this unified cloud fabric fits really nicely with the DPU and running that on the DPU and extending it really from your physical switch all the way to your host environment, specifically on the data processing unit. So if you think about what's happening as you add data processing units to your environment. So every server we believe over time is going to have data processing units. So now you'll have to manage that complexity from the physical network layer to the host layer. And so what Pluribus is really trying to do is extending the network fabric from the host from the switch to the host and really have that single pane of glass for network operators to be able to configure, provision, manage all of the complexity of the network environment. So that's really how the partnership truly started. And so it started really with extending the network fabric and now we're also working with them on security. If you sort of take that concept of isolation and security isolation, what Pluribus has within their fabric is the concept of micro segmentation. And so now you can take that extend it to the data processing unit and really have isolated micro segmentation workloads whether it's bare metal, cloud native environments, whether it's virtualized environments, whether it's public cloud, private cloud, hybrid cloud. So it really is a magical partnership between the two companies with their unified cloud fabric running on the DPU. >> You know what I love about this conversation is it reminds me of when you have these changing markets. The product gets pulled out of the market and you guys step up and create these new solutions. And I think this is a great example. So I have to ask you how do you guys differentiate what sets this apart for customers? What's in it for the customer? >> So I mentioned three things in terms of the value of what the BlueField brings. There's offloading, accelerating and isolating. And that's sort of the key core tenets of BlueField. So that, if you sort of think about what BlueField what we've done, in terms of the differentiation. We're really a robust platform for innovation. So we introduced BlueField-2 last year. We're introducing BlueField-3 which is our next generation of blue field. It'll have 5X the ARM compute capacity. It will have 400 gig line rate acceleration, 4X better crypto acceleration. So it will be remarkably better than the previous generation. And we'll continue to innovate and add, chips to our portfolio every 18 months to two years. So that's sort of one of the key areas of differentiation. The other is that if you look at NVIDIA, what we're sort of known for is really known for our AI, our artificial intelligence and our artificial intelligence software, as well as our GPU. So you look at artificial intelligence and the combination of artificial intelligence plus data processing. This really creates faster, more efficient secure AI systems from, the core of your data center, all the way out to the edge. And so with NVIDIA we really have these converged accelerators where we've combined the GPU, which does all your AI processing with your data processing with the DPU. So we have this convergence really nice convergence of that area. And I would say the third area is really around our developer environment. One of the key, one of our key motivations at NVIDIA is really to have our partner ecosystem embrace our technology and build solutions around our technology. So if you look at what we've done with the DPU we've created an SDK, which is an open SDK called DOCA. And it's an open SDK for our partners to really build and develop solutions using BlueField and using all these accelerated libraries that we expose through DOCA. And so part of our differentiation is really building this open ecosystem for our partners to take advantage and build solutions around our technology. >> What's exciting is when I hear you talk it's like you realize that there's no one general purpose network anymore. Everyone has their own super environment, super cloud or these new capabilities. They can really craft their own I'd say custom environment at scale with easy tools. And it's all kind of that again this is the new architecture Mike, you were talking about. How does customers run this effectively, cost effectively? And how do people migrate? >> I think that is the key question. So we've got this beautiful architecture. Amazon Nitro is a good example of a SmartNIC architecture that has been successfully deployed but, enterprises and Tier 2 service providers and Tier 1 service providers and governments are not Amazon. So they need to migrate there and they need this architecture to be cost of effective. And that's super key. I mean, the reality is DPU are moving fast but they're not going to be deployed everywhere on day one. Some servers will have have DPUs right away. Some servers will have DPUs in a year or two. And then there are devices that may never have DPUs. IOT gateways, or legacy servers, even mainframes. So that's the beauty of a solution that creates a fabric across both the switch and the DPU. And by leveraging the NVIDIA BlueField DPU what we really like about it is, it's open and that drives cost efficiencies. And then, with this our architectural approach effectively you get a unified solution across switch and DPU, workload independent. It doesn't matter what hypervisor it is. Integrated visibility, integrated security and that can create tremendous cost efficiencies and really extract a lot of the expense from a capital perspective out of the network as well as from an operational perspective because now I have an SDN automated solution where I'm literally issuing a command to deploy a network service, or to deploy a security policy and is deployed everywhere automatically saving the network operations team and the security operations team time. >> So let me rewind that 'cause that's super important. Got the unified cloud architecture. I'm the customer, it's implemented. What's the value again, take me through the value to me. I have a unified environment. What's the value? >> I mean the value is effectively, there's a few pieces of value. The first piece of value is I'm creating this clean demark. I'm taking networking to the host. And like I mentioned, we're not running it on the CPU. So in implementations that run networking on the CPU there's some conflict between the DevOps team who own the server, and the NetOps team who own the network because they're installing software on the CPU stealing cycles from what should be revenue generating CPUs. So now by terminating the networking on the DPU we create this real clean demark. So the DevOps folks are happy because they don't necessarily have the skills to manage network and they don't necessarily want to spend the time managing networking. They've got their network counterparts who are also happy the NetOps team because they want to control the networking. And now we've got this clean demark where the DevOps folks get the services they need and the NetOps folks get the control and agility they need. So that's a huge value. The next piece of value is distributed security. This is essential I mentioned it earlier, pushing out micro segmentation and distributed firewall basically at the application level, where I create these small segments on an application by application basis. So if a bad actor does penetrate the perimeter firewall they're contained once they get inside. 'Cause the worst thing is a bad actor penetrates perimeter firewall and can go wherever they want in wreak havoc. And so that's why this is so essential. And the next benefit obviously is this unified networking operating model. Having an operating model across switch and server, underlay and overlay, workload agnostic, making the life of the NetOps teams much easier so they can focus their time on really strategy instead of spending an afternoon deploying a single VLAN for example. >> Awesome, and I think also for my stand point I mean perimeter security is pretty much, that out there, I guess the firewall still out there exists but pretty much they're being breached all the time the perimeter. You have to have this new security model. And I think the other thing that you mentioned the separation between DevOps is cool because the infrastructure is code is about making the developers be agile and build security in from day one. So this policy aspect is huge new control plan. I think you guys have a new architecture that enables the security to be handled more flexible. That seems to be the killer feature here. >> If you look at the data processing unit, I think one of the great things about sort of this new architecture it's really the foundation for zero trust. So like you talked about the perimeter is getting breached. And so now each and every compute node has to be protected. And I think that's sort of what you see with the partnership between Pluribus and NVIDIA is the DPU is really the foundation of zero trust and Pluribus is really building on that vision with allowing sort of micro-segmentation and being able to protect each and every compute node as well as the underlying network. >> This is super exciting. This is illustration of how the market's evolving architectures are being reshaped and refactored for cloud scale and all this new goodness with data. So I got to ask how you guys go into market together. Michael, start with you. What's the relationship look like in the go to market with NVIDIA? >> We're super excited about the partnership. Obviously we're here together. We think we've got a really good solution for the market so we're jointly marketing it. Obviously we appreciate that NVIDIA's open that's sort of in our DNA, we're about a open networking. They've got other ISVs who are going to run on BlueField-2. We're probably going to run on other DPUs in the future. But right now we feel like we're partnered with the number one provider of DPUs in the world and super excited about making a splash with it. >> Oh man NVIDIA got the hot product. >> So BlueField-2 as I mentioned was GA last year, we're introducing, well we now also have the converged accelerator. So I talked about artificial intelligence our artificial intelligence software with the BlueField DPU, all of that put together on a converged accelerator. The nice thing there is you can either run those workloads, so if you have an artificial intelligence workload and an infrastructure workload, you can work on them separately on the same platform or you can actually use you can actually run artificial intelligence applications on the BlueField itself. So that's what the converged accelerator really brings to the table. So that's available now. Then we have BlueField-3 which will be available late this year. And I talked about sort of, how much better that next generation of BlueField is in comparison to BlueField-2. So we'll see BlueField-3 shipping later on this year. And then our software stack which I talked about, which is called DOCA. We're on our second version, our DOCA 1.2 we're releasing DOCA 1.3 in about two months from now. And so that's really our open ecosystem framework. So allow you to program the BlueField. So we have all of our acceleration libraries, security libraries, that's all packed into this SDK called DOCA. And it really gives that simplicity to our partners to be able to develop on top of BlueField. So as we add new generations of BlueField, next year we'll have another version and so on and so forth. DOCA is really that unified layer that allows BlueField to be both forwards compatible and backwards compatible. So partners only really have to think about writing to that SDK once. And then it automatically works with future generations of BlueField. So that's sort of the nice thing around DOCA. And then in terms of our go to market model we're working with every major OEM. Later on this year you'll see, major server manufacturers releasing BlueField enabled servers, so more to come. >> Awesome, save money, make it easier, more capabilities, more workload power. This is the future of cloud operations. >> And one thing I'll add is we are, we have a number of customers as you'll hear in the next segment that are already signed up and will be working with us for our early field trial starting late April early May. We are accepting registrations. You can go to www.pluribusnetworks.com/eft. If you're interested in signing up for being part of our field trial and providing feedback on the product >> Awesome innovation and networking. Thanks so much for sharing the news. Really appreciate, thanks so much. In a moment we'll be back to look deeper in the product the integration, security, zero trust use cases. You're watching theCUBE, the leader in enterprise tech coverage. (upbeat music)
SUMMARY :
the CMO of Pluribus Networks, So let's get into the And the way to do it is to So that's the fourth and customers are looking at this. And I appreciate the tee up. That's the BlueField NVIDIA card. And so that is the first, What is the relationship with Pluribus? DPU and running that on the DPU So I have to ask you how So that's sort of one of the And it's all kind of that again So that's the beauty of a solution that Got the unified cloud architecture. and the NetOps team who own the network that enables the security is the DPU is really the in the go to market with NVIDIA? on other DPUs in the future. So that's sort of the This is the future of cloud operations. and providing feedback on the product Thanks so much for sharing the news.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Tom | PERSON | 0.99+ |
Stefanie | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Manasi | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Pluribus | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Stephanie Chiras | PERSON | 0.99+ |
2015 | DATE | 0.99+ |
Ami Badani | PERSON | 0.99+ |
Stefanie Chiras | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
2008 | DATE | 0.99+ |
Mike Capuano | PERSON | 0.99+ |
two companies | QUANTITY | 0.99+ |
two years | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
90% | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
Mike | PERSON | 0.99+ |
RHEL | TITLE | 0.99+ |
Chicago | LOCATION | 0.99+ |
2021 | DATE | 0.99+ |
Pluribus Networks | ORGANIZATION | 0.99+ |
second version | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
next year | DATE | 0.99+ |
Ansible | ORGANIZATION | 0.99+ |
Pete Lumbis, NVIDIA & Alessandro Barbieri, Pluribus Networks
(upbeat music) >> Okay, we're back. I'm John Furrier with theCUBE and we're going to go deeper into a deep dive into unified cloud networking solution from Pluribus and NVIDIA. And we'll examine some of the use cases with Alessandro Barbieri, VP of product management at Pluribus Networks and Pete Lumbis, the director of technical marketing and video remotely. Guys thanks for coming on, appreciate it. >> Yeah thanks a lot. >> I'm happy to be here. >> So a deep dive, let's get into the what and how. Alessandro, we heard earlier about the Pluribus and NVIDIA partnership and the solution you're working together in. What is it? >> Yeah, first let's talk about the what. What are we really integrating with the NVIDIA BlueField the DPU technology? Pluribus has been shipping in volume in multiple mission critical networks, this Netvisor ONE network operating systems. It runs today on merchant silicon switches and effectively it's standard based open network operating system for data center. And the novelty about this operating system is that it integrates distributed the control plane to automate effect with SDN overlay. This automation is completely open and interoperable and extensible to other type of clouds. It's not enclosed. And this is actually what we're now porting to the NVIDIA DPU. >> Awesome, so how does it integrate into NVIDIA hardware and specifically how is Pluribus integrating its software with the NVIDIA hardware? >> Yeah, I think we leverage some of the interesting properties of the BlueField DPU hardware which allows actually to integrate our network operating system in a manner which is completely isolated and independent from the guest operating system. So the first byproduct of this approach is that whatever we do at the network level on the DPU card is completely agnostic to the hypervisor layer or OS layer running on the host. Even more, we can also independently manage this network node this switch on a NIC effectively, managed completely independently from the host. You don't have to go through the network operating system running on X86 to control this network node. So you truly have the experience effectively top of rack for virtual machine or a top of rack for Kubernetes spots, where if you allow me with analogy, instead of connecting a server NIC directly to a switchboard, now we are connecting a VM virtual interface to a virtual interface on the switch on an niche. And also as part of this integration, we put a lot of effort, a lot of emphasis in accelerating the entire data plan for networking and security. So we are taking advantage of the NVIDIA DOCA API to program the accelerators. And these you accomplish two things with that. Number one, you have much better performance. They're running the same network services on an X86 CPU. And second, this gives you the ability to free up I would say around 20, 25% of the server capacity to be devoted either to additional workloads to run your cloud applications or perhaps you can actually shrink the power footprint and compute footprint of your data center by 20% if you want to run the same number of compute workloads. So great efficiencies in the overall approach. >> And this is completely independent of the server CPU, right? >> Absolutely, there is zero code from Pluribus running on the X86. And this is why we think this enables a very clean demarcation between compute and network. >> So Pete, I got to get you in here. We heard that the DPU enable cleaner separation of DevOps and NetOps. Can you explain why that's important because everyone's talking DevSecOps, right? Now, you've got NetSecOps. This separation, why is this clean separation important? >> Yeah, I think, it's a pragmatic solution in my opinion. We wish the world was all kind of rainbows and unicorns, but it's a little messier than that. I think a lot of the DevOps stuff and that mentality and philosophy. There's a natural fit there. You have applications running on servers. So you're talking about developers with those applications integrating with the operators of those servers. Well, the network has always been this other thing and the network operators have always had a very different approach to things than compute operators. And I think that we in the networking industry have gotten closer together but there's still a gap, there's still some distance. And I think that distance isn't going to be closed and so, again, it comes down to pragmatism. And I think one of my favorite phrases is look, good fences make good neighbors. And that's what this is. >> Yeah, and it's a great point 'cause DevOps has become kind of the calling car for cloud, right? But DevOps is a simply infrastructures code and infrastructure is networking, right? So if infrastructure is code you're talking about that part of the stack under the covers, under the hood if you will. This is super important distinction and this is where the innovation is. Can you elaborate on how you see that because this is really where the action is right now? >> Yeah, exactly. And I think that's where one from the policy, the security, the zero trust aspect of this, right? If you get it wrong on that network side, all of a sudden you can totally open up those capabilities. And so security's part of that. But the other part is thinking about this at scale, right? So we're taking one top of rack switch and adding up to 48 servers per rack. And so that ability to automate, orchestrate and manage its scale becomes absolutely critical. >> Alessandro, this is really the why we're talking about here and this is scale. And again, getting it right. If you don't get it right, you're going to be really kind of up you know what? So this is a huge deal. Networking matters, security matters, automation matters, DevOps, NetOps, all coming together clean separation. Help us understand how this joint solution with NVIDIA fits into the Pluribus unified cloud networking vision because this is what people are talking about and working on right now. >> Yeah, absolutely. So I think here with this solution we're attacking two major problems in cloud networking. One, is operation of cloud networking and the second, is distributing security services in the cloud infrastructure. First, let me talk about first what are we really unifying? If we're unifying something, something must be at least fragmented or disjointed. And what is disjointed is actually the network in the cloud. If you look wholistically how networking is deployed in the cloud, you have your physical fabric infrastructure, right? Your switches and routers. You build your IP clause, fabric leaf and spine topologies. This is actually a well understood problem I would say. There are multiple vendors with let's say similar technologies, very well standardized, very well understood and almost a commodity I would say building an IP fabric these days, but this is not the place where you deploy most of your services in the cloud particularly from a security standpoint. Those services are actually now moved into the compute layer where cloud builders have to instrument a separate network virtualization layer where they deploy segmentation and security closer to the workloads. And this is where the complication arise. This high value part of the cloud network is where you have a plethora of options that they don't talk to each other and they're very dependent on the kind of hypervisor or compute solution you choose. For example, the networking API between an ESXi environment or an Hyper-V or a Zen are completely disjointed. You have multiple orchestration layers. And then when you throw in also Kubernetes in this type of architecture, you are introducing yet another level of networking. And when Kubernetes runs on top of VMs which is a prevalent approach, you actually are stuck in multiple networks on the compute layer that they eventually ran on the physical fabric infrastructure. Those are all ships in the knights effectively, right? They operate as completely disjointed and we're trying to tackle this problem first with the notion of a unified fabric which is independent from any workloads whether this fabric spans on a switch which can be connected to bare metal workload or can span all the way inside the DPU where you have your multi hypervisor compute environment. It's one API, one common network control plane and one common set of segmentation services for the network. That's problem number one. >> It's interesting I hear you talking and I hear one network among different operating models. Reminds me of the old serverless days. There's still servers but they call it serverless. Is there going to be a term network-less because at the end of the day it should be one network, not multiple operating models. This is a problem that you guys are working on, is that right? I'm just joking serverless and network-less, but the idea is it should be one thing. >> Yeah, effectively what we're trying to do is we're trying to recompose this fragmentation in terms of network cooperation across physical networking and server networking. Server networking is where the majority of the problems are because as much as you have standardized the ways of building physical networks and cloud fabrics with IP protocols and internet, you don't have that sort of operational efficiency at the server layer. And this is what we're trying to attack first with this technology. The second aspect we're trying to attack is how we distribute security services throughout the infrastructure more efficiently whether it's micro-segmentation is a stateful firewall services or even encryption. Those are all capabilities enabled by the BlueField DPU technology. And we can actually integrate those capabilities directly into the network fabric limiting dramatically at least for east west traffic the sprawl of security appliances whether virtual or physical. That is typically the way people today segment and secure the traffic in the cloud. >> Awesome. Pete, all kidding aside about network-less and serverless kind of fun play on words there, the network is one thing it's basically distributed computing, right? So I'd love to get your thoughts about this distributed security with zero trust as the driver for this architecture you guys are doing. Can you share in more detail the depth of why DPU based approach is better than alternatives? >> Yeah, I think what's beautiful and kind of what the DPU brings that's new to this model is completely isolated compute environment inside. So it's the, yo dog, I heard you like a server so I put a server inside your server. And so we provide ARM CPUs, memory and network accelerators inside and that is completely isolated from the host. The actual X86 host just thinks it has a regular niche in there, but you actually have this full control plane thing. It's just like taking your top of rack switch and shoving it inside of your compute node. And so you have not only this separation within the data plane, but you have this complete control plane separation so you have this element that the network team can now control and manage, but we're taking all of the functions we used to do at the top of rack switch and we're distributing them now. And as time has gone on we've struggled to put more and more and more into that network edge. And the reality is the network edge is the compute layer, not the top of rack switch layer. And so that provides this phenomenal enforcement point for security and policy. And I think outside of today's solutions around virtual firewalls, the other option is centralized appliances. And even if you can get one that can scale large enough, the question is, can you afford it? And so what we end up doing is we kind of hope that NVIDIA's good enough or we hope that the VXLAN tunnel's good enough. And we can't actually apply more advanced techniques there because we can't financially afford that appliance to see all of the traffic. And now that we have a distributed model with this accelerator, we could do it. >> So what's in it for the customer real quick and I think this is an interesting point you mentioned policy. Everyone in networking knows policy is just a great thing. And as you hear it being talked about up the stack as well when you start getting to orchestrating microservices and whatnot all that good stuff going on there, containers and whatnot and modern applications. What's the benefit to the customers with this approach because what I heard was more scale, more edge, deployment flexibility relative to security policies and application enablement? What's the customer get out of this architecture? What's the enablement? >> It comes down to taking again the capabilities that we're in that top of rack switch and distributing them down. So that makes simplicity smaller, blast radius' for failures smaller failure domains, maintenance on the networks and the systems become easier. Your ability to integrate across workloads becomes infinitely easier. And again, we always want to kind of separate each one of those layers so just as in say a VXLAN network, my leaf in spine don't have to be tightly coupled together. I can now do this at a different layer and so you can run a DPU with any networking in the core there. And so you get this extreme flexibility. You can start small, you can scale large. To me the possibilities are endless. >> It's a great security control plan. Really flexibility is key and also being situationally aware of any kind of threats or new vectors or whatever's happening in the network. Alessandro, this is huge upside, right? You've already identified some successes with some customers on your early field trials. What are they doing and why are they attracted to the solution? >> Yeah, I think the response from customer has been the most encouraging and exciting for us to sort of continue and work and develop this product. And we have actually learned a lot in the process. We talked to tier two, tier three cloud providers. We talked to SP, Soft Telco type of networks as well as inter large enterprise customers. In one particular case one, let me call out a couple of examples here just to give you a flavor. There is a cloud provider in Asia who is actually managing a cloud where they're offering services based on multiple hypervisors. They are native services based on Zen, but they also on ramp into the cloud workloads based on ESXi and KVM depending on what the customer picks from the menu. And they have the problem of now orchestrating through their orchestrate or integrating with Zen center, with vSphere, with OpenStack to coordinate this multiple environments. And in the process to provide security, they actually deploy virtual appliances everywhere which has a lot of cost complication and eats up into the server CPU. The promise that they saw in this technology, they call it actually game changing is actually to remove all this complexity, having a single network and distribute the micro segmentation service directly into the fabric. And overall they're hoping to get out it tremendous OPEX benefit and overall operational simplification for the cloud infrastructure. That's one important use case. Another global enterprise customer is running both ESXi and Hyper-V environment and they don't have a solution to do micro segmentation consistently across hypervisors. So again, micro segmentation is a huge driver security. Looks like it's a recurring theme talking to most of these customers. And in the Telco space, we're working with few Telco customers on the CFT program where the main goal is actually to harmonize network cooperation. They typically handle all the VNFs with their own homegrown DPDK stack. This is overly complex. It is frankly also slow and inefficient. And then they have a physical network to manage. The idea of having again one network to coordinate the provisioning of cloud services between the Telco VNFs and the rest of the infrastructure is extremely powerful on top of the offloading capability opted by the BlueField DPUs. Those are just some examples. >> That was a great use case. A lot more potential I see that with the unified cloud networking, great stuff, Pete, shout out to you 'cause at NVIDIA we've been following your success us for a long time and continuing to innovate as cloud scales and Pluribus with unified networking kind of bring it to the next level. Great stuff, great to have you guys on and again, software keeps driving the innovation and again, networking is just a part of it and it's the key solution. So I got to ask both of you to wrap this up. How can cloud operators who are interested in this new architecture and solution learn more because this is an architectural shift? People are working on this problem, they're try to think about multiple clouds, they're try to think about unification around the network and giving more security, more flexibility to their teams. How can people learn more? >> Yeah, so Alessandro and I have a talk at the upcoming NVIDIA GTC conference. So it's the week of March 21st through 24th. You can go and register for free nvidia.com/gtc. You can also watch recorded sessions if you end up watching this on YouTube a little bit after the fact. And we're going to dive a little bit more into the specifics and the details and what we're providing in the solution. >> Alessandro, how can we people learn more? >> Yeah, absolutely. People can go to the Pluribus website, www.pluribusnetworks.com/eft and they can fill up the form and they will contact Pluribus to either know more or to know more and actually to sign up for the actual early field trial program which starts at the end of April. >> Okay, well, we'll leave it there. Thank you both for joining, appreciate it. Up next you're going to hear an independent analyst perspective and review some of the research from the enterprise strategy group ESG. I'm John Furrier with theCUBE, thanks for watching. (upbeat music)
SUMMARY :
Pete Lumbis, the director and NVIDIA partnership and the solution And the novelty about So the first byproduct of this approach on the X86. We heard that the DPU and the network operators have of the calling car for cloud, right? And so that ability to into the Pluribus unified and the second, is Reminds me of the old serverless days. and secure the traffic in the cloud. as the driver for this the data plane, but you have this complete What's the benefit to the and the systems become easier. to the solution? And in the process to provide security, and it's the key solution. and the details and what we're at the end of April. and review some of the research from
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Alessandro Barbieri | PERSON | 0.99+ |
Alessandro | PERSON | 0.99+ |
Asia | LOCATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Pluribus | ORGANIZATION | 0.99+ |
Telco | ORGANIZATION | 0.99+ |
Pluribus Networks | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
20% | QUANTITY | 0.99+ |
Pete Lumbis | PERSON | 0.99+ |
First | QUANTITY | 0.99+ |
ESXi | TITLE | 0.99+ |
March 21st | DATE | 0.99+ |
ESG | ORGANIZATION | 0.99+ |
Pete | PERSON | 0.99+ |
www.pluribusnetworks.com/eft | OTHER | 0.99+ |
second aspect | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
24th | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two things | QUANTITY | 0.98+ |
one network | QUANTITY | 0.98+ |
DevOps | TITLE | 0.98+ |
end of April | DATE | 0.98+ |
second | QUANTITY | 0.97+ |
vSphere | TITLE | 0.97+ |
Soft Telco | ORGANIZATION | 0.97+ |
Kubernetes | TITLE | 0.97+ |
today | DATE | 0.97+ |
YouTube | ORGANIZATION | 0.97+ |
tier three | QUANTITY | 0.96+ |
nvidia.com/gtc | OTHER | 0.96+ |
two major problems | QUANTITY | 0.95+ |
Zen | TITLE | 0.94+ |
around 20, 25% | QUANTITY | 0.93+ |
zero code | QUANTITY | 0.92+ |
each one | QUANTITY | 0.92+ |
X86 | COMMERCIAL_ITEM | 0.92+ |
OpenStack | TITLE | 0.92+ |
NetOps | TITLE | 0.92+ |
single network | QUANTITY | 0.92+ |
ARM | ORGANIZATION | 0.91+ |
one common set | QUANTITY | 0.89+ |
one API | QUANTITY | 0.88+ |
BlueField | ORGANIZATION | 0.87+ |
one important use case | QUANTITY | 0.86+ |
zero trust | QUANTITY | 0.86+ |
tier two | QUANTITY | 0.85+ |
Hyper-V | TITLE | 0.85+ |
one common network control plane | QUANTITY | 0.83+ |
BlueField | OTHER | 0.82+ |
Number one | QUANTITY | 0.81+ |
48 servers | QUANTITY | 0.8+ |
Ian Buck, NVIDIA | AWS re:Invent 2021
>>Well, welcome back to the cubes coverage of AWS reinvent 2021. We're here joined by Ian buck, general manager and vice president of accelerated computing at Nvidia I'm. John Ford, your host of the QB. And thanks for coming on. So in video, obviously, great brand congratulates on all your continued success. Everyone who has does anything in graphics knows the GPU's are hot and you guys get great brand great success in the company, but AI and machine learning was seeing the trend significantly being powered by the GPU's and other systems. So it's a key part of everything. So what's the trends that you're seeing, uh, in ML and AI, that's accelerating computing to the cloud. Yeah, >>I mean, AI is kind of drape bragging breakthroughs innovations across so many segments, so many different use cases. We see it showing up with things like credit card, fraud prevention and product and content recommendations. Really it's the new engine behind search engines is AI. Uh, people are applying AI to things like, um, meeting transcriptions, uh, virtual calls like this using AI to actually capture what was said. Um, and that gets applied in person to person interactions. We also see it in intelligence systems assistance for a contact center, automation or chat bots, uh, medical imaging, um, and intelligence stores and warehouses and everywhere. It's really, it's really amazing what AI has been demonstrated, what it can do. And, uh, it's new use cases are showing up all the time. >>Yeah. I'd love to get your thoughts on, on how the world's evolved just in the past few years, along with cloud, and certainly the pandemics proven it. You had this whole kind of full stack mindset initially, and now you're seeing more of a horizontal scale, but yet enabling this vertical specialization in applications. I mean, you mentioned some of those apps, the new enablers, this kind of the horizontal play with enablement for specialization, with data, this is a huge shift that's going on. It's been happening. What's your reaction to that? >>Yeah, it's the innovations on two fronts. There's a horizontal front, which is basically the different kinds of neural networks or AIS as well as machine learning techniques that are, um, just being invented by researchers for, uh, and the community at large, including Amazon. Um, you know, it started with these convolutional neural networks, which are great for image processing, but as it expanded more recently into, uh, recurrent neural networks, transformer models, which are great for language and language and understanding, and then the new hot topic graph neural networks, where the actual graph now is trained as a, as a neural network, you have this underpinning of great AI technologies that are being adventure around the world in videos role is try to productize that and provide a platform for people to do that innovation and then take the next step and innovate vertically. Um, take it, take it and apply it to two particular field, um, like medical, like healthcare and medical imaging applying AI, so that radiologists can have an AI assistant with them and highlight different parts of the scan. >>Then maybe troublesome worrying, or requires more investigation, um, using it for robotics, building virtual worlds, where robots can be trained in a virtual environment, their AI being constantly trained, reinforced, and learn how to do certain activities and techniques. So that the first time it's ever downloaded into a real robot, it works right out of the box, um, to do, to activate that we co we are creating different vertical solutions, vertical stacks for products that talk the languages of those businesses, of those users, uh, in medical imaging, it's processing medical data, which is obviously a very complicated large format data, often three-dimensional boxes in robotics. It's building combining both our graphics and simulation technologies, along with the, you know, the AI training capabilities and different capabilities in order to run in real time. Those are, >>Yeah. I mean, it's just so cutting edge. It's so relevant. I mean, I think one of the things you mentioned about the neural networks, specifically, the graph neural networks, I mean, we saw, I mean, just to go back to the late two thousands, you know, how unstructured data or object store created, a lot of people realize that the value out of that now you've got graph graph value, you got graph network effect, you've got all kinds of new patterns. You guys have this notion of graph neural networks. Um, that's, that's, that's out there. What is, what is a graph neural network and what does it actually mean for deep learning and an AI perspective? >>Yeah, we have a graph is exactly what it sounds like. You have points that are connected to each other, that established relationships and the example of amazon.com. You might have buyers, distributors, sellers, um, and all of them are buying or recommending or selling different products. And they're represented in a graph if I buy something from you and from you, I'm connected to those end points and likewise more deeply across a supply chain or warehouse or other buyers and sellers across the network. What's new right now is that those connections now can be treated and trained like a neural network, understanding the relationship. How strong is that connection between that buyer and seller or that distributor and supplier, and then build up a network that figure out and understand patterns across them. For example, what products I may like. Cause I have this connection in my graph, what other products may meet those requirements, or also identifying things like fraud when, when patterns and buying patterns don't match, what a graph neural networks should say would be the typical kind of graph connectivity, the different kind of weights and connections between the two captured by the frequency half I buy things or how I rate them or give them stars as she used cases, uh, this application graph neural networks, which is basically capturing the connections of all things with all people, especially in the world of e-commerce, it's very exciting to a new application, but applying AI to optimizing business, to reducing fraud and letting us, you know, get access to the products that we want, the products that they have, our recommendations be things that, that excited us and want us to buy things >>Great setup for the real conversation that's going on here at re-invent, which is new kinds of workloads are changing. The game. People are refactoring their business with not just replatform, but actually using this to identify value and see cloud scale allows you to have the compute power to, you know, look at a note on an arc and actually code that. It's all, it's all science, all computer science, all at scale. So with that, that brings up the whole AWS relationship. Can you tell us how you're working with AWS before? >>Yeah. 80 of us has been a great partner and one of the first cloud providers to ever provide GPS the cloud, uh, we most more recently we've announced two new instances, uh, the instance, which is based on the RA 10 G GPU, which has it was supports the Nvidia RTX technology or rendering technology, uh, for real-time Ray tracing and graphics and game streaming is their highest performance graphics, enhanced replicate without allows for those high performance graphics applications to be directly hosted in the cloud. And of course runs everything else as well, including our AI has access to our AI technology runs all of our AI stacks. We also announced with AWS, the G 5g instance, this is exciting because it's the first, uh, graviton or ARM-based processor connected to a GPU and successful in the cloud. Um, this makes, uh, the focus here is Android gaming and machine learning and France. And we're excited to see the advancements that Amazon is making and AWS is making with arm and the cloud. And we're glad to be part of that journey. >>Well, congratulations. I remember I was just watching my interview with James Hamilton from AWS 2013 and 2014. He was getting, he was teasing this out, that they're going to build their own, get in there and build their own connections, take that latency down and do other things. This is kind of the harvest of all that. As you start looking at these new new interfaces and the new servers, new technology that you guys are doing, you're enabling applications. What does, what do you see this enabling as this, as this new capability comes out, new speed, more, more performance, but also now it's enabling more capabilities so that new workloads can be realized. What would you say to folks who want to ask that question? >>Well, so first off I think arm is here to stay and you can see the growth and explosion of my arm, uh, led of course, by grab a tiny to be. I spend many others, uh, and by bringing all of NVIDIA's rendering graphics, machine learning and AI technologies to arm, we can help bring that innovation. That arm allows that open innovation because there's an open architecture to the entire ecosystem. Uh, we can help bring it forward, uh, to the state of the art in AI machine learning, the graphics. Um, we all have our software that we released is both supportive, both on x86 and an army equally, um, and including all of our AI stacks. So most notably for inference the deployment of AI models. We have our, the Nvidia Triton inference server. Uh, this is the, our inference serving software where after he was trained to model, he wanted to play it at scale on any CPU or GPU instance, um, for that matter. So we support both CPS and GPS with Triton. Um, it's natively integrated with SageMaker and provides the benefit of all those performance optimizations all the time. Uh, things like, uh, features like dynamic batching. It supports all the different AI frameworks from PI torch to TensorFlow, even a generalized Python code. Um, we're activating how activating the arm ecosystem as well as bringing all those AI new AI use cases and all those different performance levels, uh, with our partnership with AWS and all the different clouds. >>And you got to making it really easy for people to use, use the technology that brings up the next kind of question I want to ask you. I mean, a lot of people are really going in jumping in the big time into this. They're adopting AI. Either they're moving in from prototype to production. There's always some gaps, whether it's knowledge, skills, gaps, or whatever, but people are accelerating into the AI and leaning into it hard. What advancements have is Nvidia made to make it more accessible, um, for people to move faster through the, through the system, through the process? >>Yeah, it's one of the biggest challenges. The other promise of AI, all the publications that are coming all the way research now, how can you make it more accessible or easier to use by more people rather than just being an AI researcher, which is, uh, uh, obviously a very challenging and interesting field, but not one that's directly in the business. Nvidia is trying to write a full stack approach to AI. So as we make, uh, discover or see these AI technologies come available, we produce SDKs to help activate them or connect them with developers around the world. Uh, we have over 150 different STKs at this point, certain industries from gaming to design, to life sciences, to earth scientist. We even have stuff to help simulate quantum computing. Um, and of course all the, all the work we're doing with AI, 5g and robotics. So, uh, we actually just introduced about 65 new updates just this past month on all those SDKs. Uh, some of the newer stuff that's really exciting is the large language models. Uh, people are building some amazing AI. That's capable of understanding the Corpus of like human understanding, these language models that are trained on literally the continent of the internet to provide general purpose or open domain chatbots. So the customer is going to have a new kind of experience with a computer or the cloud. Uh, we're offering large language, uh, those large language models, as well as AI frameworks to help companies take advantage of this new kind of technology. >>You know, each and every time I do an interview with Nvidia or talk about Nvidia my kids and their friends, they first thing they said, you get me a good graphics card. Hey, I want the best thing in their rig. Obviously the gaming market's hot and known for that, but I mean, but there's a huge software team behind Nvidia. This is a well-known your CEO is always talking about on his keynotes, you're in the software business. And then you had, do have hardware. You were integrating with graviton and other things. So, but it's a software practices, software. This is all about software. Could you share kind of more about how Nvidia culture and their cloud culture and specifically around the scale? I mean, you, you hit every, every use case. So what's the software culture there at Nvidia, >>And it is actually a bigger, we have more software people than hardware people, people don't often realize this. Uh, and in fact that it's because of we create, uh, the, the, it just starts with the chip, obviously building great Silicon is necessary to provide that level of innovation, but as it expanded dramatically from then, from there, uh, not just the Silicon and the GPU, but the server designs themselves, we actually do entire server designs ourselves to help build out this infrastructure. We consume it and use it ourselves and build our own supercomputers to use AI, to improve our products. And then all that software that we build on top, we make it available. As I mentioned before, uh, as containers on our, uh, NGC container store container registry, which is accessible for me to bus, um, to connect to those vertical markets, instead of just opening up the hardware and none of the ecosystem in develop on it, they can with a low-level and programmatic stacks that we provide with Kuda. We believe that those vertical stacks are the ways we can help accelerate and advance AI. And that's why we make as well, >>Ram a little software is so much easier. I want to get that plug for, I think it's worth noting that you guys are, are heavy hardcore, especially on the AI side. And it's worth calling out, uh, getting back to the customers who are bridging that gap and getting out there, what are the metrics they should consider as they're deploying AI? What are success metrics? What does success look like? Can you share any insight into what they should be thinking about and looking at how they're doing? >>Yeah. Um, for training, it's all about time to solution. Um, it's not the hardware that that's the cost, it's the opportunity that AI can provide your business and many, and the productivity of those data scientists, which are developing, which are not easy to come by. So, uh, what we hear from customers is they need a fast time to solution to allow people to prototype very quickly, to train a model to convergence, to get into production quickly, and of course, move on to the next or continue to refine it often. So in training is time to solution for inference. It's about our, your ability to deploy at scale. Often people need to have real time requirements. They want to run in a certain amount of latency, a certain amount of time. And typically most companies don't have a single AI model. They have a collection of them. They want, they want to run for a single service or across multiple services. That's where you can aggregate some of your infrastructure leveraging the trading infant server. I mentioned before can actually run multiple models on a single GPU saving costs, optimizing for efficiency yet still meeting the requirements for latency and the real time experience so that your customers have a good, a good interaction with the AI. >>Awesome. Great. Let's get into, uh, the customer examples. You guys have obviously great customers. Can you share some of the use cases, examples with customers, notable customers? >>Yeah. I want one great part about working in videos as a technology company. You see, you get to engage with such amazing customers across many verticals. Uh, some of the ones that are pretty exciting right now, Netflix is using the G4 instances to CLA um, to do a video effects and animation content. And, you know, from anywhere in the world, in the cloud, uh, as a cloud creation content platform, uh, we work in the energy field that Siemens energy is actually using AI combined with, um, uh, simulation to do predictive maintenance on their energy plants, um, and, and, uh, doing preventing or optimizing onsite inspection activities and eliminating downtime, which is saving a lot of money for the engine industry. Uh, we have worked with Oxford university, uh, which is Oxford university actually has over two, over 20 million artifacts and specimens and collections across its gardens and museums and libraries. They're actually using convenient GPS and Amazon to do enhance image recognition, to classify all these things, which would take literally years with, um, uh, going through manually each of these artifacts using AI, we can click and quickly catalog all of them and connect them with their users. Um, great stories across graphics, about cross industries across research that, uh, it's just so exciting to see what people are doing with our technology together with, >>And thank you so much for coming on the cube. I really appreciate Greg, a lot of great content there. We probably going to go another hour, all the great stuff going on in the video, any closing remarks you want to share as we wrap this last minute up >>Now, the, um, really what Nvidia is about as accelerating cloud computing, whether it be AI, machine learning, graphics, or headphones, community simulation, and AWS was one of the first with this in the beginning, and they continue to bring out great instances to help connect, uh, the cloud and accelerated computing with all the different opportunities integrations with with SageMaker really Ks and ECS. Uh, the new instances with G five and G 5g, very excited to see all the work that we're doing together. >>Ian buck, general manager, and vice president of accelerated computing. I mean, how can you not love that title? We want more, more power, more faster, come on. More computing. No, one's going to complain with more computing know, thanks for coming on. Thank you. Appreciate it. I'm John Farrell hosted the cube. You're watching Amazon coverage reinvent 2021. Thanks for watching.
SUMMARY :
knows the GPU's are hot and you guys get great brand great success in the company, but AI and machine learning was seeing the AI. Uh, people are applying AI to things like, um, meeting transcriptions, I mean, you mentioned some of those apps, the new enablers, Yeah, it's the innovations on two fronts. technologies, along with the, you know, the AI training capabilities and different capabilities in I mean, I think one of the things you mentioned about the neural networks, You have points that are connected to each Great setup for the real conversation that's going on here at re-invent, which is new kinds of workloads And we're excited to see the advancements that Amazon is making and AWS is making with arm and interfaces and the new servers, new technology that you guys are doing, you're enabling applications. Well, so first off I think arm is here to stay and you can see the growth and explosion of my arm, I mean, a lot of people are really going in jumping in the big time into this. So the customer is going to have a new kind of experience with a computer And then you had, do have hardware. not just the Silicon and the GPU, but the server designs themselves, we actually do entire server I want to get that plug for, I think it's worth noting that you guys are, that that's the cost, it's the opportunity that AI can provide your business and many, Can you share some of the use cases, examples with customers, notable customers? research that, uh, it's just so exciting to see what people are doing with our technology together with, all the great stuff going on in the video, any closing remarks you want to share as we wrap this last minute up Uh, the new instances with G one's going to complain with more computing know, thanks for coming on.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Ian buck | PERSON | 0.99+ |
John Farrell | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Ian Buck | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Ian buck | PERSON | 0.99+ |
Greg | PERSON | 0.99+ |
2014 | DATE | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
John Ford | PERSON | 0.99+ |
James Hamilton | PERSON | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
G five | COMMERCIAL_ITEM | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Python | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
G 5g | COMMERCIAL_ITEM | 0.99+ |
first | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Android | TITLE | 0.99+ |
Oxford university | ORGANIZATION | 0.99+ |
2013 | DATE | 0.98+ |
amazon.com | ORGANIZATION | 0.98+ |
over two | QUANTITY | 0.98+ |
two | QUANTITY | 0.98+ |
first time | QUANTITY | 0.97+ |
single service | QUANTITY | 0.97+ |
2021 | DATE | 0.97+ |
two fronts | QUANTITY | 0.96+ |
single | QUANTITY | 0.96+ |
over 20 million artifacts | QUANTITY | 0.96+ |
each | QUANTITY | 0.95+ |
about 65 new updates | QUANTITY | 0.93+ |
Siemens energy | ORGANIZATION | 0.92+ |
over 150 different STKs | QUANTITY | 0.92+ |
single GPU | QUANTITY | 0.91+ |
two new instances | QUANTITY | 0.91+ |
first thing | QUANTITY | 0.9+ |
France | LOCATION | 0.87+ |
two particular field | QUANTITY | 0.85+ |
SageMaker | TITLE | 0.85+ |
Triton | TITLE | 0.82+ |
first cloud providers | QUANTITY | 0.81+ |
NGC | ORGANIZATION | 0.77+ |
80 of | QUANTITY | 0.74+ |
past month | DATE | 0.68+ |
x86 | COMMERCIAL_ITEM | 0.67+ |
late | DATE | 0.67+ |
two thousands | QUANTITY | 0.64+ |
pandemics | EVENT | 0.64+ |
past few years | DATE | 0.61+ |
G4 | ORGANIZATION | 0.6+ |
RA | COMMERCIAL_ITEM | 0.6+ |
Kuda | ORGANIZATION | 0.59+ |
ECS | ORGANIZATION | 0.55+ |
10 G | OTHER | 0.54+ |
SageMaker | ORGANIZATION | 0.49+ |
TensorFlow | OTHER | 0.48+ |
Ks | ORGANIZATION | 0.36+ |
Breaking Analysis: How Nvidia Wins the Enterprise With AI
from the cube studios in palo alto in boston bringing you data-driven insights from the cube and etr this is breaking analysis with dave vellante nvidia wants to completely transform enterprise computing by making data centers run 10x faster at one tenth the cost and video's ceo jensen wang is crafting a strategy to re-architect today's on-prem data centers public clouds and edge computing installations with a vision that leverages the company's strong position in ai architectures the keys to this end-to-end strategy include a clarity of vision massive chip design skills a new arm-based architecture approach that integrates memory processors i o and networking and a compelling software consumption model even if nvidia is unsuccessful at acquiring arm we believe it will still be able to execute on this strategy by actively participating in the arm ecosystem however if its attempts to acquire arm are successful we believe it will transform nvidia from the world's most valuable chip company into the world's most valuable supplier of integrated computing architectures hello everyone and welcome to this week's wikibon cube insights powered by etr in this breaking analysis we'll explain why we believe nvidia is in the right position to power the world's computing centers and how it plans to disrupt the grip that x86 architectures have had on the data center for decades the data center market is in transition like the universe the cloud is expanding at an accelerated pace no longer is the cloud an opaque set of remote services i always say somewhere out there sitting in a mega data center no rather the cloud is extending to on-premises data centers data centers are moving into the cloud and they're connecting through adjacent locations that create hybrid interactions clouds are being meshed together across regions and eventually will stretch to the far edge this new definition or view of cloud will be hyper distributed and run by software kubernetes is changing the world of software development and enabling workloads to run anywhere open apis external applications expanding the digital supply chains and this expanding cloud they all increase the threat surface and vulnerability to the most sensitive information that resides within the data center and around the world zero trust has become a mandate we're also seeing ai being injected into every application and it's the technology area that we see with the most momentum coming out of the pandemic this new world will not be powered by general purpose x86 processors rather it will be supported by an ecosystem of arm-based providers in our opinion that are affecting an unprecedented increase in processor performance as we have been reporting and nvidia in our view is sitting in the poll position and is currently the favorite to dominate the next era of computing architecture for global data centers public clouds as well as the near and far edge let's talk about jensen wang's clarity of vision for this new world here's a chart that underscores some of the fundamental assumptions that he's leveraging to expand his market the first is that there's a lot of waste in the data center he claims that only half of the cpu cores deployed in the data center today actually support applications the other half are processing the infrastructure all around the applications that run the software defined data center and they're terribly under utilized nvidia's blue field three dpu the data processing unit was described in a blog post on siliconangle by analyst zias caravala as a complete mini server on a card i like that with software defined networking storage and security acceleration built in this product has the bandwidth and according to nvidia can replace 300 general purpose x86 cores jensen believes that every network chip will be intelligent programmable and capable of this type of acceleration to offload conventional cpus he believes that every server node will have this capability and enable every packed of every packet and every application to be monitored in real time all the time for intrusion and as servers move to the edge bluefield will be included as a core component in his view and this last statement by jensen is critical in our opinion he says ai is the most powerful force of our time whether you agree with that or not it's relevant because ai is everywhere an invidious position in ai and the architectures the company is building are the fundamental linchpin of its data center enterprise strategy so let's take a look at some etr spending data to see where ai fits on the priority list here's a set of data in a view that we often like to share the horizontal axis is market share or pervasiveness in the etr data but we want to call your attention to the vertical axis that's really really what really we want to pay attention today that's net score or spending momentum exiting the pandemic we've seen ai capture the number one position in the last two surveys and we think this dynamic will continue for quite some time as ai becomes the staple of digital transformations and automations an ai will be infused in every single dot you see on this chart nvidia's architectures it just so happens are tailor made for ai workloads and that is how it will enter these markets let's quantify what that means and lay out our view of how nvidia with the help of arm will go after the enterprise market here's some data from wikibon research that depicts the percent of worldwide spending on server infrastructure by workload type here are the key points first the market last year was around 78 billion dollars worldwide and is expected to approach 115 billion by the end of the decade this might even be a conservative figure and we've split the market into three broad workload categories the blue is ai and other related applications what david floyer calls matrix workloads the orange is general purpose think things like erp supply chain hcm collaboration basically oracle saps and microsoft work that's being supported today and of course many other software providers and the gray that's the area that jensen was referring to is about being wasted the offload work for networking and storage and all the software defined management in the data centers around the world okay you can see the squeeze that we think compute infrastructure is gonna gonna occur around that orange area that general-purpose workloads that we think is going to really get squeezed in the next several years on a percentage basis and on an absolute basis it's really not growing nearly as fast as the other two and video with arm in our view is well positioned to attack that blue area and the gray area those those workload offsets and the new emerging ai applications but even the orange as we've reported is under pressure as for example companies like aws and oracle they use arm-based designs to service general purpose workloads why are they doing that cost is the reason because x86 generally and intel specifically are not delivering the price performance and efficiency required to keep up with the demands to reduce data center costs and if intel doesn't respond which we believe it will but if it doesn't act arm we think will get 50 percent of the general purpose workloads by the end of the decade and with nvidia it will dominate the blue the ai and the gray the offload work when we say dominate we're talking like capture 90 percent of the available market if intel doesn't respond now intel they're not just going to sit back and let that happen pat gelsinger is well aware of this in moving intel to a new strategy but nvidia and arm are way ahead in the game in our view and as we've reported this is going to be a real challenge for intel to catch up now let's take a quick look at what nvidia is doing with relevant parts of its pretty massive portfolio here's a slide that shows nvidia's three chip strategy the company is shifting to arm-based architectures which we'll describe in more detail in a moment the slide shows at the top line nvidia's ampere architecture not to be confused with the company ampere computing nvidia is taking a gpu centric approach no surprise obvious reasons there that's their sort of stronghold but we think over time it may rethink this a little bit and lean more into npus the neural processing unit we look at what apple's doing what tesla are doing we see opportunities for companies like nvidia to really sort of go after that but we'll save that for another day nvidia has announced its grace cpu a nod to the famous computer scientist grace hopper grace is a new architecture that doesn't rely on x86 and much more efficiently uses memory resources we'll again describe this in more detail later and the bottom line there that roadmap line shows the bluefield dpu which we described is essentially a complete server on a card in this approach using arm will reduce the elapsed time to go from chip design to production by 50 we're talking about shaving years down to 18 months or less we don't have time to do a deep dive into nvidia's portfolio it's large but we want to share some things that we think are important and this next graphic is one of them this shows some of the details of nvidia's jetson architecture which is designed to accelerate those ai plus workloads that we showed earlier and the reason is that this is important in our view is because the same software supports from small to very large including edge systems and we think this type of architecture is very well suited for ai inference at the edge as well as core data center applications that use ai and as we've said before a lot of the action in ai is going to happen at the edge so this is a good example of leveraging an architecture across a wide spectrum of performance and cost now we want to take a moment to explain why the moved arm-based architectures is so critical to nvidia one of the biggest cost challenges for nvidia today is keeping the gpu utilized typical utilization of gpu is well below 20 percent here's why the left hand side of this chart shows essentially racks if you will of traditional compute and the bottlenecks that nvidia faces the processor and dram they're tied together in separate blocks imagine there are thousands thousands of cores in a rack and every time you need data that lives in another processor you have to send a request and go retrieve it it's very overhead intensive now technologies like rocky are designed to help but it doesn't solve the fundamental architectural bottleneck every gpu shown here also has its own dram and it has to communicate with the processors to get the data i.e they can't communicate with each other efficiently now the right hand side side shows where nvidia is headed start in the middle with system on chip socs cpus are packaged in with npus ipu's that's the image processing unit you know x dot dot dot x pu's the the alternative processors they're all connected with sram which is think of that as a high speed layer like an layer one cache the os for the system on a chip lives inside of this and that's where nvidia has this killer software model what they're doing is they're licensing the consumption of the operating system that's running this system on chip in this entire system and they're affecting a new and really compelling subscription model you know maybe they should just give away the chips and charge for the software like a razer blade model talk about disruptive now the outer layer is the the dpu and the shared dram and other resources like the ampere computing the company this time cpus ssds and other resources these are the processors that will manage the socs together this design is based on nvidia's three chip approach using bluefield dpu leveraging melanox that's the networking component the network enables shared dram across the cpus which will eventually be all arm based grace lives inside the system on a chip and also on the outside layers and of course the gpu lives inside the soc in a scaled-down version like for instance a rendering gpu and we show some gpus on the outer layer as well for ai workloads at least in the near term you know eventually we think they may reside solely in the system on chip but only time will tell okay so you as you can see nvidia is making some serious moves and by teaming up with arm and leaning into the arm ecosystem it plans to take the company to its next level so let's talk about how we think competition for the next era of compute stacks up here's that same xy graph that we love to show market share or pervasiveness on the horizontal tracking against next net score on the vertical net score again is spending velocity and we've cut the etr data to capture players that are that are big in compute and storage and networking we've plugged in a couple of the cloud players these are the guys that we feel are vying for data center leadership around compute aws is a very strong position we believe that more than half of its revenues comes from compute you know ec2 we're talking about more than 25 billion on a run rate basis that's huge the company designs its own silicon graviton 2 etc and is working with isvs to run general purpose workloads on arm-based graviton chips microsoft and google they're going to follow suit they're big consumers of compute they sell a lot but microsoft in particular you know they're likely to continue to work with oem partners to attack that on-prem data center opportunity but it's really intel that's the provider of compute to the likes of hpe and dell and cisco and the odms which are the odms are not shown here now hpe let's talk about them for a second they have architectures and i hate to bring it up but remember the machine i know it's the butt of many jokes especially from competitors it had been you know frankly hpe and hp they deserve some of that heat for all the fanfare and then that they they put out there and then quietly you know pulled the machine or put it out the pasture but hpe has a strong position in high performance computing and the work that it did on new computing architectures with the machine and shared memories that might be still kicking around somewhere inside of hp and could come in handy for some day in the future so hpe has some chops there plus hpe has been known hp historically has been known to design its own custom silicon so i would not count them out as an innovator in this race cisco is interesting because it not only has custom silicon designs but its entry into the compute business with ucs a decade ago was notable and they created a new way to think about integrating resources particularly compute and networking with partnerships to add in the storage piece initially it was within within emc prior to the dell acquisition but you know it continues with netapp and pure and others cisco invests they spend money investing in architectures and we expect the next generation of ucs oh ucs2 ucs 2.0 will mark another notable milestone in the company's data center business dell just had an amazing quarterly earnings report the company grew top line revenue by around 12 percent and it wasn't because of an easy compare to last year dells is simply executing despite continued softness in the legacy emc storage business laptop the laptop demand continued to soar in dell server business it's growing again but we don't see dell as an architectural innovator per se in compute rather we think the company will be content to partner with suppliers whether it's intel nvidia arm-based partners or all of the above dell we think will rely on its massive portfolio its excellent supply chain and execution ethos to compete now ibm is notable for historical reasons with its mainframe ibm created the first great compute monopoly before it unwind and wittingly handed it to intel along with microsoft we don't see ibm necessarily aspiring to retake that compute platform mantle that once once held with mainframes rather red hat in the march to hybrid cloud is the path that we think in our view is ibm's approach now let's get down to the elephants in the room intel nvidia and china inc china is of course relevant because of companies like alibaba and huawei and the chinese chinese government's desire to be self-sufficient in semiconductor technology and technology generally but our premise here is that the trends are favoring nvidia over intel in this picture because nvidia is making moves to further position itself for new workloads in the data center and compete for intel's stronghold intel is going to attempt to remake itself but it should have been doing this seven years ago what pat gelsinger is doing today intel is simply far behind and it's going to take at least a couple years for them to really start to to make inroads in this new model let's stay on the nvidia v intel comparison for a moment and take a snapshot of the two companies here's a quick chart that we put together with some basic kpis some of these figures are approximations or they're rounded so don't stress over it too much but you can see intel is an 80 billion dollar company 4x the size of nvidia but nvidia's market cap far exceeds that of intel why is that of course growth in our view it's justified due to that growth and nvidia's strategic positioning intel used to be the gross margin king but nvidia has much higher gross margins interesting now when it comes down to free cash flow intel is still dominant as it pertains to the balance sheet intel is way more capital intensive than nvidia and as it starts to build out its foundries that's going to eat into intel's cash position now what we did is we put together a little pro forma on the third column of nvidia plus arm circa let's say the end of 2022. we think they could get to a run rate that is about half the size of intel and that can propel the company's market cap to well over half a trillion dollars if they get any credit for arm they're paying 40 billion dollars for arm a company that's you know sub 2 billion the risk is that because of the arm because the arm deal is based on cash plus tons of stock it could put pressure on the market capitalization for some time arm has 90 percent gross margins because it pretty much has a pure license model so it helps the gross margin line a little bit for this in this pro forma and the balance sheet is a swag arm has said that it's not going to take on debt to do the transaction but we haven't had time to really dig into that and figure out how they're going to structure it so we took a took a swag in in what we would do with this low interest rate environment but but take that with a grain of salt we'll do more research in there the point is given the momentum and growth of nvidia its strategic position in ai is in its deep engineering they're aimed at all the right places and its potential to unlock huge value with arm on paper it looks like the horse to beat if it can execute all right let's wrap up here's a summary look the architectures on which nvidia is building its dominant ai business are evolving and nvidia is well positioned to drive a truck right to the enterprise in our view the power has shifted from intel to the arm ecosystem and nvidia is leaning in big time whereas intel it has to preserve its current business while recreating itself at the same time this is going to take a couple of years but intel potentially has the powerful backing of the us government too strategic to fail the wild card is will nvidia be successful in acquiring arm certain factions in the uk and eu are fighting the deal because they don't want the u.s dictating to whom arm can sell its technology for example the restrictions placed on huawei for many suppliers of arm-based chips based on u.s sanctions nvidia's competitors like broadcom qualcomm at all are nervous that if nvidia gets armed they will be at a competitive disadvantage they being invidious competitors and for sure china doesn't want nvidia controlling arm for obvious reasons and it will do what it can to block the deal and or put handcuffs on how business can be done in china we can see a scenario where the u.s government pressures the uk and eu regulators to let this deal go through look ai and semiconductors you can't get much more strategic than that for the u.s military and the u.s long-term competitiveness in exchange for maybe facilitating the deal the government pressures nvidia to guarantee some feed to the intel foundry business while at the same time imposing conditions that secure access to arm-based technology for nvidia's competitors and maybe as we've talked about before having them funnel business to intel's foundry actually we've talked about the us government enticing apple to do so but it could also entice nvidia's competitors to do so propping up intel's foundry business which is clearly starting from ground zero and is going to need help outside of intel's own semiconductor manufacturing internally look we don't have any inside information as to what's happening behind the scenes with the us government and so forth but on its earning call on its earnings call nvidia said they're working with regulators that are on track to complete the deal in early 2022. we'll see okay that's it for today thank you to david floyer who co-created this episode with me and remember i publish each week on wikibon.com and siliconangle.com these episodes they're all available as podcasts all you're going to do is search breaking analysis podcast and you can always connect with me on twitter at dvalante or email me at david.valante siliconangle.com i always appreciate the comments on linkedin and in the clubhouse please follow me so you can be notified when we start a room and riff on these topics and don't forget to check out etr.plus for all the survey data this is dave vellante for the cube insights powered by etr be well and we'll see you next time [Music] you
SUMMARY :
and it's the technology area that we see
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
alibaba | ORGANIZATION | 0.99+ |
nvidia | ORGANIZATION | 0.99+ |
50 percent | QUANTITY | 0.99+ |
90 percent | QUANTITY | 0.99+ |
huawei | ORGANIZATION | 0.99+ |
microsoft | ORGANIZATION | 0.99+ |
david floyer | PERSON | 0.99+ |
40 billion dollars | QUANTITY | 0.99+ |
china | LOCATION | 0.99+ |
thousands | QUANTITY | 0.99+ |
18 months | QUANTITY | 0.99+ |
apple | ORGANIZATION | 0.99+ |
david.valante | OTHER | 0.99+ |
last year | DATE | 0.99+ |
two companies | QUANTITY | 0.99+ |
boston | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
10x | QUANTITY | 0.99+ |
early 2022 | DATE | 0.99+ |
jensen | PERSON | 0.99+ |
ibm | ORGANIZATION | 0.99+ |
around 78 billion dollars | QUANTITY | 0.99+ |
third column | QUANTITY | 0.99+ |
80 billion dollar | QUANTITY | 0.99+ |
more than half | QUANTITY | 0.99+ |
uk | LOCATION | 0.99+ |
first | QUANTITY | 0.98+ |
around 12 percent | QUANTITY | 0.98+ |
a decade ago | DATE | 0.98+ |
115 billion | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
each week | QUANTITY | 0.97+ |
dells | ORGANIZATION | 0.97+ |
seven years ago | DATE | 0.97+ |
50 | QUANTITY | 0.97+ |
dell | ORGANIZATION | 0.97+ |
jensen wang | PERSON | 0.97+ |
two | QUANTITY | 0.97+ |
end of 2022 | DATE | 0.97+ |
over half a trillion dollars | QUANTITY | 0.97+ |
siliconangle.com | OTHER | 0.96+ |
intel | ORGANIZATION | 0.96+ |
Kevin Deierling, NVIDIA and Scott Tease, Lenovo | CUBE Conversation, September 2020
>> Narrator: From theCUBE studios in Palo Alto in Boston, connecting with thought leaders all around the world, this is a CUBE conversation. >> Hi, I'm Stu Miniman, and welcome to a CUBE conversation. I'm coming to you from our Boston Area studio. And we're going to be digging into some interesting news regarding networking. Some important use cases these days, in 2020, of course, AI is a big piece of it. So happy to welcome to the program. First of all, I have one of our CUBE alumni, Kevin Deierling. He's the Senior Vice President of Marketing with Nvidia, part of the networking team there. And joining him is Scott Tease, someone we've known for a while, but first time on the program, who's the General Manager of HPC and AI, for the Lenovo Data Center Group. Scott and Kevin, thanks so much for joining us. >> It's great to be here Stu. >> Yeah, thank you. >> Alright, so Kevin, as I said, you you've been on the program a number of times, first when it was just Mellanox, now of course the networking team, there's some other acquisitions that have come in. If you could just set us up with the relationship between Nvidia and Lenovo. And there's some news today that we're here to talk about too. So let's start getting into that. And then Scott, you'll jump in after Kevin. >> Yeah, so we've been a long time partner with Lenovo, on our high performance computing. And so that's the InfiniBand piece of our business. And more and more, we're seeing that AI workloads are very, very similar to HPC workloads. And so that's been a great partnership that we've had for many, many years. And now we're expanding that, and we're launching a OEM relationship with Lenovo, for our Ethernet switches. And again, with our Ethernet switches, we really take that heritage of low latency, high performance networking that we built over many years in HPC, and we bring that to Ethernet. And of course that can be with HPC, because frequently in an HPC supercomputing environment, or in an AI supercomputing environment, you'll also have an Ethernet network, either for management, or sometimes for storage. And now we can offer that together with Lenovo. So it's a great partnership. We talked about it briefly last month, and now we're coming to market, and we'll be able to offer this to the market. >> Yeah, yeah, Kevin, we're super excited about it here in Lenovo as well. We've had a great relationship over the years with Mellanox, with Nvidia Mellanox. And this is just the next step. We've shown in HPC that the days of just taking an Ethernet card, or an InfiniBand card, plugging it in the system, and having it work properly are gone. You really need a system that's engineered for whatever task the customer is going to use. And we've known that in HPC for a long time, as we move into workloads, like artificial intelligence, where networking is a critical aspect of getting these systems to communicate with one another, and work properly together. We love from HPC perspective, to use InfiniBand, but most enterprise clients are using Ethernet. So where do we go? We go to a partner that we've trusted for a very long time. And we selected the Nvidia Mellanox Ethernet switch family. And we're really excited to be able to bring that end-to-end solution to our enterprise clients, just like we've been doing for HPC for a while. >> Yeah, well Scott, maybe if you could. I'd love to hear a little bit more about kind of that customer demand that those usages there. So you think traditionally, of course, is supercomputing, as you both talked about that move from InfiniBand, to leveraging Ethernet, is something that's been talked about for quite a while now in the industry. But maybe that AI specifically, could you talk about what are the networking requirements, how similar is it? Is it 95% of the same architecture, as what you see in HPC environments? And also, I guess the big question there is, how fast are customers adopting, and rolling out those AI solutions? And what kind of scale are they getting them to today? >> So yeah, there's a lot there of good things we can talk about. So I'd say in HPC, the thing that we've learned, is that you've got to have a fabric that's up to the task. When you're testing an HPC solution, you're not looking at a single node, you're looking at a combination of servers, and storage, management, all these things have to come together, and they come together over InfiniBand fabric. So we've got this nearly a purpose built fabric that's been fantastic for the HPC community for a long time. As we start to do some of that same type of workload, but in an enterprise environment, many of those customers are not used to InfiniBand, they're used to an Ethernet fabric, something that they've got all throughout their data center. And we want to try to find a way to do was, bring a lot of that rock solid interoperability, and pre-tested capability, and bring it to our enterprise clients for these AI workloads. Anything high performance GPUs, lots of inner internode communications, worries about traffic and congestion, abnormalities in the network that you need to spot. Those things happen quite often, when you're doing these enterprise AI solutions. You need a fabric that's able to keep up with that. And the Nvidia networking is definitely going to be able to do that for us. >> Yeah well, Kevin I heard Scott mention GPUs here. So this kind of highlights one of the reasons why we've seen Nvidia expand its networking capabilities. Could you talk a little bit about that kind of expansion, the portfolio, and how these use cases really are going to highlight what Nvidia helped bring to the market? >> Yeah, we like to really focus on accelerated computing applications. And whether those are HPC applications, or now they're becoming much more broadly adopted in the enterprise. And one of the things we've done is, tight integration at a product level, between GPUs, and the networking components in our business. Whether that's the adapters, or the DPU, the data processing unit, which we've talked about before. And now even with the switches here, with our friends at Lenovo, and really bringing that all together. But most importantly, is at a platform level. And by that I mean the software. And the enterprise here has all kinds of different verticals that are going after. And we invest heavily in the software ecosystem that's built on top of the GPU, and the networking. And by integrating all of that together on a platform, we can really accelerate the time to market for enterprises that wants to leverage these modern workloads, sort of cloud native workloads. >> Yeah, please Scott, if you have some follow up there. >> Yeah, if you don't mind Stu, I just like to say, five years ago, the roadmap that we followed was the processor roadmap. We all could tell you to the week when the next Xeon processor was going to come out. And that's what drove all of our roadmaps. Since that time what we found is that the items that are making the radical, the revolutionary improvements in performance, they're attached to the processor, but they're not the processor itself. It's things like, the GPU. It's things like that, especially networking adapters. So trying to design a platform that's solely based on a CPU, and then jam these other items on top of it. It no longer works, you have to design these systems in a holistic manner, where you're designing for the GPU, you're designing for the network. And that's the beauty of having a deep partnership, like we share with Nvidia, on both the GPU side, and on the networking side, is we can do all that upfront engineering to make sure that the platform, the systems, the solution, as a whole works exactly how the customer is going to expect it to. >> Kevin, you mentioned that a big piece of this is software now. I'm curious, there's an interesting piece that your networking team has picked up, relatively recently, that the Cumulus Linux, so help us understand how that fits into the Ethernet portfolio? And would it show up in these kind of applications that we're talking about? >> Yeah, that's a great question. So you're absolutely right, Cumulus is integral to what we're doing here with Lenovo. If you looked at the heritage that Mellanox had, and Cumulus, it's all about open networking. And what we mean by that, is we really decouple the hardware, and the software. So we support multiple network operating systems on top of our hardware. And so if it's, for example, Sonic, or if it's our Onyx or Dents, which is based on switch def. But Cumulus who we just recently acquired, has been also on that same access of open networking. And so they really support multiple platforms. Now we've added a new platform with our friends at Lenovo. And really they've adopted Cumulus. So it is very much centered on, Enterprise, and really a cloud like experience in the Enterprise, where it's Linux, but it's highly automated. Everything is operationalized and automated. And so as a result of that, you get sort of the experience of the cloud, but with the economics that you get in the Enterprise. So it's kind of the best of both worlds in terms of network analytic, and all of the ability to do things that the cloud guys are doing, but fully automated, and for an Enterprise environment. >> Yeah, so Kevin, I mean, I just want to say a few things about this. We're really excited about the Cumulus acquisition here. When we started our negotiations with Mellanox, we were still planning to use Onyx. We love Onyx, it's been our IB nodes of choice. Our users love, our are architects love it. But we were trying to lean towards a more open kind of futuristic, node as we got started with this. And Cumulus is really perfect. I mean it's a Linux open source based system. We love open source in HPC. The great thing about it is, we're going to be able to take all the great learnings that we've had with Onyx over the years, and now be able to consolidate those inside of Cumulus. We think it's the perfect way to start this relationship with Nvidia networking. >> Well Scott, help us understand a little more. What you know what does this expansion of the partnership mean? If you're talking about really the full solutions that Lenovo opens in the think agile brand, as well as the hybrid and cloud solutions. Is this something then that, is it just baked into the solution, is it a reseller, what should customers, and your your channel partners understand about this? >> Yeah, so any of the Lenovo solutions that require a switch to perform the functionality needed across the solution, are going to show up with the networking from Nvidia inside of it. Reasons for that, a couple of reasons. One is even something as simple as solution management for HPC, the switch is so integral to how we do all that, how we push all those functions down, how we deploy systems. So you've got to have a switch, in a connectivity methodology, that ensures that we know how to deploy these systems. And no matter what scale they are, from a few systems up, to literally thousands of systems, we've got something that we know how to do. Then when we're we're selling these solutions, like an SAP solution, for instance. The customer is not buying a server anymore, they're buying a solution, they're buying a functionality. And we want to be able to test that in our labs to ensure that that system, that rack, leaves our factory ready to do exactly what the customer is looking for. So any of the systems that are going to be coming from us, pre configured, pre tested, are all going to have Nvidia networking inside of them. >> Yeah, and I think that's, you mentioned the hybrid cloud. I think that's really important. That's really where we cut our teeth first in InfiniBand, but also with our Ethernet solutions. And so today, we're really driving a bunch of the big hyper scalars, as well as the big clouds. And as you see things like SAP or Azure, it's really important now that you're seeing Azure stack coming into a hybrid environment, that you have the known commodity here. So we're something that we're built in to many of those different platforms, with our Spectrum ASIC, as well as our adapters. And so now the ability with Nvidia, and Lenovo together, to bring that to enterprise customers, is really important. I think it's a proven set of components that together forms a solution. And that's the real key, as Scott said, is delivering a solution, not just piece parts, we have a platform, that software, hardware, all of it integrated. >> Well, it's great to see you. We've had an existing partnership for a while. I want to give you both the opportunity, anything specific, you've been hearing kind of the customer demand leading up this. Is it people that might be transitioning from InfiniBand to Ethernet? Or is it just general market adoption of new solutions that you have out there? (speakers talk over each other) >> You go ahead and start. >> Okay, so I think that there's different networks for different workloads, is what we've seen. And InfiniBand certainly is going to continue to be the best platform out there for HPC, and often for AI. But as Scott said, the enterprise frequently is not familiar with that, and for various reasons, would like to leverage Ethernet. So I think we'll see two different cases, one where there's Ethernet with an InfiniBand network. And the other is for new enterprise workloads that are coming, that are very AI centric, modern workloads, sort of cloud native workloads. You have all of the infrastructure in place with our Spectrum ASICs, and our Connectx adapters, and now integrated with GPUs, that we'll be able to deliver solutions rather than just compliments. And that's the key. >> Yeah, I think Stu, a great example, I think of where you need that networking, like we've been used to an HPC, is when you start looking at deep learning in training, scale out training. A lot of companies have been stuck on a single workstation, because they haven't been able to figure out how to spread that workload out, and chop it up, like we've been doing in HPC, because they've been running into networking issues. They can't run over an unoptimized network. With this new technology, we're hoping to be able to do a lot of the same things that HPC customers take for granted every day, about workload management, distribution of workload, chopping jobs up into smaller portions, and feeding them out to a cluster. We're hoping that we're going to be able to do those exact same things for our enterprise clients. And it's going to look magical to them, but it's the same kind of thing we've been doing forever. With Mellanox, in the past, now Nvidia networking, we're just going to take that to the enterprise. I'm really excited about it. >> Well, it's so much flexibility. We used to look at, it would take a decade to roll out some new generations. Kevin, if you could just give us latest speeds and feeds. If I look at Ethernet, did I see that this has from n gig, all the way up to 400 gig? I think I lose track a little bit of some of the pieces. I know the industry as a whole is driving it. But where are we with the general customer adoption of some of the some of the speeds today? >> Yeah indeed, we're coming up on the 40th anniversary of the first specification of Ethernet. And we're about 4000 times faster now, 40,000 times faster at 400 gigabits, versus 10 megabits. So yeah, we're shipping today at the adapter level, 100 gig, and even 200 gig. And then at the switch level, 400 gig. And people sort of ask, "Do we really need all that performance?" The answer is absolutely. So the amount of data that the GPU can crunch, and these AI workloads, these giant neural networks, it needs massive amounts of data. And then as you're scaling out, as Scott was talking about, much along the lines of InfiniBand Ethernet needs that same level of performance, throughput, latency and offloads, and we're able to deliver. >> Yeah, so Kevin, thank you so much. Scott, I want to give you a final word here. Anything else you want your customers to understand regarding this partnerships? >> Yeah, just a quick one Stu, quick one. So we've been really fortunate in working really closely with Mellanox over the years, and with Nvidia. And now the two together, we're just excited about what the future holds. We've done some really neat things in HPC, with being one of the first watercool an InfiniBand card. We're one of the first companies to deploy Dragonfly topology. We've done some unique things where we can share a single IP adapter, across multiple users. We're looking forward to doing a lot of that same exact kind of innovation, inside of our systems as we look to Ethernet. We often think that as speeds of Ethernet continue to go higher, we may see more and more people move from InfiniBand to Ethernet. I think that now having both of these offerings inside of our lineup, is going to make it really easy for customers to choose what's best for them over time. So I'm excited about the future. >> Alright, well Kevin and Scott, thank you so much. Deep integration and customer choice, important stuff. Thank you so much for joining us. >> Thank you Stu. >> Thanks Stu. >> Alright, I'm Stu Miniman, and thank you. Thanks for watching theCUBE. (upbeat music)
SUMMARY :
leaders all around the world, for the Lenovo Data Center Group. now of course the networking team, And of course that can be with HPC, We've shown in HPC that the days Is it 95% of the same architecture, And the Nvidia networking that kind of expansion, the portfolio, And by that I mean the software. Yeah, please Scott, if you And that's the beauty of that the Cumulus Linux, and all of the ability to do things that we've had with Onyx over the years, of the partnership mean? So any of the systems that And so now the ability with Nvidia, of the customer demand leading up this. And that's the key. do a lot of the same things of some of the some of the speeds today? that the GPU can crunch, Yeah, so Kevin, thank you so much. And now the two together, Scott, thank you so much. Miniman, and thank you.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Scott | PERSON | 0.99+ |
Lenovo | ORGANIZATION | 0.99+ |
Kevin | PERSON | 0.99+ |
Kevin Deierling | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
2020 | DATE | 0.99+ |
40,000 times | QUANTITY | 0.99+ |
Onyx | ORGANIZATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Lenovo Data Center Group | ORGANIZATION | 0.99+ |
100 gig | QUANTITY | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
10 megabits | QUANTITY | 0.99+ |
95% | QUANTITY | 0.99+ |
400 gig | QUANTITY | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
September 2020 | DATE | 0.99+ |
200 gig | QUANTITY | 0.99+ |
Mellanox | ORGANIZATION | 0.99+ |
400 gigabits | QUANTITY | 0.99+ |
Scott Tease | PERSON | 0.99+ |
Cumulus | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Linux | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
Stu | PERSON | 0.99+ |
HPC | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.98+ |
two | QUANTITY | 0.98+ |
CUBE | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
five years ago | DATE | 0.98+ |
last month | DATE | 0.98+ |
InfiniBand | ORGANIZATION | 0.98+ |
two different cases | QUANTITY | 0.98+ |
Boston | LOCATION | 0.97+ |
first time | QUANTITY | 0.97+ |
Paresh Kharya & Kevin Deierling, NVIDIA | HPE Discover 2020
>> Narrator: From around the global its theCUBE, covering HPE Discover Virtual Experience, brought to you by HPE. >> Hi, I'm Stu Miniman and this is theCUBE's coverage of HPE, discover the virtual experience for 2020, getting to talk to Hp executives, their partners, the ecosystem, where they are around the globe, this session we're going to be digging in about artificial intelligence, obviously a super important topic these days. And to help me do that, I've got two guests from Nvidia, sitting in the window next to me, we have Paresh Kharya, he's director of product marketing and sitting next to him in the virtual environment is Kevin Deierling, who is this senior vice president of marketing as I mentioned both with Nvidia. Thank you both so much for joining us. >> Thank you, so great to be here. >> Great to be here. >> All right, so Paresh when you set the stage for us? AI, obviously, one of those mega trends to talk about but just, give us the stages, where Nvidia sits, where the market is, and your customers today, that they think about AI. >> Yeah, so we are basically witnessing a massive changes that are happening across every industry. And it's basically the confluence of three things. One is of course, AI, the second is 5G and IOT, and the third is the ability to process all of the data that we have, that's now possible. For AI we are now seeing really advanced models, from computer vision, to understanding natural language, to the ability to speak in conversational terms. In terms of IOT and 5G, there are billions of devices that are sensing and inferring information. And now we have the ability to act, make decisions in various industries, and finally all of the processing capabilities that we have today, at the data center, and in the cloud, as well as at the edge with the GPUs as well as advanced networking that's available, we can now make sense all of this data to help industrial transformation. >> Yeah, Kevin, you know it's interesting when you look at some of these waves of technology and we say, "Okay, there's a lot of new pieces here." You talk about 5G, it's the next generation but architecturally some of these things remind us of the past. So when I look at some of these architectures, I think about, what we've done for high performance computing for a long time, obviously, you know, Mellanox, where you came from through NVIDIA's acquisition, strong play in that environment. So, maybe give us a little bit compare, contrast, what's the same, and what's different about this highly distributed, edge compute AI, IOT environment and what's the same with what we were doing with HPC in the past. >> Yeah, so we've--Mellanox has now been a part of Nvidia for a little over a month and it's great to be part of that. We were both focused on accelerated computing and high performance computing. And to do that, what it means is the scale and the type of problems that we're trying to solve are just simply too large to fit into a single computer. So if that's the case, then you connect a lot of computers. And Jensen talked about this recently at the GTC keynote where he said that the new unit computing, it's really the data center. So it's no longer the box that sits on your desk or even in Iraq, it's the entire data center because that's the scale of the types of problems that we're solving. And so the notion of scale up and scale out, the network becomes really, really critical. And we're doing high-performance networking for a long time. When you move to the edge, instead of having, a single data center with 10,000 computers, you have 10,000 data centers, each of which as a small number of servers that is processing all of that information that's coming in. But in a sense, the problems are very, very similar, whether you're at the edge or you're doing massive HPC, scientific computing or cloud computing. And so we're excited to be part of bringing together the AI and the networking because they are really optimizing at the data center scale across the entire stack. >> All right, so it's interesting. You mentioned, Nvidia CEO, Jensen. I believe if I saw right in there, he actually could, wrote a term which I had not run across, it was the data processing unit or DPU in that, data center, as you talked about. Help us wrap our heads around this a little bit. I know my CPU, when I think about GPUs, I obviously think of Nvidia. TPUs, in the cloud and everything we're doing. So, what is DPUs? Is this just some new AI thing or, is this kind of a new architectural model? >> Yeah. I think what Jensen highlighted is that there's three key elements of this accelerated disaggregated infrastructure that the data center has becoming. And so that's the CPU, which is doing traditional single threaded workloads but for all of the accelerated workloads, you need the GPU. And that does massive parallelism deals with massive amounts of data, but to get that data into the GPU and also into the CPU, you need really an intelligent data processing because the scale and scope of GPUs and CPUs today, these are not single core entities. These are hundreds or even thousands of cores in a big system. And you need to steer the traffic exactly to the right place. You need to do it securely. You need to do it virtualized. You need to do it with containers and to do all of that, you need a programmable data processing unit. So we have something called our BlueField, which combines our latest, greatest, 100 gig and 200 gig network connectivity with Arm processors and a whole bunch of accelerators for security, for virtualization, for storage. And all of those things then feed these giant parallel engines which are the GPU. And of course the CPU, which is really the workload at the application layer for non-accelerated outs. >> Great, so Paresh, Kevin talked about, needing similar types of services, wherever the data is. I was wondering if you could really help expand for us a little bit, the implications of it AI at the edge. >> Sure, yeah, so AI is basically not just one workload. AI is many different types of models and AI also means training as well as inferences, which are very different workloads or AI printing, for example, we are seeing the models growing exponentially, think of any AI model, like a brain of a computer or like a brain, solving a particular use case a for simple models like computer vision, we have models that are smaller, bugs have computer vision but advanced models like natural language processing, they require larger brains or larger models, so on one hand we are seeing the size of the AI models increasing tremendously and in order to train these models, you need to look at computing at the scale of data center, many processors, many different servers working together to train a single model, on the other hand because of these AI models, they are so accurate today from understanding languages to speaking languages, to providing the right recommendations whether it's for products or for content that you may want to consume or advertisements and so on. These models are so effective and efficient that they are being powered by AI today. These applications are being powered by AI and each application requires a small amount of acceleration, so you need the ability to scale out or, and support many different applications. So with our newly launched MPR architecture, just couple of weeks to go that Jensen announced, in the virtual keynote for the first time, we are now able to provide both, scale up and scale out both training data analytics as well as imprints on the single architecture and that's very exciting. >> Yeah, so look at that. The other thing that's interesting is you're talking about at the edge and scale out versus scale up, the networking is critical for both of those. And there's a lot of different workloads. And as Paresh was describing, you've got different workloads that require different amounts of GPU or storage or networking. And so part of that vision of this data center as the computer is that, the DPU lets you scale independently, everything. So you can compose, you desegregate into DPUs and storage and CPUs, and then you compose exactly the computer that you need on the fly container, right, to solve the problem that you're solving right now. So these new way of programming is programming the entire data center at once and you'll go grab all of it and it'll run for a few hundred milliseconds even and then it'll come back down and recompose itself onsite. And to do that, you need this very highly efficient networking infrastructure. And the good news is we're here at HPE Discover. We've got a great partner with HPE. You know, they have our M series switches that uses the Mellanox hundred gig and now even 200 and 400 gig ethernet switches, we have all of our adapters and they have great platforms. The Apollo platform for example, is break for HPC and they have other great platforms that we're looking at with the new telco that we're doing or 5G and accelerating that. >> Yeah, and on the edge computing side, there's the edge line set of products which are very interesting, the other sort of aspect that I wanted to touch upon, is the whole software stack that's needed for the edge. So edge is different in the sense that it's not centrally managed, the edge computing devices are distributed remote locations. And so managing the workflow of running and updating software on it is important and needs to be done in a very secure manner. The second thing that's, that's very different again, for the edges, these devices are going to require connectivity. As Kevin was pointing out, the importance of networking so we also announced, a couple of weeks ago at our GTC, our EGX product that combines the Mellanox NIC and our GPUs into a single a processor, Mellanox NIC provides a fast connectivity, security, as well as the encryption and decryption capabilities, GPUs provide acceleration to run the advanced DI models, that are required for applications at the edge. >> Okay, and if I understood that, right. So, you've got these throughout the HPE the product line, HPE's got long history of making, flexible configurations, I remember when they first came out with a Blade server it was, different form factors, different connectivity options, they pushed heavily into composable infrastructure. So it sounds like this is just a kind of extending, you know, what HP has been doing for a couple of decades. >> Yeah, I think HP is a great partner there and these new platforms, the EGX, for example that was just announced, a great workload there is a 5G telco. So we'll be working with our friends at HPE to take that to market as well. And, you know, really, there's a lot of different workloads and they've got a great portfolio of products across the spectrum from regular servers. And 1U, 2U, and then all the way up to their big Apollo platform. >> Well I'm glad you brought up telco, I'm curious, are there any specific, applications or workloads that, where the low hanging fruit or the kind of the first targets that you use for AI acceleration? >> Yeah, so you know, the 5G workload is just awesome. We're introduced with the EGX, a new platform called Ariel which is a programming framework and there were lots of partners there that were part of that, including, folks like Ericsson. And the idea there is that you have a software defined hardware accelerated radio area network, so a cloud RAM and it really has all of the right attributes of the cloud and what's nice there is now you can change on the fly, the algorithms that you're using for the baseband codex without having to go climb a radio tower and change the actual physical infrastructure. So that's a critical part. Our role in that, on the networking side, we introduced the technology that's part of EGX then are connected, It's like the DX adapter, it's called 5T for 5G. And one of the things that happens is you need this time triggered transport or a telco technology. That's the 5T's for 5G. And the reason is because you're doing distributed baseband unit, distributed radio processing and the timing between each of those server nodes needs to be super precise, 20 nanosecond. It's something that simply can't be done in software. And so we did that in hardware. So instead of having an expensive FPGA, I try to synchronize all of these boxes together. We put it into our NIC and now we put that into industry standard servers HP has some fantastic servers. And then with the EGX platform, with that we can build, really scale out software to client cloud RAM. >> Awesome, Paresh, anything else on the application side you'd like to add in just about what Kevin spoke about. >> Oh yeah, so from application perspective, every industry has applications that touch on edge. If you take a look at the retail, for example, there is, you know, all the way from supply chain to inventory management, to keeping the right stock units in the shelves, making sure there is a there is no slippage or shrinkage. So to telecom, to healthcare, we are re-looking at constantly monitoring patients and taking actions for the best outcomes to manufacturing. We are looking to automate production detecting failures much early on in the production cycle and so on every industry has different applications but they all use AI. They can all leverage the computing capabilities and high-speed networking at the edge to transform their business processes. >> All right, well, it's interesting almost every time we've talked about AI, networking has come up. So, you know, Kevin, I think that probably ease up a little bit why, Nvidia, spent around $7 billion for the acquisition of Mellanox and not only was it the Mellanox acquisition, Cumulus Networks, very known in the network space for software defined really, operating system for networking but give us strategically, does this change the direction of Nvidia, how should we be thinking about Nvidia in the overall network? >> Yeah, I think the way to think about it is going back to that data center as the computer. And if you're thinking about the data center as computer then networking becomes the back plane, if you will of that data center computer and having a high performance network is really critical. And Mellanox has been a leader in that for 20 years now with our InfiniBand and our Ethernet product. But beyond that, you need a programmatic interface because one of the things that's really important in the cloud is that everything is software defined and it's containerized now and there is no better company in the world then Cumulus, really the pioneer and building Cumulus clinics, taking the Linux operating system and running that on multiple homes. So not just hardware from Mellanox but hardware from other people as well. And so that whole notion of an open networking platform more committed to, you need to support that and now you have a programmatic interface that you can drop containers on top of, Cumulus has been the leader in the Linux FRR, it's Free Range Routing, which is the core routing algorithm. And that really is at the heart of other open source network operating systems like Sonic and DENT so we see a lot of synergy here, all the analytics that Cumulus is bringing to bear with NetQ. So it's really great that they're going to be part here of the Nvidia team. >> Excellent, well thank you both much. Want to give you the final word, what should they do, HPE customers in their ecosystem know about the Nvidia and HPE partnership? >> Yeah, so I'll start you know, I think HPE has been a longtime partner and a customer of ours. If you have accelerated workloads, you need to connect those together. The HPE server portfolio is an ideal place. We can combine some of the work we're doing with our new amp years and existing GPUs and then also to connect those together with the M series, which is their internet switches that are based on our spectrum switch platforms and then all of the HPC related activities on InfiniBand, they're a great partner there. And so all of that, pulling it together, and now as at the edge, as edge becomes more and more important, security becomes more and more important and you have to go to this zero trust model, if you plug in a camera that's somebody has at the edge, even if it's on a car, you can't trust it. So everything has to become, validated authenticated, all the data needs to be encrypted. And so they're going to be a great partner because they've been a leader and building the most secure platforms in the world. >> Yeah and on the data center, server, portfolio side, we really work very closely with HP on various different lines of products and really fantastic servers from the Apollo line of a scale up servers to synergy and ProLiant line, as well as the Edgeline for the edge and on the super computing side with the pre side of things. So we really work to the fullest spectram of solutions with HP. We also work on the software side, wehere a lot of these servers, are also certified to run a full stack under a program that we call NGC-Ready so customers get phenomenal value right off the bat, they're guaranteed, to have accelerated workloads work well when they choose these servers. >> Awesome, well, thank you both for giving us the updates, lots happening, obviously in the AI space. Appreciate all the updates. >> Thanks Stu, great to talk to you, stay well. >> Thanks Stu, take care. >> All right, stay with us for lots more from HPE Discover Virtual Experience 2020. I'm Stu Miniman and thank you for watching theCUBE. (bright upbeat music)
SUMMARY :
the global its theCUBE, in the virtual environment that they think about AI. and finally all of the processing the next generation And so the notion of TPUs, in the cloud and And of course the CPU, which of it AI at the edge. for the first time, we are And the good news is we're Yeah, and on the edge computing side, the product line, HPE's across the spectrum from regular servers. and it really has all of the else on the application side and high-speed networking at the edge in the network space for And that really is at the heart about the Nvidia and HPE partnership? all the data needs to be encrypted. Yeah and on the data Appreciate all the updates. Thanks Stu, great to I'm Stu Miniman and thank
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Kevin Deierling | PERSON | 0.99+ |
Kevin | PERSON | 0.99+ |
Paresh Kharya | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
200 gig | QUANTITY | 0.99+ |
HP | ORGANIZATION | 0.99+ |
100 gig | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
10,000 computers | QUANTITY | 0.99+ |
Mellanox | ORGANIZATION | 0.99+ |
200 | QUANTITY | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Paresh | PERSON | 0.99+ |
Cumulus | ORGANIZATION | 0.99+ |
Cumulus Networks | ORGANIZATION | 0.99+ |
Iraq | LOCATION | 0.99+ |
20 years | QUANTITY | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
Ericsson | ORGANIZATION | 0.99+ |
2020 | DATE | 0.99+ |
two guests | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
third | QUANTITY | 0.99+ |
Stu | PERSON | 0.99+ |
first time | QUANTITY | 0.99+ |
around $7 billion | QUANTITY | 0.99+ |
telco | ORGANIZATION | 0.99+ |
each application | QUANTITY | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
second | QUANTITY | 0.99+ |
20 nanosecond | QUANTITY | 0.99+ |
Linux | TITLE | 0.99+ |
both | QUANTITY | 0.99+ |
NetQ | ORGANIZATION | 0.99+ |
400 gig | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
10,000 data centers | QUANTITY | 0.98+ |
second thing | QUANTITY | 0.98+ |
three key elements | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
thousands of cores | QUANTITY | 0.98+ |
three things | QUANTITY | 0.97+ |
Jensen | PERSON | 0.97+ |
Apollo | ORGANIZATION | 0.97+ |
Jensen | ORGANIZATION | 0.96+ |
single computer | QUANTITY | 0.96+ |
HPE Discover | ORGANIZATION | 0.95+ |
single model | QUANTITY | 0.95+ |
first | QUANTITY | 0.95+ |
hundred gig | QUANTITY | 0.94+ |
InfiniBand | ORGANIZATION | 0.94+ |
DENT | ORGANIZATION | 0.93+ |
GTC | EVENT | 0.93+ |
Renaud Gaubert, NVIDIA & Diane Mueller, Red Hat | KubeCon + CloudNativeCon NA 2019
>>Live from San Diego, California It's the Q covering Koopa and Cloud Native Cot brought to you by Red Cloud, Native Computing Pounding and its ecosystem March. >>Welcome back to the Cube here at Q. Khan Club native Khan, 2019 in San Diego, California Instrumental in my co host is Jon Cryer and first of all, happy to welcome back to the program. Diane Mueller, who is the technical of the tech lead of cloud native technology. I'm sorry. I'm getting the wrong That's director of community development Red Hat, because renew. Goodbye is the technical lead of cognitive technologies at in video game to the end of day one. I've got three days. I gotta make sure >>you get a little more Red Bull in the conversation. >>All right, well, there's definitely a lot of energy. Most people we don't even need Red Bull here because we're a day one. But Diane, we're going to start a day zero. So, you know, you know, you've got a good group of community of geeks when they're like Oh, yeah, let me fly in a day early and do like 1/2 day or full day of deep dives. There So the Red Hat team decided to bring everybody on a boat, I guess. >>Yeah. So, um, open ships Commons gathering for this coup con we hosted at on the inspiration Hornblower. We had about 560 people on a boat. I promised them that it wouldn't leave the dock, but we deal still have a little bit of that weight going on every time one of the big military boats came by. And so people were like a little, you know, by the end of the day, but from 8 a.m. in the morning till 8 p.m. In the evening, we just gathered had some amazing deep dives. There was unbelievable conversations onstage offstage on we had, ah, wonderful conversation with some of the new Dev ops folks that have just come on board. That's a metaphor for navigation and Coop gone. And and for events, you know, Andrew Cliche for John Willis, the inevitable Crispin Ella, who runs Open Innovation Labs, and J Bloom have all just formed the global Transformation Office. I love that title on dhe. They're gonna be helping Thio preach the gospel of Cultural Dev ops and agile transformation from a red hat office From now going on, there was a wonderful conversation. I felt privileged to actually get to moderate it and then just amazing people coming forward and sharing their stories. It was a great session. Steve Dake, who's with IBM doing all the SDO stuff? Did you know I've never seen SDO done so well, Deployment explains so well and all of the contents gonna be recorded and up on Aaron. We streamed it live on Facebook. But I'm still, like reeling from the amount of information overload. And I think that's the nice thing about doing a day zero event is that it's a smaller group of people. So we had 600 people register, but I think was 560 something. People show up and we got that facial recognition so that now when they're traveling through the hallways here with 12,000 other people, that go Oh, you were in the room. I met you there. And that's really the whole purpose for comments. Events? >>Yeah, I tell you, this is definitely one of those shows that it doesn't take long where I say, Hey, my brain is full. Can I go home. Now. You know I love your first impressions of Q Khan. Did you get to go to the day zero event And, uh, what sort of things have you been seeing? So >>I've been mostly I went to the lightning talks, which were amazing. Anything? Definitely. There. A number of shout outs to the GPU one, of course. Uh, friend in video. But I definitely enjoyed, for example, of the amazing D. M s one, the one about operators. And generally all of them were very high quality. >>Is this your first Q? Khan, >>I've been there. I've been a year. This is my third con. I've been accused in Europe in the past. Send you an >>old hat old hand at this. Well, before we get into the operator framework and I wanna love to dig into this, I just wanted to ask one more thought. Thought about open shift, Commons, The Commons in general, the relationship between open shift, the the offering. And then Okay, the comments and okay, D and then maybe the announcement about about Okay. Dee da da i o >>s. Oh, a couple of things happened yesterday. Yesterday we dropped. Okay, D for the Alfa release. So anyone who wants to test that out and try it out it's an all operators based a deployment of open shift, which is what open ship for is. It's all a slightly new architectural deployment methodology based on the operator framework, and we've been working very diligently. Thio populate operator hub dot io, which is where all of the upstream projects that have operators like the one that Reynolds has created for in the videos GP use are being hosted so that anyone could deploy them, whether on open shift or any kubernetes so that that dropped. And yesterday we dropped um, and announced Open Sourcing Quay as project quay dot io. So there's a lot of Io is going on here, but project dia dot io is, um, it's a fulfillment, really, of a commitment by Red Hat that whenever we do an acquisition and the poor folks have been their acquired by Cora West's and Cora Weston acquired by Red Hat in an IBM there. And so in the interim, they've been diligently working away to make the code available as open source. And that hit last week and, um, to some really interesting and users that are coming up and now looking forward to having them to contribute to that project as well. But I think the operator framework really has been a big thing that we've been really hearing, getting a lot of uptake on. It's been the new pattern for deploying applications or service is on getting things beyond just a basic install of a service on open shift or any kubernetes. And that's really where one of the exciting things yesterday on we were talking, you know, and I were talking about this earlier was that Exxon Mobil sent a data scientist to the open ship Commons, Audrey Resnick, who gave this amazing presentation about Jupiter Hub, deeper notebooks, deploying them and how like open shift and the advent of operators for things like GP use is really helping them enable data scientists to do their work. Because a lot of the stuff that data signs it's do is almost disposable. They'll run an experiment. Maybe they don't get the result they want, and then it just goes away, which is perfect for a kubernetes workload. But there are other things you need, like a Jeep use and work that video has been doing to enable that on open shift has been just really very helpful. And it was It was a great talk, but we were talking about it from the first day. Signs don't want to know anything about what's under the hood. They just want to run their experiments. So, >>you know, let's like to understand how you got involved in the creation of the operator. >>So generally, if we take a step back and look a bit at what we're trying to do is with a I am l and generally like EJ infrastructure and five G. We're seeing a lot of people. They're trying to build and run applications. Whether it's in data Center at the and we're trying to do here with this operator is to bring GPS to enterprise communities. And this is what we're working with. Red Hat. And this is where, for example, things like the op Agrestic A helps us a lot. So what we've built is this video Gee, few operator that space on the upper air sdk where it wants us to multiple phases to in the first space, for example, install all the components that a data scientist were generally a GPU cluster of might want to need. Whether it's the NVIDIA driver, the container runtime, the community's device again feast do is as you go on and build an infrastructure. You want to be able to have the automation that is here and, more importantly, the update part. So being able to update your different components, face three is generally being able to have a life cycle. So as you manage multiple machines, these are going to get into different states. Some of them are gonna fail, being able to get from these bad states to good states. How do you recover from them? It's super helpful. And then last one is monitoring, which is being able to actually given sites dr users. So the upper here is decay has helped us a lot here, just laying out these different state slips. And in a way, it's done the same thing as what we're trying to do for our customers. The different data scientists, which is basically get out of our way and allow us to focus on core business value. So the operator, who basically takes care of things that are pretty cool as an engineer I lost due to your election. But it doesn't really help me to focus on like my core business value. How do I do with the updates, >>you know? Can I step back one second, maybe go up a level? The problem here is that each physical machine has only ah limited number of NVIDIA. GPU is there and you've got a bunch of containers that maybe spawning on different machines. And so they have to figure out, Do I have a GPU? Can I grab one? And if I'm using it, I assume I have to reserve it and other people can't use and then I have to give it up. Is that is that the problem we're solving here? So this is >>a problem that we've worked with communities community so that like the whole resource management, it's something that is integrated almost first class, citizen in communities, being able to advertise the number of deep, use their your cluster and used and then being able to actually run or schedule these containers. The interesting components that were also recently added are, for example, the monitoring being able to see that a specific Jupiter notebook is using this much of GP utilization. So these air supercool like features that have been coming in the past two years in communities and which red hat has been super helpful, at least in these discussions pushing these different features forward so that we see better enterprise support. Yeah, >>I think the thing with with operators and the operator lifecycle management part of it is really trying to get to Day two. So lots of different methodologies, whether it's danceable or python or job or or UH, that's helm or anything else that can get you an insult of a service or an application or something. And in Stan, she ate it. But and the operator and we support all of that with SD case to help people. But what we're trying to do is bridge the to this day to stuff So Thea, you know, to get people to auto pilot, you know, and there's a whole capacity maturity model that if you go to operator hab dot io, you can see different operators are a different stages of the game. So it's been it's been interesting to work with people to see Theo ah ha moment when they realize Oh, I could do this and then I can walk away. And then if that pod that cluster dies, it'll just you know, I love the word automatically, but they, you know, it's really the goal is to help alleviate the hands on part of Day two and get more automation into the service's and applications we deploy >>right and when they when they this is created. Of course it works well with open shift, but it also works for any kubernetes >>correct operator. HAB Daddio. Everything in there runs on any kubernetes, and that's really the goal is to be ableto take stuff in a hybrid cloud model. You want to be able to run it anywhere you want, so we want people to be unable to do it anywhere. >>So if this really should be an enabler for everything that it's Vinny has been doing to be fully cloud native, Yes, >>I think completely arable here is this is a new attack. Of course, this is a bit there's a lot of complexity, and this is where we're working towards is reducing the complexity and making true that people there. Dan did that a scientist air machine learning engineers are able to focus on their core business. >>You watch all of the different service is in the different things that the data scientists are using. They don't I really want to know what's under under the hood. They would like to just open up a Jupiter Hub notebook, have everything there. They need, train their models, have them run. And then after they're done, they're done and it goes away. And hopefully they remember to turn off the Jeep, use in the woods or wherever it is, and they don't keep getting billed for it. But that's the real beauty of it is that they don't have to worry so much anymore about that. And we've got a whole nice life cycle with source to image or us to I. And they could just quickly build on deploy its been, you know, it's near and dear to my heart, the machine learning the eyesight of stuff. It is one of the more interesting, you know, it's the catchy thing, but the work was, but people are really doing it today, and it's been we had 23 weeks ago in San Francisco, we had a whole open ship comments gathering just on a I and ML and you know, it was amazing to hear. I think that's the most redeeming thing or most rewarding thing rather for people who are working on Kubernetes is to have the folks who are doing workloads come and say, Wow, you know, this is what we're doing because we don't get to see that all the time. And it was pretty amazing. And it's been, you know, makes it all worthwhile. So >>Diane Renaud, thank you so much for the update. Congratulations on the launch of the operators and look forward to hearing more in the future. >>All right >>to >>be here >>for John Troy runs to minimum. More coverage here from Q. Khan Club native Khan, 2019. Thanks for watching. Thank you.
SUMMARY :
Koopa and Cloud Native Cot brought to you by Red Cloud, California Instrumental in my co host is Jon Cryer and first of all, happy to welcome back to the program. There So the Red Hat team decided to bring everybody on a boat, And that's really the whole purpose for comments. Did you get to go to the day zero event And, uh, what sort of things have you been seeing? But I definitely enjoyed, for example, of the amazing D. I've been accused in Europe in the past. The Commons in general, the relationship between open shift, And so in the interim, you know, let's like to understand how you got involved in the creation of the So the operator, who basically takes care of things that Is that is that the problem we're solving here? added are, for example, the monitoring being able to see that a specific Jupiter notebook is using this the operator and we support all of that with SD case to help people. Of course it works well with open shift, and that's really the goal is to be ableto take stuff in a hybrid lot of complexity, and this is where we're working towards is reducing the complexity and It is one of the more interesting, you know, it's the catchy thing, but the work was, Congratulations on the launch of the operators and look forward for John Troy runs to minimum.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Audrey Resnick | PERSON | 0.99+ |
Andrew Cliche | PERSON | 0.99+ |
Diane Mueller | PERSON | 0.99+ |
Steve Dake | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Jon Cryer | PERSON | 0.99+ |
Exxon Mobil | ORGANIZATION | 0.99+ |
Diane Renaud | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
John Troy | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
1/2 day | QUANTITY | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
San Diego, California | LOCATION | 0.99+ |
first | QUANTITY | 0.99+ |
J Bloom | PERSON | 0.99+ |
Diane | PERSON | 0.99+ |
2019 | DATE | 0.99+ |
Open Innovation Labs | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
Red Cloud | ORGANIZATION | 0.99+ |
560 | QUANTITY | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
600 people | QUANTITY | 0.99+ |
three days | QUANTITY | 0.99+ |
John Willis | PERSON | 0.99+ |
8 a.m. | DATE | 0.99+ |
Crispin Ella | PERSON | 0.99+ |
Jeep | ORGANIZATION | 0.99+ |
San Diego, California | LOCATION | 0.99+ |
Cora West | ORGANIZATION | 0.99+ |
Yesterday | DATE | 0.99+ |
last week | DATE | 0.99+ |
SDO | TITLE | 0.99+ |
Dan | PERSON | 0.99+ |
8 p.m. | DATE | 0.98+ |
23 weeks ago | DATE | 0.98+ |
first impressions | QUANTITY | 0.98+ |
one second | QUANTITY | 0.98+ |
Q. Khan Club | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
Renau | PERSON | 0.98+ |
Red Bull | ORGANIZATION | 0.98+ |
Reynolds | PERSON | 0.97+ |
Aaron | PERSON | 0.97+ |
Day two | QUANTITY | 0.97+ |
March | DATE | 0.96+ |
third con. | QUANTITY | 0.96+ |
first space | QUANTITY | 0.96+ |
first day | QUANTITY | 0.95+ |
Vinny | PERSON | 0.95+ |
Cora Weston | ORGANIZATION | 0.94+ |
Thio | PERSON | 0.94+ |
Cloud | ORGANIZATION | 0.93+ |
ORGANIZATION | 0.92+ | |
first class | QUANTITY | 0.92+ |
today | DATE | 0.9+ |
about 560 people | QUANTITY | 0.9+ |
Jupiter | LOCATION | 0.89+ |
each physical machine | QUANTITY | 0.88+ |
12,000 other | QUANTITY | 0.88+ |
day zero | QUANTITY | 0.88+ |
D. M | PERSON | 0.87+ |
CloudNativeCon NA 2019 | EVENT | 0.87+ |
d Gaubert | PERSON | 0.87+ |
Thea | PERSON | 0.86+ |
python | TITLE | 0.84+ |
Native Computing Pounding | ORGANIZATION | 0.83+ |
a day | QUANTITY | 0.79+ |
day zero | EVENT | 0.78+ |
day one | QUANTITY | 0.78+ |
Koopa | ORGANIZATION | 0.76+ |
one more thought | QUANTITY | 0.74+ |
Khan | PERSON | 0.72+ |
Commons | ORGANIZATION | 0.72+ |
KubeCon + | EVENT | 0.72+ |
Jupiter Hub | ORGANIZATION | 0.71+ |
Brian Schwarz, Pure Storage & Charlie Boyle, NVIDIA | Pure Accelerate 2019
>> from Austin, Texas. It's Theo Cube, covering pure storage. Accelerate 2019. Brought to you by pure storage. >> Welcome to the Cube. The leader in live tech coverage covering up your accelerate 2019. Lisa Martin with Dave Ilan in Austin, Texas, this year. Pleased to welcome a couple of guests to the program. Please meet Charlie Boyle, VP and GM of DJ X Systems at N Video. Hey, Charlie, welcome back to the Cube, but in a long time ago and we have Brian Schwartz, VP of product management and development at your brain. Welcome. >> Thanks for having me. >> Here we are Day one of the event. Lots of News This morning here is just about to celebrate its 10th anniversary. A lot of innovation and 10 years. Nvidia partnerships. About two is two and 1/2 years old or so. Brian, let's start with you. Give us a little bit of an overview about where pure and and video are, and then let's dig into this news about the Aye aye data hub. >> Cool, it's It's been a good partnership for a couple of years now, and it really was born out of work with mutual customers. You know we brought out the flash blade product, obviously in video was in the market with DJ X is for a I, and we really started to see overlap in a bunch of initial deployments. And we really realized that there was a lot of wisdom to be gained off some of these early I deployments of capturing some of that knowledge and wisdom from those early practitioners and being able to share it with the with the wider community. So that's really kind of where the partnership was born going for a couple of years now, I've got a couple of chapters behind us and many more in the future. And obviously the eye data hub is the piece that we really talked about at this year's accelerate. >> Yeah, areas about been in the market for what? About a year and 1/2 or so Almost >> two years. >> Two years? All right, tell us a little bit about the adoption. What what customers were able to dio with this a ready infrastructure >> and point out the reason we started the partnership was our early customers that were buying dejected product from us. They were buying pure stored. Both leaders and high performance. And as they were trying to put them together, they're like, How should we do this? What's the optimal settings? They've been using storage for years. I was kind of new to them and they needed that recipe. So that's, you know, the early customer experiences turned into airy the solution, and, you know, the whole point of this to simplify. I sounds kind of scary to a lot of folks and the data scientists really just need to be productive. They don't care about infrastructure, but I t s to support this. So I t was very familiar with pure storage. They used them for years for high performance data and as they brought in the Nvidia Compute toe work with that, you know, having a solution that we both supported was super important to the I T practitioners because they knew it worked. They knew we both supported it. We stood behind it and they could get up and running in a matter of days or weeks versus 6 to 9 months if they built it >> themselves. >> You look at companies that you talk to customers. Let's let's narrow it down to those that have data scientists least one day to scientists and ask him where they are in their maturity model, if one is planning to was early threes, they got multiple use cases and four is their enterprise wide. How do you see the landscape? Are you seeing pretty aggressive adoption in those as I couched it, or is it still early? >> I mean so every customers in a different point. So there's definitely a lot of people that are still early, but we've seen a lot of production use cases. You know, everyone talks about self driving cars, but that's, you know, there's a lot behind that. But real world use cases say medicals got a ton? You know, we've got partner companies that you are looking at a reconstruction of MRI's and CT scans cutting the scan time down by 75%. You know, that's real patient outcome. You know, we've got industrial inspection, we're in Texas. People fly drones around and have a eye. Models that are built in their data center on the drone and the field operators get to re program the drones based on what they see and what is happening. Real time and re trains every night. So depending on the industry really depends on where people are in the maturity her. But you know, really, our message out to the enterprises are start now. You know, whether you've got one data scientist, you've got some community data scientists. There's no reason to wait on a because there's a use case that work somewhere in your inner. >> So so one of the key considerations to getting started. What would you say? >> So one thing I would say is, look any to your stages of maturity. Any good investment is done through some creation of business value, right? And an understanding of kind of what problem you're trying to solve and making sure it's compelling. Problem is an important one, and some industries air farther along. Like you know, one of the ones that most everybody's familiar with is the tech industry itself. Every recommendation engine you've probably ever seen on the Internet is backed by some form of a I behind it because they wanted to be super fast and, you know, customized to you as a user. So I think understanding the business value creation problem is is a really important step of it and many people go through an early stage of experimentation, data modeling really kind of, say, a prototyping stage before they go into a mass production use case. It's a very classic i t adoption curve. Just add a comment to the earlier kind of trend is it's a megatrend. Yes, not everybody is doing it in massive wide scale production today. There's some industries that are farther ahead. If you look forward over the next 15 to 20 years, there's a massive amount of Ai ai coming, and it's a It is a new form of computing, the GPU driven computing and the whole point about areas getting the ingredients right. Thio have this new set of infrastructure have storage network compute on the software stack all kind of package together to make it easier to adopt, to allow people to adopt it faster because some industries are far along and others are still in the earlier stages, >> right? So how do you help for those customers and industries that aren't self driving cards of the drones that you talked about where we use case, we all understand it and are excited about it. But for other customers in different industries. How do you help them even understand the A pipeline? And where did they start? I'm sure that varies very >> a lot. But, you know, the key point is starting a I project. You have a desired outcome from Not everything's gonna be successful, but you know Aye, aye. Projects aren't something that it's not a six month I t project or a big you know, C r m. Refresh it. Something that you could take One of our classes that we have, we do a lot of end user customer training are Deep Learning Institute. You can take 1/2 day class and actually do a deep learning project that day. And so a lot of it is understanding your data, you know, and that's where your and the data hub comes in, understanding the data that you have and then formulating a question like, What could I do if I knew this thing? That's all about a I and deep learning. It's coming up with insights that aren't natural. When you just stare at the data, how can the system understand what you want? And then what are the things that you didn't expect defined that A. I is showing you about your data, and that's really a lot of where the business value comes. And how do you know more about your customer? How do you help that customer better, eh? I can unlock things that you may not have pondered yourself. >> The other thing. I'm a huge fan of analogies when you're trying to describe a new concept of people. And there's a good analogy about Ai ai data pipelines that predates, Aye aye around data warehousing like there's been industry around, extract transformers load E T L Systems for a very long period of time. It's a very common thing for many, many people in the I T industry, and I do think there's when you think about a pipeline in a I pipeline. There's an analogy there, which you have data coming in ingress data. You're cleansing it, you're cleaning it. You're essentially trying to get some value out of it. How you do that in a eyes quite a bit different, cause it's GP use and you're looking, you know, for turning unstructured data into more structure date. It's a little different than data. Warehousing traditionally was running reports, but there's a big analogy, I think, to be used about a pipeline that is familiar to people as a way to understand the new concept. >> So that's good. I like the pipeline concept. One of the one of the counters to that would be that you know, when you think about e. T ells complicated process enterprise data warehouses that were cumbersome Do you feel like automation in the A I Pipeline? When we look back 10 years from now, we'll have maybe better things to say than we do about E D W A R e g l. >> And I think one of the things that we've seen, You know, obviously we've done a ton of work in traditional. Aye, aye, But we've also done a lot in accelerated machine learning because that's a little closer to your traditional Data analytics and one of the biggest kind of ah ha moments that I've seen customers in the past year or so. It's just how quickly, by using GPU computing, they can actually look at their data, do something useful with it, and then move on to the next thing so that rapid experimentation is all you know, what a I is about. It's not a eyes, not a one and done thing. Lots of people think Oh, I have to have a recommend er engine. And then I'm done. No, you have to keep retraining it day in and day out so that it gets better. And that's before you had accelerated. Aye, aye pipeline. Before you had accelerated data pipelines that we've been doing with cheap use. It just took too long so people didn't run those experiments. Now we're seeing people exploring Maur trying different things because when your experiment takes 10 minutes, two minutes versus two days or 10 days, you can try out your cycle time. Shorter businesses could doom or and sure, you're gonna discard a lot of results. But you're gonna find those hidden gems that weren't possible before because you just didn't have the time to do >> it. Isn't a key operational izing it as well? I mean again, one of the challenges with the analogy that you gave a needy W is fine reporting. You can operationalize it for reporting, and but the use cases weren't is rich robust, and I feel as though machine intelligence is I mean, you're not gonna help but run into it. It's gonna be part of your everyday life, your thoughts. >> It's definitely part of our everyday lives. When you talk about, you know, consumer applications of everything we all use every day just don't know it's it's, you know, the voice recognition system getting your answer right the first time. You know there's a huge investments in natural language speech right now to the point that you can ask your phone a question. It's going through searching the Web for you, getting the right answer, combining that answer, reading it back to you and giving you the Web page all in less than a second. You know, before you know that be like you talked to an I. V R system. Wait, then you go to an operator. Now people are getting such a better user experience out of a I back systems that, you know over the next few years, I think end users will start preferring to deal with those based systems rather than waiting on line for human, because it'll just get it right. It'll get you the answer you need and you're done. You save time. The company save time and you've got a better outcome. >> So there's definitely some barriers to adoption skills. Is one obvious one the other. And I wonder if Puritan video attack this problem. I'm sure you have, but I'd like some color on it. His traditional companies, which a lot of your customers, their data is in pockets. It's not at the core. You look at the aye aye leaders, you know, the Big Five data their data cos it's at the core. They're applying machine intelligence to that data. How has this modern storage that we heard about this morning affected that customers abilities to really put data at their core? >> You know, it's It's a great question, Dave and I think one of the real opportunities, particularly with Flash, is to consolidate data into a smaller number off larger kind of islands of data, because that's where you could really drive the insights. And historically, in a district in world, you would never try to consolidate your data because there was too many bad performance implications of trying to do that. So people had all these pockets, and even if you could, you probably wouldn't actually want to put the date on the same system at the same time. The difference with flashes as so much performance at the at the core of it at the foundation of it. So the concept of having a very large scale system, like 150 blade system we announced this morning is a way to put a lot of the year and be able to access it. And to Charlie's point, a lot of people they're doing constant experiment, experimentation and modeling of the data. You don't know that how the date is gonna be consumed and you need a very fast kind of wide platform to do that, Which is why it's been a good fit for us to work together >> now fall upon that. Dated by its very nature. However, Brian is distributed and we heard this morning is you're attacking that problem through in a P I framework that you don't care where it is. Cloud on Prem hybrid edge. At some point in time, your thoughts on that >> well, in again the data t be used for a I I wouldn't say it's gonna be every single piece of data inside an organization is gonna be put into the eye pipeline in a lot of cases, you could break it down again. Thio What is the problem? I'm trying to solve the business value and what is the type of data that's gonna be the best fit for it? There are a lot of common patterns for consumption in a I AA speech recognition image recognition places where you have a lot of unstructured data or it's unstructured to a computer. It's not unstructured to you. When you look at a picture, you see a lot of things in it that a computer can't see right, because you recognize what the patterns are and the whole point about a eyes. It's gonna help us get structure out of these unstructured data sets so the computer can recognize more things. You know, the speech and emotions that we as humans just take for granted. It's about having computers, being able to process and respond to that in a way that they're not really people doing today. >> Hot dog, not a hot dog. Silicon Valley >> Street light. Which one of these is not a street lights and prove you're not about to ask you about distributed environments. You know customers have so much choice for everything these days on Prem hosted SAS Public Cloud. What are some of the trends that you're seeing? I always thought that to really be able to extract a tremendous amount of value from data and to deliver a I from it you needed the cloud because you needed a massive volumes of data. Appears legacy of on print. What are some of the things that you're seeing there and how is and video you're coming together to help customers wherever this data is to really dry Valley business value from these workloads, >> I have to put comments and I'll turn over to Charlie. So one is we get asked this question a lot. Like where should I run my eye? The first thing I always tell people is, Where's your data? Gravity moving these days? That's a very large tens of terror by its hundreds of terabytes petabytes of data moving very large. That's the data is actually still ah, hard challenge today. So running your A II where your date is being generated is a good first principle. And for a lot of folks they still have a lot on premise data. That's where their systems are they're generating the systems, or it's a consolidation point from the edge or other other opportunities to run it there. So that's where your date is. Run your A I there. The second thing is about giving people flexibility. We've both made pretty big investments in the world of containerized software applications. Those things are things that can run on grammar in the cloud. So trying to use a consistent set of infrastructure and software and tooling that allows people to migrate and change over time, I think, is an important strategy not only for us but also for the end users that gives them flexibility. >> So, ideally, on Prem versus Cloud implementations shouldn't be. That shouldn't be different. Be great. It would be identical. But are they today? >> So at the lowest level, there's always technical differences, but at the layers that customers are using it, we run one software stack no matter where you're running. So if it's on one of our combined R E systems, whether it's in a cloud provider, it's the same in video software stack from our lowest end consumer of rage. He views, too. The big £350 dejected too you see back there? You know, we've got one software stack runs everywhere, And when the riders making you know, it's really Renee I where your data is And while a lot of people, if you are cloud native company, if you started that way, I'm gonna tell you to run in the cloud all day long. But most enterprises, they're some of their most valuable data is still sitting on premise. They've got decades of customer experience. They've got decades of product information that's all running in systems on Prem. And when you look at speech, speech is the biggest thing you know. They've got, you know, years of call center data that's all sitting in some offline record. What am I gonna do with that? That stuff's not in the cloud. And so you want to move the processing to that because it's impossible to move that data somewhere else and transform it because you're only gonna actually use a small fraction of that data to produce your model. But at the same time, you don't want to spend a year moving that data somewhere to process it back the truck up, put some DJ X is in front of it. And you're good to go. >> Someone's gonna beat you to finding those insides. Right? So there is no time. >> So you have another question. >> I have the last question. So you got >> so in video, you gotta be Switzerland in this game. So I'm not gonna ask you this question. But, Brian, I will ask you what? Why? You're different. I know you were first. He raced out. You got the press release out first. But now that you've been in the market for a while what up? Yours? Competitive differentiators. >> You know, there's there's really two out netted out for flash played on why we think it's a great fit for an A i N A. I use case. One is the flexibility of the performance. We call multi dimensional performance, small files, large files, meditated intensive workloads. Flash blade can do them all. It's a it's a ground up design. It's super flexible on performance. And but also more importantly, I would argue simplicity is a really hallmark of who we are. It's part of the modern date experience that we're talking about this morning. You can think about the systems. They are miniaturized supercomputers And yes, you could always build a supercomputer. People have been doing it for decades. Use Ph. D's to do it and, like most people, don't want to happen. People focused on that level of infrastructure, so we've tried to give incredible kind of capabilities in a really simple to consume platform. I joke with people. We have storage PhDs like literally people. Be cheese for storage so customers don't have to. >> Charlie, feel free to chime in on your favorite child if you want. I >> need a lot of it comes from our customers. That's how we first started with pure is our joint customers saying we need this stuff to work really fast. They're making a massive investment with us and compute. And so if you're gonna run those systems at 100% you need storage. The confusion, you know, pure is our first in there. There are longest partner in this space, and it's really our joint customers that put us together and, you know, to some extent, yes, we are Switzerland. You know, we love all of our partners, but, you know, we do incredible work with these guys all up and down the stack and that's the point to make it simple. If the customer has data we wanted to make be a simplest possible for them to run a ay, whether it's with my stuff with our cloud stuff, all of our partners, but having that deep level of integration and having some of the same shared beliefs to just make stuff simple so people can actually get value out of the data have I t get out of the way so Data scientists could just get their work done. That's what's really powerful about the partnership. >> And I imagine you know, we're out of time, but I imagine to be able to do this at the accelerated pace accelerated, I'm gonna say pun intended it wasn't but, um, cultural fed has to be pretty align. We know Piers culture is bold. Last question, Brian and we bring it home here. Talk to us about how the cultural cultures appearing and video are stars I lining to be able to enable how quickly you guys are developing together. >> Way mentioned the simplicity piece of it. The other piece that I think has been a really strong cultural fit between the companies. It's just the sheer desire to innovate and change the world to be a better place. You know, our hallmark. Our mission is to make the make the world a better place with data. And it really fits with the level of innovation that obviously the video does so like to Silicon Valley companies with wicked smart folks trying to make the world a better place, It's It's really been a good partnership. >> Echo that. That's just, you know, the rate of innovation in a I changes monthly. So if you're gonna be a good partner to your customers, you gotta change Justus fast. So our partnership has been great in that space. >> Awesome. Next time, we're out of time, But next time, come back, talk to a customer, really wanna understand it, gonna dig into some of the great things that they're extracting from you guys. So, Charlie Brian, thank you for joining David me on the Cube this afternoon. Thanks. Thanks. Thanks for David. Dante. I'm Lisa Martin. You're watching the Cube. Y'all from pure accelerate in Austin, Texas.
SUMMARY :
Brought to you by guests to the program. is just about to celebrate its 10th anniversary. And obviously the eye data hub is the What what customers were able to dio with So that's, you know, the early customer experiences turned into airy the solution, You look at companies that you talk to customers. You know, we've got partner companies that you are looking at So so one of the key considerations to getting started. Like you know, one of the ones that most everybody's familiar with is the tech of the drones that you talked about where we use case, we all understand it and are excited And how do you know more about your customer? and I do think there's when you think about a pipeline in a I pipeline. that you know, when you think about e. T ells complicated process enterprise data warehouses that were so that rapid experimentation is all you know, I mean again, one of the challenges with the analogy that you gave You know there's a huge investments in natural language speech right now to the point that you can ask You look at the aye aye leaders, you know, the Big Five data You don't know that how the date is gonna be consumed and you need a very fast However, Brian is distributed and we heard this morning a lot of cases, you could break it down again. Hot dog, not a hot dog. data and to deliver a I from it you needed the cloud because you needed a massive I have to put comments and I'll turn over to Charlie. But are they today? But at the same time, you don't want to spend a year Someone's gonna beat you to finding those insides. So you got So I'm not gonna ask you this question. And yes, you could always build a supercomputer. Charlie, feel free to chime in on your favorite child if you want. and it's really our joint customers that put us together and, you know, to some extent, yes, And I imagine you know, we're out of time, but I imagine to be able to do this at the accelerated pace accelerated, It's just the sheer desire to innovate and change the world That's just, you know, the rate of innovation in a I changes monthly. gonna dig into some of the great things that they're extracting from you guys.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Brian | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Brian Schwartz | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Brian Schwarz | PERSON | 0.99+ |
Charlie Boyle | PERSON | 0.99+ |
Dave Ilan | PERSON | 0.99+ |
Texas | LOCATION | 0.99+ |
two minutes | QUANTITY | 0.99+ |
75% | QUANTITY | 0.99+ |
Charlie | PERSON | 0.99+ |
two days | QUANTITY | 0.99+ |
10 minutes | QUANTITY | 0.99+ |
Charlie Brian | PERSON | 0.99+ |
6 | QUANTITY | 0.99+ |
Dante | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
100% | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
10 days | QUANTITY | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
10 years | QUANTITY | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Deep Learning Institute | ORGANIZATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
£350 | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
this year | DATE | 0.99+ |
DJ X Systems | ORGANIZATION | 0.99+ |
1/2 years | QUANTITY | 0.99+ |
Two years | QUANTITY | 0.99+ |
second thing | QUANTITY | 0.99+ |
9 months | QUANTITY | 0.99+ |
less than a second | QUANTITY | 0.99+ |
six month | QUANTITY | 0.99+ |
10th anniversary | QUANTITY | 0.98+ |
Switzerland | LOCATION | 0.98+ |
one | QUANTITY | 0.98+ |
N Video | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.98+ |
both | QUANTITY | 0.98+ |
four | QUANTITY | 0.97+ |
Day one | QUANTITY | 0.97+ |
first principle | QUANTITY | 0.97+ |
Echo | COMMERCIAL_ITEM | 0.97+ |
decades | QUANTITY | 0.97+ |
first time | QUANTITY | 0.96+ |
two years | QUANTITY | 0.95+ |
Puritan | ORGANIZATION | 0.95+ |
this morning | DATE | 0.95+ |
a year | QUANTITY | 0.95+ |
150 blade | QUANTITY | 0.91+ |
today | DATE | 0.91+ |
one day | QUANTITY | 0.9+ |
1/2 day class | QUANTITY | 0.88+ |
hundreds of terabytes petabytes of data | QUANTITY | 0.88+ |
first thing | QUANTITY | 0.87+ |
this afternoon | DATE | 0.87+ |
one software stack | QUANTITY | 0.86+ |
past year | DATE | 0.84+ |
John Fanelli, NVIDIA & Kevin Gray, Dell EMC | VMworld 2019
(lively music) >> Narrator: Live, from San Francisco, celebrating 10 years of high tech coverage, it's theCUBE, covering VMworld 2019! Brought to you by VMware and its ecosystem partners. >> Okay, welcome back to theCUBE's live coverage in VMworld 2019. We're in San Francisco. We're in Moscone North Lobby. I'm John Frer, my co Stu Miniman, here covering all the action of VMworld, two sets for theCUBE, our tenth year, Stu. Keeping it going. Two great guests, John Fanelli, CUBE Alumni, Vice President of Product, Virtual GPUs at NVIDIA Kevin Gray, Director of Product Marketing, Dell EMC. Thanks for coming back on. Good to see you. >> Awesome. >> Good to see you guys, too. >> NVIDIA, big news, we saw your CEO up on the keynote videoing in. Two big announcements. You got some stats on some Windows stats to talk about. Let's talk about the news first, get the news out of the way. >> Sure, at this show, NVIDIA announced our new product called NVIDIA Virtual Compute Server. So for the very first time anywhere, we're able to virtualize artificial intelligence, deep learning, machine learning, and data analytics. Of course, we did that in conjunction with our partner, VMware. This runs on top of vSphere and also in conjunction with our partner at Dell. All of this Virtual Compute Server runs on Dell VxRail, as well. >> What's the impact going to be for that? What does that mean for the customers? >> For customers, it's really going to be the on-ramp for Enterprise AI. A lot of customers, let's say they have a team of maybe eight data scientists are doing data analytics, if they want to move through GPU today, they have to buy eight GPUs. However, with our new solution, maybe they start with two GPUs and put four users on a GPU. Then as their models get bigger and their data gets bigger, they move to one user per GPU. Then ultimately, because we support multiple GPUs now as part of this, they move to a VM that has maybe four GPUs in it. We allow the enterprise to start to move on to AI and deep learning, in particular, machine learning for data analytics very easily. >> GPUs are in high demand. My son always wants the next NVIDIA, in part told me to get some GPUs from you when you came on. Ask the NVIDIA guy to get some for his gaming rig. Kidding aside, now in the enterprise, really important around some of the data crunching, this has really been a great use case. Talk about how that's changed, how people think about it, and how it's impacted traditional enterprise. >> From a data analytics perspective, the data scientists will ingest data, they'll run some machine learning on it, they'll create an inference model that they run to drive predictive business decisions. What we've done is we've GPU-accelerated the key libraries, the technologies, like PyTorch, XGBoost to use a GPU. The first announcement is about how they can now use Virtual Compute Server to do that. The second announcement is that workflow is, as I mentioned, they'll start small, and then they'll do bigger models, and eventually they want to train that scale. So what they want to do is they want to move to the cloud so they can have hundreds or thousands of GPUs. The second announcement is that NVIDIA and VMware are bringing Virtual Compute Server to VMware Cloud running on AWS with our T4 GPUs. So now I can scale virtually starting with fractional GPU to single GPU to multi GPU, and push a button with HCX and move it directly into AWS T4 accelerated cloud. >> That's the roadmap so you can get in, get the work done, scale up, that's the benefit of that. Availability, timing, when all of this is going to hit in-- >> So Virtual Compute Server is available on Friday, the 29th. We're looking at mid next year for the full suite of VMware Cloud on top of Aws T4. >> Kevin, you guys are supplier here at Dell EMC. What's the positioning there with you guys? >> We're working very closely with NVIDIA in general on all of their efforts around both AI as well as VDI too. We'll work quite a bit, most recently on the VDI front as well. We look to drive things like qualifying the devices. There's both VDI or analytics applications. >> Kevin, bring us up-to-date 'cause it's funny we were talking about this is our 10th year here at the show. I remember sitting across Howard Street here in 2010 and Dell, and HP, and IBM all claiming who had the lowest dollar per desktop as to what they were doing in VDI. It's a way different discussion here in 2019. >> Absolutely. Go ahead. >> One of the things that we've learned with NVIDIA is that it's really about the user experience. It's funny we're at a transition point now from Windows 7 to Windows 10. The last transition was Windows XP to Windows 7. What we did then is we took Windows 7, we tore everything out of it we possibly could, we made it look like XP, and we shoved it out. 10 years later, that doesn't work. Everyone's got their iPhones, their iOS devices, their Android devices. Microsoft's done a great job on Windows 10 being immersive. Now we're focused on user experience. When the VDI environment, as you move to Windows 10, you may not be aware of this, but from Windows 7 to Windows 10, it uses 50% more CPU, and you don't even get that great of a user experience. You pop a GPU in there, and you're good. Most of our customers together are working on a five-year life cycle. That means over the next five years, they're going to get 10 updates of Windows 10, and they're going to get like 60 updates of their Office applications. That means that they want to be future-proof now by putting the GPUs in to guarantee a great user experience. >> On the performance side too, obviously. In auto updates, this is the push notification world we live in. This has to built in from day one. >> Absolutely, and if you look at what Dell's doing, we really built this into both our VxRails and our VxBlocks. GPUs are just now part of it. We do these fully qualified. It stacks specifically for VDI environments as well. We're working a lot with the n-vector tools from VM which makes sure we're-- >> VDI finally made it! >> qualifying user experience. >> All these years. >> Yes, yes. In fact, we have this user experience tool called n-vector, which actually, without getting super technical for the audience, it allows you to look at the user experience based on frame-rate, latency, and image quality. We put this tool together, but Dell has really been taking a lead on testing it and promoting it to the users to really drive the cost-effectiveness. It still is about the dollar per desktop, but it's the dollar per dazzling desktop. (laughing) >> Kevin, I hear the frame-rate in there, and I've got all the remote workers, and you're saying how do I make sure that's not the gaming platform they're using because I know how important that is. >> Absolutely. There's a ton of customers that are out there that we're using. We look at folks like Guillevin as like the example of a company that's worked with us and NVIDIA to truly drive types of applications that are essential to VDI. These types of power workers doing applications like Autodesk, that user experience and that ability to support multiple users. If you look at Pat, he talked a little bit about any cloud, any application, any device. In VDI, that's really what it's about, allowing those workers to come together. >> I think the thing that the two of you mentioned, and Stu you pointed out brilliantly was that VDI is not just an IT thing anymore. It really is the expectation now that my rig, if I'm a gamer, or a young person, the younger kids, if you're under 25, if you don't have a kick-ass rig, (laughs) that's what they call it. Multiple monitors, that's the expectation, again, mobility. Work experience, workspace. >> Exactly, along those same lines, by the way. >> This is the whole category. It's not just like a VDI, this thing over here that used to be talked about as an IT thing. >> It's about the workflow. So it's how do I get my job done. We used to use words like "business worker" and "knowledge worker." It's just I'm a worker. Everybody today uses their phone that's mobile. They use their computer at home, they use their computer at work. They're all running with dual monitors. Dual monitors, sometimes dual 4K monitors. That really benefits as well from having a GPU. I know we're on TV so hopefully some of you guys are watching VDI on your GPU-accelerated. It's things like Skype, WebEX, Zoom, all the collaboration to 'em, Microsoft Teams, they all benefit from our joint solution, like the GPU. >> These new subsystems like GPUs become so critical. They're not just subsystem, they are the main part because the offload is now part of the new operating environment. >> We optimized together jointly using the n-vector tool. We optimized the server and operating environment, so that if you run into GPU, you can right-size your CPU in terms of cores, speed, etc., so that you get the best user experience at a most cost effective way. >> Also, the gaming world helps bring in the new kind of cool visualization. That's going to move into just the workflow of workers. You start to see this immersive experience, VR, ARs obviously around the corner. It's only going to get more complex, more needs for GPUs. >> Yes, in fact, we're seeing more, I think, requirements for AR and VR from business than we are actually for gaming. Don't you want to go into your auto showroom at your house and feel the fine Corinthian leather? >> We got to upgrade our CUBE game, get more GPU focused and get some tracing in there. >> Kevin, I know I've seen things from the Dell family on levering VR in the enterprise space. >> Oh, absolutely. If you look at a lot of the things that we're doing with some of the telcos around 5G. They're very interested in VR and AR. Those are areas that'll continue to use things like GPUs to help accelerate those types of applications. It really does come down to having that scalable infrastructure that's easy to manage and easy to operate. That's where I think the partnership with NVIDIA really comes together. >> Deep learning and all this stuff around data. Michael Dell always comes on theCUBE, talks about it. He sees data as the biggest opportunity and challenge. In whatever applications coming in, you got to be able to pound into that data. That's where AI's really shown... Machine learning has kind of shown that that's helping heavy lifting a lot of things that were either manual. >> Exactly. The one thing that's really great about data analytics that are GPU-accelerated is we can take a job that used to take days and bring it down to hours. Obviously, doing something faster is great, but if I take a job that used to take a week and I can do it in one day, that means I have four more days to do other things. It's almost like I'm hiring people for free because I get four more extra work days. The other thing that's really interesting as our joint solution is you can leverage that same virtual GPU technology. You can do VDI by day and at night, you run Compute. So when your users aren't at work, you migrate them off, you spin up your VMs that are doing your data analytics using our RAPIDS technology, and then you're able to get that platform running 24 by seven. >> Productivity gains just from an infrastructure. Even the user too, up and down, the productivity gains are significant. So I'll get three monitors now. I'm going to get one of those Alienware curved monitors. >> Just the difference we had, we have a suite here at the show, and just the difference, you can see such a difference when you insert the GPUs into the platform. It's just makes all the difference. >> John, I got to ask you a personal question. How many times have people asked you for a GPU? You must get that all the time? >> We do. I have a NVIDIA backpack. When I walk around, there's a lot of people that only know NVIDIA for games. So random people will always ask for that. >> I've got two sons and two daughters and they just nerd out on the GPUs. >> I think he's trying to get me to commit on camera on giving him a GPU. (laughing) I think I'm in trouble here. >> Yeah, they get the latest and greatest. Any new stuff, they're going to be happy to be the first on the block to get the GPU. It's certainly impacted on the infrastructure side, the components, the operating environment, Windows 10. Any other data you guys have to share that you think is notable around how all this is coming together working from user experience around Windows and VDI? >> I think one piece of data, again, going back to your first comment about cost per desktop. We're seeing a lot of migration to Windows 10. Customers are buying our joint solution from Dell which includes our hardware and software. They're buying that five-year life cycle, so we actually put a program in place to really drive down the cost. It's literally like $3 per month to have a GPU-accelerated virtual desktop. It's really great Value for the customers besides the great productivity. >> If you look at doing some of these workloads on premises, some of the costs can come down. We had a recent study around the VxBlock as an example. We showed that running GPUs and VDI can be up as much as 45% less on a VxBlock at scale. When you talk about the whole hybrid cloud, multi-cloud strategy, there's pluses and minuses to both. Certainly, if we look at some of the ability to start small and scale out, whether you're going HCI or you're going CI, I think there's a VDI solution there that can really drive the economics. >> The intense workloads. Is there any industries that are key for you guys in terms of verticals? >> Absolutely. So we're definitely looking at a lot of the CAD/CAM industries. We just did a certification on our platforms with Dassault's CATIA system. That's an area that we'll continue to explore as we move forward. >> I think in the workstation side of things, it's all the standard, it's automotive, it's manufacturing. Architecture is interesting. Architecture is one of those companies that has kind of an S and B profile. They have lots of offices, but they have enterprise requirements for all the hard work that they do. Then with VDI, we're very strong in financial services as well as healthcare. In fact, if you haven't seen, you should come by. We have a Bloomberg demo for financial services about the impact for traders. I have a virtualized GPU desktop. >> The speed is critical for them. Final question. Take-aways from the show this year, 2019 VMworld, Stu, we got 10 years to look back, but guys, take-aways from the show that you're going to take back from this week. >> I think there's still a lot of interest and enthusiasm. Surprisingly, there's still a lot of customers that haven't finished there migration to Windows 10 and they're coming to us saying, Oh my gosh, I only have until January, what can you do to help me? (laughing) >> Get some GPUs. Thoughts from the show. >> The multi-cloud world continues to evolve, the continued partnerships that emerge as part of this is just pretty amazing in how that's changing in things like virtual GPUs and accelerators. That experience that people have come to expect from the cloud is something, for me is a take-away. >> John Fanelli, NVIDIA, thanks for coming on. Congratulations on all the success. Kevin, Dell EMC, thanks for coming on. >> Thank you. >> Thanks for the insights. Here on theCUBE, Vmworld 2019. John Furrier, Stu Miniman, stay with us for more live coverage after this short break. (lively music)
SUMMARY :
Brought to you by VMware and its ecosystem partners. here covering all the action of VMworld, on the keynote videoing in. So for the very first time anywhere, We allow the enterprise Ask the NVIDIA guy to get some for his gaming rig. that they run to drive predictive business decisions. That's the roadmap so you can get in, on Friday, the 29th. What's the positioning there with you guys? most recently on the VDI front as well. the lowest dollar per desktop Absolutely. by putting the GPUs in to guarantee a great user experience. On the performance side too, obviously. Absolutely, and if you look at what Dell's doing, for the audience, it allows you to look and I've got all the remote workers, and that ability to support multiple users. It really is the expectation now that my rig, This is the whole category. all the collaboration to 'em, Microsoft Teams, of the new operating environment. We optimized the server and operating environment, bring in the new kind of cool visualization. and feel the fine Corinthian leather? We got to upgrade our CUBE game, on levering VR in the enterprise space. that scalable infrastructure that's easy to manage He sees data as the biggest opportunity and challenge. and at night, you run Compute. Even the user too, up and down, and just the difference, you can see such a difference You must get that all the time? that only know NVIDIA for games. and they just nerd out on the GPUs. (laughing) I think I'm in trouble here. It's certainly impacted on the infrastructure side, It's really great Value for the customers that can really drive the economics. Is there any industries that are key for you guys of the CAD/CAM industries. for all the hard work that they do. Take-aways from the show this year, that haven't finished there migration to Windows 10 Thoughts from the show. That experience that people have come to expect Congratulations on all the success. Thanks for the insights.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
$3 | QUANTITY | 0.99+ |
Michael Dell | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
2019 | DATE | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
John Fanelli | PERSON | 0.99+ |
John | PERSON | 0.99+ |
John Frer | PERSON | 0.99+ |
Kevin | PERSON | 0.99+ |
HP | ORGANIZATION | 0.99+ |
2010 | DATE | 0.99+ |
San Francisco | LOCATION | 0.99+ |
10 years | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
five-year | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
60 updates | QUANTITY | 0.99+ |
Kevin Gray | PERSON | 0.99+ |
two daughters | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
45% | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Windows 7 | TITLE | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Windows 10 | TITLE | 0.99+ |
one day | QUANTITY | 0.99+ |
Skype | ORGANIZATION | 0.99+ |
Howard Street | LOCATION | 0.99+ |
mid next year | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
iPhones | COMMERCIAL_ITEM | 0.99+ |
tenth year | QUANTITY | 0.99+ |
two GPUs | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Windows XP | TITLE | 0.99+ |
four users | QUANTITY | 0.99+ |
Dell EMC | ORGANIZATION | 0.99+ |
second announcement | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
a week | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
10th year | QUANTITY | 0.98+ |
one piece | QUANTITY | 0.98+ |
one user | QUANTITY | 0.98+ |
Windows | TITLE | 0.98+ |
this year | DATE | 0.98+ |
Pat | PERSON | 0.98+ |
Dassault | ORGANIZATION | 0.98+ |
this week | DATE | 0.98+ |
thousands | QUANTITY | 0.98+ |
eight data scientists | QUANTITY | 0.98+ |
first announcement | QUANTITY | 0.98+ |
XP | TITLE | 0.98+ |
10 years later | DATE | 0.98+ |
Stu | PERSON | 0.98+ |
first time | QUANTITY | 0.98+ |
Marty Jain, NVIDIA | DevNet Create 2019
>> live from Mountain View, California It's the queue covering definite create twenty nineteen. Brought to You by Cisco >> Welcome back to the Cube. Elisa Martin with Set Cisco Definite Create twenty nineteen at the Computer History Museum, but here all day, talking with some really great innovative folks excited to welcome to the Cube. Marty Jane, senior director of this Cisco Global Partnership and Video. Marty, It's great to have you here. >> Thank you. Good to be here. >> So I always love talking about partnerships Where what Day One of Dev. Net. Tomorrow's day to. There's been a lot of a lot of community spirit is here, so I just kind of in the spirit of partnerships, lot of collaboration that community is is really strong. Uh, before we get into kind of the details of this Cisco in video partnership first kind of thing, I wonder is all right. This is the developer community. Why the developer community within video? >> That's a great question. So if you think about way, make GP use, which is a piece of silicon graphics processing unit, and it is really only a piece of silicon until a developer comes along and develops a cool app on it. So if you think about how we go to market our large conferences called GTC, it's really developer. Focus. We have a little over a million developers in our ecosystem, and I find it very synergistic with Cisco. If you think about Suzy, we's vision. I think it's the same idea. You look at over half a million developers in their ecosystem and they want to develop collapse, and that's how your platform becomes relevant. So if you think of all the modern innovation that's coming from developers, so these are the folks that we should be talking to on a daily basis. I see a lot of commonality, a lot of synergies. In fact, we had Sisko definite come over to our conference GTC, and they they appeal to our developers. And now we're here talking to their developers and also developing some joint platforms which the the folks can use for. Like I said, the more modern *** with all the new data that's coming, whether the coyote with a machine learning automotive, smart cities, you name it, we need to be able to provide the platform to the developers >> and a number of those topics came up today, even during the keynote, Smart cities being able to utilize and accelerate work leads with a I and machine learning. They gave some great examples during the keynote of how developers can build networks. They give this cool example of I think it right off the hills of Coachella of designing a secure network for an indoor concert, designing it for an outdoor festival, Coachella and then designing it for a massive stadium like a big football game like the Super Bowl, for example. And they showed it that higher end. They showed how they're using machine, learning to zoom in on. For example, they had this little red box and you see people and what's actually in there than the machines detected was a fight and in real time, analysing this data and thence, dispatching the appropriate security to come and obviously probably take the drinks out of their hands first. But it was a really interesting, great real world example. So you guys have been partners a long time. Our you've been actually working at various companies with Cisco for a long time, but I think of Cisco and video coming together. How are you great? Something to accelerate these? Aye. Aye. And machine weren't were machine learning workloads that we're starting to see in every industry. >> You bet. Great question. So let me first comment on what you said about smart cities. I like to think of it as smart and safe cities. So actually, the first set of application will be around public safety. What the example you were giving his spot on? If you have large crowds gathering, it makes sense for us to be able to look at those clouds. Crowds? We call it intelligent video analytics or idea. In fact, we have a platform here. The Sisko i R eleven o one with a GPU added to it. So now I can wash the crowds. And if there's a fight breaking out or somebody's carrying in a weapon, you want to know somebody walks in carrying a backpack and drops it and moves on. You want to know one? Inform somebody. So what is happening is way of these millions and millions of bites of video data, >> and >> that data is not being really used today. So what we're doing is saying you know what? Let's find those pieces of intelligence and the video data and do something with it. And public safety is absolutely the highest priority. So smartest, safe city makes a lot of sense. So what we're doing is we're going to market with partners at Cisco. So what we're doing is we're saying Okay, let's design these GPS into the servers, which are connected to cameras and think about how many cameras are deployed today, probably a billion. And a lot of the video data can now be used for public safety purposes, and we basically go out and talk to large companies. We talked to governments. We talked to cities along with Sisko to go even open their eyes to what is possible today. >> Right? Because of that data is dark for so long, they don't know what they don't know. >> While most cases, what happens is you record four days of video and until something happens, nobody goes back and takes a look at it. But now we have the ability to look at the real time and cities and government's desire that very much so, >> sir example, that's such a relevant topic. I mean, they know. There's also the issue of privacy. But to your point about not just a smart city but a smart, safe city. I like that. I think it's absolutely imperative. How do you have this conversations with cities with governments about All right, this is what we want. Do we want to actually apply machine learning? So the machines are taught What that line is with privacy with those boundaries are so that a person, I'd say a lay person not in technology. Maybe is a city government official who doesn't understand the technology or need Teo will go. I get it. >> Yes. So our conversations are really about what we call you cases. So think of enterprise. A good use case would be. In fact, we work with Cisco on developing use case. You know, you always badge in into an enterprise. You have your badge, you walk in. But you also have some cases. People follow you, following you in what stops you from following me into a building. And usually people are too polite to say no, you can walk in, but we've >> all had the video training or read the manual. We know we're not >> we're not supposed to bite, but >> then you're like, I >> don't just cultural, exactly. We just can't you know that. So now we have the ability. So we trained a in a network to say, Look, if Marty's badging in, only he's allowed to walk in. And if there's a second person walks in, I want to take put Little Red Square on that face and inform security that we have had more than one person walking. So these are some of the ways. So we talk about use cases. This is one use case crowd behavior. Analytics is another use case. You know, people were walking in the backpack, dropping it. Other use case would be something like Bar to Bart loses millions of dollars year because people jumped the turnstiles and Bart didn't really have a good way of of monitoring, measuring the losses until we put a camera and captured the number of people that were jumping. The turnstiles are going in through the handicap access, okay? They were losing ten times the dollar value of what we had thought. Wow. So this is how we start the conversation with use cases, you know? And what would you like to do? Being able to count the number of cars in intersection begin with counter number of pedestrians, so you could do traffic management better. That's the language we would use with cities and governments. And then we go deeper as you go through the implementation process. >> Well, that makes perfect sense going in the use case route, because you can clearly see in that example that you mentioned with Bart a massive business outcome and an opportunity to regain a tremendous amount of resource is that they could redeploy for whether it it's new trains, new trucks, etcetera than them, not realizing we're losing how much money. I think anybody when you could put the useless in that context of this is what you can expect as an outcome. They get it >> Absolutely. That's the really the only way to start the conversation than starting from bits and bytes. And this is the This is usually the case across industries. If you think about retail, for example, you know you go to a safe way to start talking about GPS and servers. That's not the great way to start, but they do have issues with shoplifting, for example. So how do you know a person is walking in, you know, through the checkout. And they have one item. Then there's a small item right here and they walk out with this. How do you monitor that? So now you can do that with the right kind of cameras that can capture. Look there Two items, not one. How do you know where shop are stopping Which aisle is the most popular? I'Ll How do you know that? Well, now you can have cameras would say, Look, we have red zones and Green Zone so you could do those kinds of things with modern ways of doing. I >> so interesting because it's so. I mean, the examples that you gave are so disparate, but yet they make so much sense was how how you're describing it rather than going into, you know, a grocery store in talking about GPS, which they might fall over with their eyes. Doing this >> right. >> You're actually putting in the context of a real world problem they've been experiencing since the beginning of time. Don't you understand? Only goodness and this is how we can use technology. It's the safe way becomes a technology company. They don't know it. What actually started packing their bottom line. >> That's right, And so even now, you know. So I have to take that and you extend that into How do you go to market? And it's something you wanted Teo Touch on. How do you go to market with Cisco's? How does ingredients is? Could do it together, right? So think of Cisco's sales teams who are talking to all these customers every day where their retailers, financial services, federal government, health care, you name it. So what we've done is we basically sort of taking all these industries and created the top three or four use cases we know are relevant to that industry, either for safety or for saving money's. For variety of their operational reason, we have narrowed it down to three or four five use cases and each of those target industries. So what we do now with Cisco teams that we would bring them into our facility or go to them and really talkto all those use cases and train them on Hey, look, this is what we do jointly, and that makes the conversation much easier. Then they will go and present to the customer and what's the customer gets an idea far this all possible. Now that starts a deeper level technology and server and GPU engagement. So this is one way we go up and talk to different customers. What's the school's >> second? About a bit. Marcus. Cisco is so enormous, they have a billion different. I'm slightly exaggerating products with but a lot of different technologies that form many different solutions. So I imagine your Cisco expertise over many years of working with Cisco's a partner for other companies. How do you once you get to that deeper level conversation, how do you bring this different groups within Cisco together? So that that solution conversation is one that really aligns to that use case and the customer doesn't get it? >> Yeah, that's a difficult question to answer. That's like, you know your work. It's just cause a large company. But I think I also think they're also very cells driven, and that's what drives the different groups to come together. In fact, some people called me the Connector because I've been working. Cisco's so long. I know people and definite I know people in sales. I know people in the server. BU, in fact, if you think about the The platform was talking about the i r eleven o one with the jets and GPU that came as a result. I was talking to the i o t bu result talking to Dev net our situation the definite he said. You know what? This is cool are gonna do this. Then we take that to the IOC Guys is Oh, this is cool. We can take that. Put it in this platform, and then I'm next. Actually, next week I'm talking to a sale. Seaman Cisco. They cover utilities. And this platform was profit for utilities. Even think about fire monitoring in a forest. How do you do, boy thousand? The people to just watch what happens. We can take a platform like that now and really deploy it in hundreds of places which could monitor fires or the starting off a fire. But yes, bringing them together. It is no easy task. It's fun >> where you are smiling. I like that. Marty the connector. Jane, thank you >> so much for >> joining me on the kid this afternoon. Fun conversation. I enjoyed it. >> Ofcourse. Thank you. Likewise. Thank >> you, Lisa Martin for the Cube. you're watching us live, Francisco Definite. Create twenty nineteen. This is the end of day one. Stick around, John. Failure on I will be back tomorrow to cover day too. Thanks for watching.
SUMMARY :
live from Mountain View, California It's the queue covering Marty, It's great to have you here. Good to be here. So I always love talking about partnerships Where what Day One of Dev. So if you think about how we go to market our large conferences called GTC, So you So let me first comment on what you said about smart cities. So what we're doing is we're going to market with partners at Cisco. Because of that data is dark for so long, they don't know what they don't know. While most cases, what happens is you record four days of video and until something happens, How do you have this conversations with But you also have some cases. all had the video training or read the manual. And then we go deeper as you go through the implementation process. Well, that makes perfect sense going in the use case route, because you can clearly see in that example that you mentioned So now you can do that with the right I mean, the examples that you gave are so disparate, Don't you understand? So I have to take that and you extend that into How do you go to market? How do you once you get to that in fact, if you think about the The platform was talking about the i r eleven o one with the jets where you are smiling. joining me on the kid this afternoon. Thank This is the end of day one.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Marty Jain | PERSON | 0.99+ |
Marty Jane | PERSON | 0.99+ |
Elisa Martin | PERSON | 0.99+ |
Jane | PERSON | 0.99+ |
Marcus | PERSON | 0.99+ |
millions | QUANTITY | 0.99+ |
ten times | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
four days | QUANTITY | 0.99+ |
John | PERSON | 0.99+ |
tomorrow | DATE | 0.99+ |
Francisco | PERSON | 0.99+ |
Mountain View, California | LOCATION | 0.99+ |
Marty | PERSON | 0.99+ |
next week | DATE | 0.99+ |
Two items | QUANTITY | 0.99+ |
Super Bowl | EVENT | 0.99+ |
each | QUANTITY | 0.99+ |
Coachella | EVENT | 0.99+ |
one item | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
more than one person | QUANTITY | 0.99+ |
NVIDIA | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
IOC | ORGANIZATION | 0.98+ |
first set | QUANTITY | 0.98+ |
second person | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
2019 | DATE | 0.97+ |
over half a million developers | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
four use cases | QUANTITY | 0.97+ |
Sisko | ORGANIZATION | 0.96+ |
Seaman | PERSON | 0.96+ |
millions of dollars | QUANTITY | 0.96+ |
Tomorrow | DATE | 0.96+ |
five use cases | QUANTITY | 0.95+ |
a billion | QUANTITY | 0.94+ |
Little Red Square | LOCATION | 0.94+ |
twenty nineteen | QUANTITY | 0.91+ |
Bart | ORGANIZATION | 0.9+ |
one way | QUANTITY | 0.89+ |
Teo Touch | ORGANIZATION | 0.88+ |
this afternoon | DATE | 0.87+ |
thousand | QUANTITY | 0.86+ |
DevNet | ORGANIZATION | 0.85+ |
Bar to | ORGANIZATION | 0.85+ |
one use case | QUANTITY | 0.84+ |
over a million developers | QUANTITY | 0.84+ |
i R eleven | COMMERCIAL_ITEM | 0.82+ |
twenty nineteen | QUANTITY | 0.82+ |
Bart | PERSON | 0.82+ |
millions of bites | QUANTITY | 0.81+ |
four | QUANTITY | 0.79+ |
first comment | QUANTITY | 0.75+ |
second | QUANTITY | 0.74+ |
Day One | QUANTITY | 0.74+ |
Cisco Global Partnership | ORGANIZATION | 0.7+ |
day one | QUANTITY | 0.7+ |
Computer History Museum | LOCATION | 0.7+ |
Sisko | PERSON | 0.58+ |
nd of | QUANTITY | 0.56+ |
Suzy | ORGANIZATION | 0.56+ |
Teo | PERSON | 0.55+ |
Dev. Net | ORGANIZATION | 0.42+ |
GTC | EVENT | 0.32+ |
Renee Yao, NVIDIA & Bharat Badrinath, NetApp
>> Announcer: Live from Las Vegas, it's theCUBE, covering NetApp Insight 2018. Brought to you by NetApp. >> Welcome back to theCUBE, we are live. We've been here all day at NetApp Insight in Las Vegas at the Mandalay Bay. I'm Lisa Martin with Stu Miniman and we're joined by a couple of guests. One of our alumni, Bharat Badrinath, the V.P. of Product Solutions and Marketing at NetApp. Hey, Bharat, welcome back. >> Thank you, thanks for having me. >> And we've also got Renee Yao, who is a Senior Product Marketing Manager for Deep Learning and AI Systems at Nvidia. Renee, welcome to theCUBE. >> Thanks for having me. >> So guys, this is a pretty big event. NetApp's biggest customer-partner event, the keynote, standing room only this morning five thousand plus people, lot of buzz, lot of momentum. Speaking of momentum, NetApp and Nvidia just launched an interesting partnership a couple months ago. Bharat, talk to us about how NetApp is working with Nvidia to really take advantage of AI and allow your customers to do that as well. >> Sure. So, as we started talking to customers and started looking at what they were investing in, AI bubbled up, right up to the top. And given our rich history in NFS, high performance NFS, it became an obvious choice for NetApp to invest in this space. So we've been working with Nvidia for a really long time, probably close to a year, to start integrating our products with their DGX-1 supercomputer and providing it as a single package to our customers, which makes it a lot easier for them to deploy their AI instead of waiting months for testing infrastructure, which the data scientists don't want to do. We get them a pre-tested, pre-validated system and our All-Flash Fast, which has been winning multiple awards and the recent A800 announcement were perfect choice for us to integrate into this architecture for the system. >> Alright, Renee, in the keynote this morning, the Futurist, he said-- We talked about data as the new oil, he said AI is the new electricity. Maybe you can speak a little bit as to why this is so important. Having gone to a lot of shows this year, it felt like every single show I go to, I see Nvidia, arm in arm with partners, because there's a huge wave coming. >> Yes, absolutely, and I think there was this hype about data, there was this hype about AI, and I think the years of Big Data World, that's creating data, absolutely the foundation for AI, and AI as the new electricity is a very, very good analogy. And let's do some math, shall we? So Swiss Federal Railway, it's a very good customer of ours. For those of you who don't know, they're kind of like the heart or center of all the railway tracks going through, serving about 1.2 million passengers on a day-to-day basis. Securing their security is very, very important. Now, they also have a lot of switches that turn on, then the train can go by and with the tunnels and bridges and switches, so they need to make sure that these trains actually don't collide. So when one train goes by with 11 switches, that gives you 30 ways of possible routing. Two trains, 900 ways. 80 trains, 10 to the eightieth power of ways. That's more than the observed atoms in the universe. And they actually have more than 10 thousand trains. So think about, can human being possibly calculate that much data and possibilities in their brain? As smart as we all want to think we all are, they turn to DGX, and the full day of simulation on DGX-1 was only 17 seconds for them to get back results. And I think that analogy of AI as the new electricity, just talking about the speed of light, is very spot on. >> So this isn't hype anymore, this is actually reality. And you gave a really great example of how a large transportation system is using it to get almost real time information. Bharat, talk to us about NetApp storage, history, 26 years, you guys have really made a lot of pivots in terms of your digital transformation, your cultural transformation. How are you helping with, now, kind of the added power of Nvidia, helping customers to, the hype's gone, actually deploy it, live it, and benefit a business from it? >> Yeah, absolutely, I think, as you rightly pointed out, NetApp has made a lot of pivots. Right, and I think the latest journey in terms of being empowering our customers with data has been a very powerful mission for the company. We entered the Flash market a little bit later than our competitors, but we have made dramatic progress in that space. In fact, recently, based on the latest IDC report, we were number one in All-Flash market worldwide, so that is quite an accomplishment for a company which was late to the market. And having said that, that's because of the innovation engine that is still alive and well within NetApp. We're announcing, as you've seen in the conference, we're announcing a lot of new products and technology which are way ahead of what our competitors are offering, but I think it is all hinged on what our customers need. The customer benefits because, yeah, it has profound benefit of changing how customers operate, their entire operations, it can transform dramatically overnight. And as Renee pointed out, Big Data gave the foundation which collected all the data, but wasn't able to process it. But AI with the power of Nvidia and DGX is able to utilize that to create those outcomes for customers. And from our perspective, we bring two key value adds to the space. One, we're able to serve up the data at incredibly high speeds with our award-winning All-Flash systems. But more importantly, data today lives everywhere. If you think about it, edge is becoming even more important. You can't expect an autonomous car to make an instantaneous decision without the backing of data, which means it can't, everything can't reside in the cloud, it may be at the edge. Some of it may be at your data center. How do you tie all three together, edge, core, and cloud? And that's where the data fabric, the vision of data fabric that you saw today comes in the picture. So one is performance, the ability to stream up the kind of data at the speed of the new processors are demanding, at the speed the customers are demanding to make business decisions and also the edge to core to cloud, our data fabric, which is unique and unparalleled in the industry. >> Now, I'm wondering if you could both bring us inside the customers a little bit. If I think of the traditional storage customer, I need performance, I have more and more data that I need to deal with. But Renee pointed out real outcomes, which is beyond what a traditional storage person would be doing. Who are you working with at the customers-- How do they put together-- It almost sounds like you're building a car. I've got the engine, I've got all the pieces. Who helps put this whole solution together? How does the partnership on the customer's side go together? >> That's a great question. I'll give my take and you can jump on it because she's just returned from being on road shows with joint customers and prospects. So I believe it has to be a joint decision. It's not like IT does it first and the data scientists come in later. Although it may be the case in certain instances where the data scientists start the discussion and then the IT gets brought in. In an ideal case, just like building a car, you want all the teams to be sitting together, make sure they're making the right calls because every compromise you make at one end will impact the other. So you want to make sure you make the optimal decision end to end. And that's where some of our channel partners come in who kind of bridge the data scientist team and the IT team. In some cases, customers show up with data scientists and IT teams together and some, it's one after the other. >> Absolutely. We see the same thing when we're on the road show. Literally two weeks ago, in Canada, by the way, there was a snowstorm, and it was an unforeseen snowstorm, you don't get snowstorm in October-- >> Yes, even for Canada, it was unforeseen. >> Yeah, and we had a packed room of people coming to learn about AI and in the audience, we absolutely see people from the infrastructure side, from the data center side, from the data scientist side, and they realized that they really have to start talking because none of them can afford to be reactive. For example, the data scientists, we want to do the innovation. I can't just go to the infrastructure guys and say that, "Hey, this is my workload, do something about it." And the infrastructure guys don't want to hold on to that problem and then don't know what to do with it. They really need to be ahead of everything and I think the interesting thing is, among those four cities that we're at, we see customers from the government, oil and gas, transportation, health care, and just any industry you can think of, they're all here. One specific example, do you know Mike's company that actually came to us, they have about 15 petabytes of data and that's storing 20 years of historical data and they only have two staff and they were not hiring more staff. They were like, "We just want something that's "going to be able to work and we know everything, "so just give us a solution that's going to be able to "easily scale up and out and enable us to continue to "store more data, manage more data, "and get insights out of data fast." So they came to both of us, it's just a very good, natural decision. That's why we have a partnership together as well. >> So you guys talked about kind of connecting the data scientists with the infrastructure folks. Where's the business involved in this conversation? In terms of, we want to identify new products and services to deliver faster than our competition, new markets. Talk to us about, are the data scientists and the infrastructure guys and girls following business initiatives that have been set or are the business leaders involved in these joint conversations? >> Go ahead, you take it. >> Sure. So, I think we see both. We definitely see that there's top-level executives saying that this is our initiative and we have to do it. And they will make the decision that we have to refresh our infrastructure from the ground up to make sure we're supportive of our data scientists' innovation. We've also seen brilliant minds, researchers, data scientists doing amazing things and then roll it up to the VP level and then roll it up to CEO level to say that this has to be done because this-- For example, that simulation of 17 second results, it's things that people used to cannot do in their lifetime, now they can do it in seconds, that kind of innovation just cannot be ignored. >> Yeah, we see the same thing. In fact, any team that has possession of that data or is accountable for that data is the one usually driving the decisions. Because as you mine the data, as you start deploying new techniques, you realize new opportunities, which means the business gets more interested in it and vice versa. If the business is interested, they're going to look for those answers within the data that they have. >> So last thing, Renee, you were on the Women in Tech panel that ended yesterday, Bharat and I were both in the audience, and one of the things that I thought was really inspiring about your story is that you had given us, the audience, an interesting example of a TV opportunity that you were inspired to do by the CEO of Nvidia. Give our audience who didn't have a chance to see that panel a little bit, and in the last minute, of that story and how you were able to step forward and go, "I'm going to try this." >> Yeah, of course. I think that brings us back to the concept that we have at Nvidia, the speed of light concept, and you really have to learn, act, to move at the speed of light, just like our GPUs, with extreme performance. And obviously, at that speed, none of us know everything. So what Jensen, CEO, shared with us was, in an all-hands meeting internally, he told us that none of us are here qualified to do any of our jobs, maybe besides his legal counsel and CFO. And all of us are here to learn, and we need to learn as fast and as much as we can. And we can't really just let the competition determine where our limit is, but instead is by the limit of what is possible. So that is very much a fundamental mindset change in this AI revolution. >> Well thanks so much, Renee and Bharat, for stopping by and sharing with us the exciting things that you guys are doing with NetApp. We look forward to talking with you again soon. >> Thank you. >> Me too, thanks. >> For Stu Miniman, I'm Lisa Martin. You're watching theCUBE, live from NetApp Insight 2018 in Las Vegas. Stu and I will be right back with our next guests after a short break. (techno music)
SUMMARY :
Brought to you by NetApp. in Las Vegas at the Mandalay Bay. And we've also got Renee Yao, the keynote, standing room only this morning and providing it as a single package to our customers, Alright, Renee, in the keynote this morning, and AI as the new electricity is a very, very good analogy. kind of the added power of Nvidia, So one is performance, the ability to stream up How does the partnership on the customer's side go together? the optimal decision end to end. We see the same thing when we're on the road show. and they realized that they really have to start talking the data scientists with the infrastructure folks. refresh our infrastructure from the ground up If the business is interested, they're going to look for and one of the things that I thought was the speed of light concept, and you really have to learn, We look forward to talking with you again soon. Stu and I will be right back
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Renee | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Renee Yao | PERSON | 0.99+ |
Stu | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
Mike | PERSON | 0.99+ |
10 | QUANTITY | 0.99+ |
Canada | LOCATION | 0.99+ |
11 switches | QUANTITY | 0.99+ |
30 ways | QUANTITY | 0.99+ |
80 trains | QUANTITY | 0.99+ |
900 ways | QUANTITY | 0.99+ |
Swiss Federal Railway | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
one train | QUANTITY | 0.99+ |
Bharat | PERSON | 0.99+ |
Two trains | QUANTITY | 0.99+ |
Bharat Badrinath | PERSON | 0.99+ |
One | QUANTITY | 0.99+ |
October | DATE | 0.99+ |
26 years | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Mandalay Bay | LOCATION | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
more than 10 thousand trains | QUANTITY | 0.99+ |
DGX | ORGANIZATION | 0.99+ |
NetApp | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
17 seconds | QUANTITY | 0.99+ |
two weeks ago | DATE | 0.99+ |
two staff | QUANTITY | 0.98+ |
five thousand plus people | QUANTITY | 0.98+ |
two key | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
this year | DATE | 0.98+ |
Jensen | PERSON | 0.97+ |
single package | QUANTITY | 0.97+ |
Deep Learning | ORGANIZATION | 0.97+ |
NetApp | TITLE | 0.96+ |
IDC | ORGANIZATION | 0.96+ |
one | QUANTITY | 0.96+ |
NVIDIA | ORGANIZATION | 0.96+ |
about 1.2 million passengers | QUANTITY | 0.95+ |
Systems | ORGANIZATION | 0.94+ |
eightieth power | QUANTITY | 0.94+ |
first | QUANTITY | 0.94+ |
NetApp Insight | ORGANIZATION | 0.92+ |
couple months ago | DATE | 0.91+ |
this morning | DATE | 0.89+ |
about 15 p | QUANTITY | 0.89+ |
a year | QUANTITY | 0.87+ |
DGX-1 supercomputer | COMMERCIAL_ITEM | 0.87+ |
Big Data | ORGANIZATION | 0.86+ |
17 second results | QUANTITY | 0.84+ |
couple of guests | QUANTITY | 0.78+ |
theCUBE | ORGANIZATION | 0.77+ |
four cities | QUANTITY | 0.76+ |
number one | QUANTITY | 0.76+ |
three | QUANTITY | 0.75+ |
Premal Savla, NVIDIA & Tom Eby, Micron | Micron Insight'18
>> Live from San Francisco, it's theCUBE, covering Micron Insight 2018. Brought to you by Micron. >> Welcome back to San Francisco everybody. You're watching theCUBE the leader in live tech coverage. I'm Dave Vellante. He's David Floyer, and we're covering Micro Insight'18. It's all about bringing together artificial intelligence and the memory and storage requirements. We're here on the embarcadero. We've got treasure island that way. We've got the financial district over there. We've got Golden Gate bridge behind us. Tom Eby is here as senior vice president and GM of Micron's booming compute and networking business unit. Good to see you Tom. >> Great to be here. >> And Permal Savla is here. He's the director of deep learning at NVIDIA. Welcome. >> Thank you. >> So obviously some of these new emerging work loads require collaboration between folks like Micron and folks like NVIDIA. But Tom why don't you kick it off. What are some of the big trends that you're seeing in some of these alternative work loads that's driving this collaboration? >> Well a lot of what we're talking about here today is the drive of AI and machine learning work loads, and the implications for memory. Certainly there's a host of them, natural language processing, photo and image recognition, applications in medical research, applications in optimizing manufacturing like we're doing in our fabs, and there's many many more. And of course what's exciting for us is that to support those in an optimized way really does require the mating of the optimal processing architecture, things like GPUs. With the right high band width with low latency memory and storage solutions. That's what leads to great partner ships between partnerships like Micron and NVIDIA. >> David was explaining at our open the intensity of the work loads that you guys are serving, and how much more resources that requires to actually deliver the type of performance. Maybe you could talk about some of the things that you're seeing in terms of these emerging work loads. >> Yes, so at NVIDIA, we build systems for X rated computing. AI and deep learning is a very quickly expanding field at this point which needs a lot of CP horse power. What we are seeing is that different applications like you said there's image processing, whether it's video, whether it's natural language processing the amount of data that is there, that is required to do deep learning and AI around it, we break it up into two work flows. One is the training where you actually train the software, and make it intelligent enough to then go and do inference later on. So that you can go and get you results out of it at the end of it. We concentrate on this entire workflow. That's where when we are looking at it from a training perspective, the GPU gives it the processing power. But at the same time all the other components around it perform at the peak. That's where the memory comes in. That's where the storage comes in, and we need to process that data very quickly. >> Yeah, so we know from system's design that you got to have a balanced system or else you're just going to push the bottle necks around. We've learned that over the years, but so it's more than just slapping on a bunch of storage and a bunch of memory. You're doing some other deeper integration, is that correct and what is that integration? >> Yeah, I think the two companies have had a great relationship, just to talk about a couple examples. We essentially co-defined a technology called GEDR 5X, which greatly enhanced the speed of graphics technology. We gently introduced that to the marketplace with NVIDIA about 18 months ago. And then worked with them again very closely on a technology called GDDR six, which is the next generation of even faster technology. We were their launch and ran partner for their recently announced G-force RTX line of cards. It's a very deeply engaged early in the process, define the process, define the standards, jointly develop the solution. Very intimate sharing in the supply chain area. It's a great relationship for us. We're excited about how we can continue to expand and extend that relationship by going forward. >> So obviously there's the two parts of it. You said the learning part of it, and the inference part of the computing. What do you think is the difference between the two? I mean obviously at the end of the day, the inference part is critical. That's got to be the fastest response time. You have to have that in real time. Can you talk a little bit about what you're doing to really speed that up, to make that micro seconds as opposed to milliseconds? >> So from an NVIDIA perspective we build the entire end to end tools steps for training and inferencing. We have a set of libraries that we have made it openly available for all of our customers, all our partners, and all users. So that they can go download it, and do the training so they can use the different frameworks and libraries to accelerate the work that they're doing. And then transform it onto the inference spot. We have something called denser RT, which is basically denser real time. That gives the capability to get these answers very quickly. So on our D4 of the tuning, Chip said that we just announced. We can get a very high performance for our image. So any kind of image recognition or image processing that we need to do, we can do that on the systems very quickly. And we can meet, rebuild entire architectures. So it's not just about one piece. It's about the whole end to end architecture of the system. >> So we heard earlier today in the analyst briefing, the press briefing that Micron certainly in the last 40 years has changed. We're seeing a lot more diversity. Usually it'd be all about PCs. Now there's just so many alternative work loads emerging. Clearly NVIDIA is playing there as well with alternative processing capabilities. What do you guys see as some of the more exciting, emerging work loads that are going to require continued collaboration and innovation? >> Yeah, well I think to build a little bit on some of the other comments about the need for real time inference, one of the things in the area of diversity that we've found interesting. The relationship between Micron and NVIDIA in high performance memory really started around their graphics business. But we are seeing in other markets closer to the edge, in automotive, in networking and in other areas where there's a need for that real time performance. Yet there's also a need for a degree of cost effectiveness. Perhaps a little more so than in the data center. That we're seeing technologies like GDR six being applied to a much broader range of applications like automotive, like networking, like Edge AI, to provide the performance to get that real time response but in a form factor and at a cost point that's affordable for the application. >> Anything you'd add to that Permal? >> So I would also add you talked about applications, different applications that are changing right? Today we announced a new set of libraries and tools for the analytic space. That's again a big work load in the enterprise data centers, that we are trying to optimize and accelerate with machine learning. So we announced a whole set of tools which take in these large data sets that are coming in, and applying it in the data centers and using it to get answers very quickly. So that's what NVIDIA is also doing is expanding on these capabilities as we go in. And as these components and as these technologies get better it just gets our answers much more quickly. >> As exacts in the space and you guys both, you're component manufacturers, and so you sell to people who sell to end consumers. How do you get your information in that sort of pull through? Obviously you work with your customers very closely. >> Mm-hm. >> How do you get visibility to their customers? Just going to go to shows, you go do joint sales calls, how does that all work? >> Certainly some of that is in discussions with our customers and their marketing groups about what they're seeing from a customer point of view. But certainly there's other paths. One of the reasons behind the hundred million dollar venture fund that we announced today, is one of the best ways to get that advanced insight, is to be working with some of the most innovative start ups that understand what some of those end users needs might be and are developing some unique technologies. So there's a range. Working with our customers through eventually finding others, but it's important that we understand those needs because the lead time to developing the solutions both memory and processing architectures is quite well. >> Of course everybody wants to work with NVIDIA, you guys have an inundated like come on oh no we're the most. We're tied up now. Of course there's not a lot of choices here when you're talking about the levels of components that you're selling. But what's life like at NVIDIA? I mean they've been knocking down your doors to do partnerships. >> I think we've grown from being just the component to now being a complete system and an architecture. We don't only just build just a chip that the GPU was. We also build full SLCs. We also build the libraries, software, and the tools that are required to make this complete end to end solutions. We also do a lot of open source technologies because we want our customers and our end cast partners to build and take what we have and go beyond what it's capable of. That's where we end value at the end of the day. Yes, it's all of us together. We need to work together to make that much more faster as we go. >> The tuning is incredibly important. This is complicated stuff. It doesn't just work out of the box, right? So you need an ecosystem as well. >> Yes. >> Yes. >> That's what you guys have been out building. Tom, well give your final thoughts. >> Yeah well I guess to build a little bit. Certainly NVIDIA is moving up the stack in terms of the ecosystem, the software, the complete solution and I think Micron does as well. Like you commented, traditionally it was a component play. And increasingly, we're going to be building subsystems in memory and storage that occurs today on the storage side. I think we'll increasingly see that in memory, and with some of the future, very promising technologies like 30 Cross Point. >> Yeah it's the dawn of the days where everybody just gets piece parts and put them all together. They need you you guys to do more integration, and more out of the box like you say subsystems. So guys thanks very much for coming on theCUBE. Really appreciate it. >> Thank you. >> Thank you. >> Alright you're welcome, keep it right there everybody. We'll be back in San Francisco, you're watching theCUBE from Micron Insight 2018, accelerate intelligence. We'll be right back after this short break. (music)
SUMMARY :
Brought to you by Micron. and the memory and storage requirements. He's the director of What are some of the big trends that you're seeing and the implications for memory. of the work loads that you guys are serving, One is the training where you actually train the software, We've learned that over the years, We gently introduced that to the marketplace and the inference part of the computing. That gives the capability to get these answers as some of the more exciting, emerging work loads some of the other comments about the need for the data centers and using it to get answers very quickly. As exacts in the space and you guys both, because the lead time to developing the solutions that you're selling. We don't only just build just a chip that the GPU was. So you need an ecosystem as well. That's what you guys have been out building. in terms of the ecosystem, the software, and more out of the box like you say subsystems. We'll be back in San Francisco, you're watching theCUBE
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
David | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Tom | PERSON | 0.99+ |
Micron | ORGANIZATION | 0.99+ |
Tom Eby | PERSON | 0.99+ |
two companies | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Premal Savla | PERSON | 0.99+ |
two parts | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
hundred million dollar | QUANTITY | 0.98+ |
Golden Gate | LOCATION | 0.98+ |
G-force RTX | COMMERCIAL_ITEM | 0.98+ |
one piece | QUANTITY | 0.98+ |
both | QUANTITY | 0.97+ |
Chip | PERSON | 0.97+ |
Permal Savla | PERSON | 0.95+ |
Edge | TITLE | 0.93+ |
earlier today | DATE | 0.9+ |
last 40 years | DATE | 0.89+ |
about 18 months ago | DATE | 0.85+ |
six | COMMERCIAL_ITEM | 0.81+ |
two work | QUANTITY | 0.79+ |
couple examples | QUANTITY | 0.79+ |
GEDR 5X | COMMERCIAL_ITEM | 0.77+ |
Micron Insight'18 | ORGANIZATION | 0.69+ |
theCUBE | ORGANIZATION | 0.68+ |
GDDR | OTHER | 0.51+ |
GDR | OTHER | 0.47+ |
Insight 2018 | TITLE | 0.44+ |
Micro Insight'18 | ORGANIZATION | 0.44+ |
2018 | TITLE | 0.43+ |
Point | COMMERCIAL_ITEM | 0.39+ |
Micron | EVENT | 0.38+ |
D4 | COMMERCIAL_ITEM | 0.38+ |
30 | OTHER | 0.35+ |
Cross | TITLE | 0.31+ |
Insight | EVENT | 0.25+ |
Kurt Kuckein, DDN Storage, and Darrin Johnson, NVIDIA | CUBEConversation, Sept 2018
[Music] [Applause] I'll Buena Burris and welcome to another cube conversation from our fantastic studios in beautiful palo alto california today we're going to be talking about what infrastructure can do to accelerate AI and specifically we're gonna use a relationship a burgeoning relationship between PDN and nvidia to describe what we can do to accelerate AI workloads by using higher performance smarter and more focused of infrastructure for computing now to have this conversation we've got two great guests here we've got Kurt ku kind who is the senior director of marketing at ddn and also Darren Johnson is a global director of technical marketing for enterprise and NVIDIA Kurt Gerron welcome to the cube thanks for thank you very much so let's get going on this because this is a very very important topic and I think it all starts with this notion of that there is a relationship that you guys have put forward Kurt once you describe it sure well so what we're announcing today is ddn's a3i architecture powered by Nvidia so it is a full rack level solution a reference architecture that's been fully integrated and fully tested to deliver an AI infrastructure very simply very completely so if we think about how this is gonna or why this is important AI workloads clearly have a special stress on underlying technology Darin talk to us a little bit about the nature of these workloads and why in particular things like GPUs and other technologies are so important to make them go fast absolutely and as you probably know AI is all about the data whether you're doing medical imaging whether you're doing natural language processing whatever it is it's all driven by the data the more data that you have the better results that you get but to drive that data into the GPUs you need great IO and that's why we're here today to talk about ddn and the partnership of how to bring that I owe to the GPUs on our dgx platforms so if we think about what you described a lot of small files off and randomly just riveted with nonetheless very high-profile jobs that just can't stop midstream and start over absolutely and if you think about the history of high-performance computing which is very similar to a I really I owe is just that lots of files you have to get it they're low latency high throughput and that's why ddn's probably nearly twenty years of experience working in that exact same domain is perfect because you get the parallel file system which gives you that throughput gives you that low latency just helps drive the GPU so we you'd mention HPC from 20 years of experience now it used to be that HPC you'd have scientists with a bunch of graduate students setting up some of these big honkin machines but now we're moving into the commercial domain you don't have graduate students running around you don't have very low cost high quality people you're you know a lot of administrators who nonetheless good people but a lot to learn so how does this relationship actually start making or bringing AI within reach of the commercial world exactly where this reference architecture comes in right so a customer doesn't need to start from scratch they have a design now that allows them to quickly implement AI it's something that's really easily deployable we've fully integrated this solution ddn has made changes to our parallel file system appliance to integrate directly within the DG x1 environment makes that even easier to deploy from there and extract the maximum performance out of this without having to run around and tune a bunch of knobs change a bunch of settings it's really gonna work out of the box and the you know nvidia has done more than just the DG x1 it's more than hardware you've done a lot of optimization of different of AI toolkits if Sarah I'm talking what about that Darin yeah so I mean talking about the example I use researchers in the past with HPC what we have today are data scientists data scientists understand pie tours they understand tensorflow they understand the frameworks they don't want to understand the underlying filesystem networking RDMA InfiniBand any of that they just want to be able to come in run their tensorflow get the data get the results and just turn that keep turning that whether it's a single GPU or 90 Jex's or as many dejection as you want so this solution helps bring that to customers much easier so those data scientists don't have to be system administrators so a reference architecture that makes things easier but that's more than just for some of these commercial things it's also the overall ecosystem new application providers application developers how is this going to impact the aggregate ecosystem it's growing up around the need to do AI related outcomes well I think one point that Darrin was getting to you there and one of the big effects is also as these ecosystems reach a point where they're going to need to scale right there's somewhere where ddn has tons of experience right so many customers are starting off with smaller data sets they still need the performance a parallel file system in that case is going to deliver that performance but then also as they grow right going from one GPU to 90 G X's is going to be an incredible amount of both performance scalability that they're going to need from their i/o as well as probably capacity scalability and that's another thing that we've made easy with a3i is being able to scale that environment seamlessly within a single namespace so that people don't have to deal with a lot of again tuning and turning of knobs to make this stuff work really well and drive those outcomes that they need as they're successful right so in the end it is the application that's most important to both of us right it's it's not the infrastructure it's making the discoveries faster it's processing information out in the field faster it's doing analysis of the MRI faster it's you know helping the doctors helping the anybody who's using this to really make faster decisions better decisions exactly and just to add to that I mean in automotive industry you have datasets that are from 50 to 500 petabytes and you need access to all that data all the time because you're constantly training and Retraining to create better models to create better autonomous vehicles and you need you need the performance to do that ddn helps bring that to bear and with this reference architecture simplifies it so you get the value add of nvidia gpus plus its ecosystem of software plus DD on its match made in heaven Darren Johnson Nvidia Curt Koo Kien ddn thanks very much for being on the cube thank you very much and I'm Peter burrs and once again I'd like to thank you for watching this cube conversation until next time [Music]
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
Darren Johnson | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
Kurt Kuckein | PERSON | 0.99+ |
Sarah | PERSON | 0.99+ |
Sept 2018 | DATE | 0.99+ |
ddn | ORGANIZATION | 0.99+ |
nvidia | ORGANIZATION | 0.99+ |
Kurt Gerron | PERSON | 0.99+ |
Kurt | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Darrin Johnson | PERSON | 0.99+ |
today | DATE | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.98+ |
50 | QUANTITY | 0.98+ |
two great guests | QUANTITY | 0.98+ |
one point | QUANTITY | 0.96+ |
500 petabytes | QUANTITY | 0.96+ |
Curt Koo Kien | PERSON | 0.96+ |
PDN | ORGANIZATION | 0.96+ |
palo alto california | LOCATION | 0.95+ |
one GPU | QUANTITY | 0.94+ |
one | QUANTITY | 0.93+ |
DDN Storage | ORGANIZATION | 0.92+ |
Peter burrs | PERSON | 0.88+ |
nearly twenty years | QUANTITY | 0.86+ |
lots of files | QUANTITY | 0.85+ |
90 G X | QUANTITY | 0.83+ |
single namespace | QUANTITY | 0.79+ |
Burris | PERSON | 0.75+ |
single GPU | QUANTITY | 0.74+ |
DG x1 | TITLE | 0.74+ |
90 Jex | QUANTITY | 0.66+ |
a lot of small files | QUANTITY | 0.62+ |
gpus | COMMERCIAL_ITEM | 0.61+ |
Darrin | ORGANIZATION | 0.56+ |
experience | QUANTITY | 0.52+ |
9_20_18 DDN Nvidia Launch about Benchmarking with PETER & KURT KUCKEIN
(microphone not on) >> be 47 (laughter) >> Are you ready? >> Here we go, alright and, three, two... >> You know it's great to see real benchmarking data, because this is a very important domain and there is not a lot of benchmarking information out there around some of these other products that are available. But let's try to to turn that benchmarking information into business outcomes, and to do that we got, Kurt Kuckein, back from DDN. Kurt welcome back let's talk a bit about how are these high value outcomes that business seeks with AI going to be achieved as a consequence of this new performance, faster capabilities, etcetera. >> So there's a couple of considerations, the first consideration I think is just the selection of AI infrastructure itself. Right, we have customers telling us constantly that they don't know where to start. Now that they have readily available reference architectures that tell them, hey here's something you can implement get installed quickly, you're up and running, running your AI from day one. >> So the decision process for what to get is reduced. >> Exactly. >> Okay. >> Uh, number two is you're unlocking all ends of the investment with something like this right? You're maximizing the performance on the GPU side. You're maximizing the performance on the ingest side for the storage. You're maximizing the through-put of the entire system, so you're really gaining the most out of your investment there. And not just gaining the most out of the investment, but truly accelerating the application and that's the end goal right, that we're looking for with customers. Plenty of people can deliver fast storage, but it does- If it doesn't impact the application and deliver faster results, cut run times down, then what are you really gaining from having fast storage? And so that where we're focused, we're focused on application acceleration. >> So simpler architecture, faster implementation based on that, integrated capabilities, ultimately, all revealing or all resulting in, better application performance. >> Better application performance, and in the end something that's more reliable as well. >> Kurt, thanks for again for being on The Cube. >> Thanks for having me.
SUMMARY :
and to do that we got, Kurt Kuckein, back from DDN. the first consideration I think is just You're maximizing the performance on the GPU side. So simpler architecture, and in the end something that's more reliable as well.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Kurt Kuckein | PERSON | 0.99+ |
Kurt | PERSON | 0.99+ |
KURT KUCKEIN | PERSON | 0.99+ |
PETER | PERSON | 0.99+ |
first consideration | QUANTITY | 0.98+ |
two | QUANTITY | 0.97+ |
three | QUANTITY | 0.94+ |
47 | QUANTITY | 0.93+ |
DDN | ORGANIZATION | 0.91+ |
day one | QUANTITY | 0.84+ |
number two | QUANTITY | 0.79+ |
Nvidia | ORGANIZATION | 0.79+ |
Cube | COMMERCIAL_ITEM | 0.59+ |
DDN | EVENT | 0.43+ |
9_20_18 DDN Nvidia Launch AI & Storage with PETER & KURT KUCKEIN
(laughing) >> This is V-3. >> Alec, you're going to open up, we're going to cut, come to you in a second. Good luck, buddy. Okay, here we go. Alright Peter, ready? >> Yup. >> And we're coming to you in. >> Hold on guys, sorry, I lied. (laughing) V-2, V-3, there it is. Okay, ready. >> Now you're ready? >> Yup. >> You're ready ready? Okay here we go, ready and, three, two. >> Hi, I'm Peter Burris, welcome to another Cube Conversation from our wonderful studios in beautiful Palo Alto, California. Great conversation today, we're going to be talking about the relationship between AI, business, and especially some of the new infrastructure technologies in the storage part of the stack. And to join me in this endeavor is Kurt Kuckein, who's a senior director of product marketing at DDN. Kurt Kuckein, welcome to The Cube. >> Thanks, Peter, happy to be here. >> So tell us a little bit about DDN to start. >> So DDN is a storage company that's been around for 20 years. We've got a legacy in high-performance computing, and that's what we see a lot of similarities with this new AI workload. DDN is well-known in that HPC community; if you look at the top 100 supercomputers in the world we're attached to 75-percent of them and so we have a fundamental understanding of that type of scalable need that's where we're focused, we're focused on performance requirements, we're focused on scalability requirements, which can mean multiple things, right, it can mean the scaling of performance, it can mean the scaling of capacity, and we're very flexible. >> Well let me stop you and say, so you've got a lot of customers in the high-performance world, and a lot of those customers are at the vanguard of moving to some of these new AI workloads. What are customers saying? With this significant engagement that you have with the best and the brightest out there, what are they saying about this transition to AI? >> Well I think it's fascinating that we kind of have a bifurcated customer base here, where we have those traditionalists who probably have been looking at AI for over 40 years, right, and they've been exploring this idea and they've gone through the peaks and troughs in the promise of AI, and then contraction because CPUs weren't powerful enough. Now we've got this emergence of GPUs in the supercomputing world, and if you look at how the supercomputing world has expanded in the last few years, it is through investment in GPUs. And then we've got an entirely different segment, which is a much more commercial segment, and they're maybe newly invested in this AI arena, right, they don't have the legacy of 30, 40 years of research behind them, and they are trying to figure out exactly, you know, what do I do here? A lot of companies are coming to us, hey, I have an AI initiative, well what's behind it? Well, we don't know yet, but we've got to have something and they don't understand where is this infrastructure going to come from. >> So the general availability of AI technologies, and obviously Flash has been a big part of that, very high-speed networks within data centers, virtualization certainly helps as well, now opens up the possibility for using these algorithms, some of which have been around for a long time, but have required very specialized bespoke configurations of hardware, to the enterprise. That still begs the question, there are some differences between high-performance computing workloads and AI workloads. Let's start with some of the, what are the similarities, and then let's explore some of the differences. >> So the biggest similarity, I think, is just it's an intractable, hard IO problem, right, at least from the storage perspective. It requires a lot of high throughput, depending on where those IO characteristics are from, it can be very small-file, high-op-intensive type workflows, but it needs the ability of the entire infrastructure to deliver all of that seamlessly from end to end. >> So really high-performance throughput so that you can get to the data you need and keep this computing element saturated. >> Keeping the GPU saturated is really the key, that's where the huge investment is. >> So how do AI and HPC workloads differ? >> So how they're fundamentally different is often AI workloads operate on a smaller scale in terms of the amount of capacity, at least today's AI workloads. As soon as a project encounters success, what our forecast is, is those things will take off and you'll want to apply those algorithms bigger and bigger data sets. But today, you know, we encounter things like 10-terabyte data sets, 50-terabyte data sets and a lot of customers are focused only on that. But what happens when you're successful, how do you scale your current infrastructure to petabytes and multi-petabytes when you'll need it in the future? >> So when I think of HPC, I think of often very, very big batch jobs, very, very large, complex data sets. When I think about AI, like image processing or voice processing, whatever else it might be, I think of a lot of small files, randomly accessed. >> Right. >> That require nonetheless some very complex processing, that you don't want to have to restart all the time. >> Right. >> And a degree of simplicity that's required to make sure that you have the people that can do it. Have I got that right? >> You've got it right. Now one, I think, misconception is, is on the HPC side, right, that whole random small file thing has come in in the last five, 10 years and it's something DDN's been working on quite a bit, right. Our legacy was in high-performance throughput workloads, but the workloads have evolved so much on the HPC side as well, and, as you posited at the beginning, so much of it has become AI and deep-learning research >> Right, so they look a lot more alike. >> They do look a lot more alike. >> So if we think about the revolving relationship now between some of these new data-first workloads, AI-oriented, change the way the business operates types of stuff, what do you anticipate is going to be the future of the relationship between AI and storage? >> Well, what we foresee really is that the explosion in AI needs and AI capabilities is going to mimic what we already see and really drive what we see on the storage side, right? We've been showing that graph for years and years and years of just everything going up and to the right, but as AI starts working on itself and improving itself, as the collection means keep getting better and more sophisticated and have increased resolutions, whether you're talking about cameras or in life sciences, acquisition capabilities just keep getting better and better and the resolutions get better and better, it's more and more data, right? And you want to be able to expose a wide variety of data to these algorithms; that's how they're going to learn faster. And so what we see is that the data-centric part of the infrastructure is going to need to scale, even if you're starting today with a smaller workload. >> Kurt Kuckein, DDN, thanks very much for being on The Cube. >> Thanks for having me. >> And once again, this is Peter Burris with another Cube Conversation, thank you very much for watching. Until next time. (electronic whooshing)
SUMMARY :
we're going to cut, come to you in a second. Hold on guys, sorry, I lied. Okay here we go, ready and, three, two. and especially some of the new infrastructure technologies and that's what we see a lot of similarities in the high-performance world, and if you look at how the supercomputing world has expanded So the general availability of AI technologies, but it needs the ability of the entire infrastructure so that you can get to the data you need Keeping the GPU saturated is really the key, in terms of the amount of capacity, So when I think of HPC, I think of that you don't want to have to restart all the time. to make sure that you have the people that can do it. is on the HPC side, right, and the resolutions get better and better, thank you very much for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Peter | PERSON | 0.99+ |
50-terabyte | QUANTITY | 0.99+ |
Peter Burris | PERSON | 0.99+ |
10-terabyte | QUANTITY | 0.99+ |
Kurt Kuckein | PERSON | 0.99+ |
KURT KUCKEIN | PERSON | 0.99+ |
DDN | ORGANIZATION | 0.99+ |
PETER | PERSON | 0.99+ |
30, 40 years | QUANTITY | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
75-percent | QUANTITY | 0.99+ |
Alec | PERSON | 0.99+ |
over 40 years | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
today | DATE | 0.98+ |
Nvidia | ORGANIZATION | 0.95+ |
three | QUANTITY | 0.93+ |
100 supercomputers | QUANTITY | 0.92+ |
10 years | QUANTITY | 0.91+ |
20 years | QUANTITY | 0.9+ |
years | QUANTITY | 0.89+ |
Cube | COMMERCIAL_ITEM | 0.87+ |
V-2 | OTHER | 0.86+ |
V-3 | OTHER | 0.85+ |
one | QUANTITY | 0.79+ |
five | QUANTITY | 0.73+ |
The Cube | ORGANIZATION | 0.72+ |
first | QUANTITY | 0.7+ |
last few years | DATE | 0.67+ |
second | QUANTITY | 0.63+ |
Cube | ORGANIZATION | 0.55+ |
DDN | PERSON | 0.54+ |
9_20_18 | DATE | 0.45+ |
last | QUANTITY | 0.39+ |
Jim McHugh, NVIDIA and Octavian Tanase, NetApp | Accelerate Your Journey to AI
>> From Sunnyvale, California, in the heart of Silicon Valley, it's theCUBE, covering Accelerate Your Journey to AI. Brought to you by NetApp. >> Hi, I'm Peter Burris, with theCUBE and Wikibon, and we're here at the NetApp Data Visionary Center today to talk about NetApp, NVIDIA, AI, and data. We're being joined by two great guests. Jim McHugh is the Vice President and General Manager of Deep Learning Systems at NVIDIA, and Octavian Tanase is the Senior Vice President of ONTAP at NetApp. Gentlemen, welcome to theCUBE. >> Thanks for having me. >> So Jim, I want to start with you. NVIDIA's been all over the place regarding AI right now. You've had a lot of conversations with customers. What is the state of those conversations today? >> Well, I mean, it really depends on the industry that the customer's in. So, AI at at its core, is really a horizontal technology, right? It's when when we engage with a customer and their data and their vertical domain knowledge that it becomes very specialized from there. So you're seeing a lot of acceleration where there's been a lot of data, right? So it's not any secret that you're seeing a lot around autonomous driving vehicles and the activity going there. Health care, right? Because when you can marry the technology of AI with the years, and years, and years of medical research that's going on out there, incredible things come out, right? We've seen some things around looking at cancer cells, we're looking at your retina being sort of the gateway to so many health indications. We can tell you whether you have everything from Dengue fever, to malaria, to whether you're susceptible to have hypertension. All of these kind of things that we're finding, that data is actually letting us to be superhuman in our knowledge about what we're trying to accomplish. Now the exciting thing is, if you grew up like we did, in the IT industry, is you're seeing it go into mainstream companies, so you're seeing it in financial services, where they for years were, quants were very specialized, and they were writing their own apps, and now they figured out, hey, look, I could broaden this out. You're seeing it in cybersecurity, right? For years, if you wanted to check malware, what did we do? We looked up the definition in a database and said, okay, yeah, that's malware, stop it, right? But now, they're learning the characteristics of malware. They're studying the patterns of it, and that's kind of what it is. Go industry by industry, and tell me if there's enough data to show a pattern, and AI will come in and change it. >> Enough data to show a pattern? Well, that kind of introduces NetApp to the equation. A company that's been, especially more recently, very focused on the relationship between data and business value. Octavian, what has NetApp seen from customers? >> Well, we know a little bit about data. We've been the stewards of that data in the enterprise for more than 25 years, and AI comes up in every single customer conversation. They're looking to leverage AI in their digital transformation, so we see this desire to extract more value out of the data, and make better decisions, faster decisions in every sector of the industry. So, it's ubiquitous, and we are uniquely positioned to enable customers to do their data management wherever data is being created. Whether the data is created at the edge, in the traditional data center, what we call the core, or in the cloud, we enable this seamless data management via the data fabric architecture and vision. >> So, data fabric, data management, the ability to extract that, turn it into patterns. Sounds like a good partnership, Jim? >> Yeah, no, we say, data's the new source code. Really, what AI is, we're changing the way software's written. Where, instead of having humans going in, do the feature engineering and feature sets that would be required, you're letting data dictate and guide you on what the features are going to be of software. >> So right now, we've got the GPU, Graphic Data Processing revolution, you guys driving that. We've got some real advances in how data fabric works. You have come together and created a partnership. Talk a little bit about that partnership. >> Well, when we started down this journey, and it began, really, in 2012 in AI, right? So when Alex Krizhevsky discovered how to create AlexNet, NVIDIA's been focused on how do we meet the needs of the data scientists every step of the way. So beginning started around making sure they had enough compute power to solve things that they couldn't solve before. Then we started focusing on what is the software that was required, right? So how do we get them the frameworks they need? How do we integrate that? How do we get more tuned, so they could get more and more performance? Our goal has always been, if we can make the data scientists more productive, we can actually help democratize AI. As it's starting to take hold, and get more deployments, obviously we need the data. We need it to help them with the data ingest, and then deployments are starting to scale out to the point where we need to make this easy, right? We need to take the headaches of trying to figure out what are all the configurations between our product lines, but also the networking product lines, as well. We have to bring that whole, holistic picture, and do it from there. So our goal, and what we're seeing, is not only we've made the data scientists more productive, but if we can help the guys that have to do the equipment for him more productive as well, the data scientists, she and he, can get back to doing what their real core work is. They can add value, and really change a lot of the things that are going on in our lives. >> So fast, flexibility, simpler to use. Does that, kind of, capture some of the, summarize some of the strategies that NetApp has for Artificial Intelligence workloads? >> Absolutely, I think simplicity, it's one of the key attributes, because the audience for some of the infrastructure that we're deploying together, it's a data scientist, and he wants to adopt that solution with confidence, and it has to be simple to deploy. He doesn't have to think about the infrastructure. It's also important to have an integrated approach, because, again, a lot of the data will be created in the future at the core, or at edge more than in the core, and more in the cloud than in traditional data center. So that seamless data management across the edge, to the core, to the cloud, it's also important. And scalability, it's also important, because customers who look to start, perhaps, simple, with a small deployment, and have that ability to seamlessly scale. Currently, the performance of the solution that we just announced, basically beats the competition by a 4x, in terms of the performance and capability. >> So as we think about where we're going, this is a crucial partnership for both companies, and it's part of a broader ecosystem that NVIDIA's building out. How does the NetApp partnership fit into that broader ecosystem? >> Well, starting with our relationship, when the announcement we made, it should be no secret that we engaged our channel partners, right? 'Cause they are that last mile. They are those trusted advisors, a lot of times, of our customers, and going in, and we want them to add this to their portfolio, take it out to 'em, and I think we've had resounding feedback, so far, that this is something that they can definitely take, and drive out. On top of that, NVIDIA is focused on, again, this new way of writing software, right? The software that leverages the data to do the things, and so we have an ecosystem that's built around our inception program, which are thousan%ds of startups. If you add to that the thousands of startups that are coming through Sand Hill, and the investment community, that are based around NVIDIA compute, as well, all of these guys are standardizing saying, hey we need to leverage this new model. We need to go as quickly as possible, and what we've pulled together, together, is the ability for them to do that. So whether they want to do the data center, or whether they want to go with one of our joint cloud providers and do it through their service, as well. >> So a great partnership that's capable of creating a great horizontal platform. It's that last mile that does the specialization. Have I got that right? >> You had the last mile helping reach the customers who are the specialization. The customers, and their data, and their vertical domain expertise, and what the data scientists that they have bring to it. Look, they're creating the magic. We're giving them the tools to make sure they can create that magic as easy as possible. >> That's great, so one of the things, Octavian, that Jim mentioned, was industries that are able to generate significant value out of data are moving first. One of the more important industries is IT Operations, because we have a lot of devices, we're generating a lot of data. How is NetApp going to use AI in your product set to drive further levels of productivity, from a simplicity standpoint, so customers can, in fact, spend more time on creating value? >> So interestingly enough, we've been users, or practitioners, of AI for quite a while. I don't know if a lot of people in the audience know, we have a predictive analytics system called Active IQ, which is an implementation of AI in the enterprise. We take data from more than 300 thousand assets that we have deployed in the field, more than 70 billion data points every day, and we correlate that together. We put them in a data lake. We train a cluster, and we enable our customers to drive value in best practices from the data that we collect from the broader set of deployments that we have in the field, so this is something that we are sharing with our customers, in terms of blueprint, and we're looking to drive the ubiquity in the type of solutions that we enable customers to build on top of our joint infrastructure. >> Excellent, Jim McHugh, NVIDIA, Octavian Tanase, NetApp. Great partnership represented right here on theCUBE. Thanks very much for being on theCUBE tonight. >> All right. >> Thank you. >> Thank you for having us. (electronic music)
SUMMARY :
in the heart of Silicon Valley, it's theCUBE, and Octavian Tanase is the Senior What is the state of those conversations today? the gateway to so many health indications. Well, that kind of introduces NetApp to the equation. or in the cloud, we enable this seamless data management So, data fabric, data management, the ability Where, instead of having humans going in, do the feature Talk a little bit about that partnership. the data scientists, she and he, can get back to summarize some of the strategies that NetApp has So that seamless data management across the edge, How does the NetApp partnership fit The software that leverages the data to do the things, It's that last mile that does the specialization. You had the last mile helping reach One of the more important industries is IT Operations, in the type of solutions that we enable customers Thanks very much for being on theCUBE tonight. Thank you for having us.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Peter Burris | PERSON | 0.99+ |
Jim McHugh | PERSON | 0.99+ |
Jim | PERSON | 0.99+ |
2012 | DATE | 0.99+ |
Alex Krizhevsky | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Octavian | PERSON | 0.99+ |
Octavian Tanase | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
ONTAP | ORGANIZATION | 0.99+ |
more than 300 thousand assets | QUANTITY | 0.99+ |
Sunnyvale, California | LOCATION | 0.99+ |
more than 25 years | QUANTITY | 0.99+ |
Deep Learning Systems | ORGANIZATION | 0.99+ |
NetApp | ORGANIZATION | 0.99+ |
both companies | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
NetApp | TITLE | 0.99+ |
today | DATE | 0.98+ |
more than 70 billion data points | QUANTITY | 0.97+ |
Wikibon | ORGANIZATION | 0.97+ |
4x | QUANTITY | 0.97+ |
malaria | OTHER | 0.97+ |
one | QUANTITY | 0.96+ |
NetApp Data Visionary Center | ORGANIZATION | 0.96+ |
theCUBE | ORGANIZATION | 0.96+ |
tonight | DATE | 0.96+ |
two great guests | QUANTITY | 0.95+ |
Dengue fever | OTHER | 0.89+ |
Sand Hill | ORGANIZATION | 0.81+ |
first | QUANTITY | 0.8+ |
hypertension | OTHER | 0.77+ |
thousan%ds of startups | QUANTITY | 0.68+ |
IQ | OTHER | 0.68+ |
single customer | QUANTITY | 0.66+ |
AlexNet | ORGANIZATION | 0.61+ |
Vice | PERSON | 0.55+ |
years | QUANTITY | 0.54+ |
theCUBE | TITLE | 0.51+ |
Active | TITLE | 0.32+ |
Jim McHugh, NVIDIA | SAP SAPPHIRE NOW 2018
>> From Orlando, Florida it's theCUBE! Covering SAP SAPPHIRE NOW 2018, brought to you by NetApp. >> Welcome to theCUBE I'm Lisa Martin with Keith Townsend and we are in Orlando at SAP SAPPHIRE NOW 2018, where we're in the NetApp booth and talking with lots of partners and we're excited to welcome back to theCUBE, distinguished alumni Jim McHugh from NVIDIA, you are the VP and GM of Deep Learnings and "other stuff" as you said in the keynote. (all laugh) >> Yeah, and other stuff. That's a lot of responsibility! That other stuff, that, you know, that can really pile up! >> That can kill ya. Yeah, exactly. >> So here we are at SAPPHIRE you've been working with SAP in various forms for a long time, this event is enormous, lots of momentum at NVIDIA, what is NVIDIA doing with SAP? >> We're really helping SAP figure out and drive the development of their SAP Leonardo machine learning services so, machine learning, as we saw in the keynote today, with Haaso as a key component of it, and really what it's doing is it's automating a lot of the standard processes that people did, in the interactions, so whether it's closing your invoices at the end of the quarter, and that can take weeks to go through it manually, you can actually do machine learning and deep learning and do that instantaneously, so you can get a continuous close. Things like service ticketing, so when a service ticket comes in, you know, we all know, you pick up the phone, you call 'em and they collect your information, and then they pass you on to someone else that wants to confirm the information, all that can be handled just in a email, because now I know a lot about you when you send me an email I know who you are, know what company you're with, I know your problem 'cause you stated it, and I can route it, using machine learning, to the appropriate person. I can not only route it to the appropriate person I can look up in a knowledge database and say hey, have we seen this answer a question before feed that to the customer service representative, and when they start interacting with the customer they already have a lot of information about them and it's already well underway. >> So from a practical technology perspective we hear a lot about AI, machine learning, NVIDIA obviously leading the way with GPUs and enabling development frameworks to take advantage of machine learning and that compute power. But the enterprise, we'll at that and we're like you know that, we see obvious value, but I need a data scientist, I need a programmer, I need all this capability, from a technical staff perspective, to take advantage of it. How is NVIDIA, SAP, making that easier to consume? >> So most enterprises, if you're just jumpin' in and tryin' to figure it out, you would need all these people, you'd need a data scientist and someone to go through the process. 'Cause AIs, it's a new way of writing software, and you're using data to train the software, so we don't have, we don't put programmers in a room anymore and let 'em code for nine months and out pops software, you know, eventually. We give 'em more and more data, and the data scientist is training it. Well the good news is we're working with SAP and they have the data scientists, they know how SAP apps work, they know how the integration works, they know the workflows of their customers, so they're building the models and then making it available as a service, right? So when you go to the SAP cloud, you're saying I wanna actually take advantage of the SAP service for service ticketing or, you know, I wanna figure out how I can do my invoice processing better, or I'm an HR representative, and I don't wanna spend 60% of my time reading resumes, I wanna actually have an AI do it for me, and then it's a service that you can consume. There, that we do make it possible, like if you have a developer in your enterprise and you say you know what, I'm a big SAP user but I actually wanna develop a custom app or other some things I might do, then SAP makes available the Leonardo machine learning foundation and you can take advantage of that and develop a custom app. And if you have a really big problem and you wanna take it off, NVIDIA's happy to work with you directly and figure out how to solve different problems. And most of our customers are in all three of those, Right? They're consuming the services 'cause they automate things today, they're figuring out, what are the custom apps they need to build around SAP and then they're, you know, they're figuring out some of the product building products or something else that's a much bigger machine learning, deep learning problem. >> So yesterday during Bill McDermott's keynote he talked about tech for good, now there's been a lot of news recently of tech for not-so-good and data privacy, GDPR, you know, compliance going into affect last week, NVIDIA really has been an integral part of this AI renaissance, you talked about, you know, you can help loads of different customers there's so much potential with AI, as Bill McDermott said yesterday, AI to augment humanity. I can imagine, you know, life and death situations like in healthcare, can you give us an example of what you guys are doing with SAP that, you know, maybe is transforming healthcare at a particular hospital? >> Yeah, so one of the great examples I was just talking about is, what Massachusetts General is doing. Massachusetts General is one of the largest research hospitals in the United States, and they're doing a lot of work in AI, to really automate processes that, you know, when you would take your child in to figure out the bone density scan, which basically tells you the bone age of your child, and they compare it to your biological age, and that can tell you a lot of things, is it just a, you know, a growth problem, or is there something more serious to be concerned about. Well, they would do these MRIs, and then you would have to wait for days while the, the technician and the doctor would flip through a textbook from the 1950's, to determine it. Well Massachusetts General automated all that where they actually trained a neural network on all these different scans and all these different components and now you find out in minutes. So it greatly reduces the stress, right? And there's plenty of other project going on and you can see it in determination if that's a cancer cell, or, you know, so many different aspects of it, your retina happens to be an incredible venue into whether you have hypertension, whether you have Malaria, Dengue fever, so things like, you know what, maybe you shouldn't be around anywhere where you're gonna get bit by a mosquito and it's gonna pass it to your family, all that can now be handled, and you don't need expensive healthcare, you can actually take it to a clinician out in the field. So, we love all that. But if you think about the world of SAP which is the, you know, controls the data records of most companies, right? Their supply chain information, their resource information about, you know, what they have available, all that's being automated. So if we think from the production of food where we're having tractors now that they have the ability to go over a plant and say you know what, that needs insecticide or that needs weeds to be removed 'cause it's just bad for the whole component, or that's a diseased plant and I'm gonna remove it, or it just needs water so it can grow, right? That is increasing the production of food in an organic way, then we improve the distribution centers so it doesn't sit as long, right, so that we can actually have drones flying through the warehouses and knowing what needs to be moved first, go from there, we're moving to autonomous driving vehicles and, where deliveries can happen at night when there's not so much traffic, and then we can get the food as fresh as possible and deliver it. So if you think that whole distribution center and just being in the pipeline as being automated, it's doing an incredible amount of good. And then, jumping into the world of autonomous driving vehicles, it's a 10 trillion dollar business that's being changed, radically. >> So as we think about these super complex systems that we're trying to improve, we start to break them down into small components, smaller components, you end up with these scenarios, these edge scenarios, use cases where, you know, whether it's data frequency, data value, or data latency, we have to push to compute out to the edge. Can you talk about use cases where NVIDIA has pushed the technology far out to the edge to take in massive amounts of data, that effectively can't be sent back to the core or to the data center for processing, what are some of these use cases solutions? >> So it's, the world of IOT is changing as well, right, the compute power has to be where it's needed, right, and in any form, so whether that's cloud based, data center based, or at the edge and we have a great customer that is actually doing inspection, oil refineries, bridges, you know, where they spot a crack or some sort of mark where they have to go look at it, well traditionally what you do is you send out a whole team and they build up scaffolding, or they have people repel down to try to inspect it. Well now what we're doing is flying drones and sending wall crawlers up. So they find something, they get data, and then, instead of actually, like you said, putting it, you know, on a truck and taking it back to your data center or trying to figure out how to have enough bandwidth to get there, they're taking one of our products, which is a DGX station, it's basically the equivalent of a half a row of servers, but it's in a single box, water cooled, and they're putting it in vans sitting out in remote areas of Alaska, and retraining the model there on site. So, they get the latest model, they get more intelligence and they just collect it, and they can resend the drones up and then discover more about it. So it really, really is saving, and that saves a lot of money, so you have a group of really smart you know, technicians and people who understand it and a guy who can do the neural network capability instead of a whole team coming up and setting up scaffolding that would cost millions of dollars. >> That reminds me of that commercial that they showed yesterday during general session SAP commercial with Clive Owen the actor, talking about, you mentioned, you know, cracks in oil wells and things like that it just reminded me of that, and what they talked about in that video was really how invisible software, like SAP, is transforming industries, saving lives, I think I saw on their website an example of how they're leveraging AI and technology to reduce water scarcity in India or save the rhino conservation and what you just described with NVIDIA seems to be quite in alignment with the direction that SAP is going. >> Oh absolutely, yeah, I mean we believe in SAP's view of the intelligent enterprise and people gotta remember, enterprise isn't just like the corporate office whatever, enterprises are many different things, alright. Public safety, if you can think about that, that's a big thing we focus on. A really amazing thing that's going on, thinking about using drones for first responders they actually can know what's going on at the scene and when the other people are showing up they know what kind of area they're going into. Or for search and rescue, drones can cover a lot of territory and detect a human faster than a human can, right? And if you can actually find someone within the first 24 hours, chance of survival is so much higher. All of that is, you know, leveraging the exact same technology that we do for looking at our business processes, right, and it's not as, you know, dramatic, it's not gonna show up on the evening news, but honestly, streamlining our business processes, making it happen so much faster and more efficient makes businesses more efficient, you know, it's better for the company, it's better for the employees as well. >> So let's talk about, something that's, that's taboo, financial services, making money with data, or with analytics or machine learning from data, again we have to, John Furrier is here, and we have someone from NVIDIA here, and if we don't bring up blockchain in some type of way he's gonna throw something at his team, so, >> Let's give a shout out to John Furrier. (laughing) >> Give a shout out to John. But from a practical sense, let's subtract the digital currency part of machine, of blockchain, do you see applications for blockchain from a machine learning perspective? >> Yeah, I mean well, if you just boil blockchain down or for trusted networks, right? And you know you heard Bill McDermott say that on stage he called his marketplaces, or areas that he could do for an exchange, it makes total sense. If I can have a trusted way of doing things where I have a common ledger between companies and we know that it's valid, that we can each interchange with, yeah it makes complete sense, right, now we gotta get to the practical imitation of that and we have to build the trust of the companies to understand, okay this technology can take you there, and that's where I think, you know, where we come in with our technology capabilities, ensuring to people that it's reliable and work, SAP comes in with the customer relationships and trusted in what they've been doing in helping people run their business for years, and then it becomes cultural. Like all things, we can kid ourselves in technology that we'll just solve everything, it's a cultural change. I'm gonna share that common ledger, I'm gonna share that common network and feel confident in it, it's something that people have to do and, you know, my take on that always is when the accuracy is so much better, when the efficiency is so much better, when the return is so much better, we get a lot more comfortable. People used to be nervous about giving the grocery store their phone number, right, 'cause they would track their food, right? And today we're just like okay yeah here's my phone number. (Keith laughing) >> So. (laughs) >> Give you a 30 cent discount, here's my number. >> Exactly. We're so cheap. (laughing) >> So we're in the NetApp booth and you guys recently announced a reference, combined reference, AI reference architecture with NetApp, tell us a little bit more about that. >> Yeah, well the little secret behind all the things we just talked about, there's an incredible amount of data, right, and as you collect this data it's really important to store it in a way that it's accessible when you need it. And when you're doing trainings, I have a product that's called DGX-1, DGX-1 takes an incredible amount of data that helps us train these neural networks, and it's fast, and it has an insatiable desire for data. So what we've worked with NetApp is actually pool together reference architecture so that when a data scientist, who is a very valuable resource, is working on this, he's ensured that the infrastructures are gonna work together seamlessly and deliver that data to the training process. And then when you create that model, we use something that's called inference, you put it in production, and again same time, when you're having that inference running you wanna make sure that data can get to it and can interact with the data seamlessly and the reference architectures play out there as well. So our goal is, start knocking off one by one, what do the customers need to be successful? And we put a lot of effort into the GPUs, we put a lot of effort into the deep learning software that runs on top of that, we put a lot of effort into, you know, what's the models they need to use, etc. And now we have to spend a lot more time of what's their infrastructure? And make sure that's reliable because, you would hate to do all that work only to find that your infrastructure had a hiccup, and took your job down. So we're working really hard to make sure that never happens >> So I have this theory that, well I don't have the theory, David Curry came out with this theory of data has gravity, but I've come up with this additional theory, now that we look at AI, and the capability of AI and what people are and what the hyper scalers are doing in their data center is that individual companies think, have a challenge replicating in their own data center, this AI and compute now has gravity. You know, I can't well, at least before today I didn't think well I can take my data center, put it on the road, and do these massive pieces of injection on the edge, sounds like we're pushin' back on that a little bit and saying that you know what sure if it's, I don't know what the limits are, and I guess that's the question. What are the limits of what we can do on the edge when it comes to the amount of data, and portable AI to that edge? >> Well so, there's again the two aspects of it, the training takes an incredible amount of data that's why they would have to take a super computer and put it there so they could do the retraining, but, when you think about when you can have the pro-- something the size of a credit card, which is our Jetson solution, and you can install it in a drone or you can put in cameras for public safety, etc. Which is, has incredible, think about looking for a lost child or parents with Alzheimer's, you can scan through video real quick and find them, right? All because of a credit card sized processor, that's pretty impressive. But that's what's happening at the edge, we're now writing applications that are much more intelligent using AI, there are AI applications sitting at the edge that, instead of just processing the data in a way where I'm getting a average, average number of people who walked into my store, right, that's what we used to do five years ago, now we're actually using intelligent applications that are making calculated decisions, it's understanding who's coming in a store, understanding their buying/purchasing power, etc. That's extremely important in retail, because, if you wanna interact with someone and give them that, you know when they're doing self checkout, try to sell 'em one more thing, you know, did you forget the batteries that go with that, or whatever you want it to be, you only have a few seconds, right? And so you must be able to process that and have something really intelligent doing that instead of just trying to do the law of average and get a directionally correct-- and we've known this, anytime you've been on your webpage or whatever and someone recommends something you're like that doesn't have anything to do with me and then all of a sudden it started getting really good that's where they're getting more intelligent. >> When I walk into the store with my White Sox hat and then they recommend the matching jersey. I'm gonna look, gonna come lookin' for you guys at NVIDIA like wa-hey! I don't have money for a jersey, but things like that, yeah. >> We're just behind the scenes somewhere. >> Well, you title VP and GM of Deep Learning and stuff, there's a lot of stuff. (all laugh) Jim thanks so much for coming back on theCUBE sharing with us what's new at NVIDIA it sounds like the world of possibilities is endless, so exciting! >> Yeah, it is an exciting time, thank you. >> Thanks for your time, we wanna thank you for watching theCUBE, Lisa Martin with Keith Townsend from SAP SAPPHIRE 2018, thanks for watching. (bubbly music)
SUMMARY :
brought to you by NetApp. and "other stuff" as you said in the keynote. That other stuff, that, you know, That can kill ya. and then they pass you on to someone else and enabling development frameworks to take advantage of and then they're, you know, I can imagine, you know, and that can tell you a lot of things, these edge scenarios, use cases where, you know, and then, instead of actually, like you said, what you just described with NVIDIA and it's not as, you know, dramatic, Let's give a shout out to John Furrier. do you see applications for blockchain and that's where I think, you know, Give you a 30 cent discount, We're so cheap. you guys recently announced a reference, and deliver that data to the training process. and saying that you know what and you can install it in a drone and then they recommend the matching jersey. behind the scenes somewhere. Well, you title VP and GM of Deep Learning and stuff, we wanna thank you for watching theCUBE,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jim McHugh | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Massachusetts General | ORGANIZATION | 0.99+ |
Keith Townsend | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Alaska | LOCATION | 0.99+ |
David Curry | PERSON | 0.99+ |
60% | QUANTITY | 0.99+ |
Bill McDermott | PERSON | 0.99+ |
nine months | QUANTITY | 0.99+ |
Orlando | LOCATION | 0.99+ |
Clive Owen | PERSON | 0.99+ |
30 cent | QUANTITY | 0.99+ |
Jim | PERSON | 0.99+ |
United States | LOCATION | 0.99+ |
Orlando, Florida | LOCATION | 0.99+ |
yesterday | DATE | 0.99+ |
10 trillion dollar | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
White Sox | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Leonardo | ORGANIZATION | 0.99+ |
SAP | ORGANIZATION | 0.99+ |
India | LOCATION | 0.98+ |
Jetson | ORGANIZATION | 0.98+ |
SAPPHIRE | ORGANIZATION | 0.98+ |
two aspects | QUANTITY | 0.98+ |
GDPR | TITLE | 0.98+ |
millions of dollars | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
three | QUANTITY | 0.97+ |
five years ago | DATE | 0.97+ |
first responders | QUANTITY | 0.96+ |
Malaria | OTHER | 0.96+ |
single box | QUANTITY | 0.95+ |
half a row | QUANTITY | 0.95+ |
VP | PERSON | 0.95+ |
Keith | PERSON | 0.94+ |
Deep Learnings | ORGANIZATION | 0.94+ |
1950's | DATE | 0.93+ |
first 24 hours | QUANTITY | 0.93+ |
NetApp | TITLE | 0.92+ |
one more thing | QUANTITY | 0.91+ |
NOW | DATE | 0.91+ |
Dengue fever | OTHER | 0.91+ |
each | QUANTITY | 0.88+ |
SAP SAPPHIRE | TITLE | 0.88+ |
SAP | TITLE | 0.88+ |
a few seconds | QUANTITY | 0.84+ |
NetApp | ORGANIZATION | 0.81+ |
2018 | DATE | 0.81+ |
theCUBE | TITLE | 0.77+ |
hypertension | OTHER | 0.74+ |
first | QUANTITY | 0.72+ |
Matt Burr, Pure Storage & Rob Ober, NVIDIA | Pure Storage Accelerate 2018
>> Announcer: Live from the Bill Graham Auditorium in San Francisco, it's theCUBE! Covering Pure Storage Accelerate 2018 brought to you by Pure Storage. >> Welcome back to theCUBE's continuing coverage of Pure Storage Accelerate 2018, I'm Lisa Martin, sporting the clong and apparently this symbol actually has a name, the clong, I learned that in the last half an hour. I know, who knew? >> Really? >> Yes! Is that a C or a K? >> Is that a Prince orientation or, what is that? >> Yes, I'm formerly known as. >> Nice. >> Who of course played at this venue, as did Roger Daltry, and The Who. >> And I might have been staff for one of those shows. >> You could have been, yeah, could I show you to your seat? >> Maybe you're performing later. You might not even know this. We have a couple of guests joining us. We've got Matt Burr, the GM of FlashBlade, and Rob Ober, the Chief Platform Architect at NVIDIA. Guys, welcome to theCUBE. >> Hi. >> Thank you. >> Dave: Thanks for coming on. >> So, lots of excitement going on this morning. You guys announced Pure and NVIDIA just a couple of months ago, a partnership with AIRI. Talk to us about AIRI, what is it? How is it going to help organizations in any industry really democratize AI? >> Well, AIRI, so AIRI is something that we announced, the AIRI Mini today here at Accelerate 2018. AIRI was originally announced at the GTC, Global Technology Conference, for NVIDIA back in March, and what it is is, it essentially brings NVIDIA's DGX servers, connected with either Arista or Cisco switches down to the Pure Storage FlashBlade, so this is something that sits in less than half a rack in the data center, that replaces something that was probably 25 or 50 racks of compute and store, so, I think Rob and I like to talk about it as kind of a great leap forward in terms of compute potential. >> Absolutely, yeah. It's an AI supercomputer in a half rack. >> So one of the things that this morning, that we saw during the general session that Charlie talked about, and I think Matt (mumbles) kind of a really brief history of the last 10 to 20 years in storage, why is modern external storage essential for AI? >> Well, Rob, you want that one, or you want me to take it? Coming from the non storage guy, maybe? (both laugh) >> Go ahead. >> So, when you look at the structure of GPUs, and servers in general, we're talking about massively parallel compute, right? These are, we're now taking not just tens of thousands of cores but even more cores, and we're actually finding a path for them to communicate with storage that is also massively parallel. Storage has traditionally been something that's been kind of serial in nature. Legacy storage has always waited for the next operation to happen. You actually want to get things that are parallel so that you can have parallel processing, both at the compute tier, and parallel processing at the storage tier. But you need to have big network bandwidth, which was what Charlie was eluding to, when Charlie said-- >> Lisa: You like his stool? >> When Charlie was, one of his stools, or one of the legs of his stool, was talking about, 20 years ago we were still, or 10 years ago, we were at 10 gig networks, in merges of 100 gig networks has really made the data flow possible. >> So I wonder if we can unpack that. We talked a little bit to Rob Lee about this, the infrastructure for AI, and wonder if we can go deeper. So take the three legs of the stool, and you can imagine this massively parallel compute-storage-networking grid, if you will, one of our guys calls it uni-grid, not crazy about the name, but this idea of alternative processing, which is your business, really spanning this scaled out architecture, not trying to stuff as much function on a die as possible, really is taking hold, but what is the, how does that infrastructure for AI evolve from an architect's perspective? >> The overall infrastructure? I mean, it is incredibly data intensive. I mean a typical training set is terabytes, in the extreme it's petabytes, for a single run, and you will typically go through that data set again and again and again, in a training run, (mumbles) and so you have one massive set that needs to go to multiple compute engines, and the reason it's multiple compute engines is people are discovering that as they scale up the infrastructure, you actually, you get pretty much linear improvements, and you get a time to solution benefit. Some of the large data centers will run a training run for literally a month and if you start scaling it out, even in these incredibly powerful things, you can bring time to solution down, you can have meaningful results much more quickly. >> And you be a sensitive, sort of a practical application of that. Yeah there's a large hedge fund based in the U.K. called Man AHL. They're a system-based quantitative training firm, and what that means is, humans really aren't doing a lot of the training, machines are doing the vast majority if not all of the training. What the humans are doing is they're essentially quantitative analysts. The number of simulations that they can run is directly correlative to the number of trades that their machines can make. And so the more simulations you can make, the more trades you can make. The shorter your simulation time is, the more simulations that you can run. So we're talking about in a sort of a meta context, that concept applies to everything from retail and understanding, if you're a grocery store, what products are not on my shelves at a given time. In healthcare, discovering new forms of pathologies for cancer treatments. Financial services we touched on, but even broader, right down into manufacturing, right? Looking at, what are my defect rates on my lines, and if it used to take me a week to understand the efficiency of my assembly line, if I can get that down to four hours, and make adjustments in real time, that's more than just productivity, it's progress. >> Okay so, I wonder if we can talk about how you guys see AI emerging in the marketplace. You just gave an example. We were talking earlier again to Rob Lee about, it seems today to be applied and, in narrow use cases, and maybe that's going to be the norm, whether it's autonomous vehicles or facial recognition, natural language processing, how do you guys see that playing out? Whatever be, this kind of ubiquitous horizontal layer or do you think the adoption is going to remain along those sort of individual lines, if you will. >> At the extreme, like when you really look out at the future, let me start by saying that my background is processor architecture. I've worked in computer science, the whole thing is to understand problems, and create the platforms for those things. What really excited me and motivated me about AI deep learning is that it is changing computer science. It's just turning it on its head. And instead of explicitly programming, it's now implicitly programming, based on the data you feed it. And this changes everything and it can be applied to almost any use case. So I think that eventually it's going to be applied in almost any area that we use computing today. >> Dave: So another way of asking that question is how far can we take machine intelligence and your answer is pretty far, pretty far. So as processor architect, obviously this is very memory intensive, you're seeing, I was at the Micron financial analyst meeting earlier this week and listening to what they were saying about these emerging, you got T-RAM, and obviously you have Flash, people are excited about 3D cross-point, I heard it, somebody mentioned 3D cross-point on the stage today, what do you see there in terms of memory architectures and how they're evolving and what do you need as a systems architect? >> I need it all. (all talking at once) No, if I could build a GPU with more than a terabyte per second of bandwidth and more than a terabyte of capacity I could use it today. I can't build that, I can't build that yet. But I need, it's a different stool, I need teraflops, I need memory bandwidth, and I need memory capacity. And really we just push to the limit. Different types of neural nets, different types of problems, will stress different things. They'll stress the capacity, the bandwidth, or the actual compute. >> This makes the data warehousing problem seem trivial, but do you see, you know what I mean? Data warehousing, it was like always a chase, chasing the chips and snake swallowing a basketball I called it, but do you see a day that these problems are going to be solved, architecturally, it talks about, More's laws, moderating, or is this going to be this perpetual race that we're never going to get to the end of? >> So let me put things in perspective first. It's easy to forget that the big bang moment for AI and deep learning was the summer of 2012, so slightly less than six years ago. That's when Alex Ned get the seed and people went wow, this is a whole new approach, this is amazing. So a little less than six years in. I mean it is a very young, it's a young area, it is in incredible growth, the change in state of art is literally month by month right now. So it's going to continue on for a while, and we're just going to keep growing and evolving. Maybe five years, maybe 10 years, things will stabilize, but it's an exciting time right now. >> Very hard to predict, isn't it? >> It is. >> I mean who would've thought that Alexa would be such a dominant factor in voice recognition, or that a bunch of cats on the internet would lead to facial recognition. I wonder if you guys can comment, right? I mean. >> Strange beginnings. (all laughing) >> But very and, I wonder if I can ask you guys ask about the black box challenge. I've heard some companies talk about how we're going to white box everything, make it open and, but the black box problem meaning if I have to describe, and we may have talked about this, how I know that it's a dog. I struggle to do that, but a machine can do that. I don't know how it does it, probably can't tell me how it does it, but it knows, with a high degree of accuracy. Is that black box phenomenon a problem, or do we just have to get over it? >> Up to you. >> I think it's certain, I don't think it's a problem. I know that mathematicians, who are friends, it drives them crazy, because they can't tell you why it's working. So it's a intellectual problem that people just need to get over. But it's the way our brains work, right? And our brains work pretty well. There are certain areas I think where for a while there will be certain laws in place where you can't prove the exact algorithm, you can't use it, but by and large, I think the industry's going to get over it pretty fast. >> I would totally agree, yeah. >> You guys are optimists about the future. I mean you're not up there talking about how jobs are going to go away and, that's not something that you guys are worried about, and generally, we're not either. However, machine intelligence, AI, whatever you want to call it, it is very disruptive. There's no question about it. So I got to ask you guys a few fun questions. Do you think large retail stores are going to, I mean nothing's in the extreme, but do you think they'll generally go away? >> Do I think large retail stores will generally go away? When I think about retail, I think about grocery stores, and the things that are going to go away, I'd like to see standing in line go away. I would like my customer experience to get better. I don't believe that 10 years from now we're all going to live inside our houses and communicate over the internet and text and half of that be with chat mods, I just don't believe that's going to happen. I think the Amazon effect has a long way to go. I just ordered a pool thermometer from Amazon the other day, right? I'm getting old, I ordered readers from Amazon the other day, right? So I kind of think it's that spur of the moment item that you're going to buy. Because even in my own personal habits like I'm not buying shoes and returning them, and waiting five to ten times, cycle, to get there. You still want that experience of going to the store. Where I think retail will improve is understanding that I'm on my way to their store, and improving the experience once I get there. So, I think you'll see, they need to see the Amazon effect that's going to happen, but what you'll see is technology being employed to reach a place where my end user experience improves such that I want to continue to go there. >> Do you think owning your own vehicle, and driving your own vehicle, will be the exception, rather than the norm? >> It pains me to say this, 'cause I love driving, but I think you're right. I think it's a long, I mean it's going to take a while, it's going to take a long time, but I think inevitably it's just too convenient, things are too congested, by freeing up autonomous cars, things that'll go park themselves, whatever, I think it's inevitable. >> Will machines make better diagnoses than doctors? >> Matt: Oh I mean, that's not even a question. Absolutely. >> They already do. >> Do you think banks, traditional banks, will control of the payment systems? >> That's a good one, I haven't thought about-- >> Yeah, I'm not sure that's an AI related thing, maybe more of a block chain thing, but, it's possible. >> Block chain and AI, kind of cousins. >> Yeah, they are, they are actually. >> I fear a world though where we actually end up like WALLE in the movie and everybody's on these like floating chez lounges. >> Yeah lets not go there. >> Eating and drinking. No but I'm just wondering, you talked about, Matt, in terms of the number of, the different types of industries that really can verge in here. Do you see maybe the consumer world with our expectation that we can order anything on Amazon from a thermometer to a pair of glasses to shoes, as driving other industries to kind of follow what we as consumers have come to expect? >> Absolutely no question. I mean that is, consumer drives everything, right? All flash arrays were driven by you have your phone there, right? The consumerization of that device was what drove Toshiba and all the other fad manufacturers to build more NAM flash, which is what commoditized NAM flash, which what brought us faster systems, these things all build on each other, and from a consumer perspective, there are so many things that are inefficient in our world today, right? Like lets just think about your last call center experience. If you're the normal human being-- >> I prefer not to, but okay. >> Yeah you said it, you prefer not to, right? My next comment was going to be, most people's call center experiences aren't that good. But what if the call center technology had the ability to analyze your voice and understand your intonation, and your inflection, and that call center employee was being given information to react to what you were saying on the call, such that they either immediately escalated that call without you asking, or they were sent down a decision path, which brought you to a resolution that said that we know that 62% of the time if we offer this person a free month of this, that person is going to view, is going to go away a happy customer, and rate this call 10 out of 10. That is the type of things that's going to improve with voice recognition, and all of the voice analysis, and all this. >> And that really get into how far we can take machine intelligence, the things that machines, or the humans can do, that machines can't, and that list changes every year. The gap gets narrower and narrower, and that's a great example. >> And I think one of the things, going back to your, whether stores'll continue being there or not but, one of the biggest benefits of AI is recommendation, right? So you can consider it userous maybe, or on the other hand it's great service, where a lot of, something like an Amazon is able to say, I've learned about you, I've learned about what people are looking for, and you're asking for this, but I would suggest something else, and you look at that and you go, "Yeah, that's exactly what I'm looking for". I think that's really where, in the sales cycle, that's really where it gets up there. >> Can machines stop fake news? That's what I want to know. >> Probably. >> Lisa: To be continued. >> People are working on that. >> They are. There's a lot, I mean-- >> That's a big use case. >> It is not a solved problem, but there's a lot of energy going into that. >> I'd take that before I take the floating WALLE chez lounges, right? Deal. >> What if it was just for you? What if it was just a floating chez lounge, it wasn't everybody, then it would be alright, right? >> Not for me. (both laughing) >> Matt and Rob, thanks so much for stopping by and sharing some of your insights and we should have a great rest of the day at the conference. >> Great, thank you very much. Thanks for having us. >> For Dave Vellante, I'm Lisa Martin, we're live at Pure Storage Accelerate 2018 at the Bill Graham Civic Auditorium. Stick around, we'll be right back after a break with our next guest. (electronic music)
SUMMARY :
brought to you by Pure Storage. I learned that in the last half an hour. Who of course played at this venue, and Rob Ober, the Chief Platform Architect at NVIDIA. Talk to us about AIRI, what is it? I think Rob and I like to talk about it as kind of It's an AI supercomputer in a half rack. for the next operation to happen. has really made the data flow possible. and you can imagine this massively parallel and if you start scaling it out, And so the more simulations you can make, AI emerging in the marketplace. based on the data you feed it. and what do you need as a systems architect? the bandwidth, or the actual compute. in incredible growth, the change I wonder if you guys can comment, right? (all laughing) I struggle to do that, but a machine can do that. that people just need to get over. So I got to ask you guys a few fun questions. and the things that are going to go away, I think it's a long, I mean it's going to take a while, Matt: Oh I mean, that's not even a question. maybe more of a block chain thing, but, it's possible. and everybody's on these like floating to kind of follow what we as consumers I mean that is, consumer drives everything, right? information to react to what you were saying on the call, the things that machines, or the humans can do, and you look at that and you go, That's what I want to know. There's a lot, I mean-- It is not a solved problem, I'd take that before I take the Not for me. and sharing some of your insights and Great, thank you very much. at the Bill Graham Civic Auditorium.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Matt Burr | PERSON | 0.99+ |
Matt | PERSON | 0.99+ |
Charlie | PERSON | 0.99+ |
10 gig | QUANTITY | 0.99+ |
25 | QUANTITY | 0.99+ |
Rob Lee | PERSON | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Rob | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
Lisa | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
100 gig | QUANTITY | 0.99+ |
Toshiba | ORGANIZATION | 0.99+ |
Rob Ober | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
62% | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
10 | QUANTITY | 0.99+ |
March | DATE | 0.99+ |
five years | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
Alex Ned | PERSON | 0.99+ |
Roger Daltry | PERSON | 0.99+ |
AIRI | ORGANIZATION | 0.99+ |
U.K. | LOCATION | 0.99+ |
four hours | QUANTITY | 0.99+ |
ten times | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Bill Graham Civic Auditorium | LOCATION | 0.99+ |
today | DATE | 0.99+ |
less than half a rack | QUANTITY | 0.98+ |
Arista | ORGANIZATION | 0.98+ |
10 years ago | DATE | 0.98+ |
San Francisco | LOCATION | 0.98+ |
20 years ago | DATE | 0.98+ |
summer of 2012 | DATE | 0.98+ |
three legs | QUANTITY | 0.98+ |
tens of thousands of cores | QUANTITY | 0.97+ |
less than six years | QUANTITY | 0.97+ |
Man AHL | ORGANIZATION | 0.97+ |
both | QUANTITY | 0.97+ |
a week | QUANTITY | 0.96+ |
earlier this week | DATE | 0.96+ |
more than a terabyte | QUANTITY | 0.96+ |
50 racks | QUANTITY | 0.96+ |
Global Technology Conference | EVENT | 0.96+ |
this morning | DATE | 0.95+ |
more than a terabyte per second | QUANTITY | 0.95+ |
Pure | ORGANIZATION | 0.94+ |
GTC | EVENT | 0.94+ |
less than six years ago | DATE | 0.93+ |
petabytes | QUANTITY | 0.92+ |
terabytes | QUANTITY | 0.92+ |
half rack | QUANTITY | 0.92+ |
one of the legs | QUANTITY | 0.92+ |
single run | QUANTITY | 0.92+ |
a month | QUANTITY | 0.91+ |
FlashBlade | ORGANIZATION | 0.9+ |
theCUBE | ORGANIZATION | 0.88+ |
Pure Storage Accelerate 2018 | EVENT | 0.88+ |
20 years | QUANTITY | 0.87+ |
Luis Ceze & Anna Connolly, OctoML | AWS Startup Showcase S3 E1
(soft music) >> Hello, everyone. Welcome to theCUBE's presentation of the AWS Startup Showcase. AI and Machine Learning: Top Startups Building Foundational Model Infrastructure. This is season 3, episode 1 of the ongoing series covering the exciting stuff from the AWS ecosystem, talking about machine learning and AI. I'm your host, John Furrier and today we are excited to be joined by Luis Ceze who's the CEO of OctoML and Anna Connolly, VP of customer success and experience OctoML. Great to have you on again, Luis. Anna, thanks for coming on. Appreciate it. >> Thank you, John. It's great to be here. >> Thanks for having us. >> I love the company. We had a CUBE conversation about this. You guys are really addressing how to run foundational models faster for less. And this is like the key theme. But before we get into it, this is a hot trend, but let's explain what you guys do. Can you set the narrative of what the company's about, why it was founded, what's your North Star and your mission? >> Yeah, so John, our mission is to make AI sustainable and accessible for everyone. And what we offer customers is, you know, a way of taking their models into production in the most efficient way possible by automating the process of getting a model and optimizing it for a variety of hardware and making cost-effective. So better, faster, cheaper model deployment. >> You know, the big trend here is AI. Everyone's seeing the ChatGPT, kind of the shot heard around the world. The BingAI and this fiasco and the ongoing experimentation. People are into it, and I think the business impact is clear. I haven't seen this in all of my career in the technology industry of this kind of inflection point. And every senior leader I talk to is rethinking about how to rebuild their business with AI because now the large language models have come in, these foundational models are here, they can see value in their data. This is a 10 year journey in the big data world. Now it's impacting that, and everyone's rebuilding their company around this idea of being AI first 'cause they see ways to eliminate things and make things more efficient. And so now they telling 'em to go do it. And they're like, what do we do? So what do you guys think? Can you explain what is this wave of AI and why is it happening, why now, and what should people pay attention to? What does it mean to them? >> Yeah, I mean, it's pretty clear by now that AI can do amazing things that captures people's imaginations. And also now can show things that are really impactful in businesses, right? So what people have the opportunity to do today is to either train their own model that adds value to their business or find open models out there that can do very valuable things to them. So the next step really is how do you take that model and put it into production in a cost-effective way so that the business can actually get value out of it, right? >> Anna, what's your take? Because customers are there, you're there to make 'em successful, you got the new secret weapon for their business. >> Yeah, I think we just see a lot of companies struggle to get from a trained model into a model that is deployed in a cost-effective way that actually makes sense for the application they're building. I think that's a huge challenge we see today, kind of across the board across all of our customers. >> Well, I see this, everyone asking the same question. I have data, I want to get value out of it. I got to get these big models, I got to train it. What's it going to cost? So I think there's a reality of, okay, I got to do it. Then no one has any visibility on what it costs. When they get into it, this is going to break the bank. So I have to ask you guys, the cost of training these models is on everyone's mind. OctoML, your company's focus on the cost side of it as well as the efficiency side of running these models in production. Why are the production costs such a concern and where specifically are people looking at it and why did it get here? >> Yeah, so training costs get a lot of attention because normally a large number, but we shouldn't forget that it's a large, typically one time upfront cost that customers pay. But, you know, when the model is put into production, the cost grows directly with model usage and you actually want your model to be used because it's adding value, right? So, you know, the question that a customer faces is, you know, they have a model, they have a trained model and now what? So how much would it cost to run in production, right? And now without the big wave in generative AI, which rightfully is getting a lot of attention because of the amazing things that it can do. It's important for us to keep in mind that generative AI models like ChatGPT are huge, expensive energy hogs. They cost a lot to run, right? And given that model usage growth directly, model cost grows directly with usage, what you want to do is make sure that once you put a model into production, you have the best cost structure possible so that you're not surprised when it's gets popular, right? So let me give you an example. So if you have a model that costs, say 1 to $2 million to train, but then it costs about one to two cents per session to use it, right? So if you have a million active users, even if they use just once a day, it's 10 to $20,000 a day to operate that model in production. And that very, very quickly, you know, get beyond what you paid to train it. >> Anna, these aren't small numbers, and it's cost to train and cost to operate, it kind of reminds me of when the cloud came around and the data center versus cloud options. Like, wait a minute, one, it costs a ton of cash to deploy, and then running it. This is kind of a similar dynamic. What are you seeing? >> Yeah, absolutely. I think we are going to see increasingly the cost and production outpacing the costs and training by a lot. I mean, people talk about training costs now because that's what they're confronting now because people are so focused on getting models performant enough to even use in an application. And now that we have them and they're that capable, we're really going to start to see production costs go up a lot. >> Yeah, Luis, if you don't mind, I know this might be a little bit of a tangent, but, you know, training's super important. I get that. That's what people are doing now, but then there's the deployment side of production. Where do people get caught up and miss the boat or misconfigure? What's the gotcha? Where's the trip wire or so to speak? Where do people mess up on the cost side? What do they do? Is it they don't think about it, they tie it to proprietary hardware? What's the issue? >> Yeah, several things, right? So without getting really technical, which, you know, I might get into, you know, you have to understand relationship between performance, you know, both in terms of latency and throughput and cost, right? So reducing latency is important because you improve responsiveness of the model. But it's really important to keep in mind that it often leads diminishing returns. Below a certain latency, making it faster won't make a measurable difference in experience, but it's going to cost a lot more. So understanding that is important. Now, if you care more about throughputs, which is the time it takes for you to, you know, units per period of time, you care about time to solution, we should think about this throughput per dollar. And understand what you want is the highest throughput per dollar, which may come at the cost of higher latency, which you're not going to care about, right? So, and the reality here, John, is that, you know, humans and especially folks in this space want to have the latest and greatest hardware. And often they commit a lot of money to get access to them and have to commit upfront before they understand the needs that their models have, right? So common mistake here, one is not spending time to understand what you really need, and then two, over-committing and using more hardware than you actually need. And not giving yourself enough freedom to get your workload to move around to the more cost-effective choice, right? So this is just a metaphoric choice. And then another thing that's important here too is making a model run faster on the hardware directly translates to lower cost, right? So, but it takes a lot of engineers, you need to think of ways of producing very efficient versions of your model for the target hardware that you're going to use. >> Anna, what's the customer angle here? Because price performance has been around for a long time, people get that, but now latency and throughput, that's key because we're starting to see this in apps. I mean, there's an end user piece. I even seeing it on the infrastructure side where they're taking a heavy lifting away from operational costs. So you got, you know, application specific to the user and/or top of the stack, and then you got actually being used in operations where they want both. >> Yeah, absolutely. Maybe I can illustrate this with a quick story with the customer that we had recently been working with. So this customer is planning to run kind of a transformer based model for tech generation at super high scale on Nvidia T4 GPU, so kind of a commodity GPU. And the scale was so high that they would've been paying hundreds of thousands of dollars in cloud costs per year just to serve this model alone. You know, one of many models in their application stack. So we worked with this team to optimize our model and then benchmark across several possible targets. So that matching the hardware that Luis was just talking about, including the newer kind of Nvidia A10 GPUs. And what they found during this process was pretty interesting. First, the team was able to shave a quarter of their spend just by using better optimization techniques on the T4, the older hardware. But actually moving to a newer GPU would allow them to serve this model in a sub two milliseconds latency, so super fast, which was able to unlock an entirely new kind of user experience. So they were able to kind of change the value they're delivering in their application just because they were able to move to this new hardware easily. So they ultimately decided to plan their deployment on the more expensive A10 because of this, but because of the hardware specific optimizations that we helped them with, they managed to even, you know, bring costs down from what they had originally planned. And so if you extend this kind of example to everything that's happening with generative AI, I think the story we just talked about was super relevant, but the scale can be even higher, you know, it can be tenfold that. We were recently conducting kind of this internal study using GPT-J as a proxy to illustrate the experience of just a company trying to use one of these large language models with an example scenario of creating a chatbot to help job seekers prepare for interviews. So if you imagine kind of a conservative usage scenario where the model generates just 3000 words per user per day, which is, you know, pretty conservative for how people are interacting with these models. It costs 5 cents a session and if you're a company and your app goes viral, so from, you know, beginning of the year there's nobody, at the end of the year there's a million daily active active users in that year alone, going from zero to a million. You'll be spending about $6 million a year, which is pretty unmanageable. That's crazy, right? >> Yeah. >> For a company or a product that's just launching. So I think, you know, for us we see the real way to make these kind of advancements accessible and sustainable, as we said is to bring down cost to serve using these techniques. >> That's a great story and I think that illustrates this idea that deployment cost can vary from situation to situation, from model to model and that the efficiency is so strong with this new wave, it eliminates heavy lifting, creates more efficiency, automates intellect. I mean, this is the trend, this is radical, this is going to increase. So the cost could go from nominal to millions, literally, potentially. So, this is what customers are doing. Yeah, that's a great story. What makes sense on a financial, is there a cost of ownership? Is there a pattern for best practice for training? What do you guys advise cuz this is a lot of time and money involved in all potential, you know, good scenarios of upside. But you can get over your skis as they say, and be successful and be out of business if you don't manage it. I mean, that's what people are talking about, right? >> Yeah, absolutely. I think, you know, we see kind of three main vectors to reduce cost. I think one is make your deployment process easier overall, so that your engineering effort to even get your app running goes down. Two, would be get more from the compute you're already paying for, you're already paying, you know, for your instances in the cloud, but can you do more with that? And then three would be shop around for lower cost hardware to match your use case. So on the first one, I think making the deployment easier overall, there's a lot of manual work that goes into benchmarking, optimizing and packaging models for deployment. And because the performance of machine learning models can be really hardware dependent, you have to go through this process for each target you want to consider running your model on. And this is hard, you know, we see that every day. But for teams who want to incorporate some of these large language models into their applications, it might be desirable because licensing a model from a large vendor like OpenAI can leave you, you know, over provision, kind of paying for capabilities you don't need in your application or can lock you into them and you lose flexibility. So we have a customer whose team actually prepares models for deployment in a SaaS application that many of us use every day. And they told us recently that without kind of an automated benchmarking and experimentation platform, they were spending several days each to benchmark a single model on a single hardware type. So this is really, you know, manually intensive and then getting more from the compute you're already paying for. We do see customers who leave money on the table by running models that haven't been optimized specifically for the hardware target they're using, like Luis was mentioning. And for some teams they just don't have the time to go through an optimization process and for others they might lack kind of specialized expertise and this is something we can bring. And then on shopping around for different hardware types, we really see a huge variation in model performance across hardware, not just CPU vs. GPU, which is, you know, what people normally think of. But across CPU vendors themselves, high memory instances and across cloud providers even. So the best strategy here is for teams to really be able to, we say, look before you leap by running real world benchmarking and not just simulations or predictions to find the best software, hardware combination for their workload. >> Yeah. You guys sound like you have a very impressive customer base deploying large language models. Where would you categorize your current customer base? And as you look out, as you guys are growing, you have new customers coming in, take me through the progression. Take me through the profile of some of your customers you have now, size, are they hyperscalers, are they big app folks, are they kicking the tires? And then as people are out there scratching heads, I got to get in this game, what's their psychology like? Are they coming in with specific problems or do they have specific orientation point of view about what they want to do? Can you share some data around what you're seeing? >> Yeah, I think, you know, we have customers that kind of range across the spectrum of sophistication from teams that basically don't have MLOps expertise in their company at all. And so they're really looking for us to kind of give a full service, how should I do everything from, you know, optimization, find the hardware, prepare for deployment. And then we have teams that, you know, maybe already have their serving and hosting infrastructure up and ready and they already have models in production and they're really just looking to, you know, take the extra juice out of the hardware and just do really specific on that optimization piece. I think one place where we're doing a lot more work now is kind of in the developer tooling, you know, model selection space. And that's kind of an area that we're creating more tools for, particularly within the PyTorch ecosystem to bring kind of this power earlier in the development cycle so that as people are grabbing a model off the shelf, they can, you know, see how it might perform and use that to inform their development process. >> Luis, what's the big, I like this idea of picking the models because isn't that like going to the market and picking the best model for your data? It's like, you know, it's like, isn't there a certain approaches? What's your view on this? 'Cause this is where everyone, I think it's going to be a land rush for this and I want to get your thoughts. >> For sure, yeah. So, you know, I guess I'll start with saying the one main takeaway that we got from the GPT-J study is that, you know, having a different understanding of what your model's compute and memory requirements are, very quickly, early on helps with the much smarter AI model deployments, right? So, and in fact, you know, Anna just touched on this, but I want to, you know, make sure that it's clear that OctoML is putting that power into user's hands right now. So in partnership with AWS, we are launching this new PyTorch native profiler that allows you with a single, you know, one line, you know, code decorator allows you to see how your code runs on a variety of different hardware after accelerations. So it gives you very clear, you know, data on how you should think about your model deployments. And this ties back to choices of models. So like, if you have a set of choices that are equally good of models in terms of functionality and you want to understand after acceleration how are you going to deploy, how much they're going to cost or what are the options using a automated process of making a decision is really, really useful. And in fact, so I think these events can get early access to this by signing up for the Octopods, you know, this is exclusive group for insiders here, so you can go to OctoML.ai/pods to sign up. >> So that Octopod, is that a program? What is that, is that access to code? Is that a beta, what is that? Explain, take a minute and explain Octopod. >> I think the Octopod would be a group of people who is interested in experiencing this functionality. So it is the friends and users of OctoML that would be the Octopod. And then yes, after you sign up, we would provide you essentially the tool in code form for you to try out in your own. I mean, part of the benefit of this is that it happens in your own local environment and you're in control of everything kind of within the workflow that developers are already using to create and begin putting these models into their applications. So it would all be within your control. >> Got it. I think the big question I have for you is when do you, when does that one of your customers know they need to call you? What's their environment look like? What are they struggling with? What are the conversations they might be having on their side of the fence? If anyone's watching this, they're like, "Hey, you know what, I've got my team, we have a lot of data. Do we have our own language model or do I use someone else's?" There's a lot of this, I will say discovery going on around what to do, what path to take, what does that customer look like, if someone's listening, when do they know to call you guys, OctoML? >> Well, I mean the most obvious one is that you have a significant spend on AI/ML, come and talk to us, you know, putting AIML into production. So that's the clear one. In fact, just this morning I was talking to someone who is in life sciences space and is having, you know, 15 to $20 million a year cloud related to AI/ML deployment is a clear, it's a pretty clear match right there, right? So that's on the cost side. But I also want to emphasize something that Anna said earlier that, you know, the hardware and software complexity involved in putting model into production is really high. So we've been able to abstract that away, offering a clean automation flow enables one, to experiment early on, you know, how models would run and get them to production. And then two, once they are into production, gives you an automated flow to continuously updating your model and taking advantage of all this acceleration and ability to run the model on the right hardware. So anyways, let's say one then is cost, you know, you have significant cost and then two, you have an automation needs. And Anna please compliment that. >> Yeah, Anna you can please- >> Yeah, I think that's exactly right. Maybe the other time is when you are expecting a big scale up in serving your application, right? You're launching a new feature, you expect to get a lot of usage or, and you want to kind of anticipate maybe your CTO, your CIO, whoever pays your cloud bills is going to come after you, right? And so they want to know, you know, what's the return on putting this model essentially into my application stack? Am I going to, is the usage going to match what I'm paying for it? And then you can understand that. >> So you guys have a lot of the early adopters, they got big data teams, they're pushed in the production, they want to get a little QA, test the waters, understand, use your technology to figure it out. Is there any cases where people have gone into production, they have to pull it out? It's like the old lemon laws with your car, you buy a car and oh my god, it's not the way I wanted it. I mean, I can imagine the early people through the wall, so to speak, in the wave here are going to be bloody in the sense that they've gone in and tried stuff and get stuck with huge bills. Are you seeing that? Are people pulling stuff out of production and redeploying? Or I can imagine that if I had a bad deployment, I'd want to refactor that or actually replatform that. Do you see that too? >> Definitely after a sticker shock, yes, your customers will come and make sure that, you know, the sticker shock won't happen again. >> Yeah. >> But then there's another more thorough aspect here that I think we likely touched on, be worth elaborating a bit more is just how are you going to scale in a way that's feasible depending on the allocation that you get, right? So as we mentioned several times here, you know, model deployment is so hardware dependent and so complex that you tend to get a model for a hardware choice and then you want to scale that specific type of instance. But what if, when you want to scale because suddenly luckily got popular and, you know, you want to scale it up and then you don't have that instance anymore. So how do you live with whatever you have at that moment is something that we see customers needing as well. You know, so in fact, ideally what we want is customers to not think about what kind of specific instances they want. What they want is to know what their models need. Say, they know the SLA and then find a set of hybrid targets and instances that hit the SLA whenever they're also scaling, they're going to scale with more freedom, right? Instead of having to wait for AWS to give them more specific allocation for a specific instance. What if you could live with other types of hardware and scale up in a more free way, right? So that's another thing that we see customers, you know, like they need more freedom to be able to scale with whatever is available. >> Anna, you touched on this with the business model impact to that 6 million cost, if that goes out of control, there's a business model aspect and there's a technical operation aspect to the cost side too. You want to be mindful of riding the wave in a good way, but not getting over your skis. So that brings up the point around, you know, confidence, right? And teamwork. Because if you're in production, there's probably a team behind it. Talk about the team aspect of your customers. I mean, they're dedicated, they go put stuff into production, they're developers, there're data. What's in it for them? Are they getting better, are they in the beach, you know, reading the book. Are they, you know, are there easy street for them? What's the customer benefit to the teams? >> Yeah, absolutely. With just a few clicks of a button, you're in production, right? That's the dream. So yeah, I mean I think that, you know, we illustrated it before a little bit. I think the automated kind of benchmarking and optimization process, like when you think about the effort it takes to get that data by hand, which is what people are doing today, they just don't do it. So they're making decisions without the best information because it's, you know, there just isn't the bandwidth to get the information that they need to make the best decision and then know exactly how to deploy it. So I think it's actually bringing kind of a new insight and capability to these teams that they didn't have before. And then maybe another aspect on the team side is that it's making the hand-off of the models from the data science teams to the model deployment teams more seamless. So we have, you know, we have seen in the past that this kind of transition point is the place where there are a lot of hiccups, right? The data science team will give a model to the production team and it'll be too slow for the application or it'll be too expensive to run and it has to go back and be changed and kind of this loop. And so, you know, with the PyTorch profiler that Luis was talking about, and then also, you know, the other ways we do optimization that kind of prevents that hand-off problem from happening. >> Luis and Anna, you guys have a great company. Final couple minutes left. Talk about the company, the people there, what's the culture like, you know, if Intel has Moore's law, which is, you know, doubling the performance in few years, what's the culture like there? Is it, you know, more throughput, better pricing? Explain what's going on with the company and put a plug in. Luis, we'll start with you. >> Yeah, absolutely. I'm extremely proud of the team that we built here. You know, we have a people first culture, you know, very, very collaborative and folks, we all have a shared mission here of making AI more accessible and sustainable. We have a very diverse team in terms of backgrounds and life stories, you know, to do what we do here, we need a team that has expertise in software engineering, in machine learning, in computer architecture. Even though we don't build chips, we need to understand how they work, right? So, and then, you know, the fact that we have this, this very really, really varied set of backgrounds makes the environment, you know, it's say very exciting to learn more about, you know, assistance end-to-end. But also makes it for a very interesting, you know, work environment, right? So people have different backgrounds, different stories. Some of them went to grad school, others, you know, were in intelligence agencies and now are working here, you know. So we have a really interesting set of people and, you know, life is too short not to work with interesting humans. You know, that's something that I like to think about, you know. >> I'm sure your off-site meetings are a lot of fun, people talking about computer architectures, silicon advances, the next GPU, the big data models coming in. Anna, what's your take? What's the culture like? What's the company vibe and what are you guys looking to do? What's the customer success pattern? What's up? >> Yeah, absolutely. I mean, I, you know, second all of the great things that Luis just said about the team. I think one that I, an additional one that I'd really like to underscore is kind of this customer obsession, to use a term you all know well. And focus on the end users and really making the experiences that we're bringing to our user who are developers really, you know, useful and valuable for them. And so I think, you know, all of these tools that we're trying to put in the hands of users, the industry and the market is changing so rapidly that our products across the board, you know, all of the companies that, you know, are part of the showcase today, we're all evolving them so quickly and we can only do that kind of really hand in glove with our users. So that would be another thing I'd emphasize. >> I think the change dynamic, the power dynamics of this industry is just the beginning. I'm very bullish that this is going to be probably one of the biggest inflection points in history of the computer industry because of all the dynamics of the confluence of all the forces, which you mentioned some of them, I mean PC, you know, interoperability within internetworking and you got, you know, the web and then mobile. Now we have this, I mean, I wouldn't even put social media even in the close to this. Like, this is like, changes user experience, changes infrastructure. There's going to be massive accelerations in performance on the hardware side from AWS's of the world and cloud and you got the edge and more data. This is really what big data was going to look like. This is the beginning. Final question, what do you guys see going forward in the future? >> Well, it's undeniable that machine learning and AI models are becoming an integral part of an interesting application today, right? So, and the clear trends here are, you know, more and more competitional needs for these models because they're only getting more and more powerful. And then two, you know, seeing the complexity of the infrastructure where they run, you know, just considering the cloud, there's like a wide variety of choices there, right? So being able to live with that and making the most out of it in a way that does not require, you know, an impossible to find team is something that's pretty clear. So the need for automation, abstracting with the complexity is definitely here. And we are seeing this, you know, trends are that you also see models starting to move to the edge as well. So it's clear that we're seeing, we are going to live in a world where there's no large models living in the cloud. And then, you know, edge models that talk to these models in the cloud to form, you know, an end-to-end truly intelligent application. >> Anna? >> Yeah, I think, you know, our, Luis said it at the beginning. Our vision is to make AI sustainable and accessible. And I think as this technology just expands in every company and every team, that's going to happen kind of on its own. And we're here to help support that. And I think you can't do that without tools like those like OctoML. >> I think it's going to be an error of massive invention, creativity, a lot of the format heavy lifting is going to allow the talented people to automate their intellect. I mean, this is really kind of what we see going on. And Luis, thank you so much. Anna, thanks for coming on this segment. Thanks for coming on theCUBE and being part of the AWS Startup Showcase. I'm John Furrier, your host. Thanks for watching. (upbeat music)
SUMMARY :
Great to have you on again, Luis. It's great to be here. but let's explain what you guys do. And what we offer customers is, you know, So what do you guys think? so that the business you got the new secret kind of across the board So I have to ask you guys, And that very, very quickly, you know, and the data center versus cloud options. And now that we have them but, you know, training's super important. John, is that, you know, humans and then you got actually managed to even, you know, So I think, you know, for us we see in all potential, you know, And this is hard, you know, And as you look out, as And then we have teams that, you know, and picking the best model for your data? from the GPT-J study is that, you know, What is that, is that access to code? And then yes, after you sign up, to call you guys, OctoML? come and talk to us, you know, And so they want to know, you know, So you guys have a lot make sure that, you know, we see customers, you know, What's the customer benefit to the teams? and then also, you know, what's the culture like, you know, So, and then, you know, and what are you guys looking to do? all of the companies that, you know, I mean PC, you know, in the cloud to form, you know, And I think you can't And Luis, thank you so much.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Anna | PERSON | 0.99+ |
Anna Connolly | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Luis | PERSON | 0.99+ |
Luis Ceze | PERSON | 0.99+ |
John | PERSON | 0.99+ |
1 | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
15 | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
10 year | QUANTITY | 0.99+ |
6 million | QUANTITY | 0.99+ |
zero | QUANTITY | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
three | QUANTITY | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
First | QUANTITY | 0.99+ |
OctoML | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
millions | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Two | QUANTITY | 0.99+ |
$2 million | QUANTITY | 0.98+ |
3000 words | QUANTITY | 0.98+ |
one line | QUANTITY | 0.98+ |
A10 | COMMERCIAL_ITEM | 0.98+ |
OctoML | TITLE | 0.98+ |
one | QUANTITY | 0.98+ |
three main vectors | QUANTITY | 0.97+ |
hundreds of thousands of dollars | QUANTITY | 0.97+ |
both | QUANTITY | 0.97+ |
CUBE | ORGANIZATION | 0.97+ |
T4 | COMMERCIAL_ITEM | 0.97+ |
one time | QUANTITY | 0.97+ |
first one | QUANTITY | 0.96+ |
two cents | QUANTITY | 0.96+ |
GPT-J | ORGANIZATION | 0.96+ |
single model | QUANTITY | 0.95+ |
a minute | QUANTITY | 0.95+ |
about $6 million a year | QUANTITY | 0.95+ |
once a day | QUANTITY | 0.95+ |
$20,000 a day | QUANTITY | 0.95+ |
a million | QUANTITY | 0.94+ |
theCUBE | ORGANIZATION | 0.93+ |
Octopod | TITLE | 0.93+ |
this morning | DATE | 0.93+ |
first culture | QUANTITY | 0.92+ |
$20 million a year | QUANTITY | 0.92+ |
AWS Startup Showcase | EVENT | 0.9+ |
North Star | ORGANIZATION | 0.9+ |
Joseph Nelson, Roboflow | Cube Conversation
(gentle music) >> Hello everyone. Welcome to this CUBE conversation here in Palo Alto, California. I'm John Furrier, host of theCUBE. We got a great remote guest coming in. Joseph Nelson, co-founder and CEO of RoboFlow hot startup in AI, computer vision. Really interesting topic in this wave of AI next gen hitting. Joseph, thanks for coming on this CUBE conversation. >> Thanks for having me. >> Yeah, I love the startup tsunami that's happening here in this wave. RoboFlow, you're in the middle of it. Exciting opportunities, you guys are in the cutting edge. I think computer vision's been talked about more as just as much as the large language models and these foundational models are merging. You're in the middle of it. What's it like right now as a startup and growing in this new wave hitting? >> It's kind of funny, it's, you know, I kind of describe it like sometimes you're in a garden of gnomes. It's like we feel like we've got this giant headstart with hundreds of thousands of people building with computer vision, training their own models, but that's a fraction of what it's going to be in six months, 12 months, 24 months. So, as you described it, a wave is a good way to think about it. And the wave is still building before it gets to its full size. So it's a ton of fun. >> Yeah, I think it's one of the most exciting areas in computer science. I wish I was in my twenties again, because I would be all over this. It's the intersection, there's so many disciplines, right? It's not just tech computer science, it's computer science, it's systems, it's software, it's data. There's so much aperture of things going on around your world. So, I mean, you got to be batting all the students away kind of trying to get hired in there, probably. I can only imagine you're hiring regiment. I'll ask that later, but first talk about what the company is that you're doing. How it's positioned, what's the market you're going after, and what's the origination story? How did you guys get here? How did you just say, hey, want to do this? What was the origination story? What do you do and how did you start the company? >> Yeah, yeah. I'll give you the what we do today and then I'll shift into the origin. RoboFlow builds tools for making the world programmable. Like anything that you see should be read write access if you think about it with a programmer's mind or legible. And computer vision is a technology that enables software to be added to these real world objects that we see. And so any sort of interface, any sort of object, any sort of scene, we can interact with it, we can make it more efficient, we can make it more entertaining by adding the ability for the tools that we use and the software that we write to understand those objects. And at RoboFlow, we've empowered a little over a hundred thousand developers, including those in half the Fortune 100 so far in that mission. Whether that's Walmart understanding the retail in their stores, Cardinal Health understanding the ways that they're helping their patients, or even electric vehicle manufacturers ensuring that they're making the right stuff at the right time. As you mentioned, it's early. Like I think maybe computer vision has touched one, maybe 2% of the whole economy and it'll be like everything in a very short period of time. And so we're focused on enabling that transformation. I think it's it, as far as I think about it, I've been fortunate to start companies before, start, sell these sorts of things. This is the last company I ever wanted to start and I think it will be, should we do it right, the world's largest in riding the wave of bringing together the disparate pieces of that technology. >> What was the motivating point of the formation? Was it, you know, you guys were hanging around? Was there some catalyst? What was the moment where it all kind of came together for you? >> You know what's funny is my co-founder, Brad and I, we were making computer vision apps for making board games more fun to play. So in 2017, Apple released AR kit, augmented reality kit for building augmented reality applications. And Brad and I are both sort of like hacker persona types. We feel like we don't really understand the technology until we build something with it and so we decided that we should make an app that if you point your phone at a Sudoku puzzle, it understands the state of the board and then it kind of magically fills in that experience with all the digits in real time, which totally ruins the game of Sudoku to be clear. But it also just creates this like aha moment of like, oh wow, like the ability for our pocket devices to understand and see the world as good or better than we can is possible. And so, you know, we actually did that as I mentioned in 2017, and the app went viral. It was, you know, top of some subreddits, top of Injure, Reddit, the hacker community as well as Product Hunt really liked it. So it actually won Product Hunt AR app of the year, which was the same year that the Tesla model three won the product of the year. So we joked that we share an award with Elon our shared (indistinct) But frankly, so that was 2017. RoboFlow wasn't incorporated as a business until 2019. And so, you know, when we made Magic Sudoku, I was running a different company at the time, Brad was running a different company at the time, and we kind of just put it out there and were excited by how many people liked it. And we assumed that other curious developers would see this inevitable future of, oh wow, you know. This is much more than just a pedestrian point your phone at a board game. This is everything can be seen and understood and rewritten in a different way. Things like, you know, maybe your fridge. Knowing what ingredients you have and suggesting recipes or auto ordering for you, or we were talking about some retail use cases of automated checkout. Like anything can be seen and observed and we presume that that would kick off a Cambrian explosion of applications. It didn't. So you fast forward to 2019, we said, well we might as well be the guys to start to tackle this sort of problem. And because of our success with board games before, we returned to making more board game solving applications. So we made one that solves Boggle, you know, the four by four word game, we made one that solves chess, you point your phone at a chess board and it understands the state of the board and then can make move recommendations. And each additional board game that we added, we realized that the tooling was really immature. The process of collecting images, knowing which images are actually going to be useful for improving model performance, training those models, deploying those models. And if we really wanted to make the world programmable, developers waiting for us to make an app for their thing of interest is a lot less efficient, less impactful than taking our tool chain and releasing that externally. And so, that's what RoboFlow became. RoboFlow became the internal tools that we used to make these game changing applications readily available. And as you know, when you give developers new tools, they create new billion dollar industries, let alone all sorts of fun hobbyist projects along the way. >> I love that story. Curious, inventive, little radical. Let's break the rules, see how we can push the envelope on the board games. That's how companies get started. It's a great story. I got to ask you, okay, what happens next? Now, okay, you realize this new tooling, but this is like how companies get built. Like they solve their own problem that they had 'cause they realized there's one, but then there has to be a market for it. So you actually guys knew that this was coming around the corner. So okay, you got your hacker mentality, you did that thing, you got the award and now you're like, okay, wow. Were you guys conscious of the wave coming? Was it one of those things where you said, look, if we do this, we solve our own problem, this will be big for everybody. Did you have that moment? Was that in 2019 or was that more of like, it kind of was obvious to you guys? >> Absolutely. I mean Brad puts this pretty effectively where he describes how we lived through the initial internet revolution, but we were kind of too young to really recognize and comprehend what was happening at the time. And then mobile happened and we were working on different companies that were not in the mobile space. And computer vision feels like the wave that we've caught. Like, this is a technology and capability that rewrites how we interact with the world, how everyone will interact with the world. And so we feel we've been kind of lucky this time, right place, right time of every enterprise will have the ability to improve their operations with computer vision. And so we've been very cognizant of the fact that computer vision is one of those groundbreaking technologies that every company will have as a part of their products and services and offerings, and we can provide the tooling to accelerate that future. >> Yeah, and the developer angle, by the way, I love that because I think, you know, as we've been saying in theCUBE all the time, developer's the new defacto standard bodies because what they adopt is pure, you know, meritocracy. And they pick the best. If it's sell service and it's good and it's got open source community around it, its all in. And they'll vote. They'll vote with their code and that is clear. Now I got to ask you, as you look at the market, we were just having this conversation on theCUBE in Barcelona at recent Mobile World Congress, now called MWC, around 5G versus wifi. And the debate was specifically computer vision, like facial recognition. We were talking about how the Cleveland Browns were using facial recognition for people coming into the stadium they were using it for ships in international ports. So the question was 5G versus wifi. My question is what infrastructure or what are the areas that need to be in place to make computer vision work? If you have developers building apps, apps got to run on stuff. So how do you sort that out in your mind? What's your reaction to that? >> A lot of the times when we see applications that need to run in real time and on video, they'll actually run at the edge without internet. And so a lot of our users will actually take their models and run it in a fully offline environment. Now to act on that information, you'll often need to have internet signal at some point 'cause you'll need to know how many people were in the stadium or what shipping crates are in my port at this point in time. You'll need to relay that information somewhere else, which will require connectivity. But actually using the model and creating the insights at the edge does not require internet. I mean we have users that deploy models on underwater submarines just as much as in outer space actually. And those are not very friendly environments to internet, let alone 5g. And so what you do is you use an edge device, like an Nvidia Jetson is common, mobile devices are common. Intel has some strong edge devices, the Movidius family of chips for example. And you use that compute that runs completely offline in real time to process those signals. Now again, what you do with those signals may require connectivity and that becomes a question of the problem you're solving of how soon you need to relay that information to another place. >> So, that's an architectural issue on the infrastructure. If you're a tactical edge war fighter for instance, you might want to have highly available and maybe high availability. I mean, these are words that mean something. You got storage, but it's not at the edge in real time. But you can trickle it back and pull it down. That's management. So that's more of a business by business decision or environment, right? >> That's right, that's right. Yeah. So I mean we can talk through some specifics. So for example, the RoboFlow actually powers the broadcaster that does the tennis ball tracking at Wimbledon. That runs completely at the edge in real time in, you know, technically to track the tennis ball and point the camera, you actually don't need internet. Now they do have internet of course to do the broadcasting and relay the signal and feeds and these sorts of things. And so that's a case where you have both edge deployment of running the model and high availability act on that model. We have other instances where customers will run their models on drones and the drone will go and do a flight and it'll say, you know, this many residential homes are in this given area, or this many cargo containers are in this given shipping yard. Or maybe we saw these environmental considerations of soil erosion along this riverbank. The model in that case can run on the drone during flight without internet, but then you only need internet once the drone lands and you're going to act on that information because for example, if you're doing like a study of soil erosion, you don't need to be real time. You just need to be able to process and make use of that information once the drone finishes its flight. >> Well I can imagine a zillion use cases. I heard of a use case interview at a company that does computer vision to help people see if anyone's jumping the fence on their company. Like, they know what a body looks like climbing a fence and they can spot it. Pretty easy use case compared to probably some of the other things, but this is the horizontal use cases, its so many use cases. So how do you guys talk to the marketplace when you say, hey, we have generative AI for commuter vision. You might know language models that's completely different animal because vision's like the world, right? So you got a lot more to do. What's the difference? How do you explain that to customers? What can I build and what's their reaction? >> Because we're such a developer centric company, developers are usually creative and show you the ways that they want to take advantage of new technologies. I mean, we've had people use things for identifying conveyor belt debris, doing gas leak detection, measuring the size of fish, airplane maintenance. We even had someone that like a hobby use case where they did like a specific sushi identifier. I dunno if you know this, but there's a specific type of whitefish that if you grew up in the western hemisphere and you eat it in the eastern hemisphere, you get very sick. And so there was someone that made an app that tells you if you happen to have that fish in the sushi that you're eating. But security camera analysis, transportation flows, plant disease detection, really, you know, smarter cities. We have people that are doing curb management identifying, and a lot of these use cases, the fantastic thing about building tools for developers is they're a creative bunch and they have these ideas that if you and I sat down for 15 minutes and said, let's guess every way computer vision can be used, we would need weeks to list all the example use cases. >> We'd miss everything. >> And we'd miss. And so having the community show us the ways that they're using computer vision is impactful. Now that said, there are of course commercial industries that have discovered the value and been able to be out of the gate. And that's where we have the Fortune 100 customers, like we do. Like the retail customers in the Walmart sector, healthcare providers like Medtronic, or vehicle manufacturers like Rivian who all have very difficult either supply chain, quality assurance, in stock, out of stock, anti-theft protection considerations that require successfully making sense of the real world. >> Let me ask you a question. This is maybe a little bit in the weeds, but it's more developer focused. What are some of the developer profiles that you're seeing right now in terms of low-hanging fruit applications? And can you talk about the academic impact? Because I imagine if I was in school right now, I'd be all over it. Are you seeing Master's thesis' being worked on with some of your stuff? Is the uptake in both areas of younger pre-graduates? And then inside the workforce, What are some of the devs like? Can you share just either what their makeup is, what they work on, give a little insight into the devs you're working with. >> Leading developers that want to be on state-of-the-art technology build with RoboFlow because they know they can use the best in class open source. They know that they can get the most out of their data. They know that they can deploy extremely quickly. That's true among students as you mentioned, just as much as as industries. So we welcome students and I mean, we have research grants that will regularly support for people to publish. I mean we actually have a channel inside our internal slack where every day, more student publications that cite building with RoboFlow pop up. And so, that helps inspire some of the use cases. Now what's interesting is that the use case is relatively, you know, useful or applicable for the business or the student. In other words, if a student does a thesis on how to do, we'll say like shingle damage detection from satellite imagery and they're just doing that as a master's thesis, in fact most insurance businesses would be interested in that sort of application. So, that's kind of how we see uptick and adoption both among researchers who want to be on the cutting edge and publish, both with RoboFlow and making use of open source tools in tandem with the tool that we provide, just as much as industry. And you know, I'm a big believer in the philosophy that kind of like what the hackers are doing nights and weekends, the Fortune 500 are doing in a pretty short order period of time and we're experiencing that transition. Computer vision used to be, you know, kind of like a PhD, multi-year investment endeavor. And now with some of the tooling that we're working on in open source technologies and the compute that's available, these science fiction ideas are possible in an afternoon. And so you have this idea of maybe doing asset management or the aerial observation of your shingles or things like this. You have a few hundred images and you can de-risk whether that's possible for your business today. So there's pretty broad-based adoption among both researchers that want to be on the state of the art, as much as companies that want to reduce the time to value. >> You know, Joseph, you guys and your partner have got a great front row seat, ground floor, presented creation wave here. I'm seeing a pattern emerging from all my conversations on theCUBE with founders that are successful, like yourselves, that there's two kind of real things going on. You got the enterprises grabbing the products and retrofitting into their legacy and rebuilding their business. And then you have startups coming out of the woodwork. Young, seeing greenfield or pick a specific niche or focus and making that the signature lever to move the market. >> That's right. >> So can you share your thoughts on the startup scene, other founders out there and talk about that? And then I have a couple questions for like the enterprises, the old school, the existing legacy. Little slower, but the startups are moving fast. What are some of the things you're seeing as startups are emerging in this field? >> I think you make a great point that independent of RoboFlow, very successful, especially developer focused businesses, kind of have three customer types. You have the startups and maybe like series A, series B startups that you're building a product as fast as you can to keep up with them, and they're really moving just as fast as as you are and pulling the product out at you for things that they need. The second segment that you have might be, call it SMB but not enterprise, who are able to purchase and aren't, you know, as fast of moving, but are stable and getting value and able to get to production. And then the third type is enterprise, and that's where you have typically larger contract value sizes, slower moving in terms of adoption and feedback for your product. And I think what you see is that successful companies balance having those three customer personas because you have the small startups, small fast moving upstarts that are discerning buyers who know the market and elect to build on tooling that is best in class. And so you basically kind of pass the smell test of companies who are quite discerning in their purchases, plus are moving so quick they're pulling their product out of you. Concurrently, you have a product that's enterprise ready to service the scalability, availability, and trust of enterprise buyers. And that's ultimately where a lot of companies will see tremendous commercial success. I mean I remember seeing the Twilio IPO, Uber being like a full 20% of their revenue, right? And so there's this very common pattern where you have the ability to find some of those upstarts that you make bets on, like the next Ubers of the world, the smaller companies that continue to get developed with the product and then the enterprise whom allows you to really fund the commercial success of the business, and validate the size of the opportunity in market that's being creative. >> It's interesting, there's so many things happening there. It's like, in a way it's a new category, but it's not a new category. It becomes a new category because of the capabilities, right? So, it's really interesting, 'cause that's what you're talking about is a category, creating. >> I think developer tools. So people often talk about B to B and B to C businesses. I think developer tools are in some ways a third way. I mean ultimately they're B to B, you're selling to other businesses and that's where your revenue's coming from. However, you look kind of like a B to C company in the ways that you measure product adoption and kind of go to market. In other words, you know, we're often tracking the leading indicators of commercial success in the form of usage, adoption, retention. Really consumer app, traditionally based metrics of how to know you're building the right stuff, and that's what product led growth companies do. And then you ultimately have commercial traction in a B to B way. And I think that that actually kind of looks like a third thing, right? Like you can do these sort of funny zany marketing examples that you might see historically from consumer businesses, but yet you ultimately make your money from the enterprise who has these de-risked high value problems you can solve for them. And I selfishly think that that's the best of both worlds because I don't have to be like Evan Spiegel, guessing the next consumer trend or maybe creating the next consumer trend and catching lightning in a bottle over and over again on the consumer side. But I still get to have fun in our marketing and make sort of fun, like we're launching the world's largest game of rock paper scissors being played with computer vision, right? Like that's sort of like a fun thing you can do, but then you can concurrently have the commercial validation and customers telling you the things that they need to be built for them next to solve commercial pain points for them. So I really do think that you're right by calling this a new category and it really is the best of both worlds. >> It's a great call out, it's a great call out. In fact, I always juggle with the VC. I'm like, it's so easy. Your job is so easy to pick the winners. What are you talking about its so easy? I go, just watch what the developers jump on. And it's not about who started, it could be someone in the dorm room to the boardroom person. You don't know because that B to C, the C, it's B to D you know? You know it's developer 'cause that's a human right? That's a consumer of the tool which influences the business that never was there before. So I think this direct business model evolution, whether it's media going direct or going direct to the developers rather than going to a gatekeeper, this is the reality. >> That's right. >> Well I got to ask you while we got some time left to describe, I want to get into this topic of multi-modality, okay? And can you describe what that means in computer vision? And what's the state of the growth of that portion of this piece? >> Multi modality refers to using multiple traditionally siloed problem types, meaning text, image, video, audio. So you could treat an audio problem as only processing audio signal. That is not multimodal, but you could use the audio signal at the same time as a video feed. Now you're talking about multi modality. In computer vision, multi modality is predominantly happening with images and text. And one of the biggest releases in this space is actually two years old now, was clip, contrastive language image pre-training, which took 400 million image text pairs and basically instead of previously when you do classification, you basically map every single image to a single class, right? Like here's a bunch of images of chairs, here's a bunch of images of dogs. What clip did is used, you can think about it like, the class for an image being the Instagram caption for the image. So it's not one single thing. And by training on understanding the corpora, you basically see which words, which concepts are associated with which pixels. And this opens up the aperture for the types of problems and generalizability of models. So what does this mean? This means that you can get to value more quickly from an existing trained model, or at least validate that what you want to tackle with a computer vision, you can get there more quickly. It also opens up the, I mean. Clip has been the bedrock of some of the generative image techniques that have come to bear, just as much as some of the LLMs. And increasingly we're going to see more and more of multi modality being a theme simply because at its core, you're including more context into what you're trying to understand about the world. I mean, in its most basic sense, you could ask yourself, if I have an image, can I know more about that image with just the pixels? Or if I have the image and the sound of when that image was captured or it had someone describe what they see in that image when the image was captured, which one's going to be able to get you more signal? And so multi modality helps expand the ability for us to understand signal processing. >> Awesome. And can you just real quick, define clip for the folks that don't know what that means? >> Yeah. Clip is a model architecture, it's an acronym for contrastive language image pre-training and like, you know, model architectures that have come before it captures the almost like, models are kind of like brands. So I guess it's a brand of a model where you've done these 400 million image text pairs to match up which visual concepts are associated with which text concepts. And there have been new releases of clip, just at bigger sizes of bigger encoding's, of longer strings of texture, or larger image windows. But it's been a really exciting advancement that OpenAI released in January, 2021. >> All right, well great stuff. We got a couple minutes left. Just I want to get into more of a company-specific question around culture. All startups have, you know, some sort of cultural vibe. You know, Intel has Moore's law doubles every whatever, six months. What's your culture like at RoboFlow? I mean, if you had to describe that culture, obviously love the hacking story, you and your partner with the games going number one on Product Hunt next to Elon and Tesla and then hey, we should start a company two years later. That's kind of like a curious, inventing, building, hard charging, but laid back. That's my take. How would you describe the culture? >> I think that you're right. The culture that we have is one of shipping, making things. So every week each team shares what they did for our customers on a weekly basis. And we have such a strong emphasis on being better week over week that those sorts of things compound. So one big emphasis in our culture is getting things done, shipping, doing things for our customers. The second is we're an incredibly transparent place to work. For example, how we think about giving decisions, where we're progressing against our goals, what problems are biggest and most important for the company is all open information for those that are inside the company to know and progress against. The third thing that I'd use to describe our culture is one that thrives with autonomy. So RoboFlow has a number of individuals who have founded companies before, some of which have sold their businesses for a hundred million plus upon exit. And the way that we've been able to attract talent like that is because the problems that we're tackling are so immense, yet individuals are able to charge at it with the way that they think is best. And this is what pairs well with transparency. If you have a strong sense of what the company's goals are, how we're progressing against it, and you have this ownership mentality of what can I do to change or drive progress against that given outcome, then you create a really healthy pairing of, okay cool, here's where the company's progressing. Here's where things are going really well, here's the places that we most need to improve and work on. And if you're inside that company as someone who has a preponderance to be a self-starter and even a history of building entire functions or companies yourself, then you're going to be a place where you can really thrive. You have the inputs of the things where we need to work on to progress the company's goals. And you have the background of someone that is just necessarily a fast moving and ambitious type of individual. So I think the best way to describe it is a transparent place with autonomy and an emphasis on getting things done. >> Getting shit done as they say. Getting stuff done. Great stuff. Hey, final question. Put a plug out there for the company. What are you going to hire? What's your pipeline look like for people? What jobs are open? I'm sure you got hiring all around. Give a quick plug for the company what you're looking for. >> I appreciate you asking. Basically you're either building the product or helping customers be successful with the product. So in the building product category, we have platform engineering roles, machine learning engineering roles, and we're solving some of the hardest and most impactful problems of bringing such a groundbreaking technology to the masses. And so it's a great place to be where you can kind of be your own user as an engineer. And then if you're enabling people to be successful with the products, I mean you're working in a place where there's already such a strong community around it and you can help shape, foster, cultivate, activate, and drive commercial success in that community. So those are roles that tend themselves to being those that build the product for developer advocacy, those that are account executives that are enabling our customers to realize commercial success, and even hybrid roles like we call it field engineering, where you are a technical resource to drive success within customer accounts. And so all this is listed on roboflow.com/careers. And one thing that I actually kind of want to mention John that's kind of novel about the thing that's working at RoboFlow. So there's been a lot of discussion around remote companies and there's been a lot of discussion around in-person companies and do you need to be in the office? And one thing that we've kind of recognized is you can actually chart a third way. You can create a third way which we call satellite, which basically means people can work from where they most like to work and there's clusters of people, regular onsite's. And at RoboFlow everyone gets, for example, $2,500 a year that they can use to spend on visiting coworkers. And so what's sort of organically happened is team numbers have started to pull together these resources and rent out like, lavish Airbnbs for like a week and then everyone kind of like descends in and works together for a week and makes and creates things. And we call this lighthouses because you know, a lighthouse kind of brings ships into harbor and we have an emphasis on shipping. >> Yeah, quality people that are creative and doers and builders. You give 'em some cash and let the self-governing begin, you know? And like, creativity goes through the roof. It's a great story. I think that sums up the culture right there, Joseph. Thanks for sharing that and thanks for this great conversation. I really appreciate it and it's very inspiring. Thanks for coming on. >> Yeah, thanks for having me, John. >> Joseph Nelson, co-founder and CEO of RoboFlow. Hot company, great culture in the right place in a hot area, computer vision. This is going to explode in value. The edge is exploding. More use cases, more development, and developers are driving the change. Check out RoboFlow. This is theCUBE. I'm John Furrier, your host. Thanks for watching. (gentle music)
SUMMARY :
Welcome to this CUBE conversation You're in the middle of it. And the wave is still building the company is that you're doing. maybe 2% of the whole economy And as you know, when you it kind of was obvious to you guys? cognizant of the fact that I love that because I think, you know, And so what you do is issue on the infrastructure. and the drone will go and the marketplace when you say, in the sushi that you're eating. And so having the And can you talk about the use case is relatively, you know, and making that the signature What are some of the things you're seeing and pulling the product out at you because of the capabilities, right? in the ways that you the C, it's B to D you know? And one of the biggest releases And can you just real quick, and like, you know, I mean, if you had to like that is because the problems Give a quick plug for the place to be where you can the self-governing begin, you know? and developers are driving the change.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Brad | PERSON | 0.99+ |
Joseph | PERSON | 0.99+ |
Joseph Nelson | PERSON | 0.99+ |
January, 2021 | DATE | 0.99+ |
John Furrier | PERSON | 0.99+ |
Medtronic | ORGANIZATION | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
2019 | DATE | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
400 million | QUANTITY | 0.99+ |
Evan Spiegel | PERSON | 0.99+ |
24 months | QUANTITY | 0.99+ |
2017 | DATE | 0.99+ |
RoboFlow | ORGANIZATION | 0.99+ |
15 minutes | QUANTITY | 0.99+ |
Rivian | ORGANIZATION | 0.99+ |
12 months | QUANTITY | 0.99+ |
20% | QUANTITY | 0.99+ |
Cardinal Health | ORGANIZATION | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Barcelona | LOCATION | 0.99+ |
Wimbledon | EVENT | 0.99+ |
roboflow.com/careers | OTHER | 0.99+ |
first | QUANTITY | 0.99+ |
second segment | QUANTITY | 0.99+ |
each team | QUANTITY | 0.99+ |
six months | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
both worlds | QUANTITY | 0.99+ |
2% | QUANTITY | 0.99+ |
two years later | DATE | 0.98+ |
Mobile World Congress | EVENT | 0.98+ |
Ubers | ORGANIZATION | 0.98+ |
third way | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
a week | QUANTITY | 0.98+ |
Magic Sudoku | TITLE | 0.98+ |
second | QUANTITY | 0.98+ |
Nvidia | ORGANIZATION | 0.98+ |
Sudoku | TITLE | 0.98+ |
MWC | EVENT | 0.97+ |
today | DATE | 0.97+ |
billion dollar | QUANTITY | 0.97+ |
one single thing | QUANTITY | 0.97+ |
over a hundred thousand developers | QUANTITY | 0.97+ |
four | QUANTITY | 0.97+ |
third | QUANTITY | 0.96+ |
Elon | ORGANIZATION | 0.96+ |
third thing | QUANTITY | 0.96+ |
Tesla | ORGANIZATION | 0.96+ |
Jetson | COMMERCIAL_ITEM | 0.96+ |
Elon | PERSON | 0.96+ |
RoboFlow | TITLE | 0.96+ |
ORGANIZATION | 0.95+ | |
Twilio | ORGANIZATION | 0.95+ |
twenties | QUANTITY | 0.95+ |
Product Hunt AR | TITLE | 0.95+ |
Moore | PERSON | 0.95+ |
both researchers | QUANTITY | 0.95+ |
one thing | QUANTITY | 0.94+ |
Supercloud Applications & Developer Impact | Supercloud2
(gentle music) >> Okay, welcome back to Supercloud 2, live here in Palo Alto, California for our live stage performance. Supercloud 2 is our second Supercloud event. We're going to get these out as fast as we can every couple months. It's our second one, you'll see two and three this year. I'm John Furrier, my co-host, Dave Vellante. A panel here to break down the Supercloud momentum, the wave, and the developer impact that we bringing back Vittorio Viarengo, who's a VP for Cross-Cloud Services at VMware. Sarbjeet Johal, industry influencer and Analyst at StackPayne, his company, Cube alumni and Influencer. Sarbjeet, great to see you. Vittorio, thanks for coming back. >> Nice to be here. >> My pleasure. >> Vittorio, you just gave a keynote where we unpacked the cross-cloud services, what VMware is doing, how you guys see it, not just from VMware's perspective, but VMware looking out broadly at the industry and developers came up and you were like, "Developers, developer, developers", kind of a goof on the Steve Ballmer famous meme that everyone's seen. This is a huge star, sorry, I mean a big piece of it. The developers are the canary in the coal mines. They're the ones who are being asked to code the digital transformation, which is fully business transformation and with the market the way it is right now in terms of the accelerated technology, every enterprise grade business model's changing. The technology is evolving, the builders are kind of, they want go faster. I'm saying they're stuck in a way, but that's my opinion, but there's a lot of growth. >> Yeah. >> The impact, they got to get released up and let it go. Those developers need to accelerate faster. It's been a big part of productivity, and the conversations we've had. So developer impact is huge in Supercloud. What's your, what do you guys think about this? We'll start with you, Sarbjeet. >> Yeah, actually, developers are the masons of the digital empires I call 'em, right? They lay every brick and build all these big empires. On the left side of the SDLC, or the, you know, when you look at the system operations, developer is number one cost from economic side of things, and from technology side of things, they are tech hungry people. They are developers for that reason because developer nights are long, hours are long, they forget about when to eat, you know, like, I've been a developer, I still code. So you want to keep them happy, you want to hug your developers. We always say that, right? Vittorio said that right earlier. The key is to, in this context, in the Supercloud context, is that developers don't mind mucking around with platforms or APIs or new languages, but they hate the infrastructure part. That's a fact. They don't want to muck around with servers. It's friction for them, it is like they don't want to muck around even with the VMs. So they want the programmability to the nth degree. They want to automate everything, so that's how they think and cloud is the programmable infrastructure, industrialization of infrastructure in many ways. So they are happy with where we are going, and we need more abstraction layers for some developers. By the way, I have this sort of thinking frame for last year or so, not all developers are same, right? So if you are a developer at an ISV, you behave differently. If you are a developer at a typical enterprise, you behave differently or you are forced to behave differently because you're not writing software.- >> Well, developers, developers have changed, I mean, Vittorio, you and I were talking earlier on the keynote, and this is kind of the key point is what is a developer these days? If everything is software enabled, I mean, even hardware interviews we do with Nvidia, and Amazon and other people building silicon, they all say the same thing, "It's software on a chip." So you're seeing the role of software up and down the stack and the role of the stack is changing. The old days of full stack developer, what does that even mean? I mean, the cloud is a half a stack kind of right there. So, you know, developers are certainly more agile, but cloud native, I mean VMware is epitome of operations, IT operations, and the Tan Zoo initiative, you guys started, you went after the developers to look at them, and ask them questions, "What do you need?", "How do you transform the Ops from virtualization?" Again, back to your point, so this hardware abstraction, what is software, what is cloud native? It's kind of messy equation these days. How do you guys grokel with that? >> I would argue that developers don't want the Supercloud. I dropped that up there, so, >> Dave: Why not? >> Because developers, they, once they get comfortable in AWS or Google, because they're doing some AI stuff, which is, you know, very trendy right now, or they are in IBM, any of the IPA scaler, professional developers, system developers, they love that stuff, right? Yeah, they don't, the infrastructure gets in the way, but they're just, the problem is, and I think the Supercloud should be driven by the operators because as we discussed, the operators have been left behind because they're busy with day-to-day jobs, and in most cases IT is centralized, developers are in the business units. >> John: Yeah. >> Right? So they get the mandate from the top, say, "Our bank, they're competing against". They gave teenagers or like young people the ability to do all these new things online, and Venmo and all this integration, where are we? "Oh yeah, we can do it", and then build it, and then deploy it, "Okay, we caught up." but now the operators are back in the private cloud trying to keep the backend system running and so I think the Supercloud is needed for the primarily, initially, for the operators to get in front of the developers, fit in the workflow, but lay the foundation so it is secure.- >> So, so I love this thinking because I love the rift, because the rift points to what is the target audience for the value proposition and if you're a developer, Supercloud enables you so you shouldn't have to deal with Supercloud. >> Exactly. >> What you're saying is get the operating environment or operating system done properly, whether it's architecture, building the platform, this comes back to architecture platform conversations. What is the future platform? Is it a vendor supplied or is it customer created platform? >> Dave: So developers want best to breed, is what you just said. >> Vittorio: Yeah. >> Right and operators, they, 'cause developers don't want to deal with governance, they don't want to deal with security, >> No. >> They don't want to deal with spinning up infrastructure. That's the role of the operator, but that's where Supercloud enables, to John's point, the developer, so to your question, is it a platform where the platform vendor is responsible for the architecture, or there is it an architectural standard that spans multiple clouds that has to emerge? Based on what you just presented earlier, Vittorio, you are the determinant of the architecture. It's got to be open, but you guys determine that, whereas the nirvana is, "Oh no, it's all open, and it just kind of works." >> Yeah, so first of all, let's all level set on one thing. You cannot tell developers what to do. >> Dave: Right, great >> At least great developers, right? Cannot tell them what to do. >> Dave: So that's what, that's the way I want to sort of, >> You can tell 'em what's possible. >> There's a bottle on that >> If you tell 'em what's possible, they'll test it, they'll look at it, but if you try to jam it down their throat, >> Yeah. >> Dave: You can't tell 'em how to do it, just like your point >> Let me answer your answer the question. >> Yeah, yeah. >> So I think we need to build an architect, help them build an architecture, but it cannot be proprietary, has to be built on what works in the cloud and so what works in the cloud today is Kubernetes, is you know, number of different open source project that you need to enable and then provide, use this, but when I first got exposed to Kubernetes, I said, "Hallelujah!" We had a runtime that works the same everywhere only to realize there are 12 different distributions. So that's where we come in, right? And other vendors come in to say, "Hey, no, we can make them all look the same. So you still use Kubernetes, but we give you a place to build, to set those operation policy once so that you don't create friction for the developers because that's the last thing you want to do." >> Yeah, actually, coming back to the same point, not all developers are same, right? So if you're ISV developer, you want to go to the lowest sort of level of the infrastructure and you want to shave off the milliseconds from to get that performance, right? If you're working at AWS, you are doing that. If you're working at scale at Facebook, you're doing that. At Twitter, you're doing that, but when you go to DMV and Kansas City, you're not doing that, right? So your developers are different in nature. They are given certain parameters to work with, certain sort of constraints on the budget side. They are educated at a different level as well. Like they don't go to that end of the degree of sort of automation, if you will. So you cannot have the broad stroking of developers. We are talking about a citizen developer these days. That's a extreme low, >> You mean Low-Code. >> Yeah, Low-Code, No-code, yeah, on the extreme side. On one side, that's citizen developers. On the left side is the professional developers, when you say developers, your mind goes to the professional developers, like the hardcore developers, they love the flexibility, you know, >> John: Well app, developers too, I mean. >> App developers, yeah. >> You're right a lot of, >> Sarbjeet: Infrastructure platform developers, app developers, yes. >> But there are a lot of customers, its a spectrum, you're saying. >> Yes, it's a spectrum >> There's a lot of customers don't want deal with that muck. >> Yeah. >> You know, like you said, AWS, Twitter, the sophisticated developers do, but there's a whole suite of developers out there >> Yeah >> That just want tools that are abstracted. >> Within a company, within a company. Like how I see the Supercloud is there shouldn't be anything which blocks the developers, like their view of the world, of the future. Like if you're blocked as a developer, like something comes in front of you, you are not developer anymore, believe me, (John laughing) so you'll go somewhere else >> John: First of all, I'm, >> You'll leave the company by the way. >> Dave: Yeah, you got to quit >> Yeah, you will quit, you will go where the action is, where there's no sort of blockage there. So like if you put in front of them like a huge amount of a distraction, they don't like it, so they don't, >> Well, the idea of a developer, >> Coming back to that >> Let's get into 'cause you mentioned platform. Get year in the term platform engineering now. >> Yeah. >> Platform developer. You know, I remember back in, and I think there's still a term used today, but when I graduated my computer science degree, we were called "Software engineers," right? Do people use that term "Software engineering", or is it "Software development", or they the same, are they different? >> Well, >> I think there's a, >> So, who's engineering what? Are they engineering or are they developing? Or both? Well, I think it the, you made a great point. There is a factor of, I had the, I was blessed to work with Adam Bosworth, that is the guy that created some of the abstraction layer, like Visual Basic and Microsoft Access and he had so, he made his whole career thinking about this layer, and he always talk about the professional developers, the developers that, you know, give him a user manual, maybe just go at the APIs, he'll build anything, right, from system engine, go down there, and then through obstruction, you get the more the procedural logic type of engineers, the people that used to be able to write procedural logic and visual basic and so on and so forth. I think those developers right now are a little cut out of the picture. There's some No-code, Low-Code environment that are maybe gain some traction, I caught up with Adam Bosworth two weeks ago in New York and I asked him "What's happening to this higher level developers?" and you know what he is told me, and he is always a little bit out there, so I'm going to use his thought process here. He says, "ChapGPT", I mean, they will get to a point where this high level procedural logic will be written by, >> John: Computers. >> Computers, and so we may not need as many at the high level, but we still need the engineers down there. The point is the operation needs to get in front of them >> But, wait, wait, you seen the ChatGPT meme, I dunno if it's a Dilbert thing where it's like, "Time to tic" >> Yeah, yeah, yeah, I did that >> "Time to develop the code >> Five minutes, time to decode", you know, to debug the codes like five hours. So you know, the whole equation >> Well, this ChatGPT is a hot wave, everyone's been talking about it because I think it illustrates something that's NextGen, feels NextGen, and it's just getting started so it's going to get better. I mean people are throwing stones at it, but I think it's amazing. It's the equivalent of me seeing the browser for the first time, you know, like, "Wow, this is really compelling." This is game-changing, it's not just keyword chat bots. It's like this is real, this is next level, and I think the Supercloud wave that people are getting behind points to that and I think the question of Ops and Dev comes up because I think if you limit the infrastructure opportunity for a developer, I think they're going to be handicapped. I mean that's a general, my opinion, the thesis is you give more aperture to developers, more choice, more capabilities, more good things could happen, policy, and that's why you're seeing the convergence of networking people, virtualization talent, operational talent, get into the conversation because I think it's an infrastructure engineering opportunity. I think this is a seminal moment in a new stack that's emerging from an infrastructure, software virtualization, low-code, no-code layer that will be completely programmable by things like the next Chat GPT or something different, but yet still the mechanics and the plumbing will still need engineering. >> Sarbjeet: Oh yeah. >> So there's still going to be more stuff coming on. >> Yeah, we have, with the cloud, we have made the infrastructure programmable and you give the programmability to the programmer, they will be very creative with that and so we are being very creative with our infrastructure now and on top of that, we are being very creative with the silicone now, right? So we talk about that. That's part of it, by the way. So you write the code to the particle's silicone now, and on the flip side, the silicone is built for certain use cases for AI Inference and all that. >> You saw this at CES? >> Yeah, I saw at CES, the scenario is this, the Bosch, I spoke to Bosch, I spoke to John Deere, I spoke to AWS guys, >> Yeah. >> They were showcasing their technology there and I was spoke to Azure guys as well. So the Bosch is a good example. So they are building, they are right now using AWS. I have that interview on camera, I will put it some sometime later on there online. So they're using AWS on the back end now, but Bosch is the number one, number one or number two depending on what day it is of the year, supplier of the componentry to the auto industry, and they are creating a platform for our auto industry, so is Qualcomm actually by the way, with the Snapdragon. So they told me that customers, their customers, BMW, Audi, all the manufacturers, they demand the diversity of the backend. Like they don't want all, they, all of them don't want to go to AWS. So they want the choice on the backend. So whatever they cook in the middle has to work, they have to sprinkle the data for the data sovereign side because they have Chinese car makers as well, and for, you know, for other reasons, competitive reasons and like use. >> People don't go to, aw, people don't go to AWS either for political reasons or like competitive reasons or specific use cases, but for the most part, generally, I haven't met anyone who hasn't gone first choice with either, but that's me personally. >> No, but they're building. >> Point is the developer wants choice at the back end is what I'm hearing, but then finish that thought. >> Their developers want the choice, they want the choice on the back end, number one, because the customers are asking for, in this case, the customers are asking for it, right? But the customers requirements actually drive, their economics drives that decision making, right? So in the middle they have to, they're forced to cook up some solution which is vendor neutral on the backend or multicloud in nature. So >> Yeah, >> Every >> I mean I think that's nirvana. I don't think, I personally don't see that happening right now. I mean, I don't see the parody with clouds. So I think that's a challenge. I mean, >> Yeah, true. >> I mean the fact of the matter is if the development teams get fragmented, we had this chat with Kit Colbert last time, I think he's going to come on and I think he's going to talk about his keynote in a few, in an hour or so, development teams is this, the cloud is heterogenous, which is great. It's complex, which is challenging. You need skilled engineering to manage these clouds. So if you're a CIO and you go all in on AWS, it's hard. Then to then go out and say, "I want to be completely multi-vendor neutral" that's a tall order on many levels and this is the multicloud challenge, right? So, the question is, what's the strategy for me, the CIO or CISO, what do I do? I mean, to me, I would go all in on one and start getting hedges and start playing and then look at some >> Crystal clear. Crystal clear to me. >> Go ahead. >> If you're a CIO today, you have to build a platform engineering team, no question. 'Cause if we agree that we cannot tell the great developers what to do, we have to create a platform engineering team that using pieces of the Supercloud can build, and let's make this very pragmatic and give examples. First you need to be able to lay down the run time, okay? So you need a way to deploy multiple different Kubernetes environment in depending on the cloud. Okay, now we got that. The second part >> That's like table stakes. >> That are table stake, right? But now what is the advantage of having a Supercloud service to do that is that now you can put a policy in one place and it gets distributed everywhere consistently. So for example, you want to say, "If anybody in this organization across all these different buildings, all these developers don't even know, build a PCI compliant microservice, They can only talk to PCI compliant microservice." Now, I sleep tight. The developers still do that. Of course they're going to get their hands slapped if they don't encrypt some messages and say, "Oh, that should have been encrypted." So number one. The second thing I want to be able to say, "This service that this developer built over there better satisfy this SLA." So if the SLA is not satisfied, boom, I automatically spin up multiple instances to certify the SLA. Developers unencumbered, they don't even know. So this for me is like, CIO build a platform engineering team using one of the many Supercloud services that allow you to do that and lay down. >> And part of that is that the vendor behavior is such, 'cause the incentive is that they don't necessarily always work together. (John chuckling) I'll give you an example, we're going to hear today from Western Union. They're AWS shop, but they want to go to Google, they want to use some of Google's AI tools 'cause they're good and maybe they're even arguably better, but they're also a Snowflake customer and what you'll hear from them is Amazon and Snowflake are working together so that SageMaker can be integrated with Snowflake but Google said, "No, you want to use our AI tools, you got to use BigQuery." >> Yeah. >> Okay. So they say, "Ah, forget it." So if you have a platform engineering team, you can maybe solve some of that vendor friction and get competitive advantage. >> I think that the future proximity concept that I talk about is like, when you're doing one thing, you want to do another thing. Where do you go to get that thing, right? So that is very important. Like your question, John, is that your point is that AWS is ahead of the pack, which is true, right? They have the >> breadth of >> Infrastructure by a lot >> infrastructure service, right? They breadth of services, right? So, how do you, When do you bring in other cloud providers, right? So I believe that you should standardize on one cloud provider, like that's your primary, and for others, bring them in on as needed basis, in the subsection or sub portfolio of your applications or your platforms, what ever you can. >> So yeah, the Google AI example >> Yeah, I mean, >> Or the Microsoft collaboration software example. I mean there's always or the M and A. >> Yeah, but- >> You're going to get to run Windows, you can run Windows on Amazon, so. >> By the way, Supercloud doesn't mean that you cannot do that. So the perfect example is say that you're using Azure because you have a SQL server intensive workload. >> Yep >> And you're using Google for ML, great. If you are using some differentiated feature of this cloud, you'll have to go somewhere and configure this widget, but what you can abstract with the Supercloud is the lifecycle manage of the service that runs on top, right? So how does the service get deployed, right? How do you monitor performance? How do you lifecycle it? How you secure it that you can abstract and that's the value and eventually value will win. So the customers will find what is the values, obstructing in making it uniform or going deeper? >> How about identity? Like take identity for instance, you know, that's an opportunity to abstract. Whether I use Microsoft Identity or Okta, and I can abstract that. >> Yeah, and then we have APIs and standards that we can use so eventually I think where there is enough pain, the right open source will emerge to solve that problem. >> Dave: Yeah, I can use abstract things like object store, right? That's pretty simple. >> But back to the engineering question though, is that developers, developers, developers, one thing about developers psychology is if something's not right, they say, "Go get fixing. I'm not touching it until you fix it." They're very sticky about, if something's not working, they're not going to do it again, right? So you got to get it right for developers. I mean, they'll maybe tolerate something new, but is the "juice worth the squeeze" as they say, right? So you can't go to direct say, "Hey, it's, what's a work in progress? We're going to get our infrastructure together and the world's going to be great for you, but just hang tight." They're going to be like, "Get your shit together then talk to me." So I think that to me is the question. It's an Ops question, but where's that value for the developer in Supercloud where the capabilities are there, there's less friction, it's simpler, it solves the complexity problem. I don't need these high skilled labor to manage Amazon. I got services exposed. >> That's what we talked about earlier. It's like the Walmart example. They basically, they took away from the developer the need to spin up infrastructure and worry about all the governance. I mean, it's not completely there yet. So the developer could focus on what he or she wanted to do. >> But there's a big, like in our industry, there's a big sort of flaw or the contention between developers and operators. Developers want to be on the cutting edge, right? And operators want to be on the stability, you know, like we want governance. >> Yeah, totally. >> Right, so they want to control, developers are like these little bratty kids, right? And they want Legos, like they want toys, right? Some of them want toys by way. They want Legos, they want to build there and they want make a mess out of it. So you got to make sure. My number one advice in this context is that do it up your application portfolio and, or your platform portfolio if you are an ISV, right? So if you are ISV you most probably, you're building a platform these days, do it up in a way that you can say this portion of our applications and our platform will adhere to what you are saying, standardization, you know, like Kubernetes, like slam dunk, you know, it works across clouds and in your data center hybrid, you know, whole nine yards, but there is some subset on the next door systems of innovation. Everybody has, it doesn't matter if you're DMV of Kansas or you are, you know, metaverse, right? Or Meta company, right, which is Facebook, they have it, they are building something new. For that, give them some freedom to choose different things like play with non-standard things. So that is the mantra for moving forward, for any enterprise. >> Do you think developers are happy with the infrastructure now or are they wanting people to get their act together? I mean, what's your reaction, or you think. >> Developers are happy as long as they can do their stuff, which is running code. They want to write code and innovate. So to me, when Ballmer said, "Developer, develop, Developer, what he meant was, all you other people get your act together so these developers can do their thing, and to me the Supercloud is the way for IT to get there and let developer be creative and go fast. Why not, without getting in trouble. >> Okay, let's wrap up this segment with a super clip. Okay, we're going to do a sound bite that we're going to make into a short video for each of you >> All right >> On you guys summarizing why Supercloud's important, why this next wave is relevant for the practitioners, for the industry and we'll turn this into an Instagram reel, YouTube short. So we'll call it a "Super clip. >> Alright, >> Sarbjeet, you want, you want some time to think about it? You want to go first? Vittorio, you want. >> I just didn't mind. (all laughing) >> No, okay, okay. >> I'll do it again. >> Go back. No, we got a fresh one. We'll going to already got that one in the can. >> I'll go. >> Sarbjeet, you go first. >> I'll go >> What's your super clip? >> In software systems, abstraction is your friend. I always say that. Abstraction is your friend, even if you're super professional developer, abstraction is your friend. We saw from the MFC library from C++ days till today. Abstract, use abstraction. Do not try to reinvent what's already being invented. Leverage cloud, leverage the platform side of the cloud. Not just infrastructure service, but platform as a service side of the cloud as well, and Supercloud is a meta platform built on top of these infrastructure services from three or four or five cloud providers. So use that and embrace the programmability, embrace the abstraction layer. That's the key actually, and developers who are true developers or professional developers as you said, they know that. >> Awesome. Great super clip. Vittorio, another shot at the plate here for super clip. Go. >> Multicloud is awesome. There's a reason why multicloud happened, is because gave our developers the ability to innovate fast and ever before. So if you are embarking on a digital transformation journey, which I call a survival journey, if you're not innovating and transforming, you're not going to be around in business three, five years from now. You have to adopt the Supercloud so the developer can be developer and keep building great, innovating digital experiences for your customers and IT can get in front of it and not get in trouble together. >> Building those super apps with Supercloud. That was a great super clip. Vittorio, thank you for sharing. >> Thanks guys. >> Sarbjeet, thanks for coming on talking about the developer impact Supercloud 2. On our next segment, coming up right now, we're going to hear from Walmart enterprise architect, how they are building and they are continuing to innovate, to build their own Supercloud. Really informative, instructive from a practitioner doing it in real time. Be right back with Walmart here in Palo Alto. Thanks for watching. (gentle music)
SUMMARY :
the Supercloud momentum, and developers came up and you were like, and the conversations we've had. and cloud is the and the role of the stack is changing. I dropped that up there, so, developers are in the business units. the ability to do all because the rift points to What is the future platform? is what you just said. the developer, so to your question, You cannot tell developers what to do. Cannot tell them what to do. You can tell 'em your answer the question. but we give you a place to build, and you want to shave off the milliseconds they love the flexibility, you know, platform developers, you're saying. don't want deal with that muck. that are abstracted. Like how I see the Supercloud is So like if you put in front of them you mentioned platform. and I think there's the developers that, you The point is the operation to decode", you know, the browser for the first time, you know, going to be more stuff coming on. and on the flip side, the middle has to work, but for the most part, generally, Point is the developer So in the middle they have to, the parody with clouds. I mean the fact of the matter Crystal clear to me. in depending on the cloud. So if the SLA is not satisfied, boom, 'cause the incentive is that So if you have a platform AWS is ahead of the pack, So I believe that you should standardize or the M and A. you can run Windows on Amazon, so. So the perfect example is abstract and that's the value Like take identity for instance, you know, the right open source will Dave: Yeah, I can use abstract things and the world's going to be great for you, the need to spin up infrastructure on the stability, you know, So that is the mantra for moving forward, Do you think developers are happy and to me the Supercloud is for each of you for the industry you want some time to think about it? I just didn't mind. got that one in the can. platform side of the cloud. Vittorio, another shot at the the ability to innovate thank you for sharing. the developer impact Supercloud 2.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
BMW | ORGANIZATION | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Sarbjeet | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Bosch | ORGANIZATION | 0.99+ |
Vittorio | PERSON | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Audi | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Steve Ballmer | PERSON | 0.99+ |
Qualcomm | ORGANIZATION | 0.99+ |
Adam Bosworth | PERSON | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
New York | LOCATION | 0.99+ |
Vittorio Viarengo | PERSON | 0.99+ |
Kit Colbert | PERSON | 0.99+ |
Ballmer | PERSON | 0.99+ |
four | QUANTITY | 0.99+ |
Sarbjeet Johal | PERSON | 0.99+ |
five hours | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Palo Alto, California | LOCATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Five minutes | QUANTITY | 0.99+ |
NextGen | ORGANIZATION | 0.99+ |
StackPayne | ORGANIZATION | 0.99+ |
Visual Basic | TITLE | 0.99+ |
second part | QUANTITY | 0.99+ |
12 different distributions | QUANTITY | 0.99+ |
CES | EVENT | 0.99+ |
First | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Kansas City | LOCATION | 0.99+ |
second one | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Kansas | LOCATION | 0.98+ |
first time | QUANTITY | 0.98+ |
Windows | TITLE | 0.98+ |
last year | DATE | 0.98+ |
Breaking Analysis: Google's Point of View on Confidential Computing
>> From theCUBE studios in Palo Alto in Boston, bringing you data-driven insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante. >> Confidential computing is a technology that aims to enhance data privacy and security by providing encrypted computation on sensitive data and isolating data from apps in a fenced off enclave during processing. The concept of confidential computing is gaining popularity, especially in the cloud computing space where sensitive data is often stored and of course processed. However, there are some who view confidential computing as an unnecessary technology in a marketing ploy by cloud providers aimed at calming customers who are cloud phobic. Hello and welcome to this week's Wikibon CUBE Insights powered by ETR. In this Breaking Analysis, we revisit the notion of confidential computing, and to do so, we'll invite two Google experts to the show, but before we get there, let's summarize briefly. There's not a ton of ETR data on the topic of confidential computing. I mean, it's a technology that's deeply embedded into silicon and computing architectures. But at the highest level, security remains the number one priority being addressed by IT decision makers in the coming year as shown here. And this data is pretty much across the board by industry, by region, by size of company. I mean we dug into it and the only slight deviation from the mean is in financial services. The second and third most cited priorities, cloud migration and analytics, are noticeably closer to cybersecurity in financial services than in other sectors, likely because financial services has always been hyper security conscious, but security is still a clear number one priority in that sector. The idea behind confidential computing is to better address threat models for data in execution. Protecting data at rest and data and transit have long been a focus of security approaches, but more recently, silicon manufacturers have introduced architectures that separate data and applications from the host system. Arm, Intel, AMD, Nvidia and other suppliers are all on board, as are the big cloud players. Now the argument against confidential computing is that it narrowly focuses on memory encryption and it doesn't solve the biggest problems in security. Multiple system images updates different services and the entire code flow aren't directly addressed by memory encryption, rather to truly attack these problems, many believe that OSs need to be re-engineered with the attacker and hacker in mind. There are so many variables and at the end of the day, critics say the emphasis on confidential computing made by cloud providers is overstated and largely hype. This tweet from security researcher Rodrigo Branco sums up the sentiment of many skeptics. He says, "Confidential computing is mostly a marketing campaign for memory encryption. It's not driving the industry towards the hard open problems. It is selling an illusion." Okay. Nonetheless, encrypting data in use and fencing off key components of the system isn't a bad thing, especially if it comes with the package essentially for free. There has been a lack of standardization and interoperability between different confidential computing approaches. But the confidential computing consortium was established in 2019 ostensibly to accelerate the market and influence standards. Notably, AWS is not part of the consortium, likely because the politics of the consortium were probably a conundrum for AWS because the base technology defined by the the consortium is seen as limiting by AWS. This is my guess, not AWS's words, and but I think joining the consortium would validate a definition which AWS isn't aligned with. And two, it's got a lead with this Annapurna acquisition. This was way ahead with Arm integration and so it probably doesn't feel the need to validate its competitors. Anyway, one of the premier members of the confidential computing consortium is Google, along with many high profile names including Arm, Intel, Meta, Red Hat, Microsoft, and others. And we're pleased to welcome two experts on confidential computing from Google to unpack the topic, Nelly Porter is head of product for GCP confidential computing and encryption, and Dr. Patricia Florissi is the technical director for the office of the CTO at Google Cloud. Welcome Nelly and Patricia, great to have you. >> Great to be here. >> Thank you so much for having us. >> You're very welcome. Nelly, why don't you start and then Patricia, you can weigh in. Just tell the audience a little bit about each of your roles at Google Cloud. >> So I'll start, I'm owning a lot of interesting activities in Google and again security or infrastructure securities that I usually own. And we are talking about encryption and when encryption and confidential computing is a part of portfolio in additional areas that I contribute together with my team to Google and our customers is secure software supply chain. Because you need to trust your software. Is it operate in your confidential environment to have end-to-end story about if you believe that your software and your environment doing what you expect, it's my role. >> Got it. Okay. Patricia? >> Well, I am a technical director in the office of the CTO, OCTO for short, in Google Cloud. And we are a global team. We include former CTOs like myself and senior technologists from large corporations, institutions and a lot of success, we're startups as well. And we have two main goals. First, we walk side by side with some of our largest, more strategic or most strategical customers and we help them solve complex engineering technical problems. And second, we are devise Google and Google Cloud engineering and product management and tech on there, on emerging trends and technologies to guide the trajectory of our business. We are unique group, I think, because we have created this collaborative culture with our customers. And within OCTO, I spend a lot of time collaborating with customers and the industry at large on technologies that can address privacy, security, and sovereignty of data in general. >> Excellent. Thank you for that both of you. Let's get into it. So Nelly, what is confidential computing? From Google's perspective, how do you define it? >> Confidential computing is a tool and it's still one of the tools in our toolbox. And confidential computing is a way how we would help our customers to complete this very interesting end-to-end lifecycle of the data. And when customers bring in the data to cloud and want to protect it as they ingest it to the cloud, they protect it at rest when they store data in the cloud. But what was missing for many, many years is ability for us to continue protecting data and workloads of our customers when they running them. And again, because data is not brought to cloud to have huge graveyard, we need to ensure that this data is actually indexed. Again, there is some insights driven and drawn from this data. You have to process this data and confidential computing here to help. Now we have end to end protection of our customer's data when they bring the workloads and data to cloud, thanks to confidential computing. >> Thank you for that. Okay, we're going to get into the architecture a bit, but before we do, Patricia, why do you think this topic of confidential computing is such an important technology? Can you explain, do you think it's transformative for customers and if so, why? >> Yeah, I would maybe like to use one thought, one way, one intuition behind why confidential commuting matters, because at the end of the day, it reduces more and more the customer's thresh boundaries and the attack surface. That's about reducing that periphery, the boundary in which the customer needs to mind about trust and safety. And in a way, is a natural progression that you're using encryption to secure and protect the data. In the same way that we are encrypting data in transit and at rest, now we are also encrypting data while in use. And among other beneficials, I would say one of the most transformative ones is that organizations will be able to collaborate with each other and retain the confidentiality of the data. And that is across industry, even though it's highly focused on, I wouldn't say highly focused, but very beneficial for highly regulated industries. It applies to all of industries. And if you look at financing for example, where bankers are trying to detect fraud, and specifically double finance where you are, a customer is actually trying to get a finance on an asset, let's say a boat or a house, and then it goes to another bank and gets another finance on that asset. Now bankers would be able to collaborate and detect fraud while preserving confidentiality and privacy of the data. >> Interesting. And I want to understand that a little bit more but I'm going to push you a little bit on this, Nelly, if I can because there's a narrative out there that says confidential computing is a marketing ploy, I talked about this upfront, by cloud providers that are just trying to placate people that are scared of the cloud. And I'm presuming you don't agree with that, but I'd like you to weigh in here. The argument is confidential computing is just memory encryption and it doesn't address many other problems. It is over hyped by cloud providers. What do you say to that line of thinking? >> I absolutely disagree, as you can imagine, with this statement, but the most importantly is we mixing multiple concepts, I guess. And exactly as Patricia said, we need to look at the end-to-end story, not again the mechanism how confidential computing trying to again, execute and protect a customer's data and why it's so critically important because what confidential computing was able to do, it's in addition to isolate our tenants in multi-tenant environments the cloud covering to offer additional stronger isolation. They called it cryptographic isolation. It's why customers will have more trust to customers and to other customers, the tenant that's running on the same host but also us because they don't need to worry about against threats and more malicious attempts to penetrate the environment. So what confidential computing is helping us to offer our customers, stronger isolation between tenants in this multi-tenant environment, but also incredibly important, stronger isolation of our customers, so tenants from us. We also writing code, we also software providers will also make mistakes or have some zero days. Sometimes again us introduced, sometimes introduced by our adversaries. But what I'm trying to say by creating this cryptographic layer of isolation between us and our tenants and amongst those tenants, we're really providing meaningful security to our customers and eliminate some of the worries that they have running on multi-tenant spaces or even collaborating to gather this very sensitive data knowing that this particular protection is available to them. >> Okay, thank you. Appreciate that. And I think malicious code is often a threat model missed in these narratives. Operator access, yeah, maybe I trust my clouds provider, but if I can fence off your access even better, I'll sleep better at night. Separating a code from the data, everybody's, Arm, Intel, AMD, Nvidia, others, they're all doing it. I wonder if, Nelly, if we could stay with you and bring up the slide on the architecture. What's architecturally different with confidential computing versus how operating systems and VMs have worked traditionally. We're showing a slide here with some VMs, maybe you could take us through that. >> Absolutely. And Dave, the whole idea for Google and now industry way of dealing with confidential computing is to ensure that three main property is actually preserved. Customers don't need to change the code. They can operate on those VMs exactly as they would with normal non-confidential VMs, but to give them this opportunity of lift and shift or no changing their apps and performing and having very, very, very low latency and scale as any cloud can, something that Google actually pioneer in confidential computing. I think we need to open and explain how this magic was actually done. And as I said, it's again the whole entire system have to change to be able to provide this magic. And I would start with we have this concept of root of trust and root of trust where we will ensure that this machine, when the whole entire post has integrity guarantee, means nobody changing my code on the most low level of system. And we introduce this in 2017 called Titan. It was our specific ASIC, specific, again, inch by inch system on every single motherboard that we have that ensures that your low level former, your actually system code, your kernel, the most powerful system is actually proper configured and not changed, not tampered. We do it for everybody, confidential computing included. But for confidential computing, what we have to change, we bring in AMD, or again, future silicon vendors and we have to trust their former, their way to deal with our confidential environments. And that's why we have obligation to validate integrity, not only our software and our former but also former and software of our vendors, silicon vendors. So we actually, when we booting this machine, as you can see, we validate that integrity of all of the system is in place. It means nobody touching, nobody changing, nobody modifying it. But then we have this concept of AMD secure processor, it's special ASICs, best specific things that generate a key for every single VM that our customers will run or every single node in Kubernetes or every single worker thread in our Hadoop or Spark capability. We offer all of that. And those keys are not available to us. It's the best keys ever in encryption space because when we are talking about encryption, the first question that I'm receiving all the time, where's the key, who will have access to the key? Because if you have access to the key then it doesn't matter if you encrypted or not. So, but the case in confidential computing provides so revolutionary technology, us cloud providers, who don't have access to the keys. They sitting in the hardware and they head to memory controller. And it means when hypervisors that also know about these wonderful things saying I need to get access to the memories that this particular VM trying to get access to, they do not decrypt the data, they don't have access to the key because those keys are random, ephemeral and per VM, but the most importantly, in hardware not exportable. And it means now you would be able to have this very interesting role that customers or cloud providers will not be able to get access to your memory. And what we do, again, as you can see our customers don't need to change their applications, their VMs are running exactly as it should run and what you're running in VM, you actually see your memory in clear, it's not encrypted, but God forbid is trying somebody to do it outside of my confidential box. No, no, no, no, no, they would not be able to do it. Now you'll see cyber and it's exactly what combination of these multiple hardware pieces and software pieces have to do. So OS is also modified. And OS is modified such way to provide integrity. It means even OS that you're running in your VM box is not modifiable and you, as customer, can verify. But the most interesting thing, I guess, how to ensure the super performance of this environment because you can imagine, Dave, that encrypting and it's additional performance, additional time, additional latency. So we were able to mitigate all of that by providing incredibly interesting capability in the OS itself. So our customers will get no changes needed, fantastic performance and scales as they would expect from cloud providers like Google. >> Okay, thank you. Excellent. Appreciate that explanation. So, again, the narrative on this as well, you've already given me guarantees as a cloud provider that you don't have access to my data, but this gives another level of assurance, key management as they say is key. Now humans aren't managing the keys, the machines are managing them. So Patricia, my question to you is, in addition to, let's go pre confidential computing days, what are the sort of new guarantees that these hardware-based technologies are going to provide to customers? >> So if I am a customer, I am saying I now have full guarantee of confidentiality and integrity of the data and of the code. So if you look at code and data confidentiality, the customer cares and they want to know whether their systems are protected from outside or unauthorized access, and that recovered with Nelly, that it is. Confidential computing actually ensures that the applications and data internals remain secret, right? The code is actually looking at the data, the only the memory is decrypting the data with a key that is ephemeral and per VM and generated on demand. Then you have the second point where you have code and data integrity, and now customers want to know whether their data was corrupted, tampered with or impacted by outside actors. And what confidential computing ensures is that application internals are not tampered with. So the application, the workload as we call it, that is processing the data, it's also, it has not been tampered and preserves integrity. I would also say that this is all verifiable. So you have attestation and these attestation actually generates a log trail and the log trail guarantees that, provides a proof that it was preserved. And I think that the offer's also a guarantee of what we call ceiling, this idea that the secrets have been preserved and not tampered with, confidentiality and integrity of code and data. >> Got it. Okay, thank you. Nelly, you mentioned, I think I heard you say that the applications, it's transparent, you don't have to change the application, it just comes for free essentially. And we showed some various parts of the stack before. I'm curious as to what's affected, but really more importantly, what is specifically Google's value add? How do partners participate in this, the ecosystem, or maybe said another way, how does Google ensure the compatibility of confidential computing with existing systems and applications? >> And a fantastic question by the way. And it's very difficult and definitely complicated world because to be able to provide these guarantees, actually a lot of work was done by community. Google is very much operate in open, so again, our operating system, we working with operating system repository OSs, OS vendors to ensure that all capabilities that we need is part of the kernels, are part of the releases and it's available for customers to understand and even explore if they have fun to explore a lot of code. We have also modified together with our silicon vendors a kernel, host kernel to support this capability and it means working this community to ensure that all of those patches are there. We also worked with every single silicon vendor as you've seen, and that's what I probably feel that Google contributed quite a bit in this whole, we moved our industry, our community, our vendors to understand the value of easy to use confidential computing or removing barriers. And now I don't know if you noticed, Intel is pulling the lead and also announcing their trusted domain extension, very similar architecture. And no surprise, it's, again, a lot of work done with our partners to, again, convince, work with them and make this capability available. The same with Arm this year, actually last year, Arm announced their future design for confidential computing. It's called Confidential Computing Architecture. And it's also influenced very heavily with similar ideas by Google and industry overall. So it's a lot of work in confidential computing consortiums that we are doing, for example, simply to mention, to ensure interop, as you mentioned, between different confidential environments of cloud providers. They want to ensure that they can attest to each other because when you're communicating with different environments, you need to trust them. And if it's running on different cloud providers, you need to ensure that you can trust your receiver when you are sharing your sensitive data workloads or secret with them. So we coming as a community and we have this attestation sig, the, again, the community based systems that we want to build and influence and work with Arm and every other cloud providers to ensure that we can interrupt and it means it doesn't matter where confidential workloads will be hosted, but they can exchange the data in secure, verifiable and controlled by customers way. And to do it, we need to continue what we are doing, working open, again, and contribute with our ideas and ideas of our partners to this role to become what we see confidential computing has to become, it has to become utility. It doesn't need to be so special, but it's what we want it to become. >> Let's talk about, thank you for that explanation. Let's talk about data sovereignty because when you think about data sharing, you think about data sharing across the ecosystem and different regions and then of course data sovereignty comes up. Typically public policy lags, the technology industry and sometimes is problematic. I know there's a lot of discussions about exceptions, but Patricia, we have a graphic on data sovereignty. I'm interested in how confidential computing ensures that data sovereignty and privacy edicts are adhered to, even if they're out of alignment maybe with the pace of technology. One of the frequent examples is when you delete data, can you actually prove that data is deleted with a hundred percent certainty? You got to prove that and a lot of other issues. So looking at this slide, maybe you could take us through your thinking on data sovereignty. >> Perfect. So for us, data sovereignty is only one of the three pillars of digital sovereignty. And I don't want to give the impression that confidential computing addresses it all. That's why we want to step back and say, hey, digital sovereignty includes data sovereignty where we are giving you full control and ownership of the location, encryption and access to your data. Operational sovereignty where the goal is to give our Google Cloud customers full visibility and control over the provider operations, right? So if there are any updates on hardware, software stack, any operations, there is full transparency, full visibility. And then the third pillar is around software sovereignty where the customer wants to ensure that they can run their workloads without dependency on the provider's software. So they have sometimes is often referred as survivability, that you can actually survive if you are untethered to the cloud and that you can use open source. Now let's take a deep dive on data sovereignty, which by the way is one of my favorite topics. And we typically focus on saying, hey, we need to care about data residency. We care where the data resides because where the data is at rest or in processing, it typically abides to the jurisdiction, the regulations of the jurisdiction where the data resides. And others say, hey, let's focus on data protection. We want to ensure the confidentiality and integrity and availability of the data, which confidential computing is at the heart of that data protection. But it is yet another element that people typically don't talk about when talking about data sovereignty, which is the element of user control. And here, Dave, is about what happens to the data when I give you access to my data. And this reminds me of security two decades ago, even a decade ago, where we started the security movement by putting firewall protections and login accesses. But once you were in, you were able to do everything you wanted with the data. An insider had access to all the infrastructure, the data and the code. And that's similar because with data sovereignty we care about whether it resides, where, who is operating on the data. But the moment that the data is being processed, I need to trust that the processing of the data will abide by user control, by the policies that I put in place of how my data is going to be used. And if you look at a lot of the regulation today and a lot of the initiatives around the International Data Space Association, IDSA, and Gaia-X, there is a movement of saying the two parties, the provider of the data and the receiver of the data are going to agree on a contract that describes what my data can be used for. The challenge is to ensure that once the data crosses boundaries, that the data will be used for the purposes that it was intended and specified in the contract. And if you actually bring together, and this is the exciting part, confidential computing together with policy enforcement, now the policy enforcement can guarantee that the data is only processed within the confines of a confidential computing environment, that the workload is cryptographically verified that there is the workload that was meant to process the data and that the data will be only used when abiding to the confidentiality and integrity safety of the confidential computing environment. And that's why we believe confidential computing is one necessary and essential technology that will allow us to ensure data sovereignty, especially when it comes to user control. >> Thank you for that. I mean it was a deep dive, I mean brief, but really detailed. So I appreciate that, especially the verification of the enforcement. Last question, I met you two because as part of my year end prediction post, you guys sent in some predictions and I wasn't able to get to them in the predictions post. So I'm thrilled that you were able to make the time to come on the program. How widespread do you think the adoption of confidential computing will be in 23 and what's the maturity curve look like, this decade in your opinion? Maybe each of you could give us a brief answer. >> So my prediction in five, seven years, as I started, it'll become utility. It'll become TLS as of, again, 10 years ago we couldn't believe that websites will have certificates and we will support encrypted traffic. Now we do and it's become ubiquity. It's exactly where confidential computing is getting and heading, I don't know we deserve yet. It'll take a few years of maturity for us, but we will be there. >> Thank you. And Patricia, what's your prediction? >> I will double that and say, hey, in the future, in the very near future, you will not be able to afford not having it. I believe as digital sovereignty becomes evermore top of mind with sovereign states and also for multi national organizations and for organizations that want to collaborate with each other, confidential computing will become the norm. It'll become the default, if I say, mode of operation. I like to compare that today is inconceivable. If we talk to the young technologists, it's inconceivable to think that at some point in history, and I happen to be alive that we had data at rest that was not encrypted, data in transit that was not encrypted, and I think that will be inconceivable at some point in the near future that to have unencrypted data while in use. >> And plus I think the beauty of the this industry is because there's so much competition, this essentially comes for free. I want to thank you both for spending some time on Breaking Analysis. There's so much more we could cover. I hope you'll come back to share the progress that you're making in this area and we can double click on some of these topics. Really appreciate your time. >> Anytime. >> Thank you so much. >> In summary, while confidential computing is being touted by the cloud players as a promising technology for enhancing data privacy and security, there are also those, as we said, who remain skeptical. The truth probably lies somewhere in between and it will depend on the specific implementation and the use case as to how effective confidential computing will be. Look, as with any new tech, it's important to carefully evaluate the potential benefits, the drawbacks, and make informed decisions based on the specific requirements in the situation and the constraints of each individual customer. But the bottom line is silicon manufacturers are working with cloud providers and other system companies to include confidential computing into their architectures. Competition, in our view, will moderate price hikes. And at the end of the day, this is under the covers technology that essentially will come for free. So we'll take it. I want to thank our guests today, Nelly and Patricia from Google, and thanks to Alex Myerson who's on production and manages the podcast. Ken Schiffman as well out of our Boston studio, Kristin Martin and Cheryl Knight help get the word out on social media and in our newsletters. And Rob Hof is our editor-in-chief over at siliconangle.com. Does some great editing for us, thank you all. Remember all these episodes are available as podcasts. Wherever you listen, just search Breaking Analysis podcast. I publish each week on wikibon.com and siliconangle.com where you can get all the news. If you want to get in touch, you can email me at david.vellante@siliconangle.com or dm me @DVellante. And you can also comment on my LinkedIn post. Definitely you want to check out etr.ai for the best survey data in the enterprise tech business. I know we didn't hit on a lot today, but there's some amazing data and it's always being updated, so check that out. This is Dave Vellante for theCUBE Insights, powered by ETR. Thanks for watching and we'll see you next time on Breaking Analysis. (upbeat music)
SUMMARY :
bringing you data-driven and at the end of the day, Just tell the audience a little and confidential computing Got it. and the industry at large for that both of you. in the data to cloud into the architecture a bit, and privacy of the data. people that are scared of the cloud. and eliminate some of the we could stay with you and they head to memory controller. So, again, the narrative on this as well, and integrity of the data and of the code. how does Google ensure the compatibility and ideas of our partners to this role One of the frequent examples and that the data will be only used of the enforcement. and we will support encrypted traffic. And Patricia, and I happen to be alive beauty of the this industry and the constraints of
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Nelly | PERSON | 0.99+ |
Patricia | PERSON | 0.99+ |
International Data Space Association | ORGANIZATION | 0.99+ |
Alex Myerson | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
IDSA | ORGANIZATION | 0.99+ |
Rodrigo Branco | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Nvidia | ORGANIZATION | 0.99+ |
2019 | DATE | 0.99+ |
2017 | DATE | 0.99+ |
Kristin Martin | PERSON | 0.99+ |
Nelly Porter | PERSON | 0.99+ |
Ken Schiffman | PERSON | 0.99+ |
Rob Hof | PERSON | 0.99+ |
Cheryl Knight | PERSON | 0.99+ |
last year | DATE | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
two parties | QUANTITY | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
Patricia Florissi | PERSON | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
second point | QUANTITY | 0.99+ |
david.vellante@siliconangle.com | OTHER | 0.99+ |
Meta | ORGANIZATION | 0.99+ |
second | QUANTITY | 0.99+ |
third | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Arm | ORGANIZATION | 0.99+ |
each | QUANTITY | 0.99+ |
two experts | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
first question | QUANTITY | 0.99+ |
Gaia-X | ORGANIZATION | 0.99+ |
two decades ago | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
this year | DATE | 0.99+ |
seven years | QUANTITY | 0.99+ |
OCTO | ORGANIZATION | 0.99+ |
zero days | QUANTITY | 0.98+ |
10 years ago | DATE | 0.98+ |
each week | QUANTITY | 0.98+ |
today | DATE | 0.97+ |
Breaking Analysis: Google's PoV on Confidential Computing
>> From theCUBE Studios in Palo Alto in Boston, bringing you data-driven insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante. >> Confidential computing is a technology that aims to enhance data privacy and security, by providing encrypted computation on sensitive data and isolating data, and apps that are fenced off enclave during processing. The concept of, I got to start over. I fucked that up, I'm sorry. That's not right, what I said was not right. On Dave in five, four, three. Confidential computing is a technology that aims to enhance data privacy and security by providing encrypted computation on sensitive data, isolating data from apps and a fenced off enclave during processing. The concept of confidential computing is gaining popularity, especially in the cloud computing space, where sensitive data is often stored and of course processed. However, there are some who view confidential computing as an unnecessary technology in a marketing ploy by cloud providers aimed at calming customers who are cloud phobic. Hello and welcome to this week's Wikibon Cube Insights powered by ETR. In this Breaking Analysis, we revisit the notion of confidential computing, and to do so, we'll invite two Google experts to the show. But before we get there, let's summarize briefly. There's not a ton of ETR data on the topic of confidential computing, I mean, it's a technology that's deeply embedded into silicon and computing architectures. But at the highest level, security remains the number one priority being addressed by IT decision makers in the coming year as shown here. And this data is pretty much across the board by industry, by region, by size of company. I mean we dug into it and the only slight deviation from the mean is in financial services. The second and third most cited priorities, cloud migration and analytics are noticeably closer to cybersecurity in financial services than in other sectors, likely because financial services has always been hyper security conscious, but security is still a clear number one priority in that sector. The idea behind confidential computing is to better address threat models for data in execution. Protecting data at rest and data in transit have long been a focus of security approaches, but more recently, silicon manufacturers have introduced architectures that separate data and applications from the host system, ARM, Intel, AMD, Nvidia and other suppliers are all on board, as are the big cloud players. Now, the argument against confidential computing is that it narrowly focuses on memory encryption and it doesn't solve the biggest problems in security. Multiple system images, updates, different services and the entire code flow aren't directly addressed by memory encryption. Rather to truly attack these problems, many believe that OSs need to be re-engineered with the attacker and hacker in mind. There are so many variables and at the end of the day, critics say the emphasis on confidential computing made by cloud providers is overstated and largely hype. This tweet from security researcher Rodrigo Bronco, sums up the sentiment of many skeptics. He says, "Confidential computing is mostly a marketing campaign from memory encryption. It's not driving the industry towards the hard open problems. It is selling an illusion." Okay. Nonetheless, encrypting data in use and fencing off key components of the system isn't a bad thing, especially if it comes with the package essentially for free. There has been a lack of standardization and interoperability between different confidential computing approaches. But the confidential computing consortium was established in 2019 ostensibly to accelerate the market and influence standards. Notably, AWS is not part of the consortium, likely because the politics of the consortium were probably a conundrum for AWS because the base technology defined by the consortium is seen as limiting by AWS. This is my guess, not AWS' words. But I think joining the consortium would validate a definition which AWS isn't aligned with. And two, it's got to lead with this Annapurna acquisition. It was way ahead with ARM integration, and so it's probably doesn't feel the need to validate its competitors. Anyway, one of the premier members of the confidential computing consortium is Google, along with many high profile names, including Aem, Intel, Meta, Red Hat, Microsoft, and others. And we're pleased to welcome two experts on confidential computing from Google to unpack the topic. Nelly Porter is Head of Product for GCP Confidential Computing and Encryption and Dr. Patricia Florissi is the Technical Director for the Office of the CTO at Google Cloud. Welcome Nelly and Patricia, great to have you. >> Great to be here. >> Thank you so much for having us. >> You're very welcome. Nelly, why don't you start and then Patricia, you can weigh in. Just tell the audience a little bit about each of your roles at Google Cloud. >> So I'll start, I'm owning a lot of interesting activities in Google and again, security or infrastructure securities that I usually own. And we are talking about encryption, end-to-end encryption, and confidential computing is a part of portfolio. Additional areas that I contribute to get with my team to Google and our customers is secure software supply chain because you need to trust your software. Is it operate in your confidential environment to have end-to-end security, about if you believe that your software and your environment doing what you expect, it's my role. >> Got it. Okay, Patricia? >> Well, I am a Technical Director in the Office of the CTO, OCTO for short in Google Cloud. And we are a global team, we include former CTOs like myself and senior technologies from large corporations, institutions and a lot of success for startups as well. And we have two main goals, first, we walk side by side with some of our largest, more strategic or most strategical customers and we help them solve complex engineering technical problems. And second, we advice Google and Google Cloud Engineering, product management on emerging trends and technologies to guide the trajectory of our business. We are unique group, I think, because we have created this collaborative culture with our customers. And within OCTO I spend a lot of time collaborating with customers in the industry at large on technologies that can address privacy, security, and sovereignty of data in general. >> Excellent. Thank you for that both of you. Let's get into it. So Nelly, what is confidential computing from Google's perspective? How do you define it? >> Confidential computing is a tool and one of the tools in our toolbox. And confidential computing is a way how we would help our customers to complete this very interesting end-to-end lifecycle of the data. And when customers bring in the data to cloud and want to protect it as they ingest it to the cloud, they protect it at rest when they store data in the cloud. But what was missing for many, many years is ability for us to continue protecting data and workloads of our customers when they run them. And again, because data is not brought to cloud to have huge graveyard, we need to ensure that this data is actually indexed. Again, there is some insights driven and drawn from this data. You have to process this data and confidential computing here to help. Now we have end-to-end protection of our customer's data when they bring the workloads and data to cloud thanks to confidential computing. >> Thank you for that. Okay, we're going to get into the architecture a bit, but before we do Patricia, why do you think this topic of confidential computing is such an important technology? Can you explain? Do you think it's transformative for customers and if so, why? >> Yeah, I would maybe like to use one thought, one way, one intuition behind why confidential computing matters because at the end of the day, it reduces more and more the customer's thrush boundaries and the attack surface. That's about reducing that periphery, the boundary in which the customer needs to mind about trust and safety. And in a way is a natural progression that you're using encryption to secure and protect data in the same way that we are encrypting data in transit and at rest. Now, we are also encrypting data while in the use. And among other beneficials, I would say one of the most transformative ones is that organizations will be able to collaborate with each other and retain the confidentiality of the data. And that is across industry, even though it's highly focused on, I wouldn't say highly focused but very beneficial for highly regulated industries, it applies to all of industries. And if you look at financing for example, where bankers are trying to detect fraud and specifically double finance where a customer is actually trying to get a finance on an asset, let's say a boat or a house, and then it goes to another bank and gets another finance on that asset. Now bankers would be able to collaborate and detect fraud while preserving confidentiality and privacy of the data. >> Interesting and I want to understand that a little bit more but I got to push you a little bit on this, Nellie if I can, because there's a narrative out there that says confidential computing is a marketing ploy I talked about this up front, by cloud providers that are just trying to placate people that are scared of the cloud. And I'm presuming you don't agree with that, but I'd like you to weigh in here. The argument is confidential computing is just memory encryption, it doesn't address many other problems. It is over hyped by cloud providers. What do you say to that line of thinking? >> I absolutely disagree as you can imagine Dave, with this statement. But the most importantly is we mixing a multiple concepts I guess, and exactly as Patricia said, we need to look at the end-to-end story, not again, is a mechanism. How confidential computing trying to execute and protect customer's data and why it's so critically important. Because what confidential computing was able to do, it's in addition to isolate our tenants in multi-tenant environments the cloud offering to offer additional stronger isolation, they called it cryptographic isolation. It's why customers will have more trust to customers and to other customers, the tenants running on the same host but also us because they don't need to worry about against rats and more malicious attempts to penetrate the environment. So what confidential computing is helping us to offer our customers stronger isolation between tenants in this multi-tenant environment, but also incredibly important, stronger isolation of our customers to tenants from us. We also writing code, we also software providers, we also make mistakes or have some zero days. Sometimes again us introduce, sometimes introduced by our adversaries. But what I'm trying to say by creating this cryptographic layer of isolation between us and our tenants and among those tenants, we really providing meaningful security to our customers and eliminate some of the worries that they have running on multi-tenant spaces or even collaborating together with very sensitive data knowing that this particular protection is available to them. >> Okay, thank you. Appreciate that. And I think malicious code is often a threat model missed in these narratives. You know, operator access. Yeah, maybe I trust my cloud's provider, but if I can fence off your access even better, I'll sleep better at night separating a code from the data. Everybody's ARM, Intel, AMD, Nvidia and others, they're all doing it. I wonder if Nell, if we could stay with you and bring up the slide on the architecture. What's architecturally different with confidential computing versus how operating systems and VMs have worked traditionally? We're showing a slide here with some VMs, maybe you could take us through that. >> Absolutely, and Dave, the whole idea for Google and now industry way of dealing with confidential computing is to ensure that three main property is actually preserved. Customers don't need to change the code. They can operate in those VMs exactly as they would with normal non-confidential VMs. But to give them this opportunity of lift and shift though, no changing the apps and performing and having very, very, very low latency and scale as any cloud can, some things that Google actually pioneer in confidential computing. I think we need to open and explain how this magic was actually done, and as I said, it's again the whole entire system have to change to be able to provide this magic. And I would start with we have this concept of root of trust and root of trust where we will ensure that this machine within the whole entire host has integrity guarantee, means nobody changing my code on the most low level of system, and we introduce this in 2017 called Titan. So our specific ASIC, specific inch by inch system on every single motherboard that we have that ensures that your low level former, your actually system code, your kernel, the most powerful system is actually proper configured and not changed, not tempered. We do it for everybody, confidential computing included, but for confidential computing is what we have to change, we bring in AMD or future silicon vendors and we have to trust their former, their way to deal with our confidential environments. And that's why we have obligation to validate intelligent not only our software and our former but also former and software of our vendors, silicon vendors. So we actually, when we booting this machine as you can see, we validate that integrity of all of this system is in place. It means nobody touching, nobody changing, nobody modifying it. But then we have this concept of AMD Secure Processor, it's special ASIC best specific things that generate a key for every single VM that our customers will run or every single node in Kubernetes or every single worker thread in our Hadoop spark capability. We offer all of that and those keys are not available to us. It's the best case ever in encryption space because when we are talking about encryption, the first question that I'm receiving all the time, "Where's the key? Who will have access to the key?" because if you have access to the key then it doesn't matter if you encrypted or not. So, but the case in confidential computing why it's so revolutionary technology, us cloud providers who don't have access to the keys, they're sitting in the hardware and they fed to memory controller. And it means when hypervisors that also know about this wonderful things saying I need to get access to the memories, that this particular VM I'm trying to get access to. They do not decrypt the data, they don't have access to the key because those keys are random, ephemeral and per VM, but most importantly in hardware not exportable. And it means now you will be able to have this very interesting world that customers or cloud providers will not be able to get access to your memory. And what we do, again as you can see, our customers don't need to change their applications. Their VMs are running exactly as it should run. And what you've running in VM, you actually see your memory clear, it's not encrypted. But God forbid is trying somebody to do it outside of my confidential box, no, no, no, no, no, you will now be able to do it. Now, you'll see cyber test and it's exactly what combination of these multiple hardware pieces and software pieces have to do. So OS is also modified and OS is modified such way to provide integrity. It means even OS that you're running in your VM box is not modifiable and you as customer can verify. But the most interesting thing I guess how to ensure the super performance of this environment because you can imagine Dave, that's increasing and it's additional performance, additional time, additional latency. So we're able to mitigate all of that by providing incredibly interesting capability in the OS itself. So our customers will get no changes needed, fantastic performance and scales as they would expect from cloud providers like Google. >> Okay, thank you. Excellent, appreciate that explanation. So you know again, the narrative on this is, well, you've already given me guarantees as a cloud provider that you don't have access to my data, but this gives another level of assurance, key management as they say is key. Now humans aren't managing the keys, the machines are managing them. So Patricia, my question to you is in addition to, let's go pre-confidential computing days, what are the sort of new guarantees that these hardware based technologies are going to provide to customers? >> So if I am a customer, I am saying I now have full guarantee of confidentiality and integrity of the data and of the code. So if you look at code and data confidentiality, the customer cares and they want to know whether their systems are protected from outside or unauthorized access, and that we covered with Nelly that it is. Confidential computing actually ensures that the applications and data antennas remain secret. The code is actually looking at the data, only the memory is decrypting the data with a key that is ephemeral, and per VM, and generated on demand. Then you have the second point where you have code and data integrity and now customers want to know whether their data was corrupted, tempered with or impacted by outside actors. And what confidential computing ensures is that application internals are not tempered with. So the application, the workload as we call it, that is processing the data is also has not been tempered and preserves integrity. I would also say that this is all verifiable, so you have attestation and this attestation actually generates a log trail and the log trail guarantees that provides a proof that it was preserved. And I think that the offers also a guarantee of what we call sealing, this idea that the secrets have been preserved and not tempered with, confidentiality and integrity of code and data. >> Got it. Okay, thank you. Nelly, you mentioned, I think I heard you say that the applications is transparent, you don't have to change the application, it just comes for free essentially. And we showed some various parts of the stack before, I'm curious as to what's affected, but really more importantly, what is specifically Google's value add? How do partners participate in this, the ecosystem or maybe said another way, how does Google ensure the compatibility of confidential computing with existing systems and applications? >> And a fantastic question by the way, and it's very difficult and definitely complicated world because to be able to provide these guarantees, actually a lot of work was done by community. Google is very much operate and open. So again our operating system, we working this operating system repository OS is OS vendors to ensure that all capabilities that we need is part of the kernels are part of the releases and it's available for customers to understand and even explore if they have fun to explore a lot of code. We have also modified together with our silicon vendors kernel, host kernel to support this capability and it means working this community to ensure that all of those pages are there. We also worked with every single silicon vendor as you've seen, and it's what I probably feel that Google contributed quite a bit in this world. We moved our industry, our community, our vendors to understand the value of easy to use confidential computing or removing barriers. And now I don't know if you noticed Intel is following the lead and also announcing a trusted domain extension, very similar architecture and no surprise, it's a lot of work done with our partners to convince work with them and make this capability available. The same with ARM this year, actually last year, ARM announced future design for confidential computing, it's called confidential computing architecture. And it's also influenced very heavily with similar ideas by Google and industry overall. So it's a lot of work in confidential computing consortiums that we are doing, for example, simply to mention, to ensure interop as you mentioned, between different confidential environments of cloud providers. They want to ensure that they can attest to each other because when you're communicating with different environments, you need to trust them. And if it's running on different cloud providers, you need to ensure that you can trust your receiver when you sharing your sensitive data workloads or secret with them. So we coming as a community and we have this at Station Sig, the community-based systems that we want to build, and influence, and work with ARM and every other cloud providers to ensure that they can interop. And it means it doesn't matter where confidential workloads will be hosted, but they can exchange the data in secure, verifiable and controlled by customers really. And to do it, we need to continue what we are doing, working open and contribute with our ideas and ideas of our partners to this role to become what we see confidential computing has to become, it has to become utility. It doesn't need to be so special, but it's what what we've wanted to become. >> Let's talk about, thank you for that explanation. Let's talk about data sovereignty because when you think about data sharing, you think about data sharing across the ecosystem in different regions and then of course data sovereignty comes up, typically public policy, lags, the technology industry and sometimes it's problematic. I know there's a lot of discussions about exceptions but Patricia, we have a graphic on data sovereignty. I'm interested in how confidential computing ensures that data sovereignty and privacy edicts are adhered to, even if they're out of alignment maybe with the pace of technology. One of the frequent examples is when you delete data, can you actually prove the data is deleted with a hundred percent certainty, you got to prove that and a lot of other issues. So looking at this slide, maybe you could take us through your thinking on data sovereignty. >> Perfect. So for us, data sovereignty is only one of the three pillars of digital sovereignty. And I don't want to give the impression that confidential computing addresses it at all, that's why we want to step back and say, hey, digital sovereignty includes data sovereignty where we are giving you full control and ownership of the location, encryption and access to your data. Operational sovereignty where the goal is to give our Google Cloud customers full visibility and control over the provider operations, right? So if there are any updates on hardware, software stack, any operations, there is full transparency, full visibility. And then the third pillar is around software sovereignty, where the customer wants to ensure that they can run their workloads without dependency on the provider's software. So they have sometimes is often referred as survivability that you can actually survive if you are untethered to the cloud and that you can use open source. Now, let's take a deep dive on data sovereignty, which by the way is one of my favorite topics. And we typically focus on saying, hey, we need to care about data residency. We care where the data resides because where the data is at rest or in processing need to typically abides to the jurisdiction, the regulations of the jurisdiction where the data resides. And others say, hey, let's focus on data protection, we want to ensure the confidentiality, and integrity, and availability of the data, which confidential computing is at the heart of that data protection. But it is yet another element that people typically don't talk about when talking about data sovereignty, which is the element of user control. And here Dave, is about what happens to the data when I give you access to my data, and this reminds me of security two decades ago, even a decade ago, where we started the security movement by putting firewall protections and logging accesses. But once you were in, you were able to do everything you wanted with the data. An insider had access to all the infrastructure, the data, and the code. And that's similar because with data sovereignty, we care about whether it resides, who is operating on the data, but the moment that the data is being processed, I need to trust that the processing of the data we abide by user's control, by the policies that I put in place of how my data is going to be used. And if you look at a lot of the regulation today and a lot of the initiatives around the International Data Space Association, IDSA and Gaia-X, there is a movement of saying the two parties, the provider of the data and the receiver of the data going to agree on a contract that describes what my data can be used for. The challenge is to ensure that once the data crosses boundaries, that the data will be used for the purposes that it was intended and specified in the contract. And if you actually bring together, and this is the exciting part, confidential computing together with policy enforcement. Now, the policy enforcement can guarantee that the data is only processed within the confines of a confidential computing environment, that the workload is in cryptographically verified that there is the workload that was meant to process the data and that the data will be only used when abiding to the confidentiality and integrity safety of the confidential computing environment. And that's why we believe confidential computing is one necessary and essential technology that will allow us to ensure data sovereignty, especially when it comes to user's control. >> Thank you for that. I mean it was a deep dive, I mean brief, but really detailed. So I appreciate that, especially the verification of the enforcement. Last question, I met you two because as part of my year-end prediction post, you guys sent in some predictions and I wasn't able to get to them in the predictions post, so I'm thrilled that you were able to make the time to come on the program. How widespread do you think the adoption of confidential computing will be in '23 and what's the maturity curve look like this decade in your opinion? Maybe each of you could give us a brief answer. >> So my prediction in five, seven years as I started, it will become utility, it will become TLS. As of freakin' 10 years ago, we couldn't believe that websites will have certificates and we will support encrypted traffic. Now we do, and it's become ubiquity. It's exactly where our confidential computing is heeding and heading, I don't know we deserve yet. It'll take a few years of maturity for us, but we'll do that. >> Thank you. And Patricia, what's your prediction? >> I would double that and say, hey, in the very near future, you will not be able to afford not having it. I believe as digital sovereignty becomes ever more top of mind with sovereign states and also for multinational organizations, and for organizations that want to collaborate with each other, confidential computing will become the norm, it will become the default, if I say mode of operation. I like to compare that today is inconceivable if we talk to the young technologists, it's inconceivable to think that at some point in history and I happen to be alive, that we had data at rest that was non-encrypted, data in transit that was not encrypted. And I think that we'll be inconceivable at some point in the near future that to have unencrypted data while we use. >> You know, and plus I think the beauty of the this industry is because there's so much competition, this essentially comes for free. I want to thank you both for spending some time on Breaking Analysis, there's so much more we could cover. I hope you'll come back to share the progress that you're making in this area and we can double click on some of these topics. Really appreciate your time. >> Anytime. >> Thank you so much, yeah. >> In summary, while confidential computing is being touted by the cloud players as a promising technology for enhancing data privacy and security, there are also those as we said, who remain skeptical. The truth probably lies somewhere in between and it will depend on the specific implementation and the use case as to how effective confidential computing will be. Look as with any new tech, it's important to carefully evaluate the potential benefits, the drawbacks, and make informed decisions based on the specific requirements in the situation and the constraints of each individual customer. But the bottom line is silicon manufacturers are working with cloud providers and other system companies to include confidential computing into their architectures. Competition in our view will moderate price hikes and at the end of the day, this is under-the-covers technology that essentially will come for free, so we'll take it. I want to thank our guests today, Nelly and Patricia from Google. And thanks to Alex Myerson who's on production and manages the podcast. Ken Schiffman as well out of our Boston studio. Kristin Martin and Cheryl Knight help get the word out on social media and in our newsletters, and Rob Hoof is our editor-in-chief over at siliconangle.com, does some great editing for us. Thank you all. Remember all these episodes are available as podcasts. Wherever you listen, just search Breaking Analysis podcast. I publish each week on wikibon.com and siliconangle.com where you can get all the news. If you want to get in touch, you can email me at david.vellante@siliconangle.com or DM me at D Vellante, and you can also comment on my LinkedIn post. Definitely you want to check out etr.ai for the best survey data in the enterprise tech business. I know we didn't hit on a lot today, but there's some amazing data and it's always being updated, so check that out. This is Dave Vellante for theCUBE Insights powered by ETR. Thanks for watching and we'll see you next time on Breaking Analysis. (subtle music)
SUMMARY :
bringing you data-driven and at the end of the day, and then Patricia, you can weigh in. contribute to get with my team Okay, Patricia? Director in the Office of the CTO, for that both of you. in the data to cloud into the architecture a bit, and privacy of the data. that are scared of the cloud. and eliminate some of the we could stay with you and they fed to memory controller. to you is in addition to, and integrity of the data and of the code. that the applications is transparent, and ideas of our partners to this role One of the frequent examples and a lot of the initiatives of the enforcement. and we will support encrypted traffic. And Patricia, and I happen to be alive, the beauty of the this industry and at the end of the day,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Nelly | PERSON | 0.99+ |
Patricia | PERSON | 0.99+ |
Alex Myerson | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
International Data Space Association | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
AWS' | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Rob Hoof | PERSON | 0.99+ |
Cheryl Knight | PERSON | 0.99+ |
Nelly Porter | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Nvidia | ORGANIZATION | 0.99+ |
IDSA | ORGANIZATION | 0.99+ |
Rodrigo Bronco | PERSON | 0.99+ |
2019 | DATE | 0.99+ |
Ken Schiffman | PERSON | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
AMD | ORGANIZATION | 0.99+ |
2017 | DATE | 0.99+ |
ARM | ORGANIZATION | 0.99+ |
Aem | ORGANIZATION | 0.99+ |
Nellie | PERSON | 0.99+ |
Kristin Martin | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
two parties | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
Patricia Florissi | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
Meta | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
third | QUANTITY | 0.99+ |
Gaia-X | ORGANIZATION | 0.99+ |
second point | QUANTITY | 0.99+ |
two experts | QUANTITY | 0.99+ |
david.vellante@siliconangle.com | OTHER | 0.99+ |
second | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
first question | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
theCUBE Studios | ORGANIZATION | 0.99+ |
two decades ago | DATE | 0.99+ |
'23 | DATE | 0.99+ |
each | QUANTITY | 0.99+ |
a decade ago | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
zero days | QUANTITY | 0.98+ |
four | QUANTITY | 0.98+ |
OCTO | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |