Santosh Rao, NetApp | Accelerate Your Journey to AI
>> From Sunnyvale California, in the heart of Silicon Valley. It's theCUBE, covering, Accelerate Your Journey to AI, Brought to you by NetApp. >> Hi I'm Peter Burris, welcome to another conversation here from the Data Visionary Center at NetApp's headquarters in beautiful Sunnyvale California. I'm being joined today by Santosh Rao. Santosh is the Senior Technical Director at NetApp, Specifically Santosh we're going to talk about some of the challenges and opportunities associated with AI and how NetApp is making that possible. Welcome to theCUBE. >> Thank you Peter, I'm excited to be here. Thank you for that. >> So, Santosh what is your role at Netapp? Why don't we start there. >> Wonderful, glad to be here, my name is Santosh Rao, I'm a Senior Technical Director at NetApp, part of the Product Operations group, and I've been here 10 years. My role is to drive up new lines of opportunity for NetApp, build up new product businesses. The most recent one has been AI. So I've been focused on bootstrapping and incubating the AI effort at NetApp for the last nine months now. Been excited to be part of this effort now. >> So nine months of talking, both internally, but spending time with customers too. What are customers telling you that are NetApp's opportunities, and what NetApp has to do to respond to those opportunities? >> That's a great question. We are seeing a lot of focus around expanding the digital transformation to really get value out of the data, and start looking at AI, and Deep Learning in particular, as a way to prove the ROI on the opportunities that they've had. AI and deep learning requires a tremendous amount of data. We're actually fascinated to see the amount of data sets that customers are starting to look at. A petabyte of data is sort of the minimum size of data set. So when you think about petabyte-scale data lakes. The first think you want to think about is how do you optimize the TCO for the solution. NetApp is seen as a leader in that, just because of our rich heritage of storage efficiency. A lot of these are video image and audio files, and so you're seeing a lot of unstructured data in general, and we're a leader in NFS as well. So a lot of that starts to come together from a NetApp perspective. And that's where customers see us as the leader in NFS, the leader in files, and the leader in storage efficiency, all coming together. >> And you want to join that together with some leadership, especially in GPU's, so that leads to NVIDIA. So you've announced an interesting partnership between NetApp and NVIDIA. How did that factor into your products, and where do you think that goes? >> It's kind of interesting how that came about, because when you look at the industry it's a small place. Some of the folks driving the NVIDIA leadership have been working with us in the past, when we've bootstrapped converged infrastructures with other vendors. We're known to have been a 10 year metro vendor in the converged infrastructure space. The way this came about was NVIDIA is clearly a leader in the GPU and AI acceleration from a computer perspective. But they're also seen as a long history of GPU virtualization and GPU graphics acceleration. When they look at NetApp, what NetApp brings to NVIDIA is just the converged infrastructure, the maturity of that solution, the depth that we have in the enterprise and the rich partner ecosystem. All of that starts to come together, and some of the players in this particular case, have had aligned in the past working on virtualization based conversion infrastructures in the past. It's an exciting time, we're really looking forward to working closely with NVIDIA. >> So NVIDIA brings these lighting fast machines, optimized for some of the new data types, data forms, data structures associated with AI. But they got to be fed, got to get the data to them. What is NetApp doing from a standpoint of the underlying hardware to improve the overall performance, and insure that these solutions really scream for customers? >> Yeah, it's kind of interesting, because when you look at how customers are designing this. They're thinking about digital transformation as, "What is the flow of that data? "What am I doing to create new sensors "and endpoints that create data? "How do I flow the data in? "How do I forecast how much data I'm going to "create quarter over quarter, year over year? "How many endpoints? what is the resolution of the data?" And then as that starts to come into the data center, they got to think about, where are the bottlenecks. So you start looking at a wide range of bottlenecks. You look at the edge data aggregation, then you start looking at network bandwidth to push data into the core data centers. You got to think smart about some of these things. For example, no matter how much network bandwidth you throw at it, you want to reduce the amount of data you're moving. Smart data movement technologies like SnapMirror, which NetApp brings to the table, are some things that we uniquely enable compared to others. The fact of the matter is when you take a common operating system, like ONTAP, and you can lear it across the Edge, Core and Cloud, that gives us some unnatural advantages. We can do things that you can't do in a silo. You've got a commodities server trying to push data, and having to do raw full copies of data into the data center. So we think smart data movement is a huge opportunity. When you look at the core, obviously it's a workhorse, and you've got the random sampling of data into this hardware. And we think the A800 is a workhorse built for AI. It is a best of a system in terms of performance, it does about 25 gigabytes per second just on a dual controller pair. You'll recall that we spent several number of years building out the foundation of Clustered ONTAP to allow us to scale to gigantic sizes. So 24 node or 12 controller pad A800 gets us to over 300 gigabytes per second, and over 11 million IOPS if you think about that. That's over about four to six times greater than anybody else in the industry. So when you think about NVIDIA investment in DGX and they're performance investment they've made there. We think only NetApp can keep up with that, in terms of performance. >> So 11 million IOPS, phenomenal performance for today. But the future is going to demand ever more. Where do you think these trends go? >> Well nobody really knows for sure. The most exciting part of this journey, is nobody knows where this is going. This is where you need to future proof customers, and you need to enable the technology to have sufficient legs, and the architecture to have sufficient legs. That no matter how it evolves and where customers go, the vendors working with customers can go there with them. And actually when customers look at NetApp and say, "You guys are working with the Cloud partners, "you're now working with NVIDIA. "And in the past you worked with a "variety of data source vendors. "So we think we can work with NetApp because, "you're not affiliated to any one of them, "and yet you're giving us that full range of solutions." So we think that performance is going to be key. Acceleration of compute workloads is going to demand orders of magnitude performance improvement. We think data set efficiencies and storage efficiencies is absolutely key. And we think you got to really look at PCO, because customers want to build these great solutions for the business, but they can't afford it unless vendors give them viable options. So it's really up to partners like NVIDIA and NetApp to work together to give customers the best of breed solutions that reduce the TCO, accelerate compute, accelerate the data pipeline, and yet, bring the cost of the overall solution down, and make it simple to deploy and pre integrated. These are the things customers are looking for and we think we have the best bet at getting there. >> So that leads to... Great summary, but that leads to some interesting observations on what customers should be basing their decisions on. What would you say are the two or three most crucial things that customers need to think about right now as a conceptualized, where to go with their AI application, or AI workloads, their AI projects and initiatives? >> So when customers are designing and building these solutions, they're thinking the entire data lifecycle. "How am I getting this new type of "data for digital transformation? "What is the ingestion architecture? "What are my data aggregation endpoints for ingestion? "How am I going to build out my AI data sources? "What are the types of data? "Am I collecting sensor data? Is it a variety of images? "Am I going to add in audio transcription? "Is there video feeds that come in over time?" So customers are having to think about the entire digital experience, the types of data, because that leads to the selection of data sources. For example, if you're going to be learning sensor data, you want to be looking at maybe graph databases. If you want to be learning log data, you're going to be looking at log analytics over time, as well as AI. You're going to look at video image and audio accordingly. Architecting these solutions requires an understanding of, what is your digital experience? How does that evolve over time? What is the right and optimal data source to learn that data, so that you get the best experience from a search, from an indexing, from a tiering, from analytics and AI? And then, what is the flow of that data? And how do you architect it for a global experience? How do you build out these data centers where you're not having to copy all data maybe, into your global headquarters. If you're a global company with presence across multiple Geo's, how do you architect for regional data centers to be self contained? Because we're looking at exabyte scale opportunities in some of these. I think that's pretty much the two or three things that I'd say, across the entire gamut of space here. >> Excellent, turning that then into some simple observations about the fact that data still is physical. There's latency issues, there's the cost of bandwidth issues. There's other types of issues. This notion of Edge, Core, Cloud. How do you see the ONTAP operating system, the ONTAP product set, facilitating being able to put data where it needs to be, while at the same time creating the options that a customer needs to use data as they need to use it? >> The fact of the matter is, these things cannot be achieved overnight. It takes a certain amount of foundational work, that, frankly, takes several years. The fact that ONTAP can run on small, form factor hardware at the edge is a journey that we started several years ago. The fact that ONTAP can run on commodity white box hardware, has been a journey that we have run over the last three, four years. Same thing in the Cloud, we have virtualized ONTAP to the point that it can run on all hyperscalers and now we are in the process of consuming ONTAP as a service, where you don't even know that it is an infrastructure product, or has been. So the process of building an Edge, Core, and Cloud data pipeline leverages the investments that we've made over time. When you think about the scale of compute, data and performance needed, that's a five to six year journey in Clustered ONTAP, if you look at NetApp's past. These are all elements that are coming together from a product and solution perspective. But the reality is that leveraging years and years of investment that NetApp engineering has made. In a a way that the industry really did not invest in the same areas. So when we compare and contrast what NetApp has done versus the rest of the industry. At a time when people were building monolithic engineered systems, we were building software defined architectures. At a time when they were building tightly cobbled system for traditional enterprise, we were building flexible, scale out systems, that assumed that you would want to scale in modular increments. Now as the world has shifted from enterprise into third platform and Webscale. We're finding all those investments NetApp made over the years is really starting to pay off for us. >> Including some of the investments in how AI can be used to handle how ONTAP operates at each of those different levels of scale. >> Absolutely, yes. >> Sontash Rao, Technical Director at NetApp, talking about AI, some of the new changes in the relationships between AI and storage. Thanks very much for being on theCUBE. >> Thank you, appreciate it.
SUMMARY :
Brought to you by NetApp. Santosh is the Senior Technical Director at NetApp, Thank you Peter, I'm excited to be here. Why don't we start there. the AI effort at NetApp for the last nine months now. What are customers telling you that are So a lot of that starts to come especially in GPU's, so that leads to NVIDIA. All of that starts to come together, What is NetApp doing from a standpoint of the The fact of the matter is when you But the future is going to demand ever more. and the architecture to have sufficient legs. Great summary, but that leads to some because that leads to the selection of data sources. observations about the fact that data The fact of the matter is, Including some of the investments in how AI can in the relationships between AI and storage.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Justin Warren | PERSON | 0.99+ |
Sanjay Poonen | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Clarke | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
Jeff Frick | PERSON | 0.99+ |
Dave Volante | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Diane Greene | PERSON | 0.99+ |
Michele Paluso | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Sam Lightstone | PERSON | 0.99+ |
Dan Hushon | PERSON | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
Teresa Carlson | PERSON | 0.99+ |
Kevin | PERSON | 0.99+ |
Andy Armstrong | PERSON | 0.99+ |
Michael Dell | PERSON | 0.99+ |
Pat Gelsinger | PERSON | 0.99+ |
John | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Lisa Martin | PERSON | 0.99+ |
Kevin Sheehan | PERSON | 0.99+ |
Leandro Nunez | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
EMC | ORGANIZATION | 0.99+ |
GE | ORGANIZATION | 0.99+ |
NetApp | ORGANIZATION | 0.99+ |
Keith | PERSON | 0.99+ |
Bob Metcalfe | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
90% | QUANTITY | 0.99+ |
Sam | PERSON | 0.99+ |
Larry Biagini | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
Brendan | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Peter | PERSON | 0.99+ |
Clarke Patterson | PERSON | 0.99+ |