Justin Emerson, Pure Storage | SuperComputing 22
(soft music) >> Hello, fellow hardware nerds and welcome back to Dallas Texas where we're reporting live from Supercomputing 2022. My name is Savannah Peterson, joined with the John Furrier on my left. >> Looking good today. >> Thank you, John, so are you. It's been a great show so far. >> We've had more hosts, more guests coming than ever before. >> I know. >> Amazing, super- >> We've got a whole thing going on. >> It's been a super computing performance. >> It, wow. And, we'll see how many times we can say super on this segment. Speaking of super things, I am in a very unique position right now. I am a flanked on both sides by people who have been doing content on theCUBE for 12 years. Yes, you heard me right, our next guest was on theCUBE 12 years ago, the third event, was that right, John? >> Man: First ever VM World. >> Yeah, the first ever VM World, third event theCUBE ever did. We are about to have a lot of fun. Please join me in welcoming Justin Emerson of Pure Storage. Justin, welcome back. >> It's a pleasure to be here. It's been too long, you never call, you don't write. (Savannah laughs) >> Great to see you. >> Yeah, likewise. >> How fun is this? Has the set evolved? Is everything looking good? >> I mean, I can barely remember what happened last week, so. (everyone laughs) >> Well, I remember lot's changed that VM world. You know, Paul Moritz was the CEO if you remember at that time. His actual vision actually happened but not the way, for VMware, but the industry, the cloud, he called the software mainframe. We were kind of riffing- >> It was quite the decade. >> Unbelievable where we are now, how we got here, but not where we're going to be. And you're with Pure Storage now which we've been, as you know, covering as well. Where's the connection into the supercomputing? Obviously storage performance, big part of this show. >> Right, right. >> What's the take? >> Well, I think, first of all it's great to be back at events in person. We were talking before we went on, and it's been so great to be back at live events now. It's been such a drought over the last several years, but yeah, yeah. So I'm very glad that we're doing in person events again. For Pure, this is an incredibly important show. You know, the product that I work with, with FlashBlade is you know, one of our key areas is specifically in this high performance computing, AI machine learning kind of space. And so we're really glad to be here. We've met a lot of customers, met a lot of other folks, had a lot of really great conversations. So it's been a really great show for me. And also just seeing all the really amazing stuff that's around here, I mean, if you want to find, you know, see what all the most cutting edge data center stuff that's going to be coming down the pipe, this is the place to do it. >> So one of the big themes of the show for us and probably, well, big theme of your life, is balancing power efficiency. You have a product in this category, Direct Flash. Can you tell us a little bit more about that? >> Yeah, so Pure as a storage company, right, what do we do differently from everybody else? And if I had to pick one thing, right, I would talk about, it's, you know, as the name implies, we're an all, we're purely flash, we're an all flash company. We've always been, don't plan to be anything else. And part of that innovation with Direct Flash is the idea of rather than treating a solid state disc as like a hard drive, right? Treat it as it actually is, treat it like who it really is and that's a very different kind of thing. And so Direct Flash is all about bringing native Flash interfaces to our product portfolio. And what's really exciting for me as a FlashBlade person, is now that's also part of our FlashBlade S portfolio, which just launched in June. And so the benefits of that are our myriad. But, you know, talking about efficiency, the biggest difference is that, you know, we can use like 90% less DRAM in our drives, which you know, everything uses, everything that you put in a drive uses power, it adds cost and all those things and so that really gives us an efficiency edge over everybody else and at a show like this, where, I mean, you walk the aisles and there's there's people doing liquid cooling and so much immersion stuff, and the reason they're doing that is because power is just increasing everywhere, right? So if you can figure out how do we use less power in some areas means you can shift that budget to other places. So if you can talk to a customer and say, well, if I could shrink your power budget for storage by two thirds or even, save you two-thirds of power, how many more accelerators, how many more CPUs, how much more work could you actually get done? So really exciting. >> I mean, less power consumption, more power and compute. >> Right. >> Kind of power center. So talk about the AI implications, where the use cases are. What are you seeing here? A lot of simulations, a lot of students, again, dorm room to the boardroom we've been saying here on theCUBE this is a great broad area, where's the action in the ML and the AI for you guys? >> So I think, not necessarily storage related but I think that right now there's this enormous explosion of custom silicon around AI machine learning which I as a, you said welcome hardware nerds at the beginning and I was like, ah, my people. >> We're all here, we're all here in Dallas. >> So wonderful. You know, as a hardware nerd we're talking about conferences, right? Who has ever attended hot chips and there's so much really amazing engineering work going on in the silicon space. It's probably the most exciting time for, CPU and accelerator, just innovation in, since the days before X 86 was the defacto standard, right? And you could go out and buy a different workstation with 16 different ISAs. That's really the most exciting thing, I walked past so many different places where you know, our booth is right next to Havana Labs with their gout accelerator, and they're doing this cute thing with one of the AI image generators in their booth, which is really cute. >> Woman: We're going to have to go check that out. >> Yeah, but that to me is like one of the more exciting things around like innovation at a, especially at a show like this where it's all about how do we move forward, the state of the art. >> What's different now than just a few years ago in terms of what's opening up the creativity for people to look at things that they could do with some of the scale that's different now. >> Yeah well, I mean, every time the state of the art moves forward what it means is, is that the entry level gets better, right? So if the high end is going faster, that means that the mid-range is going faster, and that means the entry level is going faster. So every time it pushes the boundary forward, it's a rising tide that floats all boats. And so now, the kind of stuff that's possible to do, if you're a student in a dorm room or if you're an enterprise, the world, the possible just keeps expanding dramatically and expanding almost, you know, geometrically like the amount of data that we are, that we have, as a storage guy, I was coming back to data but the amount of data that we have and the amount of of compute that we have, and it's not just about the raw compute, but also the advances in all sorts of other things in terms of algorithms and transfer learning and all these other things. There's so much amazing work going on in this area and it's just kind of this Kay Green explosion of innovation in the area. >> I love that you touched on the user experience for the community, no matter the level that you're at. >> Yeah. >> And I, it's been something that's come up a lot here. Everyone wants to do more faster, always, but it's not just that, it's about making the experience and the point of entry into this industry more approachable and digestible for folks who may not be familiar, I mean we have every end of the ecosystem here, on the show floor, where does Pure Storage sit in the whole game? >> Right, so as a storage company, right? What AI is all about deriving insights from data, right? And so everyone remembers that magazine cover data's the new oil, right? And it's kind of like, okay, so what do you do with it? Well, how do you derive value from all of that data? And AI machine learning and all of this supercomputing stuff is about how do we take all this data? How do we innovate with it? And so if you want data to innovate with, you need storage. And so, you know, our philosophy is that how do we make the best storage platforms that we can using the best technology for our customers that enable them to do really amazing things with AI machine learning and we've got different products, but, you know at the show here, what we're specifically showing off is our new flashlight S product, which, you know, I know we've had Pure folks on theCUBE before talking about FlashBlade, but for viewers out there, FlashBlade is our our scale out unstructured data platform and AI and machine learning and supercomputing is all about unstructured data. It's about sensor data, it's about imaging, it's about, you know, photogrammetry, all this other kinds of amazing stuff. But, you got to land all that somewhere. You got to process that all somewhere. And so really high performance, high throughput, highly scalable storage solutions are really essential. It's an enabler for all of the amazing other kinds of engineering work that goes on at a place like Supercomputing. >> It's interesting you mentioned data's oil. Remember in 2010, that year, our first year of theCUBE, Hadoop World, Hadoop just started to come on the scene, which became, you know kind of went away and, but now you got, Spark and Databricks and Snowflake- >> Justin: And it didn't go away, it just changed, right? >> It just got refactored and right size, I think for what the people wanted it to be easy to use but there's more data coming. How is data driving innovation as you bring, as people see clearly the more data's coming? How is data driving innovation as you guys look at your products, your roadmap and your customer base? How is data driving innovation for your customers? >> Well, I think every customer who has been, you know collecting all of this data, right? Is trying to figure out, now what do I do with it? And a lot of times people collect data and then it will end up on, you know, lower slower tiers and then suddenly they want to do something with it. And it's like, well now what do I do, right? And so there's all these people that are reevaluating you know, we, when we developed FlashBlade we sort of made this bet that unstructured data was going to become the new tier one data. It used to be that we thought unstructured data, it was emails and home directories and all that stuff the kind of stuff that you didn't really need a really good DR plan on. It's like, ah, we could, now of course, as soon as email goes down, you realize how important email is. But, the perspectives that people had on- >> Yeah, exactly. (all laughing) >> The perspectives that people had on unstructured data and it's value to the business was very different and so now- >> Good bet, by the way. >> Yeah, thank you. So now unstructured data is considered, you know, where companies are going to derive their value from. So it's whether they use the data that they have to build better products whether it's they use the data they have to develop you know, improvements in processes. All those kinds of things are data driven. And so all of the new big advancements in industry and in business are all about how do I derive insights from data? And so machine learning and AI has something to do with that, but also, you know, it all comes back to having data that's available. And so, we're working very hard on building platforms that customers can use to enable all of this really- >> Yeah, it's interesting, Savannah, you know, the top three areas we're covering for reinventing all the hyperscale events is data. How does it drive innovation and then specialized solutions to make customers lives easier? >> Yeah. >> It's become a big category. How do you compose stuff and then obviously compute, more and more compute and services to make the performance goes. So those seem to be the three hot areas. So, okay, data's the new oil refineries. You've got good solutions. What specialized solutions do you see coming out because once people have all this data, they might have either large scale, maybe some edge use cases. Do you see specialized solutions emerging? I mean, obviously it's got DPU emerging which is great, but like, do you see anything else coming out at that people are- >> Like from a hardware standpoint. >> Or from a customer standpoint, making the customer's lives easier? So, I got a lot of data flowing in. >> Yeah. >> It's never stopping, it keeps powering in. >> Yeah. >> Are there things coming out that makes their life easier? Have you seen anything coming out? >> Yeah, I think where we are as an industry right now with all of this new technology is, we're really in this phase of the standards aren't quite there yet. Everybody is sort of like figuring out what works and what doesn't. You know, there was this big revolution in sort of software development, right? Where moving towards agile development and all that kind of stuff, right? The way people build software change fundamentally this is kind of like another wave like that. I like to tell people that AI and machine learning is just a different way of writing software. What is the output of a training scenario, right? It's a model and a model is just code. And so I think that as all of these different, parts of the business figure out how do we leverage these technologies, what it is, is it's a different way of writing software and it's not necessarily going to replace traditional software development, but it's going to augment it, it's going to let you do other interesting things and so, where are things going? I think we're going to continue to start coalescing around what are the right ways to do things. Right now we talk about, you know, ML Ops and how development and the frameworks and all of this innovation. There's so much innovation, which means that the industry is moving so quickly that it's hard to settle on things like standards and, or at least best practices you know, at the very least. And that the best practices are changing every three months. Are they really best practices right? So I think, right, I think that as we progress and coalesce around kind of what are the right ways to do things that's really going to make customers' lives easier. Because, you know, today, if you're a software developer you know, we build a lot of software at Pure Storage right? And if you have people and developers who are familiar with how the process, how the factory functions, then their skills become portable and it becomes easier to onboard people and AI is still nothing like that right now. It's just so, so fast moving and it's so- >> Wild West kind of. >> It's not standardized. It's not industrialized, right? And so the next big frontier in all of this amazing stuff is how do we industrialize this and really make it easy to implement for organizations? >> Oil refineries, industrial Revolution. I mean, it's on that same trajectory. >> Yeah. >> Yeah, absolutely. >> Or industrial revolution. (John laughs) >> Well, we've talked a lot about the chaos and sort of we are very much at this early stage stepping way back and this can be your personal not Pure Storage opinion if you want. >> Okay. >> What in HPC or AIML I guess it all falls under the same umbrella, has you most excited? >> Ooh. >> So I feel like you're someone who sees a lot of different things. You've got a lot of customers, you're out talking to people. >> I think that there is a lot of advancement in the area of natural language processing and I think that, you know, we're starting to take things just like natural language processing and then turning them into vision processing and all these other, you know, I think the, the most exciting thing for me about AI is that there are a lot of people who are, you are looking to use these kinds of technologies to make technology more inclusive. And so- >> I love it. >> You know the ability for us to do things like automate captioning or the ability to automate descriptive, audio descriptions of video streams or things like that. I think that those are really,, I think they're really great in terms of bringing the benefits of technology to more people in an automated way because the challenge has always been bandwidth of how much a human can do. And because they were so difficult to automate and what AI's really allowing us to do is build systems whether that's text to speech or whether that's translation, or whether that's captioning or all these other things. I think the way that AI interfaces with humans is really the most interesting part. And I think the benefits that it can bring there because there's a lot of talk about all of the things that it does that people don't like or that they, that people are concerned about. But I think it's important to think about all the really great things that maybe don't necessarily personally impact you, but to the person who's not cited or to the person who you know is hearing impaired. You know, that's an enormously valuable thing. And the fact that those are becoming easier to do they're becoming better, the quality is getting better. I think those are really important for everybody. >> I love that you brought that up. I think it's a really important note to close on and you know, there's always the kind of terminator, dark side that we obsess over but that's actually not the truth. I mean, when we think about even just captioning it's a tool we use on theCUBE. It's, you know, we see it on our Instagram stories and everything else that opens the door for so many more people to be able to learn. >> Right? >> And the more we all learn, like you said the water level rises together and everything is magical. Justin, it has been a pleasure to have you on board. Last question, any more bourbon tasting today? >> Not that I'm aware of, but if you want to come by I'm sure we can find something somewhere. (all laughing) >> That's the spirit, that is the spirit of an innovator right there. Justin, thank you so much for joining us from Pure Storage. John Furrier, always a pleasure to interview with you. >> I'm glad I can contribute. >> Hey, hey, that's the understatement of the century. >> It's good to be back. >> Yeah. >> Hopefully I'll see you guys in, I'll see you guys in 2034. >> No. (all laughing) No, you've got the Pure Accelerate conference. We'll be there. >> That's right. >> We'll be there. >> Yeah, we have our Pure Accelerate conference next year and- >> Great. >> Yeah. >> I love that, I mean, feel free to, you know, hype that. That's awesome. >> Great company, great runs, stayed true to the mission from day one, all Flash, continue to innovate congratulations. >> Yep, thank you so much, it's pleasure being here. >> It's a fun ride, you are a joy to talk to and it's clear you're just as excited as we are about hardware, so thanks a lot Justin. >> My pleasure. >> And thank all of you for tuning in to this wonderfully nerdy hardware edition of theCUBE live from Dallas, Texas, where we're at, Supercomputing, my name's Savannah Peterson and I hope you have a wonderful night. (soft music)
SUMMARY :
and welcome back to Dallas Texas It's been a great show so far. We've had more hosts, more It's been a super the third event, was that right, John? Yeah, the first ever VM World, It's been too long, you I mean, I can barely remember for VMware, but the industry, the cloud, as you know, covering as well. and it's been so great to So one of the big the biggest difference is that, you know, I mean, less power consumption, in the ML and the AI for you guys? nerds at the beginning all here in Dallas. places where you know, have to go check that out. Yeah, but that to me is like one of for people to look at and the amount of of compute that we have, I love that you touched and the point of entry It's an enabler for all of the amazing but now you got, Spark and as you guys look at your products, the kind of stuff that Yeah, exactly. And so all of the new big advancements Savannah, you know, but like, do you see a hardware standpoint. the customer's lives easier? It's never stopping, it's going to let you do And so the next big frontier I mean, it's on that same trajectory. (John laughs) a lot about the chaos You've got a lot of customers, and I think that, you know, or to the person who you and you know, there's always And the more we all but if you want to come by that is the spirit of an Hey, hey, that's the Hopefully I'll see you guys We'll be there. free to, you know, hype that. all Flash, continue to Yep, thank you so much, It's a fun ride, you and I hope you have a wonderful night.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Paul Moritz | PERSON | 0.99+ |
Justin | PERSON | 0.99+ |
Justin Emerson | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Savannah Peterson | PERSON | 0.99+ |
Savannah | PERSON | 0.99+ |
Dallas | LOCATION | 0.99+ |
June | DATE | 0.99+ |
John Furrier | PERSON | 0.99+ |
12 years | QUANTITY | 0.99+ |
2010 | DATE | 0.99+ |
Kay Green | PERSON | 0.99+ |
Dallas, Texas | LOCATION | 0.99+ |
third event | QUANTITY | 0.99+ |
Dallas Texas | LOCATION | 0.99+ |
last week | DATE | 0.99+ |
12 years ago | DATE | 0.99+ |
two-thirds | QUANTITY | 0.99+ |
First | QUANTITY | 0.98+ |
VM World | EVENT | 0.98+ |
first | QUANTITY | 0.98+ |
two thirds | QUANTITY | 0.98+ |
Havana Labs | ORGANIZATION | 0.98+ |
Pure Accelerate | EVENT | 0.98+ |
next year | DATE | 0.98+ |
today | DATE | 0.98+ |
both sides | QUANTITY | 0.98+ |
Pure Storage | ORGANIZATION | 0.97+ |
first year | QUANTITY | 0.97+ |
16 different ISAs | QUANTITY | 0.96+ |
FlashBlade | TITLE | 0.96+ |
three hot areas | QUANTITY | 0.94+ |
three | QUANTITY | 0.94+ |
Snowflake | ORGANIZATION | 0.93+ |
one | QUANTITY | 0.93+ |
2034 | DATE | 0.93+ |
one thing | QUANTITY | 0.93+ |
Supercomputing | ORGANIZATION | 0.9+ |
90% less | QUANTITY | 0.89+ |
theCUBE | ORGANIZATION | 0.86+ |
agile | TITLE | 0.84+ |
VM world | EVENT | 0.84+ |
few years ago | DATE | 0.81+ |
day one | QUANTITY | 0.81+ |
Hadoop World | ORGANIZATION | 0.8+ |
VMware | ORGANIZATION | 0.79+ |
ORGANIZATION | 0.78+ | |
Spark and | ORGANIZATION | 0.77+ |
Hadoop | ORGANIZATION | 0.74+ |
years | DATE | 0.73+ |
last | DATE | 0.73+ |
three months | QUANTITY | 0.69+ |
FlashBlade | ORGANIZATION | 0.68+ |
Direct Flash | TITLE | 0.67+ |
year | DATE | 0.65+ |
tier one | QUANTITY | 0.58+ |
Supercomputing | TITLE | 0.58+ |
Direct | TITLE | 0.56+ |
Flash | ORGANIZATION | 0.55+ |
86 | TITLE | 0.55+ |
aces | QUANTITY | 0.55+ |
Pure | ORGANIZATION | 0.51+ |
Databricks | ORGANIZATION | 0.5+ |
2022 | ORGANIZATION | 0.5+ |
X | EVENT | 0.45+ |
Prakash Darji, Pure Storage
(upbeat music) >> Hello, and welcome to the special Cube conversation that we're launching in conjunction with Pure Accelerate. Prakash Darji is here, is the general manager of Digital Experience. They actually have a business unit dedicated to this at Pure Storage. Prakash, welcome back, good to see you. >> Yeah Dave, happy to be here. >> So a few weeks back, you and I were talking about the Shift 2 and as a service economy and which is a good lead up to Accelerate, held today, we're releasing this video in LA. This is the fifth in person Accelerate. It's got a new tagline techfest so you're making it fun, but still hanging out to the tech, which we love. So this morning you guys made some announcements expanding the portfolio. I'm really interested in your reaffirmed commitment to Evergreen. That's something that got this whole trend started in the introduction of Evergreen Flex. What is that all about? What's your vision for Evergreen Flex? >> Well, so look, this is one of the biggest moments that I think we have as a company now, because we introduced Evergreen and that was and probably still is one of the largest disruptions to happen to the industry in a decade. Now, Evergreen Flex takes the power of modernizing performance and capacity to storage beyond the box, full stop. So we first started on a project many years ago to say, okay, how can we bring that modernization concept to our entire portfolio? That means if someone's got 10 boxes, how do you modernize performance and capacity across 10 boxes or across maybe FlashBlade and FlashArray. So with Evergreen Flex, we first are starting to hyper disaggregate performance and capacity and the capacity can be moved to where you need it. So previously, you could have thought of a box saying, okay, it has this performance or capacity range or boundary, but let's think about it beyond the box. Let's think about it as a portfolio. My application needs performance or capacity for storage, what if I could bring the resources to it? So with Evergreen Flex within the QLC family with our FlashBlade and our FlashArray QLC projects, you could actually move QLC capacity to where you need it. And with FlashArray X and XL or TLC family, you could move capacity to where you need it within that family. Now, if you're enabling that, you have to change the business model because the capacity needs to get build where you use it. If you use it in a high performance tier, you could build at a high performance rate. If you use it as a lower performance tier, you could build at a lower performance rate. So we changed the business model to enable this technology flexibility, where customers can buy the hardware and they get a pay per use consumption model for the software and services, but this enables the technology flexibility to use your capacity wherever you need. And we're just continuing that journey of hyper disaggregated. >> Okay, so you solve the problem of having to allocate specific capacity or performance to a particular workload. You can now spread that across whatever products in the portfolio, like you said, you're disaggregating performance and capacity. So that's very cool. Maybe you could double click on that. You obviously talk to customers about doing this. They were in pain a little bit, right? 'Cause they had this sort of stovepipe thing. So talk a little bit about the customer feedback that led you here. >> Well, look, let's just say today if you're an application developer or you haven't written your app yet, but you know you're going to. Well, you need that at least say I need something, right? So someone's going to ask you what kind of storage do you need? How many IOPS, what kind of performance capacity, before you've written your code. And you're going to buy something and you're going to spend that money. Now at that point, you're going to go write your application, run it on that box and then say, okay, was I right or was I wrong? And you know what? You were guessing before you wrote the software. After you wrote the software, you can test it and decide what you need, how it's going to scale, et cetera. But if you were wrong, you already bought something. In a hyper disaggregated world, that capacity is not a sunk cost, you can use it wherever you want. You can use capacity of somewhere else and bring it over there. So in the world of application development and in the world of storage, today people think about, I've got a workload, it's SAP, it's Oracle, I've built this custom app. I need to move it to a tier of storage, a performance class. Like you think about the application and you think about moving the application. And it takes time to move the application, takes performance, takes loan, it's a scheduled event. What if you said, you know what? You don't have to do any of that. You just move the capacity to where you need it, right? >> Yep. >> So the application's there and you actually have the ability to instantaneously move the capacity to where you need it for the application. And eventually, where we're going is we're looking to do the same thing across the performance hearing. So right now, the biggest benefit is the agility and flexibility a customer has across their fleet. So Evergreen was great for the customer with one array, but Evergreen Flex now brings that power to the entire fleet. And that's not tied to just FlashArray or FlashBlade. We've engineered a data plane in our direct flash fabric software to be able to take on the personality of the system it needs to go into. So when a data pack goes into a FlashBlade, that data pack is optimized for use in that scale out architecture with the metadata for FlashBlade. When it goes into a FlashArray C it's optimized for that metadata structure. So our Purity software has made this transformative to be able to do this. And we created a business model that allowed us to take advantage of this technology flexibility. >> Got it. Okay, so you got this mutually interchangeable performance and capacity across the portfolio beautiful. And I want to come back to sort of the Purity, but help me understand how this is different from just normal Evergreen, existing evergreen options. You mentioned the one array, but help us understand that more fully. >> Well, look, so in addition to this, like we had Evergreen Gold historically. We introduced Evergreen Flex and we had Pure as a service. So you had kind of two spectrums previously. You had Evergreen Gold on one hand, which modernized the performance and capacity of a box. You had Pure as a service that said don't worry about the box, tell me how many IOPS you have and will run and operate and manage that service for you. I think we've spoken about that previously on theCUBE. >> Yep. >> Now, we have this model where it's not just about the box, we have this model where we say, you know what, it's your fleet. You're going to run and operate and manage your fleet and you could move the capacity to where you need it. So as we started thinking about this, we decided to unify our entire portfolio of sub software and subscription services under the Evergreen brand. Evergreen Gold we're renaming to Evergreen Forever. We've actually had seven customers just crossed a decade of updates Forever Evergreen within a box. So Evergreen Forever is about modernizing a box. Evergreen Flex is about modernizing your fleet and Evergreen one, which is our rebrand of Pure as a service is about modernizing your labor. Instead of you worrying about it, let us do it for you. Because if you're an application developer and you're trying to figure out, where should I put my capacity? Where should I do it? You can just sign up for the IOPS you need and let us actually deliver and move the components to where you need it for performance, capacity, management, SLAs, et cetera. So as we think about this, for us this is a spectrum and a continuum of where you're at in the modernization journey to software subscription and services. >> Okay, got it. So why did you feel like now was the right time for the rebranding and the renaming convention, what's behind? What was the thing? Take us inside the internal conversations and the chalkboard discussion? >> Well, look, the chalkboard discussion's simple. It's everything was built on the Evergreen stateless architecture where within a box, right? We disaggregated the performance and capacity within the box already, 10 years ago within Evergreen. And that's what enabled us to build Pure as a service. That's why I say like when companies say they built a service, I'm like it's not a service if you have to do a data migration. You need a stateless architecture that's disaggregated. You can almost think of this as the anti hyper-converge, right? That's going the other way. It's hyper disaggregated. >> Right. >> And that foundation is true for our whole portfolio. That was fundamental, the Evergreen architecture. And then if Gold is modernizing a box and Flex is modernizing your fleet and your portfolio and Pure as a service is modernizing the labor, it is more of a continuation in the spectrum of how do you ensure you get better with age, right? And it's like one of those things when you think about a car. Miles driven on a car means your car's getting older and it doesn't necessarily get better with age, right? What's interesting when you think about the human body, yeah, you get older and some people deteriorate with age and some people it turns out for a period of time, you pick up some muscle mass, you get a little bit older, you get a little bit wiser and you get a little bit better with age for a while because you're putting in the work to modernize, right? But where in infrastructure and hardware and technology are you at the point where it always just gets better with age, right? We've introduced that concept 10 years ago. And we've now had proven industry success over a decade, right? As I mentioned, our first seven customers who've had a decade of Evergreen update started with an FA-300 way back when, and since then performance and capacity has been getting better over time with Evergreen Forever. So this is the next 10 years of it getting better and better for the company and not just tying it to the box because now we've grown up, we've got customers with like large fleets. I think one of our customers just hit 900 systems, right? >> Wow. >> So when you have 900 systems, right? And you're running a fleet you need to think about, okay, how am I using these resources? And in this day and age in that world, power becomes a big thing because if you're using resources inefficiently and the cost of power and energy is up, you're going to be in a world of hurt. So by using Flex where you can move the capacity to where it's needed, you're creating the most efficient operating environment, which is actually the lowest power consumption environment as well. >> Right. >> So we're really excited about this journey of modernizing, but that rebranding just became kind of a no brainer to us because it's all part of the spectrum on your journey of whether you're a single array customer, you're a fleet customer, or you don't want to even run, operate and manage. You can actually just say, you know what, give me the guarantee in the SLA. So that's the spectrum that informed the rebranding. >> Got it. Yeah, so to your point about the human body, all you got to do is look at Tom Brady's NFL combine videos and you'll see what a transformation. Fine wine is another one. I like the term hyper disaggregated because that to me is consistent with what's happening with the cloud and edge. We're building this hyper distributed or disaggregated system. So I want to just understand a little bit about you mentioned Purity so there's this software obviously is the enabler here, but what's under the covers? Is it like a virtualizer or megaload balancer, metadata manager, what's the tech behind this? >> Yeah, so we'll do a little bit of a double tech, right? So we have this concept of drives where in Purity, we build our own software for direct flash that takes the NAND and we do the NAND management as we're building our drives in Purity software. Now ,that advantage gives us the ability to say how should this drive behave? So in a FlashArray C system, it can behave as part of a FlashArray C and its usable capacity that you can write because the metadata and some of the system information is in NVRAM as part of the controller, right? So you have some metadata capability there. In a legend architecture for example, you have a distributed Blade architecture. So you need parts of that capacity to operate almost like a single layer chip where you can actually have metadata operations independent of your storage operations that operate like QLC. So we actually manage the NAND in a very very different way based on the persona of the system it's going into, right? So this capacity to make it usable, right? It's like saying a competitor could go ahead name it, Dell that has power max in Isilon, HPE that has single store and three power and nimble and like you name, like can you really from a technology standpoint say your capacity can be used anywhere or all these independent systems. Everyone's thinking about the world like a system, like here's this system, here's that system, here's that system. And your capacity is locked into a system. To be able to unlock that capacity to the system, you need to behave differently with the media type in the operating environment you're going into and that's what Purity does, right? So we are doing that as part of our direct Flex software around how we manage these drives to enable this. >> Well, it's the same thing in the cloud precaution, right? I mean, you got different APIs and primitive for object, for block, for file. Now, it's all programmable infrastructure so that makes it easier, but to the point, it's still somewhat stovepipe. So it's funny, it's good to see your commitment to Evergreen, I think you're right. You lay down the gauntlet a decade plus ago. First everybody ignored you and then they kind of laughed at you, then they criticized you, and then they said, oh, then you guys reached the escape velocity. So you had a winning hand. So I'm interested in that sort of progression over the past decade where you're going, why this is so important to your customers, where you're trying to get them ultimately. >> Well, look, the thing that's most disappointing is if I bought 100 terabytes still have to re-buy it every three or five years. That seems like a kind of ridiculous proposition, but welcome to storage. You know what I mean? That's what most people do with Evergreen. We want to end data migrations. We want to make sure that every software updates, hardware updates, non disruptive. We want to make it easy to deploy and run at scale for your fleet. And eventually we want everyone to move to our Evergreen one, formerly Pure as a service where we can run and operate and manage 'cause this is all about trust. We're trying to create trust with the customer to say, trust us, to run and operate and scale for you and worry about your business because we make tech easy. And like think about this hyper disaggregated if you go further. If you're going further with hyper disaggregated, you can think about it as like performance and capacity is your Lego building blocks. Now for anyone, I have a son, he wants to build a Lego Death Star. He didn't have that manual, he's toast. So when you move to at scale and you have this hyper disaggregated world and you have this unlimited freedom, you have unlimited choice. It's the problem of the cloud today, too much choice, right? There's like hundreds of instances of this, what do I even choose? >> Right. >> Well, so the only way to solve that problem and create simplicity when you have so much choice is put data to work. And that's where Pure one comes in because we've been collecting and we can scan your landscape and tell you, you should move these types of resources here and move those types of resources there, right? In the past, it was always about you should move this application there or you should move this application there. We're actually going to turn the entire industry on it's head. It's not like applications and data have gravity. So let's think about moving resources to where that are needed versus saying resources are a fixed asset, let's move the applications there. So that's a concept that's new to the industry. Like we're creating that concept, we're introducing that concept because now we have the technology to make that reality a new efficient way of running storage for the world. Like this is that big for the company. >> Well, I mean, a lot of the failures in data analytics and data strategies are a function of trying to jam everything into a single monolithic system and hyper centralize it. Data by its very nature is distributed. So hyper disaggregated fits that model and the pendulum's clearly swinging to that. Prakash, great to have you, purestorage.com I presume is where I can learn more? >> Oh, absolutely. We're super excited and our pent up by demand I think in this space is huge so we're looking forward to bringing this innovation to the world. >> All right, hey, thanks again. Great to see you, I appreciate you coming on and explaining this new model and good luck with it. >> All right, thank you. >> All right, and thanks for watching. This is David Vellante, and appreciate you watching this Cube conversation, we'll see you next time. (upbeat music)
SUMMARY :
is the general manager So this morning you guys capacity to where you need it. in the portfolio, like you So someone's going to ask you the capacity to where you and capacity across the the box, tell me how many IOPS you have capacity to where you need it. and the chalkboard discussion? if you have to do a data migration. and technology are you at the point So when you have 900 systems, right? So that's the spectrum that disaggregated because that to me and like you name, like can you really So you had a winning hand. and you have this hyper and create simplicity when you have and the pendulum's to bringing this innovation to the world. appreciate you coming on and appreciate you watching
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David Vellante | PERSON | 0.99+ |
Evergreen | ORGANIZATION | 0.99+ |
Prakash | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
LA | LOCATION | 0.99+ |
10 boxes | QUANTITY | 0.99+ |
10 boxes | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
Accelerate | ORGANIZATION | 0.99+ |
Prakash Darji | PERSON | 0.99+ |
today | DATE | 0.99+ |
Tom Brady | PERSON | 0.99+ |
900 systems | QUANTITY | 0.99+ |
100 terabytes | QUANTITY | 0.99+ |
Lego | ORGANIZATION | 0.99+ |
Pure Accelerate | ORGANIZATION | 0.99+ |
five years | QUANTITY | 0.99+ |
seven customers | QUANTITY | 0.99+ |
first seven customers | QUANTITY | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
10 years ago | DATE | 0.98+ |
Evergreen Gold | ORGANIZATION | 0.98+ |
Evergreen Forever | ORGANIZATION | 0.98+ |
First | QUANTITY | 0.98+ |
one array | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
fifth | QUANTITY | 0.97+ |
purestorage.com | OTHER | 0.95+ |
single | QUANTITY | 0.95+ |
Forever Evergreen | ORGANIZATION | 0.94+ |
first | QUANTITY | 0.93+ |
Evergreen Flex | ORGANIZATION | 0.93+ |
single layer | QUANTITY | 0.93+ |
FlashArray C | TITLE | 0.91+ |
single store | QUANTITY | 0.91+ |
two spectrums | QUANTITY | 0.9+ |
a decade plus ago | DATE | 0.9+ |
TLC | ORGANIZATION | 0.89+ |
NFL | ORGANIZATION | 0.89+ |
single array | QUANTITY | 0.88+ |
three | QUANTITY | 0.87+ |
FA-300 | COMMERCIAL_ITEM | 0.87+ |
SAP | ORGANIZATION | 0.85+ |
hundreds of instances | QUANTITY | 0.83+ |
past | DATE | 0.83+ |
over a decade | QUANTITY | 0.82+ |
double | QUANTITY | 0.8+ |
Shift 2 | TITLE | 0.79+ |
Purity | TITLE | 0.79+ |
FlashBlade | COMMERCIAL_ITEM | 0.78+ |
Death Star | COMMERCIAL_ITEM | 0.78+ |
Miles | QUANTITY | 0.77+ |
next 10 years | DATE | 0.73+ |
Pure | COMMERCIAL_ITEM | 0.73+ |
Isilon | LOCATION | 0.73+ |
every three | QUANTITY | 0.73+ |
this morning | DATE | 0.72+ |
a decade | QUANTITY | 0.71+ |
Purity | ORGANIZATION | 0.71+ |
a few weeks back | DATE | 0.71+ |
Pure | ORGANIZATION | 0.69+ |
Matt Burr, Pure Storage
(Intro Music) >> Hello everyone and welcome to this special cube conversation with Matt Burr who is the general manager of FlashBlade at Pure Storage. Matt, how you doing? Good to see you. >> I'm doing great. Nice to see you again, Dave. >> Yeah. You know, welcome back. We're going to be broadcasting this is at accelerate. You guys get big news. Of course, FlashBlade S we're going to dig into it. The famous FlashBlade now has new letter attached to it. Tell us what it is, what it's all about. >> (laughing) >> You know, it's easy to say. It's just the latest and greatest version of the FlashBlade, but obviously it's a lot more than that. We've had a lot of success with FlashBlade kind of across the board in particular with Meta and their research super cluster, which is one of the largest AI super clusters in the world. But, it's not enough to just build on the thing that you had, right? So, with the FlashBlade S, we've increased modularity, we've done things like, building co-design software and hardware and leveraging that into something that increases, or it actually doubles density, performance, power efficiency. On top of that, you can scale independently, storage, networking, and compute, which is pretty big deal because it gives you more flexibility, gives you a little more granularity around performance or capacity, depending on which direction you want to go. And we believe that, kind of the end of this is fundamentally the, I guess, the way to put it is sort of the highest performance and capacity optimization, unstructured data platform on the market today without the need for, kind of, an expensive data tier of cash or expected data cash and tier. So we're pretty excited about, what we've ended up with here. >> Yeah. So I think sometimes people forget, about how much core engineering Meta does. Facebook, you go on Facebook and play around and post things, but yeah, their backend cloud is just amazing. So talk a little bit more about the problem targets for FlashBlade. I mean, it's pretty wide scope and we're going to get into that, but what's the core of that. >> Yeah. We've talked about that extensively in the past, the use cases kind of generally remain the same. I know, we'll probably explore this a little bit more deeply, but you know, really what we're talking about here is performance and scalability. We have written essentially an unlimited Metadata software level, which gives us the ability to expand, we're already starting to think about computing an exabyte scale. Okay. So, the problem that the customer has of, Hey, I've got a Greenfield, object environment, or I've got a file environment and my 10 K and 7,500 RPM disc is just spiraling out of control in my environment. It's an environmental problem. It's a management problem, we have effectively, simplified the process of bringing together highly performant, very large multi petabyte to eventually exabyte scale unstructured data systems. >> So people are obviously trying to inject machine intelligence, AI, ML into applications, bring data into applications, bringing those worlds closer together. Analytics is obviously exploding. You see some other things happening in the news, read somewhere, protection and the like, where does FlashBlade fit in terms of FlashBlade S in some terms of some of these new use cases. >> All those things, we're only going wider and broader. So, we've talked in the past about having a having a horizontal approach to this market. The unstructured data market has often had vertical specificity. You could see successful infrastructure companies in oil and gas that may not play median entertainment, where you see, successful companies that play in media entertainment, but don't play well in financial services, for example. We're sort of playing the long game here with this and we're focused on, bringing an all Q L C architecture that combines our traditional kind of pure DFM with the software that is, now I guess seven years hardened from the original FlashBlade system. And so, when we look at customers and we look at kind of customers in three categories, right, we have customers that sort of fit into a very traditional, more than three, but kind of make bucketized this way, customers that fit into kind of this EDA HPC space, then you have that sort of data protection, which I believe kind of ransomware falls under that as well. The world has changed, right? So customers want their data back faster. Rapid restore is a real thing, right? We have customers that come to us and say, anybody can back up my data, but if I want to get something back fast and I mean in less than a week or a couple days, what do I do? So we can solve that problem. And then as you sort of accurately pointed out where you started, there is the AI ML side of things where the Invidia relationship that we have, right. DGX is are a pretty powerful weapon in that market and solving those problems. But they're not cheap. And keeping those DGX's running all the time requires an extremely efficient underpinning of a flash system. And we believe we have that market as well. >> It's interesting when pure was first coming out as a startup, you obviously had some cool new tech, but you know, your stack wasn't as hard. And now you've got seven years under your belt. The last time you were on the cube, we talked about some of the things that you guys were doing differently. We talked about UFFO, unified fast file and object. How does this new product, FlashBlade S, compare to some previous generations of FlashBlade in terms of solving unstructured data and some of these other trends that we've been talking about? >> Yeah. I touched on this a little bit earlier, but I want to go a little bit deeper on this concept of modularity. So for those that are familiar with Pure Storage, we have what's called the evergreen storage program. It's not as much a program as it is an engineering philosophy. The belief that everything we build should be modular in nature so that we can have essentially a chassi that has an a 100% modular components inside of it. Such that we can upgrade all of those features, non disruptively from one version to the next, you should think about that as you know, if you have an iPhone, when you go get a new iPhone, what do you do with your old iPhone? You either throw it away or you sell it. Well, imagine if your iPhone just got newer and better each time you renewed your, whatever it is, two year or three year subscription with apple. That's effectively what we have as a core philosophy, core operating engineering philosophy within pure. That is now a completely full and robust program with this instantiation of the FlashBlade S. And so kind of what that means is, for a customer I'm future proofed for X number of years, knowing that we have a run rate of being able to keep customers on the flash array side from the FA 400 all the way through the flash array X and Excel, which is about a 10 year time span. So, that then, and of itself sort of starts to play into customers that have concerns around ESG. Right? Last time I checked power space and cooling, still mattered in data center. So although I have people that tell me all the time, power space clearly doesn't matter anymore, but I know at the end of the day, most customers seem to say that it does, you're not throwing away refrigerator size pieces of equipment that once held spinning disc, something that's a size of a microwave that's populated with DFMs with all LC flash that you can actually upgrade over time. So if you want to scale more performance, we can do that through adding CPU. If you want to scale more capacity, we can do that through adding more And we're in control of those parameters because we're building our own DFM, our direct fabric modules on our own storage notes, if you will. So instead of relying on the consumer packaging of an SSD, we're upgrading our own stuff and growing it as we can. So again, on the ESG side, I think for many customers going into the next decade, it's going to be a huge deal. >> Yeah. Interesting comments, Matt. I mean, I don't know if you guys invented it, but you certainly popularize the idea of, no Fort lift upgrades and sort of set the industry on its head when you guys really drove that evergreen strategy and kind of on that note, you guys talk about simplicity. I remember last accelerate went deep with cause on your philosophy of keeping things simple, keeping things uncomplicated, you guys talk about using better science to do that. And you a lot of talk these days about outcomes. How does FlashBlade S support those claims and what do you guys mean by better science? >> Yeah. You know, better science is kind of a funny term. It was an internal term that I was on a sales call actually. And the customer said, well, I understand the difference between these two, but could you tell me how we got there and I was a little stumped on the answer. And I just said, well, I think we have better scientists and that kind of morphed into better science, a good example of that is our Metadata architecture, right? So our scalable Metadata allows us to avoid having that cashing tier, that other architectures have to rely on in order to anticipate, which files are going to need to be in read cash and read misses become very expensive. Now, a good follow up question there, not to do your job, but it's the question that I always get is, well, when you're designing your own hardware and your own software, what's the real material advantage of that? Well, the real material advantage of that is that you are in control of the combination and the interaction of those two things you don't give up the sort of the general purpose nature, if you will, of the performance characteristics that come along with things like commodity, you get a very specific performance profile. That's tailored to the software that's being married to it. Now in some instances you could say, well, okay, does that really matter? Well, when you start to talking about 20, 40, 50, 100, 500, petabyte data sets, every percentage matters. And so those individual percentages equate to space savings. They equate to power and cooling savings. We believe that we're going to have industry best dollars per lot. We're going to have industry best, kind of dollar PRU. So really the whole kind of game here is a round scale. >> Yeah. I mean, look, there's clearly places for the pure software defined. And then when cloud first came out, everybody said, oh, build the cloud and commodity, they don't build custom art. Now you see all the hyper scalers building custom software, custom hardware and software integration, custom Silicon. So co-innovation between hardware and software. It seems pretty as important, if not more important than ever, especially for some of these new workloads who knows what the edge is going to bring. What's the downside of not having that philosophy in your view? Is it just, you can't scale to the degree that you want, you can't support the new workloads or performance? What should customers be thinking about there? >> I think the downside plays in two ways. First is kind of the future and at scale, as I alluded to earlier around cost and just savings over time. Right? So if you're using a you know a commodity SSD, there's packaging around that SSD that is wasteful both in terms of- It's wasteful in the environmental sense and wasteful in the sort of computing performance sense. So that's kind of one thing. On the second side, it's easier for us to control the controllables around reliability when you can eliminate the number of things that actually sit in that workflow and by workflow, I mean when a right is acknowledged from a host and it gets down to the media, the more control you have over that, the more reliability you have over that piece. >> Yeah. I know. And we talked about ESG earlier. I know you guys, I'm going to talk a little bit about more news from accelerate within Invidia. You've certainly heard Jensen talk about the wasted CPU cycles in the data center. I think he's forecasted, 25 to 30% of the cycles are wasted on doing things like storage offload, or certainly networking and security. So now it sort of confirms your ESG thought, we can do things more efficiently, but as it relates to Invidia and some of the news around AIRI's, what is the AI RI? What's that stand for? What's the high level overview of AIRI. >> So the AIRI has been really successful for both us and Invidia. It's a really great partnership we're appreciative of the partnership. In fact, Tony pack day will be speaking here at accelerate. So, really looking forward to that, Look, there's a couple ways to look at this and I take the macro view on this. I know that there's a equally as good of a micro example, but I think the macro is really kind of where it's at. We don't have data center space anymore, right? There's only so many data centers we can build. There's only so much power we can create. We are going to reach a point in time where municipalities are going to struggle against the businesses that are in their municipalities for power. And now you're essentially bidding big corporations against people who have an electric bill. And that's only going to last so long, you know who doesn't win in that? The big corporation doesn't win in that. Because elected officials will have to find a way to serve the people so that they can get power. No matter how skewed we think that may be. That is the reality. And so, as we look at this transition, that first decade of disc to flash transition was really in the block world. The second decade, which it's really fortunate to have a multi decade company, of course. But the second decade of riding that wave from disk to flash is about improving space, power, efficiency, and density. And we sort of reach that, it's a long way of getting to the point about iMedia where these AI clusters are extremely powerful things. And they're only going to get bigger, right? They're not going to get smaller. It's not like anybody out there saying, oh, it's a Thad, or, this isn't going to be something that's going to yield any results or outcomes. They yield tremendous outcomes in healthcare. They yield tremendous outcomes in financial services. They use tremendous outcome in cancer research, right? These are not things that we as a society are going to give up. And in fact, we're going to want to invest more on them, but they come at a cost and one of the resources that is required is power. And so when you look at what we've done in particular with Invidia. You found something that is extremely power efficient that meets the needs of kind of going back to that macro view of both the community and the business. It's a win-win. >> You know and you're right. It's not going to get smaller. It's just going to continue to in momentum, but it could get increasingly distributed. And you think about, I talked about the edge earlier. You think about AI inferencing at the edge. I think about Bitcoin mining, it's very distributed, but it consumes a lot of power and so we're not exactly sure what the next level architecture is, but we do know that science is going to be behind it. Talk a little bit more about your Invidia relationship, because I think you guys were the first, I might be wrong about this, but I think you were the first storage company to announce a partnership with Invidia several years ago, probably four years ago. How is this new solution with a AIRI slash S building on that partnership? What can we expect with Invidia going forward? >> Yeah. I think what you can expect to see is putting the foot on the gas on kind of where we've been with Invidia. So, as I mentioned earlier Meta is by some measurements, the world's largest research super cluster, they're a huge Invidia customer and built on pure infrastructure. So we see kind of those types of well reference architectures, not that everyone's going to have a Meta scale reference architecture, but the base principles of what they're solving for are the base principles of what we're going to begin to see in the enterprise. I know that begin sounds like a strange word because there's already a big business in DGX. There's already a sizable business in performance, unstructured data. But those are only going to get exponentially bigger from here. So kind of what we see is a deepening and a strengthening of the of the relationship and opportunity for us to talk, jointly to customers that are going to be building these big facilities and big data centers for these types of compute related problems and talking about efficiency, right? DGX are much more efficient and Flash Blades are much more efficient. It's a great pairing. >> Yeah. I mean you're definitely, a lot of AI today is modeling in the cloud, seeing HPC and data just slam together all kinds of new use cases. And these types of partnerships are the only way that we're going to solve the future problems and go after these future opportunities. I'll give you a last word you got to be excited with accelerate, what should people be looking for, add accelerate and beyond. >> You know, look, I am really excited. This is going on my 12th year at Pure Storage, which has to be seven or eight accelerates whenever we started this thing. So it's a great time of the year, maybe take a couple off because of because of COVID, but I love reconnecting in particular with partners and customers and just hearing kind of what they have to say. And this is kind of a nice one. This is four years or five years worth of work for my team who candidly I'm extremely proud of for choosing to take on some of the solutions that they, or excuse me, some of the problems that they chose to take on and find solutions for. So as accelerate roles around, I think we have some pretty interesting evolutions of the evergreen program coming to be announced. We have some exciting announcements in the other product arenas as well, but the big one for this event is FlashBlade. And I think that we will see. Look, no one's going to completely control this transition from disc to flash, right? That's a that's a macro trend. But there are these points in time where individual companies can sort of accelerate the pace at which it's happening. And that happens through cost, it happens through performance. My personal belief is this will be one of the largest points of those types of acceleration in this transformation from disc to flash and unstructured data. This is such a leap. This is essentially the equivalent of us going from the 400 series on the block side to the X, for those that you're familiar with the flash array lines. So it's a huge, huge leap for us. I think it's a huge leap for the market. And look, I think you should be proud of the company you work for. And I am immensely proud of what we've created here. And I think one of the things that is a good joy in life is to be able to talk to customers about things you care about. I've always told people my whole life, inefficiency is the bane of my existence. And I think we've rooted out ton of inefficiency with this product and looking forward to going and reclaiming a bunch of data center space and power without sacrificing any performance. >> Well congratulations on making it into the second decade. And I'm looking forward to the orange and the third decade, Matt Burr, thanks so much for coming back in the cubes. It's good to see you. >> Thanks, Dave. Nice to see you as well. We appreciate it. >> All right. And thank you for watching. This is Dave Vellante for the Cube. And we'll see you next time. (outro music)
SUMMARY :
Good to see you. to see you again, Dave. We're going to be broadcasting kind of the end of this the problem targets for FlashBlade. in the past, the use cases kind of happening in the news, We have customers that come to us and say, that you guys were doing differently. that tell me all the time, and kind of on that note, the general purpose nature, if you will, to the degree that you want, First is kind of the future and at scale, and some of the news around AIRI's, that meets the needs of I talked about the edge earlier. of the of the relationship are the only way that we're going to solve of the company you work for. and the third decade, Nice to see you as well. This is Dave Vellante for the Cube.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Matt Burr | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Invidia | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
100% | QUANTITY | 0.99+ |
25 | QUANTITY | 0.99+ |
AIRI | ORGANIZATION | 0.99+ |
seven years | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
10 K | QUANTITY | 0.99+ |
four years | QUANTITY | 0.99+ |
seven | QUANTITY | 0.99+ |
Excel | TITLE | 0.99+ |
three year | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
12th year | QUANTITY | 0.99+ |
7,500 RPM | QUANTITY | 0.99+ |
Matt | PERSON | 0.99+ |
two year | QUANTITY | 0.99+ |
apple | ORGANIZATION | 0.99+ |
less than a week | QUANTITY | 0.99+ |
first decade | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
seven years | QUANTITY | 0.99+ |
second side | QUANTITY | 0.99+ |
eight | QUANTITY | 0.99+ |
second decade | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
40 | QUANTITY | 0.99+ |
four years ago | DATE | 0.99+ |
more than three | QUANTITY | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
100 | QUANTITY | 0.98+ |
next decade | DATE | 0.98+ |
two ways | QUANTITY | 0.98+ |
50 | QUANTITY | 0.98+ |
one version | QUANTITY | 0.98+ |
several years ago | DATE | 0.98+ |
30% | QUANTITY | 0.98+ |
two | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
Tony | PERSON | 0.97+ |
two things | QUANTITY | 0.97+ |
500 | QUANTITY | 0.97+ |
Pure Storage | ORGANIZATION | 0.97+ |
FlashBlade | TITLE | 0.97+ |
today | DATE | 0.94+ |
third decade | QUANTITY | 0.94+ |
FlashBlade | EVENT | 0.94+ |
a couple days | QUANTITY | 0.9+ |
first storage company | QUANTITY | 0.88+ |
each time | QUANTITY | 0.88+ |
ESG | ORGANIZATION | 0.87+ |
Jensen | PERSON | 0.85+ |
DGX | ORGANIZATION | 0.85+ |
FlashBlade S | TITLE | 0.85+ |
three categories | QUANTITY | 0.85+ |
FlashBlade S | COMMERCIAL_ITEM | 0.82+ |
about a 10 year | QUANTITY | 0.82+ |
400 series | QUANTITY | 0.78+ |
Pure//Launch | Pure Storage
(electronic music) >> The cloud is evolving. You know, it's no longer just a set of remote services accessed through a public cloud. Rather, it's expanding to on-premises, to multiple premises, across clouds, and eventually out to the edge. The challenge for customers is how to treat these locations as one. The opportunity for technology companies is to make that as simple as possible from an operational perspective. Welcome to this CUBE program where we're featuring Pure Storage in its latest innovations in bringing infrastructure and applications more closely together, fusing them, if you will. And today, we have a two-part program. First, we're going to hear from Rob Lee who's the CTO of Pure Storage and then my colleague John Walls is going to talk to Scott Sinclair of Enterprise Strategy Group. Scott will provide his expert analysis on infrastructure modernization and what to expect in today's changing world. So joining me right now is Rob Lee, CTO of Pure Storage. Welcome, Rob, good to see you. >> Good to see you again too, Dave. >> So take us through the announcements from today at a high level. What's most exciting about what you're delivering? >> Yeah, absolutely. So as you know, many announcement today, many things to discuss. But overall, I think what's most exciting is it's the expansion of our ability to help customers along the modern data journey. We've always thought of the journey to modern data as being formed by three pillars, if you will, certainly, modernizing infrastructure, modernizing operations and applications. And today's announcements are really in that kind of middle category of, like you said, bringing infrastructures and applications a lot more closely together. We've been modernizing infrastructure since day one, probably, people best know us for that and today's announcements are really about tackling that operations piece, bringing infrastructure and code and applications more closely together. So when we think about Pure Fusion, for example, that's really a huge step forward in how we're enabling our customers to manage large fleets of infrastructure, products, and components to deliver those services in a more automated, more tightly-integrated, seamlessly transparently delivered way to the applications that they serve, whether these services are being delivered by many different arrays in one location, many different arrays in different data center locations, or between the premise, on-premise environment and the cloud environment. Likewise, on the application front, when we think about today's announcements in Portworx Data Services, that's really all about how do we make the run and operate steps of a lot of the application building blocks that cloud-native developers are using and relying on, the database applications that are most poplar in open source, Cassandra, Mongo, so on and so forth, how dow we make the run and operate pieces of those applications a lot more intuitive, a lot more easily deployed, scaled, managed, monitored for those app developers? And so a ton of momentum. It's a big step forward on that front. And then right in the middle, when we think about today's announcements in Pure One, that's really all about how do we create more visibility, connecting the monitoring and management of the infrastructure running the apps and bring those closer together? So when we think about the visibility, we're now able to deliver for Portworx topologies allowing developers and DevOps teams to look at the entire tech stack, if you will, of a container environment from the application to the containers, to the Kubernetes cluster, to the compute nodes, all the way down to the storage, and be able to see everything that's going on, the root cause of any sort of problems that come up, that again, that's all in service of bringing infrastructure and applications a lot more closely together. So that's really how I view it and like I said, that's really the next step in our journey of helping customers modernize between infrastructure, operations, and their applications. >> Okay, so you got the control plane piece which is all about the operating model, you've got Pure One, you mentioned that which is for monitoring, you've got the Portworx piece which brings sort of development and deployment together in both infrastructure as code and better understanding of that full stack of, like you say, from applications through the clusters, the containers, all the way down to the storage. So I feel like it's not even the storage anymore. I mean, it's cloud. (chuckling) >> It is and you know, I chuckle a little bit because at the end of the day, we deliver storage but what customers are looking for is, and what they value and what they care about is their data. Now obviously, the storage is in service of the data and what we're doing with today's announcements is, again, just making it, extending our reach, helping customers work with their data a couple more steps down the road beyond just serving the bits and bytes of the storage but now getting into how do we connect the data that's sitting on our storage more quickly, get it, you know, in the hands of developers and the applications more seamlessly and more fluidly across these different environments. >> How does this news fit into Pure's evolution as a company? I mean, I don't see it as a pivot because a pivot's like, okay, we're going to go from here and now we're doin' this? >> Rob: Yeah, we were doing this, now we're doing that, right. >> And so it's more like a reinvention or a progression of the vision and the strategy. Can you talk to that? >> Absolutely. You know what, I think between those two words, I would say it's a progression, it's a next step in the journey as opposed to a reinvention. And again, I go back to, you know, I go back to the difference between storage and data and how customers are using data. We've been on a long-term path, long-term journey to continue to help customers modernize how they work with data, the results they're able to drive from the data. We got our start in infrastructure and just, you know, if you want to do bleeding edge things with data, you're not going to do it on decades-old infrastructure. So let's fix that component first, that's how we got our start. Today's announcement are really the next couple of steps along that journey. How do we make the core infrastructure more easily delivered, more flexible to operate, more automated in the hands of not just the DevOps teams, the IT teams, but the application developers? How do we deliver infrastructure more seamlessly as code? Well, why is that important? It's important because what customers are looking for out of their data is both speeds and feeds, the traditional kind of measures, bandwidth, iOps, latency, that sort of thing, but they're looking for speed of agility. You look at the modern application space around how data's being processed, it's a very, very fast-moving application space. The databases that are being used today may be different than the ones being used three months from now or six months from now. And so developers, application teams are looking for a ton more flexibility, a ton more agility than they were three, five, 10, 15 years ago. The other aspect is simplicity and reliability. As you know, that's a core component of everything we do. Our core products, you know, our arrays, our storage appliances, we're very well-known for the simplicity and reliability we drive at the individual product level. Well, as we scale and look at larger environments, as we look at customers' expectations for what they expect from a cloud-like service, there's the next level of scale and how we deliver that simplicity and reliability. And what do I mean by that? Well, a large enterprise customer who wants to operate like a cloud, wants to be able to manage large fleets of infrastructure resources, be able to package them up, deliver infrastructure services to their internal customers, they want to be able to do it in a self-service, policy-driven, easy to control, easy to manage way and that's the next level of fleet level simplicity and that's really what Pure Fusion is about is allowing operators that control plane to specify those attributes and how that service should be delivered. Same thing with Portworx, if we think about simplicity and reliability, containers, cloud-native applications, micro services, a lot of benefits there. A very fast-moving space, you can mix and match components, put them together very easily, but what goes hand in hand with that is now a need for a greater degree of simplicity 'cause you have more moving parts, and a greater need for reliability because, well now, you're not just serving one application but 30 or 40 working in unison. And that's really what we're after with Portworx and Portworx Data Services and the evolution of that family. So getting back to your original question, I really look at today's announcements as not a pivot, not a reinvention, but the next logical steps in our long-term journey to help customers modernize everything they do around data. >> Right, thanks for that, Rob. Hey, I want to switch topics. So virtually every infrastructure player now has an as-a-service offering and there're lots of claims out there about who was first, who's the best, et cetera. What's Pure's position on this topic? You claim you're ahead of the pack in delivering subscription and as-a-service offerings in the storage industry. You certainly were first with Evergreen. That was sort of a real change in how folks delivered. What about as-a-service and Pure as-a-service? What gives you confidence that you have the right approach and you're lead in the industry in this regard? >> Yeah, absolutely. I mean, I think of, first and foremost, we think of everything we do at Pure as a service and whether that's delivering products and helping customers to run and operate in an as-a-service model internally, or whether it's Pure taking on more of that run and operate as-a-service, ourselves, with Pure as a service. And so the second part of your question which is what is it that sets us apart, what are we doing differently, what gives us confidence that this is the right path, well, fundamentally, I think the difference is obviously this is a, you know, a hotter topic in the industry of late, but I think the difference is between us and the competitive set is we really look at this as a product and technology-led philosophy and strategy and we have since day one. And I think that's different than a lot of others in the industry who look at it as a little bit more of a packaging exercise between financial services, professional services, wrap it up in T(s) and C(s) and you call it a service. And what do I mean by that? So, you know, if you look internally at Pure, everything we do we think of as a service. We have a business unit organized around it, we have an engineering team, significant resources dedicated to it and building out service offerings. When we think about why this is technology-led, I think of a service. For something to be thought of as a service, it's got to be flexible, it's got to be adaptable. I've got to be able to grow as a customer and evolve as I need, whether that's changing needs in terms of performance and capacity, I've got to be able to do that without being locked into day-one, rigid kind of static some lands of having the capacity planned or plan out what my user's going to look like 18 months from now. I've got to be able to move and evolve and grow without disruption, right? You know, it's not a service if you're going to make me do a data migration or take a down time. And so when I net all that out, what are the things that you need the attributes that you need to be able to deliver a service? Well, you need a product set that is going to be able to be highly malleable, highly flexible, highly evolvable. You need something that's going to be able to cover the entire gamut of needs, whether it's price performance, tiers, you know, high performance capacity, lower cost, price points. You need something that's got a rich set of capabilities whether it's access protocols, file block object, whether it's data protection properties, you know, replications, snapshots, ransomware protection. So you need that full suite of capabilities but in order to deliver it as a service and enable me, as a customer, to seamlessly grow and change, that's got to be delivered on a very tight set of technology that can be repurposed and configured in different ways. You can't do this on 17 different products (chuckling) and expect me to change and move every single time that I have a service need change. And so when I net that out, that puts us in an absolutely differentiated position to be able to deliver this because again, everything we do is based on two core product families, Portworx adds a third. We're able to deliver all of the major storage protocols, all of the data protection capabilities across all of the price performance and service tiers, and we're able to do this on a very tight code base. And as you know, everything we do is completely non-disruptive so all of the elements really add up in our favor. And like I said, this is a huge area of a strategic focus for us. >> So these offerings, they're all part of the service-driven component of your portfolio, is that correct? >> Absolutely, yep. >> Great. You talk all the time about modern data experiences, modern application, the modern data changing the way customers think about infrastructure. What exactly does that mean and how are you driving that? >> Well, I think it means a couple of different things, but if I were to net it out, it's a greater demand for agility, a greater demand for flexibility and optionality. And if we look at why that is, you know, when I talk to customers, as they think about an infrastructure, largely, they think about their existing application demands and needs, what they're spending 90% of their time and budget dealing with today, and then the new stuff that they're getting more and more pressured to go off and build and support which is oftentimes the more strategic initiatives that they have to serve, so they're kind of balancing both worlds. And in the new world of modern applications, it's much more dynamic, meaning the application sets that are being deployed are changing all the time, the environments and what the infrastructure needs to deliver has to change more quickly in terms of scaling up, down, growing, it has to be a lot more elastic, and has much more variance. And what I mean by that is you look at a modern, cloud-native, micro services architecture-type application, it's really, you know, 20, 30, 40 different applications all working in concert with one another under the hood. This is a very different infrastructure demand than your more traditional application set. Back in the day, you have an Oracle application, you go design an environment for that. It's a big exercise, but once you put it in place, it has its own lifecycle. These days with modern applications, it's not just one application, it's 20 or 30, you've got to support all of them working in unison, you don't want to build separate infrastructures for each piece, and that set of 20 or 30 applications is changing very rapidly as open source ecosystem moves forward, as the application space moves forward. And so when customers think about the change in demands and infrastructure, this is kind of what they're thinking about and having to juggle. And so that, at the end of the day, drives them to demand much more flexibility in their infrastructure being able to use it for many different purposes, much more agility being able to adapt very, very quickly, and much more variance or dynamic range, the ability to support many different needs on the same set of infrastructure. And this is where we see very, very strong demand indicators and we're very invested in meeting these needs because they fit very well with our core product principles. >> Great, thank you for that. I really like that answer because it's not just a bunch of slideware mumbo-jumbo. You actually put some substance on it. Rob, we're going to have to leave it there. Thanks so much for joining us today. >> Thank you. >> And look forward to havin' you back soon. Now, in a moment, Scott Sinclair who's a senior analyst at Enterprise Strategy Group speaks with theCUBE's John Walls to give you the independent analyst's take. You're watching theCUBE, your global leader in high tech coverage. (techno music) >> Agility is what all digital organizations strive for, and for almost the entirety of the enterprise storage industry, agility and storage aren't words you'd often hear together. Since the founding of Pure Storage, we've been laser focused on taking what's painful about traditional enterprise storage and making it better. We imagined a world where consumers self-service the provisioning of their storage resources to match the performance and data protection capabilities that their applications require. No endless back and forth between application owners and storage teams, just true on-demand self-service. At the same time, imagine all of the complex storage management operations required to make this possible being automated through software. From the placement of the initial workload to storage adjusting with the unpredictable needs of an application and seamlessly migrating and rebalancing the fleet as needed, all with zero down time and no manual intervention. And finally, imagine almost limitless scale that adjusts to meet your business' data management needs over time. This is what we believe the future of enterprise storage looks like. >> Today, we are announcing Pure Fusion, a leap forward in enterprise storage, marrying the best parts of the public cloud with the storage experience and capabilities you've come to expect from Pure. By bringing the simplicity and scalability of the cloud operating model with on-demand consumption and automated provisioning, organizations can deliver an enterprise-grade managed, self-service storage model that unifies fleets of arrays and optimizes storage pulls on the fly. End users will be able to rapidly consume volumes, file systems, and advanced data services like replication without waiting for backend manual work making storage hardware truly invisible. And organizations will be able to scale seamlessly across block, file, and object workloads, leveraging the power of the entire Pure Storage family, including FlashArray, Pure Cloud Block Store, FlashBlade, and Portworx. (electronic music) >> It is time to take a look at what Pure's up to from a slightly different perspective. To help us do that is Scott Sinclair who's a senior analyst at the Enterprise Strategy Group. And Scott, thanks for joining us here, on theCUBE. Good to see ya today. >> Great to see you. >> All right, so let's jump into this. First, we'll get to the announcement in just a little bit. First off, in terms of Pure's strategy, as you've been watching this company evolve over years now, how has it evolved? And then we'll go to the announcements and how that fits into the strategy. But first off, let's just take them from your point of view where have they been and how are they doin'? >> You know, many people know of Pure or maybe they don't know of their history as an all-Flash array. I think Pure has always been, ever since they entered the IT industry as a pioneer, they're one of the early ones that said look, we're going all in on the all-Flash array business and a focus on Flash technology. Then they were early pioneers in things like Evergreen and things like storage-as-a-service capabilities for on-premises storage. And the entire time, they've had a really almost streamline focus on ease of use which, you know, from the outside, I think everyone talks about ease of use and making things simple for IT, but Pure has really made that almost like core as part of not only their product and their design but also part of their culture. And one of the things, and we'll get into this a little bit as we talk about the announcements, but, you know, if you look at these announcements and where Pure's going, they're trying to expand that culture, that DNA around ease of use or simplicity, and expanding it beyond just storage or IT operations, and really trying to see okay, how do we make the entire digital initiative process or the larger IT operations journey simpler. And I think that's part of where Pure is going is not just storage but focusing more on apps, operations, and data, and making it easier for the entire experience. >> So how do the announcements we're talking about, well, there're three phases here, and again, we'll unpack those separately, but in general, how do the announcements then, you think, fit into that strategy and fit into their view and your view, really, of the market trends? >> I think one of the big trends is, you know, IT in terms for most businesses is, it's not just an enabler anymore. IT's actually in the driver's seat. We see in our research at ESGU, we just did this study and I'm going to glance over my notes as I'm kind of talking, but we see one of the things is more than half of businesses are identifying some portion of their revenue is coming from digital products or digital services. So data is part of the revenue chain for a majority of organizations according to what we're seeing in our research. And so what that does is it puts IT right in that core, you know, that core delivery model of where the faster IT can operate, the faster organizations can realize these revenues opportunities. So what is that doing to IT organizations? Well first off, it makes their life a lot harder, it makes demands continue to increase. But also, this old adage or this old narrative that IT's about availability, it's about resiliency, it's about keeping the lights on and ensuring that the business doesn't go down, well none of that goes away. But now, IT organizations are being measured on their ability to accelerate operations. And in this world where everything's becoming more, you know, more complex, there're more demands, organizations are becoming more distributed, application demands are becoming more diverse and they're growing in breadth. All of this means that more pressure is falling not only on the IT operations but also on the infrastructure providers like Pure Storage to step up and make things even simpler with things like automation and simplification which, you know, we're going to talk about, but to help accelerate those operations. >> Yeah, I mean, if you're DevOps these days, I mean, and you're talkin' about kind of these quandaries that people are in, but what are these specific challenges do you think, on the enterprise level here, that Pure is addressing? >> Well so for example, you talked about developers and driving into that in particular, I want to say let's see, glance at my notes here, about two-thirds of organizations say they're under pressure to accelerate IT initiatives due to pressures specifically from DevOps teams as well as line of business teams. So what does that mean? It means that as organizations build up and try to accelerate either their revenue creation via the creation of software or products, or things of that, that drive, that support a DevOps team, maybe it's improving customer experience for example, as well as other line of business teams such as analytics and trying to provide better insights and better decision making off of data, what that means is this traditional process of IT operations of where you submit a trouble ticket and then it takes, after a few days, something happens and they start doing analysis in terms of basically what ends up being multiple days or multiple weeks, to end up to basically provision storage, it just takes too long. And so in these announcements what we're seeing is Pure delivering solutions that are all about automating the backend services and delivering storage in a way that is designed to be easily and quickly consumed by the new consumers of IT, the developers, the line of business teams via APIs where you can write to a standard API and it goes across basically lots of different technologies and happens very quickly where a lot of the backend processes are automated, and essentially, making the storage invisible to these new consumers. And all of that just delivers value because what these groups are doing is now they can access and get the resources that they need and they don't have to know about what's happening behind the scenes which, candidly, they don't really know much about, right now, and they don't really care. >> Right. (chuckling) That's right. Yeah, what I don't see, what I don't know won't hurt me. And it can, as we know, it can. So let's look at the announcements. Pure Fusion, I think we were hearing about that just a little bit before, earlier in the interview that Dave was conducting, but let's talk about Pure Fusion and your thoughts on that. >> Pure Fusion is what I was talking about a little bit where they're abstracting a lot of the storage capabilities and presenting it as an API, a consistent API that allows developers to provision things very quickly and where a lot of the backend services are automated and, you know, essentially invisible to the developer. And that is, I mean, it addresses where, you know, I kind of talk about this with some of the data that we just, you know, some of our research stats that we just discussed, but it's where a lot of organizations are going. The bottom line is, we used to, in a world where IT services weren't growing as fast and where everything had to be resilient and available, you could put a lot of personnel power or personal hours focused on okay, making sure every box and everything was checked prior to doing a new implementation.and all that was designed to reduce risk and possibly optimize the environment and reduce cost. Now in this world of acceleration what we've seen is organizations need faster responsiveness from the IT organization. Well that's all well and good, but the problem is it's difficult to do all those backend processes and make sure that data's fully being protected or making sure that everything is happening behind the scenes the way it should be. And so this is, again, just mounting more and more pressure. So with things like Pure Fusion what they're doing is they're essentially automating a lot of that on the backend and really simplifying it and making it so storage, or IT administrators can provide access to their line of business, to development teams to leverage infrastructure a lot faster while still ensuring that all those backend services, all those operations still happen. Portworx Data Services also announced and we're hearing it from Dave, for that perspective may be a game-changer in terms of storage. So your take on that and Portworx? >> You know, I really like Portworx. I've been following them ever since prior to the acquisition. One of the things that they were very early on is understanding the impact of micro services on the industry and really, the importance of designing infrastructure around for that environment. I think what they're doing around data services is really intriguing. I think it's really intriguing, first off, for Pure as a company because it elevates their visibility to a new audience and a new persona that may not have been familiar with them. As organizations are looking at, you know, one of the things that they're doing with this data services is essentially delivering a database-as-a-service platform where you can go provision and stand up databases very quickly and again, similar to we talked about fusion, a lot of those backend processes are automated. Really fascinating, again, aligns directly with this acceleration need that we talked about. So, you know, a huge value, but it's really fascinating for Pure because it opens them up to, you know, hey, there's this whole new world of possible consumers that where they're, that they can get experience to really, the ease of use that Pure is known for a lot of the capabilities that Portworx is known for, but also just increase really the value that Pure is able to deliver to some of these modern enterprises. >> And just to add, briefly, on the enhancements that Pure One also being announced today. Your take on those? >> I like that as well. I think one of the things if I kind of go through the list is a lot of insights and intelligence in terms of new app, sizing applications for the environment if I remember correctly, and more, you know, better capabilities to help ensure that your environment is optimized which candidly is a top challenge around IT organizations. We talk about, again, I keep hitting on this need to move faster, faster, faster. One of the big disconnects that we've seen and we saw it very early when organizations were moving to, for example, public cloud services, is this disconnect towards for this individual app, how many resources do I really need and I think that's something that, you know, vendors like Pure need to start integrating more and more intelligence. And that's, my understanding is they're doing with Pure One which is really impressive. >> I hope it's all it takes. Scott, we appreciate the time. Thank you for your insights into what has been a big day for Pure Storage. But thank you again for the time. Scott Sinclair at the Enterprise Strategy Group, senior analyst, there. Let's go back to Dave Vellante now with more on theCUBE. (electronic music) >> Thanks for watching this CUBE program made possible by Pure Storage. I want to say in summary, you know, sometimes it's hard to squint through all the vendor noise on cloud and as-a-service, and all the buzz words, and acronyms in the marketplace. But as I said at the top, the cloud is changing, it's evolving, it's expanding to new locations. The operating model is increasingly defining the cloud. There's so much opportunity to build value on top of the massive infrastructure build-out from the hyperscalers to $100 billion in CapEx last year, alone. This is not just true for technology vendors, but organizations are building their own layer to take advantage of the cloud. Now, of course, technology's critical so when you're evaluating technology solutions, look for the following. First, the ability of the solution to simplify your life. Can it abstract the underlying complexity of a cloud, multiple clouds, connect to on-prem workloads in an experience that is substantially identical, irrespective of location? Does the solution leverage cloud-native technologies and innovations and primitives and APIs or is it just a hosted stack that's really not on the latest technology curve, whether that's processor technology or virtualization, or machine learning, streaming, open source tech, et cetera? Third, how programmable is the infrastructure? Does it make developers more productive? Does it accelerate time to value? Does it minimize rework and increase the quality of your output? And four, what's the business impact? Will customers stand up and talk about the solution and how it contributed to their digital transformation by flexibly supporting emerging data-intensive workloads and evolving as their business rapidly changed? These are some of the important markers that we would suggest you monitor. Pure is obviously driving hard to optimize these and other areas, so watch closely and make your own assessment as to how, what they and others are building will fit into your business. Now as always, this content is available on demand on theCUBE.net, so definitely check that out. This I Dave Vellante for John Walls and the entire CUBE team, thanks for watching, everybody. We'll see ya next time. (soft electronic music)
SUMMARY :
and eventually out to the edge. what you're delivering? and the cloud environment. all the way down to the storage. and bytes of the storage Rob: Yeah, we were doing this, of the vision and the strategy. and that's the next level in the storage industry. and change, that's got to be and how are you driving that? the ability to support have to leave it there. John Walls to give you the and rebalancing the fleet as of the public cloud with at the Enterprise Strategy Group. and how that fits into the strategy. And the entire time, they've had a really and I'm going to glance over my and get the resources that earlier in the interview a lot of that on the backend for a lot of the capabilities And just to add, One of the big disconnects that we've seen Scott Sinclair at the and acronyms in the marketplace.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Scott | PERSON | 0.99+ |
Scott Sinclair | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Rob Lee | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Portworx | ORGANIZATION | 0.99+ |
John Walls | PERSON | 0.99+ |
Rob | PERSON | 0.99+ |
Scott Sinclair | PERSON | 0.99+ |
90% | QUANTITY | 0.99+ |
20 | QUANTITY | 0.99+ |
two words | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
ESGU | ORGANIZATION | 0.99+ |
$100 billion | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
17 different products | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
Today | DATE | 0.99+ |
Evergreen | ORGANIZATION | 0.99+ |
each piece | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
three months | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
30 | QUANTITY | 0.99+ |
30 applications | QUANTITY | 0.99+ |
Pure | ORGANIZATION | 0.99+ |
Portworx Data Services | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
40 | QUANTITY | 0.98+ |
Third | QUANTITY | 0.98+ |
CUBE | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.98+ |
Enterprise Strategy Group | ORGANIZATION | 0.98+ |
Enterprise Strategy Group | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
both worlds | QUANTITY | 0.97+ |
one location | QUANTITY | 0.97+ |
six months | QUANTITY | 0.97+ |
10 | QUANTITY | 0.97+ |
Cassandra | TITLE | 0.96+ |
one application | QUANTITY | 0.96+ |
Ajay Singh, Pure Storage | CUBEconversation
(upbeat music) >> The Cloud essentially turned the data center into an API and ushered in the era of programmable infrastructure, no longer do we think about deploying infrastructure in rigid silos with a hardened, outer shell, rather infrastructure has to facilitate digital business strategies. And what this means is putting data at the core of your organization, irrespective of its physical location. It also means infrastructure generally and storage specifically must be accessed as sets of services that can be discovered, deployed, managed, secured, and governed in a DevOps model or OpsDev, if you prefer. Now, this has specific implications as to how vendor product strategies will evolve and how they'll meet modern data requirements. Welcome to this Cube conversation, everybody. This is Dave Vellante. And with me to discuss these sea changes is Ajay Singh, the Chief Product Officer of Pure Storage, Ajay welcome. >> Thank you, David, gald to be on. >> Yeah, great to have you, so let's talk about your role at Pure. I think you're the first CPO, what's the vision there? >> That's right, I just joined up Pure about eight months ago from VMware as the chief product officer and you're right, I'm the first our chief product officer at Pure. And at VMware I ran the Cloud management business unit, which was a lot about automation and infrastructure as code. And it's just great to join Pure, which has a phenomenal all flash product set. I kind of call it the iPhone or flash story super easy to use. And how do we take that same ease of use, which is a heart of a Cloud operating principle, and how do we actually take it up to really deliver a modern data experience, which includes infrastructure and storage as code, but then even more beyond that and how do you do modern operations and then modern data services. So super excited to be at Pure. And the vision, if you may, at the end of the day, is to provide, leveraging this moderate experience, a connected and effortless experience data experience, which allows customers to ultimately focus on what matters for them, their business, and by really leveraging and managing and winning with their data, because ultimately data is the new oil, if you may, and if you can mine it, get insights from it and really drive a competitive edge in the digital transformation in your head, and that's what be intended to help our customers to. >> So you joined earlier this year kind of, I guess, middle of the pandemic really I'm interested in kind of your first 100 days, what that was like, what key milestones you set and now you're into your second a 100 plus days. How's that all going? What can you share with us in and that's interesting timing because the effects of the pandemic you came in in a kind of post that, so you had experience from VMware and then you had to apply that to the product organization. So tell us about that sort of first a 100 days and the sort of mission now. >> Absolutely, so as we talked about the vision, around the modern data experience, kind of have three components to it, modernizing the infrastructure and really it's kudos to the team out of the work we've been doing, a ton of work in modernizing the infrastructure, I'll briefly talk to that, then modernizing the data, much more than modernizing the operations. I'll talk to that as well. And then of course, down the pike, modernizing data services. So if you think about it from modernizing the infrastructure, if you think about Pure for a minute, Pure is the first company that took flash to mainstream, essentially bringing what we call consumer simplicity to enterprise storage. The manual for the products with the front and back of a business card, that's it, you plug it in, boom, it's up and running, and then you get proactive AI driven support, right? So that was kind of the heart of Pure. Now you think about Pure again, what's unique about Pure has been a lot of our competition, has dealt with flash at the SSD level, hey, because guess what? All this software was built for hard drive. And so if I can treat NAND as a solid state drive SSD, then my software would easily work on it. But with Pure, because we started with flash, we released went straight to the NAND level, and as opposed to kind of the SSD layer, and what that does is it gives you greater efficiency, greater reliability and create a performance compared to an SSD, because you can optimize at the chip level as opposed to at the SSD module level. That's one big advantage that Pure has going for itself. And if you look at the physics, in the industry for a minute, there's recent data put out by Wikibon early this year, effectively showing that by the year 2026, flash on a dollar per terabyte basis, just the economics of the semiconductor versus the hard disk is going to be cheaper than hard disk. So this big inflection point is slowly but surely coming that's going to disrupt the hardest industry, already the high end has been taken over by flash, but hybrid is next and then even the long tail is coming up over there. And so to end to that extent our lead, if you may, the introduction of QLC NAND, QLC NAND powerful competition is barely introducing, we've been at it for a while. We just recently this year in my first a 100 days, we introduced the flasher AC, C40 and C60 drives, which really start to open up our ability to go after the hybrid story market in a big way. It opens up a big new market for us. So great work there by the team,. Also at the heart of it. If you think about it in the NAND side, we have our flash array, which is a scale-up latency centric architecture and FlashBlade which is a scale-out throughput architecture, all operating with NAND. And what that does is it allows us to cover both structured data, unstructured data, tier one apps and tier two apps. So pretty broad data coverage in that journey to the all flash data center, slowly but surely we're heading over there to the all flash data center based on demand economics that we just talked about, and we've done a bunch of releases. And then the team has done a bunch of things around introducing and NVME or fabric, the kind of thing that you expect them to do. A lot of recognition in the industry for the team or from the likes of TrustRadius, Gartner, named FlashRay, the Carton Peer Insights, the customer choice award and primary storage in the MQ. We were the leader. So a lot of kudos and recognition coming to the team as a result, Flash Blade just hit a billion dollars in cumulative revenue, kind of a leader by far in kind of the unstructured data, fast file an object marketplace. And then of course, all the work we're doing around what we say, ESG, environmental, social and governance, around reducing carbon footprint, reducing waste, our whole notion of evergreen and non-disruptive upgrades. We also kind of did a lot of work in that where we actually announced that over 2,700 customers have actually done non-disruptive upgrades over the technology. >> Yeah a lot to unpack there. And a lot of this sometimes you people say, oh, it's the plumbing, but the plumbing is actually very important too. 'Cause we're in a major inflection point, when we went from spinning disk to NAND. And it's all about volumes, you're seeing this all over the industry now, you see your old boss, Pat Gelsinger, is dealing with this at Intel. And it's all about consumer volumes in my view anyway, because thanks to Steve Jobs, NAND volumes are enormous and what two hard disk drive makers left in the planet. I don't know, maybe there's two and a half, but so those volumes drive costs down. And so you're on that curve and you can debate as to when it's going to happen, but it's not an if it's a when. Let me, shift gears a little bit. Because Cloud, as I was saying, it's ushered in this API economy, this as a service model, a lot of infrastructure companies have responded. How are you thinking at Pure about the as a service model for your customers? What's the strategy? How is it evolving and how does it differentiate from the competition? >> Absolutely, a great question. It's kind of segues into the second part of the moderate experience, which is how do you modernize the operations? And that's where automation as a service, because ultimately, the Cloud has validated and the address of this model, right? People are looking for outcomes. They care less about how you get there. They just want the outcome. And the as a service model actually delivers these outcomes. And this whole notion of infrastructure as code is kind of the start of it. Imagine if my infrastructure for a developer is just a line of code, in a Git repository in a program that goes through a CICD process and automatically kind of is configured and set up, fits in with the Terraform, the Ansibles, all that different automation frameworks. And so what we've done is we've gone down the path of really building out what I think is modern operations with this ability to have storage as code, disability, in addition modern operations is not just storage scored, but also we've got recently introduced some comprehensive ransomware protection, that's part of modern operations. There's all the threat you hear in the news or ransomware. We introduced what we call safe mode snapshots that allow you to recover in literally seconds. When you have a ransomware attack, we also have in the modern operations Pure one, which is maybe the leader in AI driven support to prevent downtime. We actually call you 80% of the time and fix the problems without you knowing about it. That's what modern operations is all about. And then also Martin operations says, okay, you've got flash on your on-prem side, but even maybe using flash in the public Cloud, how can I have seamless multi-Cloud experience in our Cloud block store we've introduced around Amazon, AWS and Azure allows one to do that. And then finally, for modern applications, if you think about it, this whole notion of infrastructure's code, as a service, software driven storage, the Kubernetes infrastructure enables one to really deliver a great automation framework that enables to reduce the labor required to manage the storage infrastructure and deliver it as code. And we have, kudos to Charlie and the Pure storage team before my time with the acquisition of Portworx, Portworx today is truly delivers true storage as code orchestrated entirely through Kubernetes and in a multi-Cloud hybrid situation. So it can run on EKS, GKE, OpenShift rancher, Tansu, recently announced as the leader by giggle home for enterprise Kubernetes storage. We were really proud about that asset. And then finally, the last piece are Pure as a service. That's also all outcome oriented, SLS. What matters is you sign up for SLS, and then you get those SLS, very different from our competition, right? Our competition tends to be a lot more around financial engineering, hey, you can buy it OPEX versus CapEx. And, but you get the same thing with a lot of professional services, we've really got, I'd say a couple of years and lead on, actually delivering and managing with SRE engineers for the SLA. So a lot of great work there. We recently also introduced Cisco FlashStack, again, flash stack as a service, again, as a service, a validation of that. And then finally, we also recently did a announcement with Aquaponics, with their bare metal as a service where we are a key part of their bare metal as a service offering, again, pushing the kind of the added service strategy. So yes, big for us, that's where the buck is skating, half the enterprises, even on prem, wanting to consume things in the Cloud operating model. And so that's where we're putting it lot. >> I see, so your contention is, it's not just this CapEx to OPEX, that's kind of the, during the economic downturn of 2007, 2008, the economic crisis, that was the big thing for CFOs. So that's kind of yesterday's news. What you're saying is you're creating a Cloud, like operating model, as I was saying upfront, irrespective of physical location. And I see that as your challenge, the industry's challenge, be, if I'm going to effect the digital transformation, I don't want to deal with the Cloud primitives. I want you to hide the underlying complexity of that Cloud. I want to deal with higher level problems, but so that brings me to digital transformation, which is kind of the now initiative, or I even sometimes call it the mandate. There's not a one size fits all for digital transformation, but I'm interested in your thoughts on the must take steps, universal steps that everybody needs to think about in a digital transformation journey. >> Yeah, so ultimately the digital transformation is all about how companies are gain a competitive edge in this new digital world or that the company are, and the competition are changing the game on, right? So you want to make sure that you can rapidly try new things, fail fast, innovate and invest, but speed is of the essence, agility and the Cloud operating model enables that agility. And so what we're also doing is not only are we driving agility in a multicloud kind of data, infrastructure, data operation fashion, but we also taking it a step further. We were also on the journey to deliver modern data services. Imagine on a Pure on-prem infrastructure, along with your different public Clouds that you're working on with the Kubernetes infrastructures, you could, with a few clicks run Kakfa as a service, TensorFlow as a service, Mongo as a service. So me as a technology team can truly become a service provider and not just an on-prem service provider, but a multi-Cloud service provider. Such that these services can be used to analyze the data that you have, not only your data, your partner data, third party public data, and how you can marry those different data sets, analyze it to deliver new insights that ultimately give you a competitive edge in the digital transformation. So you can see data plays a big role there. The data is what generates those insights. Your ability to match that data with partner data, public data, your data, the analysis on it services ready to go, as you get the digital, as you can do the insights. You can really start to separate yourself from your competition and get on the leaderboard a decade from now when this digital transformation settles down. >> All right, so bring us home, Ajay, summarize what does a modern data strategy look like and how does it fit into a digital business or a digital organization? >> So look, at the end of the day, data and analysis, both of them play a big role in the digital transformation. And it really comes down to how do I leverage this data, my data, partner data, public data, to really get that edge. And that links back to a vision. How do we provide that connected and effortless, modern data experience that allows our customers to focus on their business? How do I get the edge in the digital transformation? But easily leveraging, managing and winning with their data. And that's the heart of where Pure is headed. >> Ajay Singh, thanks so much for coming inside theCube and sharing your vision. >> Thank you, Dave, it was a real pleasure. >> And thank you for watching this Cube conversation. This is Dave Vellante and we'll see you next time. (upbeat music)
SUMMARY :
in the era of programmable Yeah, great to have you, And the vision, if you the pandemic you came in in kind of the unstructured data, And a lot of this sometimes and the address of this model, right? of 2007, 2008, the economic crisis, the data that you have, And that's the heart of and sharing your vision. was a real pleasure. And thank you for watching
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Ajay Singh | PERSON | 0.99+ |
Charlie | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Pat Gelsinger | PERSON | 0.99+ |
Ajay | PERSON | 0.99+ |
Steve Jobs | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Pure | ORGANIZATION | 0.99+ |
TrustRadius | ORGANIZATION | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
2008 | DATE | 0.99+ |
2007 | DATE | 0.99+ |
first | QUANTITY | 0.99+ |
CapEx | ORGANIZATION | 0.99+ |
Aquaponics | ORGANIZATION | 0.99+ |
Portworx | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
OPEX | ORGANIZATION | 0.99+ |
Martin | PERSON | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
both | QUANTITY | 0.99+ |
100 plus days | QUANTITY | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
second part | QUANTITY | 0.99+ |
over 2,700 customers | QUANTITY | 0.99+ |
Wikibon | ORGANIZATION | 0.98+ |
second | QUANTITY | 0.98+ |
first 100 days | QUANTITY | 0.98+ |
billion dollars | QUANTITY | 0.98+ |
this year | DATE | 0.97+ |
Kubernetes | TITLE | 0.97+ |
Cisco | ORGANIZATION | 0.96+ |
two and a half | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
Mongo | ORGANIZATION | 0.96+ |
Tansu | ORGANIZATION | 0.95+ |
Azure | ORGANIZATION | 0.95+ |
early this year | DATE | 0.94+ |
earlier this year | DATE | 0.94+ |
100 days | QUANTITY | 0.94+ |
FlashRay | ORGANIZATION | 0.93+ |
first company | QUANTITY | 0.93+ |
tier two apps | QUANTITY | 0.93+ |
C60 | COMMERCIAL_ITEM | 0.92+ |
pandemic | EVENT | 0.92+ |
OpenShift | ORGANIZATION | 0.91+ |
SLS | TITLE | 0.91+ |
2026 | DATE | 0.91+ |
Carton | ORGANIZATION | 0.91+ |
three components | QUANTITY | 0.9+ |
today | DATE | 0.88+ |
Cloud | TITLE | 0.88+ |
a minute | QUANTITY | 0.87+ |
SRE | ORGANIZATION | 0.86+ |
Cloud block | TITLE | 0.86+ |
two hard disk drive | QUANTITY | 0.86+ |
EKS | ORGANIZATION | 0.85+ |
Kubernetes | ORGANIZATION | 0.82+ |
about eight months ago | DATE | 0.82+ |
Ansibles | ORGANIZATION | 0.8+ |
GKE | ORGANIZATION | 0.79+ |
Kakfa | ORGANIZATION | 0.79+ |
a decade | DATE | 0.77+ |
tier one apps | QUANTITY | 0.76+ |
Peer Insights | TITLE | 0.75+ |
Git | TITLE | 0.75+ |
TensorFlow | ORGANIZATION | 0.71+ |
one big advantage | QUANTITY | 0.7+ |
Murli Thirumale, Pure Storage | CUBE Conversations, May 2021
(bright upbeat music) >> Hey, welcome to theCUBE's coverage of Pure Accelerate 2021. I'm Lisa Martin, please stay welcoming back one of our alumni Murli Thirumale is here, the VP & GM of the Cloud Native Business Unit at Pure Storage, Murli, welcome back. >> Lisa, it's great to be back at theCUBE, looking forward to discussion. >> Likewise, so it's been about six months or so since the Portworx acquisition by Pure Storage, give us a lay of the land, what's been going on? What are some of the successes, early wins, and some of the lessons that you've learned? >> Yeah, this is my third time being in Cloud, being a serial entrepreneur. So I've seen this movie before, and I have to say that this is really a lot of good anticipation followed by actually a lot of good stuff that has happened since, so it's been really a great ride so far. And when, let me start with the beginning, what the fundamental goal of the acquisition were, right? The couple of major goals, and then I can talk about how that integration is going. Really, I think from our viewpoint, from the Portworx viewpoint, the goal of the acquisition, from our view, was really to help turbocharge in our growth, we had really a very, very good product that was well accepted and established at customers, doing well as far as industry acceptance was concerned. And frankly, we had some great reference customers and some great installs expanding pretty well. Our issue was really how fast can we turbocharge that growth because as everybody knows, for a startup, the expensive part of an expansion is really on the go-to-market and sales side. And frankly, the timing for this was critical for us because the market had moved from the Kubernetes' market, has moved from sort of the innovator stage to the early majority stage. So from the Pure side, I think this made a lot of sense for them, because they have been looking for how they can expand their subscription models, how they can move to add more value from the array based business that there really have been a wonderful disruptor and to add more value up the stack, and that was the premise of the acquisition. One of the things that I paid a lot of attention to, as anybody does in acquisitions, is not just the strategy but really to understand if there was a culture fit between the teams, because a lot of the times acquisitions don't work because of the poor culture fit. So now let me kind of fast forward little bit and say, "Hey, what we know looking back in about six to eight months into it, how has it turning out so far?" And things have been just absolutely wonderful. Let me actually start with the culture fit, because that often is ignored and is one of the most important parts, right? The resonance in the culture between the two companies is just off the charts, right? It actually starts with what I would call a dramatic kind of customer first orientation, it's something we always had at Portworx. I always used to tell our customers with a startup you end up kind of, you buy the product, but you get the team, right? That's what happens with early stage startups, but Pure is sort of the same way, they are very focused on customer. So the customer focus is a very very useful thing that pulls us together. The second thing that's been really heartwarming to see has been really the focus on product excellence. Pure made it's dramatic entry into the market using Flash, and being the best Flash-based solution, and now they've expanded into many, many different areas. And Portworx also had a focus on product excellence, and so that has kind of moved the needle forward for both of us. And then I think the third thing is really a focus on the team winning, and not just an individual, right? And look, in these COVID times, this has been a tough year for everybody, I think it's, to some extent, even as we onboard new people, it's the culture of the team, the ability to bring new people onboard, and buy the culture, and make progress, all of that is really a function of how well the team is, 'we' is greater than 'me' type of a model, and I think that both these three values of customer first, high focus on product excellence, and the value in the team, including the resellers and the customers as part of the team, has really been the cornerstone, I think, of our success in the integration. >> That's outstanding because, like you said, this is not your first rodeo launching, coming out of stealth and launching and getting acquired, but doing so during one of the most challenging times in the last 100 years in our history while aligning cultures, I think that says a lot about the leadership on the Portworx side and the Pure side. >> I have to say, right? This is one of those amazing things, many people now that having been acquired can say this, really, most of the diligence, the transactions, all of that were done over Zoom, right? So, and then of course, everything since then is we're still in Zoom paradise. And so I think it really is a testament to the modern tools and stuff that we have that enable that. Now, let me talk a little bit about the content of what has happened, right? So strategically, I think the three areas that I think we've had huge synergy and seeing the benefits are first and foremost on the product side. A little later, I'd like to talk a little bit about some of the announcements we're making, but essentially, Pure had this outstanding core storage infrastructure product, well-known in the industry, very much Flash-oriented, part of the whole all Flash era now. And Portworx really came in with the idea of driving Kubernetes and Cloud Native workloads, which are really the majority of modern workloads. And what we found since then is that the integration of having really a more complete stack, which is really centered around what used to be an IT infrastructure of purchase, and what is in fact, for Kubernetes, a more DevOps oriented purchase. And that kind of a combination of being able to provide that combo in one package is something that we've been working very hard on in the last six months. And I'll mention some of the announcements, but we have a number of integrations with FlashArray and FlashBlade and other Pure products that we're able to highlight. So product integration for sure has been an area of some focus, but against a lot of progress. The second one is really customer synergy. I kind of described to our team when we got acquired, I said it's, for us, it's, being acquired by Pure is like strapping a rocket ship to ourselves as a small company, because we now have access to a huge customer footprint. Pure has over 8,000 customers, hugely amazingly high, almost unbelievable NPS score with customers, one of the best in the IT industry. And I think we are finding that with the deployment of containers becoming more ubiquitous, right? 80, 90% of customers in the enterprise are adopting Kubernetes and Containers. And therefore these 8,000 customers are a big huge target, they got a big target sign for both of us to be able to leverage. And so we've had a number of things that we're doing to address and use the Pure sales team to get access to them. The Pure channel of course is also part of that, Pure is 100% channel organization, which is great. So I think the synergy on the customer side with being able to have a solution that works for infrastructure and for DevOps has been a big area. In this day and age, Kubernetes is an area, for many of your listeners who are very, very familiar with Kubernetes, customers struggle, not just with day zero, but day one, day two, day three, right? It's how do you put it in production. And support, and integrating, and the use of Kubernetes and containers, putting that stack together is a big area. So support is a big area of pain for customers, and it's an area that, again, for a Portworx viewpoint, now we've expanded our footprint with a great support organization that we can bring to bear 24 by seven around the globe. Portworx is running on a lot of mission critical applications in big industries like finance and retail, and these types of things, really, support is a big area. And then the last thing I will just say is the use cases are usually synergistic, right? And we'll talk a little bit more about use cases as we go along here, but really there's legacy apps, right? In an interesting way, there's 80% of, IT spending is still on legacy apps, if you will, in that stack. However, 80% of all the new applications are being deployed on this modern app stack, right? >> Right. >> With all these open-source type of products and technologies. And most of that stack, most of the modern app stack is containerized. The 80, 85% of those applications really are where customers have chosen containers and Kubernetes as the as the mechanism to deliver those apps. And therefore Pure products like FlashBlade were very, very focused with fast recovery for these kind of modern apps, which are the stack of AI, and personalization, and all the modern digital apps. And I think those things can align well with the Portworx offering. So really around the areas of culture, customers, product synergy, support, and finally use cases, are all kind of been areas of huge progress for us. >> It also seems to me that the Portworx acquisition gives Pure a foray, a new buying center with respect to DevOps, talk to me a little bit about that as an opportunity for Pure. >> Yeah, the modern world is one where the enterprise itself has segmented into whole lot of new areas of spending and infrastructure ownership, right? And in the old days it used to be the network, storage, compute, and apps, sort of the old model of the world. And of course the app model has moved on, and then certainly there's a lot of different ways, web apps, the three tier apps, and the web apps, and so on. But the infrastructure world has morphed really into a bunch of other sub-segments, and some of it is still traditional hardware, but then even that is being cloudified, right? Because a lot of companies like Pure have taken their hardware array offerings and are offering that as a cloud-like offering where you can purchase it as a service, and in fact, Pure is offering a set of solutions called Evergreen that allow you to not even, you're just under subscription, you get your hardware refresh bundled in, very, very innovative. So you have now new buying centers coming in, in addition to the old traditional IT, there is sort of this whole, what used to be in the old ways called middleware, now has kind of morphed into this DevSecOps set of folks, right? Which is DevOps it's ITOps, and even security is a big part of that, the CISO Organization has that kind of segment. And so these buying centers often have new budgets, right? It turns out that, for example, to contrast, the Portworx budget really comes from entirely different budget, right? Our top two budget sources are usually CIO initiatives, they're not from the traditional storage budget, it comes from things like move to cloud or business transformation. And those set of folks, that set of customers, is really born in a different era, so to speak. You know, Lisa, they come, and I come from the old world, so I would say that I'm kind of more of an oldie, hopefully a Goldie, but an oldie. These folks are born in the post-DevOps, post-cloud, post-open-source world, right? They are used to brand new tools, get-ops, the way that everything's run on the cloud, it's on demand. So what we bring to Pure is really the ability to take their initiatives, which were around infrastructure, and cloudifying infrastructure to now adding two layers on top of that, right? So what Portworx adds to Pure is the access to the new automation layer of middleware. Kubernetes is nothing but really an automation of model for containers and for infrastructure now. And then the third layer is on top of us, is what I would call SaaS, the SaaSified layer, and as a service layer. And so we bring the opportunity to get those SaaS-like budgets, the DevOps budgets, and the DevOps and the SaaS kind of buyers, and together the business has very different models to it. In addition to not just a different technologies, the buying behavior is different, it's based on a consumption model, it's a subscription business. So it really is a change for new budgets, new buyers, and new financial models, which is a subscription model, which as you know, is valued much more highly by Wall Street nowadays compared to say some of the older hardware models. >> Well, Murli, when we talk about storage, we talk about data or the modern data experience. The more and more data that's being produced, the more value potentially there is for organizations, I think we saw, we learned several lessons in the last year, and one of them is that being able to glean insights from data in real-time or near real-time is, for many businesses, no longer a nice to have, it's really table stakes, it was for survival of getting through COVID, it is now in terms of identification of new business models, but it elevates the data conversation up to the C-suite, the board going, "Is our data protected? Is it secure? Can we access it?" And, "How do we deliver a modern data experience to our customers and to our internal employees?" So with that modern data experience, and maybe the elevation about conversation lengths, talk to me about some of the things that you're announcing at Accelerate with respect to Portworx. >> Yeah, so there are two sets of announcements. To be honest actually, this is a pretty exciting time for us, we're in theCUBE Cone time and the Accelerate time. And so let me kind of draw a circle around both those sets of announcements, if you will, right? So let's start perhaps with just the sets of things that we are announcing at Accelerate, right? This is kind of the first things that are coming up right now. And I'll tell you, there are some very, very exciting things that we're doing. So the majority of the announcements are centered around a release that we have called 2.8, so Portworx says, "We've been in the market now for well over five years with the product that really has been well deployed in very large global 2K enterprises." So the three or four major announcements, one of them is what I was talking about earlier, the integration of true Kubernetes applications running on Pure Storage. So we have a Cloud Native, a Native implementation of Portworx running on FlashArray and FlashBlade, where essentially when users now provision a container volume to Portworx, the storage volumes are magically created on FlashArray and FlashBlade, right? It's the idea of, without having to interface, so a DevOps engineer can deploy storage as code by provisioning volumes using Kubernetes without having to go issue a trouble ticket or a service ticket for a PureArray. And Portworx essentially access a layer between Kubernetes and the PureArray, and we allow configuration of volumes on the storage volumes of the PureArray directly. So essentially now on FlashArray, these volumes now receive the full suite of Portworx Storage Management features, including Kubernetes DR, backup, security, auto scaling, and migration. So that is a first version of this integration, right? The second one, it's, I am, is a personal favorite of mine, it's very, very exciting, right? When we came into Pure, we discovered that Pure already had this software solution called Pure as a service, it was essentially a Pure1 service that allowed for continuous call home, and log and diagnostic information, really an awesome window for customers to be able to see what their array utilization is like, complete observability, end-to-end on capacity, what's coming up, and allowed for proactive addressing of outages, or issues, or being able to kind of see it before it happen. The good news now is Portworx is integrated with Pure1, and so now customers have a unified observability stack for their Kubernetes applications using Portworx and FlashArray and FlashBlade in the Pure1 portal. So we are in the Pure1 portal now really providing end-to-end troubleshooting of issues and deployment, so very, very exciting, something that I think is a major step forward, right? >> Absolutely, well that single pane of glass is critical for management, so many companies waste a lot of time and resources managing disparate disconnected systems. And again, the last year has taught us so many businesses, there wasn't time, because there's going to be somebody right behind you that's going to be faster and more nimble, and has that single pane of glass unified view to be able to make better decisions. Last question, really, before we wrap here. >> Yeah. >> I can hear your momentum, I can feel your momentum through Zoom here. Talk to me about what's next, 'cause I know that when the acquisition happened about, we said six months or so ago, you said, "This is a small step in the Portworx journey." So what's ahead? >> Lisa, great question. I can state 10 things, but let me kind of step up a little bit at the 10,000 foot level, right? In one sense, I think no company gets to declare victory in this ongoing battle and we're just getting started. But if I had to kind of say, "What are some of the major teams that we have been part of and have been able to make happen in addition to take advantage of?" Pure obviously took advantage of the Flash wave, and they moved to all Flash, that's been a major disruptor with Pure being the lead. For Portworx, it has been really the move to containers and data management in an automated form, right? Kubernetes has become sort of not just a container orchestrator looking North, but looking southbound, is orchestrating infrastructure, we are in the throws of that revolution. But if you think about it, the other thing that's happening is all of this is in the service of, if you're a CIO, you're in the service of lines of businesses asking for a way to run their applications in a multicloud way, run their applications faster. And that is really the, as a service revolution, and it feels a little silly to almost talk about it as a service in that it's this late in the Cloud era, but the reality is that's just beginning, right? As a service revolution dramatically changed the IaaS business, the infrastructure business. But if you look at it, data services as a, data as a service is something that is what our customers are doing, so our customers are taking Pure hardware, Portworx software, and then they are building them into a platform as a service, things like databases as a service. And what we are doing, you will see some announcements from us in the second half of this year, terribly exciting, I just can't wait for it, where we're going to be actually moving forward to allow our customers to more quickly get to data services at the push of a button, so to speak, right? So- >> Excellent. >> The idea of database as a service to offer messaging as a service, search as a service, streaming as a service, and then finally some ML kind of AI as a service, these five categories of data services are what you should be expecting to see from Portworx and Pure going forward in the next half. >> Big potential there to really kick the door wide open on the total adjustable market. Well, Murli, it's been great to have you on the program, I can't wait to have you on next 'cause I know that there's so much more, like I said, I can feel your momentum through our virtual experience here. Thank you so much for joining us and giving us the lay of the land of what's been happening with the Portworx acquisition and all of the momentum and excitement that is about to come, we appreciate your time. >> Thank you, Lisa. Cheers to a great reduced COVID second half of the year. >> Oh, cheers to that. >> Yeah cheers, thanks. >> From Murli Thirumale, I'm Lisa Martin, you're watching theCUBE's coverage of Pure Accelerate. (bright upbeat music)
SUMMARY :
of the Cloud Native Business Lisa, it's great to be back at theCUBE, and so that has kind of moved the needle on the Portworx side and the Pure side. of the announcements, most of the modern app the Portworx acquisition is really the ability to and maybe the elevation This is kind of the first things And again, the last year has taught us step in the Portworx journey." advantage of the Flash wave, forward in the next half. and all of the momentum and excitement COVID second half of the year. coverage of Pure Accelerate.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
Portworx | ORGANIZATION | 0.99+ |
Murli Thirumale | PERSON | 0.99+ |
80% | QUANTITY | 0.99+ |
May 2021 | DATE | 0.99+ |
two companies | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
Murli Thirumale | PERSON | 0.99+ |
100% | QUANTITY | 0.99+ |
Murli | PERSON | 0.99+ |
24 | QUANTITY | 0.99+ |
third layer | QUANTITY | 0.99+ |
10 things | QUANTITY | 0.99+ |
two layers | QUANTITY | 0.99+ |
Portworx | TITLE | 0.99+ |
last year | DATE | 0.99+ |
FlashArray | TITLE | 0.99+ |
8,000 customers | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
over 8,000 customers | QUANTITY | 0.99+ |
80, 85% | QUANTITY | 0.99+ |
Accelerate | ORGANIZATION | 0.99+ |
10,000 foot | QUANTITY | 0.99+ |
third time | QUANTITY | 0.99+ |
two sets | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
second thing | QUANTITY | 0.98+ |
first version | QUANTITY | 0.98+ |
third thing | QUANTITY | 0.98+ |
one package | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
Flash | TITLE | 0.98+ |
second one | QUANTITY | 0.97+ |
three tier | QUANTITY | 0.97+ |
PureArray | TITLE | 0.97+ |
one sense | QUANTITY | 0.97+ |
theCUBE | ORGANIZATION | 0.97+ |
four major announcements | QUANTITY | 0.97+ |
Pure1 | TITLE | 0.97+ |
Pure | ORGANIZATION | 0.97+ |
80, 90% | QUANTITY | 0.97+ |
Zoom | ORGANIZATION | 0.97+ |
Flash wave | EVENT | 0.96+ |
FlashBlade | TITLE | 0.96+ |
five categories | QUANTITY | 0.96+ |
first things | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.96+ |
Pure | TITLE | 0.95+ |
three areas | QUANTITY | 0.95+ |
three values | QUANTITY | 0.95+ |
seven | QUANTITY | 0.95+ |
CB Bohn, Principal Data Engineer, Microfocus | The Convergence of File and Object
>> Announcer: From around the globe it's theCUBE. Presenting the Convergence of File and Object brought to you by Pure Storage. >> Okay now we're going to get the customer perspective on object and we'll talk about the convergence of file and object, but really focusing on the object pieces this is a content program that's being made possible by Pure Storage and it's co-created with theCUBE. Christopher CB Bohn is here. He's a lead architect for MicroFocus the enterprise data warehouse and principal data engineer at MicroFocus. CB welcome good to see you. >> Thanks Dave good to be here. >> So tell us more about your role at Microfocus it's a pan Microfocus role because we know the company is a multi-national software firm it acquired the software assets of HP of course including Vertica tell us where you fit. >> Yeah so Microfocus is you know, it's like I can says it's wide, worldwide company that it sells a lot of software products all over the place to governments and so forth. And it also grows often by acquiring other companies. So there is there the problem of integrating new companies and their data. And so what's happened over the years is that they've had a number of different discreet data systems so you've got this data spread all over the place and they've never been able to get a full complete introspection on the entire business because of that. So my role was come in, design a central data repository and an enterprise data warehouse, that all reporting could be generated against. And so that's what we're doing and we selected Vertica as the EDW system and Pure Storage FlashBlade as the communal repository. >> Okay so you obviously had experience with with Vertica in your previous role, so it's not like you were starting from scratch, but paint a picture of what life was like before you embarked on this sort of consolidated approach to your data warehouse. Was it just dispared data all over the place? A lot of M and A going on, where did the data live? >> CB: So >> Right so again the data is all over the place including under people's desks and just dedicated you know their own private SQL servers, It, a lot of data in a Microfocus is one on SQL server, which has pros and cons. Cause that's a great transactional database but it's not really good for analytics in my opinion. So but a lot of stuff was running on that, they had one Vertica instance that was doing some select reporting. Wasn't a very powerful system and it was what they call Vertica enterprise mode where it had dedicated nodes which had the compute and storage in the same locus on each server okay. So Vertica Eon mode is a whole new world because it separates compute from storage. Okay and at first was implemented in AWS so that you could spin up you know different numbers of compute nodes and they all share the same communal storage. But there has been a demand for that kind of capability, but in an on-prem situation. Okay so Pure storage was the first vendor to come along and have an S3 emulation that was actually workable. And so Vertica worked with Pure Storage to make that all happen and that's what we're using. >> Yeah I know back when back from where we used to do face-to-face, we would be at you know Pure Accelerate, Vertica was always there it stopped by the booth, see what they're doing so tight integration there. And you mentioned Eon mode and the ability to scale, storage and compute independently. And so and I think Vertica is the only one I know they were the first, I'm not sure anybody else does that both for cloud and on-prem, but so how are you using Eon mode, are you both in AWS and on-prem are you exclusively cloud? Maybe you could describe that a little bit. >> Right so there's a number of internal rules at Microfocus that you know there's, it's not AWS is not approved for their business processes. At least not all of them, they really wanted to be on-prem and all the transactional systems are on-prem. And so we wanted to have the analytics OLAP stuff close to the OLTP stuff right? So that's why they called there, co-located very close to each other. And so we could, what's nice about this situation is that these S3 objects, it's an S3 object store on the Pure Flash Blade. We could copy those over if we needed it to AWS and we could spin up a version of Vertica there, and keep going. It's like a tertiary GR strategy cause we actually have a, we're setting up a second, Flash Blade Vertica system geo located elsewhere for backup and we can get into it if you want to talk about how the latest version of the Pure software for the Flash Blade allows synchronization across network boundaries of those Flash Blade which is really nice because if, you know there's a giant sinkhole opens up under our Koll of facility and we lose that thing then we just have to switch to DNS. And we were back in business of the DR. And then the third one was to go, we could copy those objects over to AWS and be up and running there. So we're feeling pretty confident about being able to weather whatever comes along. >> Yeah I'm actually very interested in that conversation but before we go there. you mentioned you want, you're going to have the old lab close to the OLTP, was that for latency reasons, data movement reasons, security, all of the above. >> Yeah it's really all of the above because you know we are operating under the same sub-net. So to gain access to that data, you know you'd have to be within that VPN environment. We didn't want to going out over the public internet. Okay so and just for latency reasons also, you know we have a lot of data and we're continually doing ETL processes into Vertica from our production data, transactional databases. >> Right so they got to be approximate. So I'm interested in so you're using the Pure Flash Blade as an object store, most people think, oh object simple but slow. Not the case for you is that right? >> Not the case at all >> Why is that. >> This thing had hoop It's ripping, well you have to understand about Vertica and the way it stores data. It stores data in what they call storage containers. And those are immutable, okay on disc whether it's on AWS or if you had a enterprise mode Vertica, if you do an update or delete it actually has to go and retrieve that object container from disc and it destroys it and rebuilds it, okay which is why you don't, you want to avoid updates and deletes with vertica because the way it gets its speed is by sorting and ordering and encoding the data on disk. So it can read it really fast. But if you do an operation where you're deleting or updating a record in the middle of that, then you've got to rebuild that entire thing. So that actually matches up really well with S3 object storage because it's kind of the same way, it gets destroyed and rebuilt too okay. So that matches up very well with Vertica and we were able to design the system so that it's a panda only. Now we have some reports that we're running in SQL server. Okay which we're taking seven days. So we moved that to Vertica from SQL server and we rewrote the queries, which were had, which had been written in TC SQL with a bunch of loops and so forth and we were to get, this is amazing it went from seven days to two seconds, to generate this report. Which has tremendous value to the company because it would have to have this long cycle of seven days to get a new introspection in what they call the knowledge base. And now all of a sudden it's almost on demand two seconds to generate it. That's great and that's because of the way the data is stored. And the S3 you asked about, oh you know it, it's slow, well not in that context. Because what happens really with Vertica Eon mode is that it can, they have, when you set up your compute nodes, they have local storage also which is called the depot. It's kind of a cache okay. So the data will be drawn from the Flash Blade and cached locally. And that was, it was thought when they designed that, oh you know it's that'll cut down on the latency. Okay but it turns out that if you have your compute nodes close meaning minimal hops to the Flash Blade that you can actually tell Vertica, you know don't even bother caching that stuff just read it directly on the fly from the from the Flash Blade and the performance is still really good. It depends on your situation. But I know for example a major telecom company that uses the same topologies we're talking about here they did the same thing. They just dropped the cache cause the Flash Blade was able to deliver the data fast enough. >> So that's, you're talking about that's speed of light issues and just the overhead of switching infrastructure is that, it's eliminated and so as a result you can go directly to the storage array? >> That's correct yeah, it's like, it's fast enough that it's almost as if it's local to the compute node. But every situation is different depending on your needs. If you've got like a few tables that are heavily used, then yeah put them in the cache because that'll be probably a little bit faster. But if you're have a lot of ad hoc queries that are going on, you know you may exceed the storage of the local cache and then you're better off having it just read directly from the, from the Flash Blade. >> Got it so it's >> Okay. >> It's an append only approach. So you're not >> Right >> Overwriting on a record, so but then what you have automatically re index and that's the intelligence of the system. how does that work? >> Oh this is where we did a little bit of magic. There's not really anything like magic but I'll tell you what it is I mean. ( Dave laughing) Vertica does not have indexes. They don't exist. Instead I told you earlier that it gets a speed by sorting and encoding the data on disk and ordering it right. So when you've got an append-only situation, the natural question is well if I have a unique record, with let's say ID one, two, three, what happens if I append a new version of that, what happens? Well the way Vertica operates is that there's a thing called a projection which is actually like a materialized columnar data store. And you can have a, what they call a top-K projection, which says only put in this projection the records that meet a certain condition. So there's a field that we like to call a discriminator field which is like okay usually it's the latest update timestamp. So let's say we have record one, two, three and it had yesterday's date and that's the latest version. Now a new version comes in. When the data at load time vertical looks at that and then it looks in the projection and says does this exist already? If it doesn't then it adds it. If it does then that one now goes into that projection okay. And so what you end up having is a projection that is the latest snapshot of the data, which would be like, oh that's the reality of what the table is today okay. But inherent in that is that you now have a table that has all the change history of those records, which is awesome. >> Yeah. >> Because, you often want to go back and revisit, you know what it will happen to you. >> But that materialized view is the most current and the system knows that at least can (murmuring). >> Right so we then create views that draw off from that projection so that our users don't have to worry about any of that. They just get oh and say select from this view and they're getting the latest greatest snapshot of what the reality of the data is right now. But if they want to go back and say, well how did this data look two days ago? That's an easy query for them to do also. So they get the best of both worlds. >> So could you just plug any flash array into your system and achieve the same results or is there anything really unique about Pure? >> Yeah well they're the only ones that have got I think really dialed in the S3 object form because I don't think AWS actually publishes every last detail of that S3 spec. Okay so it had, there's a certain amount of reverse engineering they had to do I think. But they got it right. When we've, a couple maybe a year and a half ago or so there they were like at 99%, but now they worked with Vertica people to make sure that that object format was true to what it should be. So that it works just as if Vertica doesn't care, if it is on AWS or if it's on Pure Flash Blade because Pure did a really good job of dialing in that format and so Vertica doesn't care. It just knows S3, doesn't know what it doesn't care where it's going it just works. >> So the essentially vendor R and D abstracted that complexity so you didn't have to rewrite the application is that right? >> Right, so you know when Vertica ships it's software, you don't get a specific version for Pure or AWS, it's all in one package, and then when you configure it, it knows oh okay well, I'm just pointed at the, you know this port, on the Pure storage Flash Blade, and it just works. >> CB what's your data team look like? How is it evolving? You know a lot of customers I talked to they complain that they struggled to get value out of the data and they don't have the expertise, what does your team look like? How is it, is it changing or did the pandemic change things at all? I wonder if you could bring us up to date on that? >> Yeah but in some ways Microfocus has an advantage in that it's such a widely dispersed across the world company you know it's headquartered in the UK, but I deal with people I'm in the Bay Area, we have people in Mexico, Romania, India. >> Okay enough >> All over the place yeah all over the place. So when this started, it was actually a bigger project it got scaled back, it was almost to the point where it was going to be cut. Okay, but then we said, well let's try to do almost a skunkworks type of thing with reduced staff. And so we're just like a hand. You could count the number of key people on this on one hand. But we got it all together, and it's been a traumatic transformation for the company. Now there's, it's one approval and admiration from the highest echelons of this company that, hey this is really providing value. And the company is starting to get views into their business that they didn't have before. >> That's awesome, I mean, I've watched Microfocus for years. So to me they've always had a, their part of their DNA is private equity I mean they're sharp investors, they do great M and A >> CB: Yeah >> They know how to drive value and they're doing modern M and A, you know, we've seen what they what wait, what they did with SUSE, obviously driving value out of Vertica, they've got a really, some sharp financial people there. So that's they must have loved the the Skunkworks, fast ROI you know, small denominator, big numerator. (laughing) >> Well I think that in this case, smaller is better when you're doing development. You know it's a two-minute cooks type of thing and if you've got people who know what they're doing, you know I've got a lot of experience with Vertica, I've been on the advisory board for Vertica for a long time. >> Right And you know I was able to learn from people who had already, we're like the second or third company to do a Pure Flash Blade Vertica installation, but some of the best companies after they've already done it we are members of the advisory board also. So I learned from the best, and we were able to get this thing up and running quickly and we've got you know, a lot of other, you know handful of other key people who know how to write SQL and so forth to get this up and running quickly. >> Yeah so I mean, look it Pure is a fit I mean I sound like a fan boy, but Pure is all about simplicity, so is object. So that means you don't have to ra, you know worry about wrangling storage and worrying about LANs and all that other nonsense and file names but >> I have burned by hardware in the past you know, where oh okay they built into a price and so they cheap out on stuff like fans or other things in these components fail and the whole thing goes down, but this hardware is super good quality. And so I'm happy with the quality of that we're getting. >> So CB last question. What's next for you? Where do you want to take this initiative? >> Well we are in the process now of, we're when, so I designed a system to combine the best of the Kimball approach to data warehousing and the inland approach okay. And what we do is we bring over all the data we've got and we put it into a pristine staging layer. Okay like I said it's a, because it's append-only, it's essentially a log of all the transactions that are happening in this company, just as they appear okay. And then from the Kimball side of things we're designing the data marts now. So that's what the end users actually interact with. So we're taking the, we're examining the transactional systems to say, how are these business objects created? What's the logic there and we're recreating those logical models in Vertica. So we've done a handful of them so far, and it's working out really well. So going forward we've got a lot of work to do, to create just about every object that the company needs. >> CB you're an awesome guest really always a pleasure talking to you and >> Thank you. >> congratulations and good luck going forward stay safe. >> Thank you, you too Dave. >> All right thank you. And thank you for watching the Convergence of File and Object. This is Dave Vellante for theCUBE. (soft music)
SUMMARY :
brought to you by Pure Storage. but really focusing on the object pieces it acquired the software assets of HP all over the place to Okay so you obviously so that you could spin up you know and the ability to scale, and we can get into it if you want to talk security, all of the above. Yeah it's really all of the above Not the case for you is that right? And the S3 you asked about, storage of the local cache So you're not and that's the intelligence of the system. and that's the latest version. you know what it will happen to you. and the system knows that at least the data is right now. in the S3 object form and then when you configure it, I'm in the Bay Area, And the company is starting to get So to me they've always had loved the the Skunkworks, I've been on the advisory a lot of other, you know So that means you don't have to by hardware in the past you know, Where do you want to take this initiative? object that the company needs. congratulations and good And thank you for watching
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Mexico | LOCATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
MicroFocus | ORGANIZATION | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
UK | LOCATION | 0.99+ |
seven days | QUANTITY | 0.99+ |
Romania | LOCATION | 0.99+ |
99% | QUANTITY | 0.99+ |
HP | ORGANIZATION | 0.99+ |
Microfocus | ORGANIZATION | 0.99+ |
two-minute | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
two seconds | QUANTITY | 0.99+ |
India | LOCATION | 0.99+ |
Kimball | ORGANIZATION | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
each server | QUANTITY | 0.99+ |
CB Bohn | PERSON | 0.99+ |
yesterday | DATE | 0.99+ |
two days ago | DATE | 0.99+ |
first | QUANTITY | 0.99+ |
Christopher CB Bohn | PERSON | 0.98+ |
SQL | TITLE | 0.98+ |
Vertica | TITLE | 0.98+ |
a year and a half ago | DATE | 0.98+ |
both worlds | QUANTITY | 0.98+ |
Pure Flash Blade | COMMERCIAL_ITEM | 0.98+ |
both | QUANTITY | 0.98+ |
vertica | TITLE | 0.98+ |
Bay Area | LOCATION | 0.97+ |
one | QUANTITY | 0.97+ |
Flash Blade | COMMERCIAL_ITEM | 0.97+ |
third one | QUANTITY | 0.96+ |
CB | PERSON | 0.96+ |
one package | QUANTITY | 0.96+ |
today | DATE | 0.95+ |
Pure storage Flash Blade | COMMERCIAL_ITEM | 0.95+ |
first vendor | QUANTITY | 0.95+ |
pandemic | EVENT | 0.94+ |
S3 | TITLE | 0.94+ |
marts | DATE | 0.92+ |
Skunkworks | ORGANIZATION | 0.91+ |
SUSE | ORGANIZATION | 0.89+ |
three | QUANTITY | 0.87+ |
S3 | COMMERCIAL_ITEM | 0.87+ |
third company | QUANTITY | 0.84+ |
Pure Flash Blade Vertica | COMMERCIAL_ITEM | 0.83+ |
Pure Storage Convergence of File and Object FULL SHOW V1
we're running what i would call a little mini series and we're exploring the convergence of file and object storage what are the key trends why would you want to converge file an object what are the use cases and architectural considerations and importantly what are the business drivers of uffo so-called unified fast file and object in this program you'll hear from matt burr who is the gm of pure's flashblade business and then we'll bring in the perspectives of a solutions architect garrett belsner who's from cdw and then the analyst angle with scott sinclair of the enterprise strategy group esg he'll share some cool data on our power panel and then we'll wrap with a really interesting technical conversation with chris bond cb bond who is a lead data architect at microfocus and he's got a really cool use case to share with us so sit back and enjoy the program from around the globe it's thecube presenting the convergence of file and object brought to you by pure storage we're back with the convergence of file and object a special program made possible by pure storage and co-created with the cube so in this series we're exploring that convergence between file and object storage we're digging into the trends the architectures and some of the use cases for unified fast file and object storage uffo with me is matt burr who's the vice president and general manager of flashblade at pure storage hello matt how you doing i'm doing great morning dave how are you good thank you hey let's start with a little 101 you know kind of the basics what is unified fast file and object yeah so look i mean i think you got to start with first principles talking about the rise of unstructured data so um when we think about unstructured data you sort of think about the projections 80 of data by 2025 is going to be unstructured data whether that's machine generated data or um you know ai and ml type workloads uh you start to sort of see this um i don't want to say it's a boom uh but it's sort of a renaissance for unstructured data if you will we move away from you know what we've traditionally thought of as general purpose nas and and file shares to you know really things that focus on uh fast object taking advantage of s3 cloud native applications that need to integrate with applications on site um you know ai workloads ml workloads tend to look to share data across you know multiple data sets and you really need to have a platform that can deliver both highly performant and scalable fast file and object from one system so talk a little bit more about some of the drivers that you know bring forth that need to unify file an object yeah i mean look you know there's a there's there's a real challenge um in managing you know bespoke uh bespoke infrastructure or architectures around general purpose nas and daz etc so um if you think about how a an architect sort of looks at an application they might say well okay i need to have um you know fast daz storage proximal to the application um but that's going to require a tremendous amount of dams which is a tremendous amount of drives right hard drives are you know historically pretty pretty pretty unwieldy to manage because you're replacing them relatively consistently at multi-petabyte scale um so you start to look at things like the complexity of daz you start to look at the complexity of general purpose nas and you start to just look at quite frankly something that a lot of people don't really want to talk about anymore but actual data center space right like consolidation matters the ability to take you know something that's the size of a microwave like a modern flash blade or a modern um you know uffo device uh replaces something that might be you know the size of three or four or five refrigerators so matt what why is is now the right time for this i mean for years nobody really paid much attention to object s3 already obviously changed you know that course most of the world's data is still stored in file formats and you get there with nfs or smb why is now the time to think about unifying object and file well because we're moving to things like a contactless society um you know the the things that we're going to do are going to just require a tremendous amount more compute power network um and quite frankly storage throughput and you know i can give you two sort of real primary examples here right you know warehouses are being you know taken over by robots if you will um it's not a war it's a it's a it's sort of a friendly advancement in you know how do i how do i store a box in a warehouse and you know we have we have a customer who focuses on large sort of big box distribution warehousing and you know a box that carried a an object two weeks ago might have a different box size two weeks later well that robot needs to know where the space is in the data center in order to put it but also needs to be able to process hey i don't want to put the thing that i'm going to access the most in the back of the warehouse i'm going to put that thing in the front of the warehouse all of those types of data you know sort of real time you can think of the robot as almost an edge device is processing in real time unstructured data in its object right so it's sort of the emergence of these new types of workloads and i give you the opposite example the other end of the spectrum is ransomware right you know today you know we'll talk to customers and they'll say quite commonly hey if you know anybody can sell me a backup device i need something that can restore quickly um if you had the ability to restore something in 270 terabytes an hour or 250 terabytes an hour uh that's much faster when you're dealing with a ransomware attack you want to get your data back quickly you know so i want to add i was going to ask you about that later but since you brought it up what is the right i guess call it architecture for for for ransomware i mean how and explain like how unified object and file which appointment i get the fast recovery but how how would you recommend a customer uh go about architecting a ransomware proof you know system yeah well you know with with flashblade and and with flasharray there's an actual feature called called safe mode and that safe mode actually protects uh the snapshots and and the data from uh sort of being a part of the of the ransomware event and so if you're in a type of ransomware situation like this you're able to leverage safe mode and you say okay what happens in a ransomware attack is you can't get access to your data and so you know the bad guy the perpetrator is basically saying hey i'm not going to give you access to your data until you pay me you know x in bitcoin or whatever it might be right um with with safe mode those snapshots are actually protected outside of the ransomware blast zone and you can bring back those snapshots because what's your alternative if you're not doing something like that your alternative is either to pay and unlock your data or you have to start retouring restoring excuse me from tape or slow disk that could take you days or weeks to get your data back so leveraging safe mode um you know in either the flash for the flash blade product uh is a great way to go about architecting against ransomware i got to put my my i'm thinking like a customer now so safe mode so that's an immutable mode right can't change the data um is it can can an administrator go in and change that mode can you turn it off do i still need an air gap for example what would you recommend there yeah so there there are still um uh you know sort of our back or roll back role-based access control policies uh around who can access that safe mode and who can right okay so uh anyway subject for a different day i want to i want to actually bring up uh if you don't object a topic that i think used to be really front and center and it now be is becoming front and center again i mean wikibon just produced a research note forecasting the future of flash and hard drives and those of you who follow us know we've done this for quite some time and you can if you could bring up the chart here you you could and we see this happening again it was originally we forecast the the the death of of quote-unquote high spin speed disc drives which is kind of an oxymoron but you can see on here on this chart this hard disk had a magnificent journey but they peaked in volume in manufacturing volume in 2010 and the reason why that is is so important is that volumes now are steadily dropping you can see that and we use wright's law to explain why this is a problem and wright's law essentially says that as you your cumulative manufacturing volume doubles your cost to manufacture decline by a constant percentage now i won't go too much detail on that but suffice it to say that flash volumes are growing very rapidly hdd volumes aren't and so flash because of consumer volumes can take advantage of wright's law and that constant reduction and that's what's really important for the next generation which is always more expensive to build uh and so this kind of marks the beginning of the end matt what do you think what what's the future hold for spinning disc in your view uh well i can give you the answer on two levels on a personal level uh it's why i come to work every day uh you know the the eradication or or extinction of an inefficient thing um you know i like to say that uh inefficiency is the bane of my existence uh and i think hard drives are largely inefficient and i'm willing to accept the sort of long-standing argument that um you know we've seen this transition in block right and we're starting to see it repeat itself in in unstructured data and i'm going to accept the argument that cost is a vector here and it most certainly is right hdds have been considerably cheaper uh than than than flash storage um you know even to this day uh you know up up to this point right but we're starting to approach the point where you sort of reach a a 3x sort of um you know differentiator between the cost of an hdd and an std and you know that really is that point in time when uh you begin to pick up a lot of volume and velocity and so you know that tends to map directly to you know what you're seeing here which is you know a a slow decline uh which i think is going to become even more rapid kind of probably starting around next year um where you start to see sds excuse me ssds uh you know really replacing hdds uh at a much more rapid clip particularly on the unstructured data side and it's largely around cost the the workloads that we talked about robots and warehouses or you know other types of advanced machine learning and artificial intelligence type applications and workflows you know they require a degree of performance that a hard drive just can't deliver we are we are seeing sort of the um creative innovative uh disruption of an entire industry right before our eyes it's a fun thing to live through yeah and and we would agree i mean it doesn't the premise there is that it doesn't have to be less expensive we think it will be by you know the second half or early second half of this decade but even if it's a we think around a 3x delta the value of of ssd relative to spinning disk is going to overwhelm just like with your laptop you know it got to the point where you said why would i ever have a spinning disc in my laptop we see the same thing happening here um and and so and we're talking about you know raw capacity you know put in compression and d-dupe and everything else that you really can't do with spinning discs because of the performance issues you can do with flash okay let's come back to uffo can we dig into the challenges specifically that that this solves for customers give me give us some examples yeah so you know i mean if we if we think about the examples um you know the the robotic one um i think is is is the one that i think is the marker for you know kind of of of the the modern side of of of what we see here um but what we're you know what we're what we're seeing from a trend perspective which you know not everybody's deploying robots right um you know there's there's many companies that are you know that aren't going to be in either the robotic business uh or or even thinking about you know sort of future type oriented type things but what they are doing is green field applications are being built on object um generally not on not on file and and not on block and so you know the rise of of object as sort of the the sort of let's call it the the next great protocol for um you know for uh for for modern workloads right this is this is that that modern application coming to the forefront and that could be anything from you know financial institutions you know right down through um you we've even see it and seen it in oil and gas uh we're also seeing it across across healthcare uh so you know as as as companies take the opportunity as industries to take this opportunity to modernize you know they're modernizing not on things that are are leveraging you know um you know sort of archaic disk technology they're they're they're really focusing on on object but they still have file workflows that they need to that they need to be able to support and so having the ability to be able to deliver those things from one device in a capacity orientation or a performance orientation uh while at the same time dramatically simplifying uh the overall administration of your environment both physically and non-physically is a key driver so the great thing about object is it's simple it's a kind of a get put metaphor um it's it scales out you know because it's got metadata associated with the data uh and and it's cheap uh the drawback is you don't necessarily associate it with high performance and and and as well most applications don't you know speak in that language they speak in the language of file you know or as you mentioned block so i i see real opportunities here if i have some some data that's not necessarily frequently accessed you know every day but yet i want to then whether end of quarter or whatever it is i want to i want to or machine learning i want to apply some ai to that data i want to bring it in and then apply a file format uh because for performance reasons is that right maybe you could unpack that a little bit yeah so um you know we see i mean i think you described it well right um but i don't think object necessarily has to be slow um and nor does it have to be um you know because when you think about you brought up a good point with metadata right being able to scale to a billions of objects being able to scale to billions of objects excuse me is of value right um and i think people do traditionally associate object with slow but it's not necessarily slow anymore right we we did a sort of unofficial survey of of of our of our customers and our employee base and when people described object they thought of it as like law firms and storing a word doc if you will um and that that's just you know i think that there's a lack of understanding or a misnomer around what modern what modern object has become and perform an object particularly at scale when we're talking about billions of objects you know that's the next frontier right um is it at pace performance wise with you know the other protocols no uh but it's making leaps and grounds so you talked a little bit more about some of the verticals that you see i mean i think when i think of financial services i think transaction processing but of course they have a lot of tons of unstructured data are there any patterns you're seeing by by vertical market um we're you know we're not that's the interesting thing um and you know um as a as a as a as a company with a with a block heritage or a block dna those patterns were pretty easy to spot right there were a certain number of databases that you really needed to support oracle sql some postgres work et cetera then kind of the modern databases around cassandra and things like that you knew that there were going to be vmware environments you know you could you could sort of see the trends and where things were going unstructured data is such a broader horizontal thing right so you know inside of oil and gas for example you have you know um you have specific applications and bespoke infrastructures for those applications um you know inside of media entertainment you know the same thing the the trend that we're seeing the commonality that we're seeing is the modernization of you know object as a starting point for all the all the net new workloads within within those industry verticals right that's the most common request we see is what's your object roadmap what's your you know what's your what's your object strategy you know where do you think where do you think object is going so um there isn't any um you know sort of uh there's no there's no path uh it's really just kind of a wide open field in front of us with common requests across all industries so the amazing thing about pure just as a kind of a little you know quasi you know armchair historian the industry is pure was really the only company in many many years to be able to achieve escape velocity break through a billion dollars i mean three part couldn't do it isilon couldn't do it compellent couldn't do it i could go on but pure was able to achieve that as an independent company and so you become a leader you look at the gartner magic quadrant you're a leader in there i mean if you've made it this far you've got to have some chops and so of course it's very competitive there are a number of other storage suppliers that have announced products that unify object and file so i'm interested in how pure differentiates why pure um it's a great question um and it's one that uh you know having been a long time puritan uh you know i take pride in answering um and it's actually a really simple answer um it's it's business model innovation and technology right the the technology that goes behind how we do what we do right and i don't mean the product right innovation is product but having a better support model for example um or having on the business model side you know evergreen storage right where we sort of look at your relationship to us as a subscription right um you know we're going to sort of take the thing that that you've had and we're going to modernize that thing in place over time such that you're not rebuying that same you know terabyte or you know petabyte of storage that you've that you that you've paid for over time so um you know sort of three legs of the stool uh that that have made you know pure clearly differentiated i think the market has has recognized that um you're right it's it's hard to break through to a billion dollars um but i look forward to the day that you know we we have two billion dollar products and i think with uh you know that rise in in unstructured data growing to 80 by 2025 and you know the massive transition that you know you guys have noted in in in your hdd slide i think it's a huge opportunity for us on you know the other unstructured data side of the house you know the other thing i'd add matt i've talked to cause about this is is it's simplicity first i've asked them why don't you do this why don't you do it and the answer is always the same is that adds complexity and we we put simplicity for the customer ahead of everything else and i think that served you very very well what about the economics of of unified file an object i mean if you bring in additional value presumably there's a there there's a cost to that but there's got to be also a business case behind it what kind of impact have you seen uh with customers yeah i mean look i'll i'll i'll go back to something i mentioned earlier which is just the reclamation of floor space and power and cooling right um you know there's a you know there's people people people want to search for kind of the the sexier element if you will when it comes to looking at how we how you derive value from something but the reality is if you're reducing your power consumption by you know by by a material percentage power bills matter in big in big data centers um you know customers typically are are facing you know a paradigm of well i i want to go to the cloud but you know the clouds are not being more expensive than i thought it was going to be or you know i figured out what i can use in the cloud i thought it was going to be everything but it's not going to be everything so hybrid's where we're landing but i want to be out of the data center business and i don't want to have a team of 20 storage people to match you know to administer my storage um you know so there's sort of this this very tangible value around you know hey if i could manage um you know multiple petabytes with one full-time engineer uh because the system uh to yoran kaz's point was radically simpler to administer didn't require someone to be running around swapping drives all the time would that be a value the answer is yes 100 of the time right and then you start to look at okay all right well on the uffo side from a product perspective hey if i have to manage a you know bespoke environment for this application if i have to manage a bespoke environment for this application and a bespoke environment for this application and this book environment for this application i'm managing four different things and can i actually share data across those four different things there's ways to share data but most customers it just gets too complex how do you even know what your what your gold.master copy is of data if you have it in four different places or you try to have it in four different places and it's four different siloed infrastructures so when you get to the sort of the side of you know how do we how do you measure value in uffo it's actually being able to have all of that data concentrated in one place so that you can share it from application to application got it i'm interested we use a couple minutes left i'm interested in the the update on flashblade you know generally but also i have a specific question i mean look getting file right is hard enough uh you just announced smb support for flashblade i'm interested in you know how that fits in i think it's kind of obvious with file and object converging but give us the update on on flashblade and maybe you could address that specific question yeah so um look i mean we're we're um you know tremendously excited about the growth of flashblade uh you know we we we found workloads we never expected to find um you know the rapid restore workload was one that was actually brought to us from from from a customer actually and has become you know one of our one of our top two three four you know workloads so um you know we're really happy with the trend we've seen in it um and you know mapping back to you know thinking about hdds and ssds you know we're well on a path to building a billion dollar business here so you know we're very excited about that um but to your point you know you don't just snap your fingers and get there right um you know we've learned that doing file and object uh is is harder than block um because there's more things that you have to go do for one you're basically focused on three protocols s b nfs and s3 not necessarily in that order um but to your point about smb uh you know we we are uh on the path through to releasing um you know smb uh full full native smb support in in the system that will allow us to uh service customers we have a limitation with some customers today where they'll have an s b portion of their nfs workflow um and we do great on the nfs side um but you know we didn't we didn't have the ability to plug into the s p component of their workflow so that's going to open up a lot of opportunity for us um on on that front um and you know we continue to you know invest significantly across the board in in areas like security which is you know become more than just a hot button you know today security's always been there but it feels like it's blazing hot today um and so you know going through the next couple years we'll be looking at uh you know developing some some um you know pretty material security elements of the product as well so uh well on a path to a billion dollars is the net on that and uh you know we're we're fortunate to have have smb here and we're looking forward to introducing that to to those customers that have you know nfs workloads today with an s p component yeah nice tailwind good tam expansion strategy matt thanks so much really appreciate you coming on the program we appreciate you having us and uh thanks much dave good to see you [Music] okay we're back with the convergence of file and object in a power panel this is a special content program made possible by pure storage and co-created with the cube now in this series what we're doing is we're exploring the coming together of file and object storage trying to understand the trends that are driving this convergence the architectural considerations that users should be aware of and which use cases make the most sense for so-called unified fast file in object storage and with me are three great guests to unpack these issues garrett belsner is the data center solutions architect he's with cdw scott sinclair is a senior analyst at enterprise strategy group he's got deep experience on enterprise storage and brings that independent analyst perspective and matt burr is back with us gentlemen welcome to the program thank you hey scott let me let me start with you uh and get your perspective on what's going on the market with with object the cloud a huge amount of unstructured data out there that lives in files give us your independent view of the trends that you're seeing out there well dave you know where to start i mean surprise surprise date is growing um but one of the big things that we've seen is we've been talking about data growth for what decades now but what's really fascinating is or changed is because of the digital economy digital business digital transformation whatever you call it now people are not just storing data they actually have to use it and so we see this in trends like analytics and artificial intelligence and what that does is it's just increasing the demand for not only consolidation of massive amounts of storage that we've seen for a while but also the demand for incredibly low latency access to that storage and i think that's one of the things that we're seeing that's driving this need for convergence as you put it of having multiple protocols consolidated onto one platform but also the need for high performance access to that data thank you for that a great setup i got like i wrote down three topics that we're going to unpack as a result of that so garrett let me let me go to you maybe you can give us the perspective of what you see with customers is is this is this like a push where customers are saying hey listen i need to converge my file and object or is it more a story where they're saying garrett i have this problem and then you see unified file and object as a solution yeah i think i think for us it's you know taking that consultative approach with our customers and really kind of hearing pain around some of the pipelines the way that they're going to market with data today and kind of what are the problems that they're seeing we're also seeing a lot of the change driven by the software vendors as well so really being able to support a disaggregated design where you're not having to upgrade and maintain everything as a single block has really been a place where we've seen a lot of customers pivot to where they have more flexibility as they need to maintain larger volumes of data and higher performance data having the ability to do that separate from compute and cache and those other layers are is really critical so matt i wonder if if you could you know follow up on that so so gary was talking about this disaggregated design so i like it you know distributed cloud etc but then we're talking about bringing things together in in one place right so square that circle how does this fit in with this hyper-distributed cloud edge that's getting built out yeah you know i mean i i could give you the easy answer on that but i could also pass it back to garrett in the sense that you know garrett maybe it's important to talk about um elastic and splunk and some of the things that you're seeing in in that world and and how that i think the answer to dave's question i think you can give you can give a pretty qualified answer relative what your customers are seeing oh that'd be great please yeah absolutely no no problem at all so you know i think with um splunk kind of moving from its traditional design and classic design whatever you want you want to call it up into smart store um that was kind of one of the first that we saw kind of make that move towards kind of separating object out and i think you know a lot of that comes from their own move to the cloud and updating their code to basically take advantage of object object in the cloud uh but we're starting to see you know with like vertica eon for example um elastic other folks taking that same type of approach where in the past we were building out many 2u servers we were jamming them full of uh you know ssds and nvme drives that was great but it doesn't really scale and it kind of gets into that same problem that we see with you know hyper convergence a little bit where it's you know you're all you're always adding something maybe that you didn't want to add um so i think it you know again being driven by software is really kind of where we're seeing the world open up there but that whole idea of just having that as a hub and a central place where you can then leverage that out to other applications whether that's out to the edge for machine learning or ai applications to take advantage of it i think that's where that convergence really comes back in but i think like scott mentioned earlier it's really folks are now doing things with the data where before i think they were really storing it trying to figure out what are we going to actually do with it when we need to do something with it so this is making it possible yeah and dave if i could just sort of tack on to the end of garrett's answer there you know in particular vertica with neon mode the ability to leverage sharded subclusters give you um you know sort of an advantage in terms of being able to isolate performance hot spots you an advantage to that is being able to do that on a flashblade for example so um sharded subclusters allow you to sort of say i'm you know i'm going to give prioritization to you know this particular element of my application and my data set but i can still share those share that data across those across those subclusters so um you know as you see you know vertica advance with eon mode or you see splunk advance with with smart store you know these are all sort of advancements that are you know it's a chicken in the egg thing um they need faster storage they need you know sort of a consolidated data storage data set um and and that's what sort of allows these things to drive forward yeah so vertica eon mode for those who don't know it's the ability to separate compute and storage and scale independently i think i think vertica if they're if they're not the only one they're one of the only ones i think they might even be the only one that does that in the cloud and on-prem and that sort of plays into this distributed you know nature of this hyper-distributed cloud i sometimes call it and and i'm interested in the in the data pipeline and i wonder scott if we could talk a little bit about that maybe we're unified object and file i mean i'm envisioning this this distributed mesh and then you know uffo is sort of a node on that that i i can tap when i need it but but scott what are you seeing as the state of infrastructure as it relates to the data pipeline and the trends there yeah absolutely dave so when i think data pipeline i immediately gravitate to analytics or or machine learning initiatives right and so one of the big things we see and this is it's an interesting trend it seems you know we continue to see increased investment in ai increased interest and people think and as companies get started they think okay well what does that mean well i got to go hire a data scientist okay well that data scientist probably needs some infrastructure and what they end what often happens in these environments is where it ends up being a bespoke environment or a one-off environment and then over time organizations run into challenges and one of the big challenges is the data science team or people whose jobs are outside of it spend way too much time trying to get the infrastructure to to keep up with their demands and predominantly around data performance so one of the one of the ways organizations that especially have artificial intelligence workloads in production and we found this in our research have started mitigating that is by deploying flash all across the data pipeline we have we have data on this sorry interrupt but yeah if you could bring up that that chart that would be great um so take us through this uh uh scott and share with us what we're looking at here yeah absolutely so so dave i'm glad you brought this up so we did this study um i want to say late last year uh one of the things we looked at was across artificial intelligence environments now one thing that you're not seeing on this slide is we went through and we asked all around the data pipeline and we saw flash everywhere but i thought this was really telling because this is around data lakes and when when or many people think about the idea of a data lake they think about it as a repository it's a place where you keep maybe cold data and what we see here is especially within production environments a pervasive use of flash storage so i think that 69 of organizations are saying their data lake is mostly flash or all flash and i think we have zero percent that don't have any flash in that environment so organizations are finding out that they that flash is an essential technology to allow them to harness the value of their data so garrett and then matt i wonder if you could chime in as well we talk about digital transformation and i sometimes call it you know the coveted forced march to digital transformation and and i'm curious as to your perspective on things like machine learning and the adoption and scott you may have a perspective on this as well you know we had to pivot we had to get laptops we had to secure the end points you know and vdi those became super high priorities what happened to you know injecting ai into my applications and and machine learning did that go in the back burner was that accelerated along with the need to digitally transform garrett i wonder if you could share with us what you saw with with customers last year yeah i mean i think we definitely saw an acceleration um i think folks are in in my market are still kind of figuring out how they inject that into more of a widely distributed business use case but again this data hub and allowing folks to now take advantage of this data that they've had in these data lakes for a long time i agree with scott i mean many of the data lakes that we have were somewhat flash accelerated but they were typically really made up of you know large capacity slower spinning near-line drive accelerated with some flash but i'm really starting to see folks now look at some of those older hadoop implementations and really leveraging new ways to look at how they consume data and many of those redesigned customers are coming to us wanting to look at all flash solutions so we're definitely seeing it we're seeing an acceleration towards folks trying to figure out how to actually use it in more of a business sense now or before i feel it goes a little bit more skunk works kind of people dealing with uh you know in a much smaller situation maybe in the executive offices trying to do some testing and things scott you're nodding away anything you can add in here yeah so first off it's great to get that confirmation that the stuff we're seeing in our research garrett's seeing you know out in the field and in the real world um but you know as it relates to really the past year it's been really fascinating so one of the things we study at esg is i.t buying intentions what are things what are initiatives that companies plan to invest in and at the beginning of 2020 we saw a heavy interest in machine learning initiatives then you transition to the middle of 2020 in the midst of covid some organizations continued on that path but a lot of them had the pivot right how do we get laptops to everyone how do we continue business in this new world well now as we enter into 2021 and hopefully we're coming out of this uh you know the pandemic era um we're getting into a world where organizations are pivoting back towards these strategic investments around how do i maximize the usage of data and actually accelerating those because they've seen the importance of of digital business initiatives over the past year yeah matt i mean when we exited 2019 we saw a narrowing of experimentation and our premise was you know that that organizations are going to start now operationalizing all their digital transformation experiments and and then we had a you know 10 month petri dish on on digital so what do you what are you seeing in this regard a 10 month petri dish is an interesting way to interesting way to describe it um you know we saw another there's another there's another candidate for pivot in there around ransomware as well right um you know security entered into the mix which took people's attention away from some of this as well i mean look i'd like to bring this up just a level or two um because what we're actually talking about here is progress right and and progress isn't is an inevitability um you know whether it's whether whether you believe that it's by 2025 or you or you think it's 2035 or 2050 it doesn't matter we're on a forced march to the eradication of disk and that is happening in many ways uh you know in many ways um due to some of the things that garrett was referring to and what scott was referring to in terms of what are customers demands for how they're going to actually leverage the data that they have and that brings me to kind of my final point on this which is we see customers in three phases there's the first phase where they say hey i have this large data store and i know there's value in there i don't know how to get to it or i have this large data store and i've started a project to get value out of it and we failed those could be customers that um you know marched down the hadoop path early on and they they got some value out of it um but they realized that you know hdfs wasn't going to be a modern protocol going forward for any number of reasons you know the first being hey if i have gold.master how do i know that i have gold.4 is consistent with my gold.master so data consistency matters and then you have the sort of third group that says i have these large data sets i know how to extract value from them and i'm already on to the verticas the elastics you know the splunks etc um i think those folks are the folks that that ladder group are the folks that kept their their their projects going because they were already extracting value from them the first two groups we we're seeing sort of saying the second half of this year is when we're going to begin really being picking up on these on these types of initiatives again well thank you matt by the way for for hitting the escape key because i think value from data really is what this is all about and there are some real blockers there that i kind of want to talk about you mentioned hdfs i mean we were very excited of course in the early days of hadoop many of the concepts were profound but at the end of the day it was too complicated we've got these hyper-specialized roles that are that are you know serving the business but it still takes too long it's it's too hard to get value from data and one of the blockers is infrastructure that the complexity of that infrastructure really needs to be abstracted taking up a level we're starting to see this in in cloud where you're seeing some of those abstraction layers being built from some of the cloud vendors but more importantly a lot of the vendors like pew are saying hey we can do that heavy lifting for you uh and we you know we have expertise in engineering to do cloud native so i'm wondering what you guys see uh maybe garrett you could start us off and other students as some of the blockers uh to getting value from data and and how we're going to address those in the coming decade yeah i mean i i think part of it we're solving here obviously with with pure bringing uh you know flash to a market that traditionally was utilizing uh much slower media um you know the other thing that i that i see that's very nice with flashblade for example is the ability to kind of do things you know once you get it set up a blade at a time i mean a lot of the things that we see from just kind of more of a you know simplistic approach to this like a lot of these teams don't have big budgets and being able to kind of break them down into almost a blade type chunk i think has really kind of allowed folks to get more projects and and things off the ground because they don't have to buy a full expensive system to run these projects so that's helped a lot i think the wider use cases have helped a lot so matt mentioned ransomware you know using safe mode as a place to help with ransomware has been a really big growth spot for us we've got a lot of customers very interested and excited about that and the other thing that i would say is bringing devops into data is another thing that we're seeing so kind of that push towards data ops and really kind of using automation and infrastructure as code as a way to now kind of drive things through the system the way that we've seen with automation through devops is really an area we're seeing a ton of growth with from a services perspective guys any other thoughts on that i mean we're i'll tee it up there we are seeing some bleeding edge which is somewhat counterintuitive especially from a cost standpoint organizational changes at some some companies uh think of some of the the the internet companies that do uh music uh for instance and adding podcasts etc and those are different data products we're seeing them actually reorganize their data architectures to make them more distributed uh and actually put the domain heads the business heads in charge of the the data and the data pipeline and that is maybe less efficient but but it's again some of these bleeding edge what else are you guys seeing out there that might be yes some harbingers of the next decade uh i'll go first um you know i think specific to um the the construct that you threw out dave one of the things that we're seeing is um you know the the application owner maybe it's the devops person but it's you know maybe it's it's it's the application owner through the devops person they're they're becoming more technical in their understanding of how infrastructure um interfaces with their with their application i think um you know what what we're seeing on the flashblade side is we're having a lot more conversations with application people than um just i.t people it doesn't mean that the it people aren't there the it people are still there for sure they have to deliver the service etc um but you know the days of of i.t you know building up a catalog of services and a business owner subscribing to one of those services you know picking you know whatever sort of fits their need um i don't think that constru i think that's the construct that changes going forward the application owner is becoming much more prescriptive about what they want the infrastructure to fit how they want the infrastructure to fit into their application and that's a big change and and for for um you know certainly folks like like garrett and cdw um you know they do a good job with this being able to sort of get to the application owner and bring those two sides together there's a tremendous amount of value there for us it's been a little bit of a retooling we've traditionally sold to the i.t side of the house and um you know we've had to teach ourselves how to go talk the language of of applications so um you know i think you pointed out a good a good a good construct there and and you know that that application owner taking playing a much bigger role in what they're expecting uh from the performance of it infrastructure i think is is is a key is a key change interesting i mean that definitely is a trend that's put you guys closer to the business where the the infrastructure team is is serving the business as opposed to sometimes i talk to data experts and they're frustrated uh especially data owners or or data product builders who are frustrated that they feel like they have to beg beg the the data pipeline team to get you know new data sources or get data out how about the edge um you know maybe scott you can kick us off i mean we're seeing you know the emergence of edge use cases ai inferencing at the edge a lot of data at the edge what are you seeing there and and how does this unified object i'll bring us back to that and file fit wow dave how much time do we have um two minutes first of all scott why don't you why don't you just tell everybody what the edge is yeah you got it figured out all right how much time do you have matt at the end of the day and that that's that's a great question right is if you take a step back and i think it comes back today of something you mentioned it's about extracting value from data and what that means is when you extract value from data what it does is as matt pointed out the the influencers or the users of data the application owners they have more power because they're driving revenue now and so what that means is from an i.t standpoint it's not just hey here are the services you get use them or lose them or you know don't throw a fit it is no i have to i have to adapt i have to follow what my application owners mean now when you bring that back to the edge what it means is is that data is not localized to the data center i mean we just went through a nearly 12-month period where the entire workforce for most of the companies in this country had went distributed and business continued so if business is distributed data is distributed and that means that means in the data center that means at the edge that means that the cloud that means in all other places in tons of places and what it also means is you have to be able to extract and utilize data anywhere it may be and i think that's something that we're going to continue to and continue to see and i think it comes back to you know if you think about key characteristics we've talked about things like performance and scale for years but we need to start rethinking it because on one hand we need to get performance everywhere but also in terms of scale and this ties back to some of the other initiatives and getting value from data it's something i call that the massive success problem one of the things we see especially with with workloads like machine learning is businesses find success with them and as soon as they do they say well i need about 20 of these projects now all of a sudden that overburdens it organizations especially across across core and edge and cloud environments and so when you look at environments ability to meet performance and scale demands wherever it needs to be is something that's really important you know so dave i'd like to um just sort of tie together sort of two things that um i think that i heard from scott and garrett that i think are important and it's around this concept of scale um you know some of us are old enough to remember the day when kind of a 10 terabyte blast radius was too big of a blast radius for people to take on or a terabyte of storage was considered to be um you know an exemplary budget environment right um now we sort of think as terabytes kind of like we used to think of as gigabytes in some ways um petabyte like you don't have to explain anybody what a petabyte is anymore um and you know what's on the horizon and it's not far are our exabyte type data set workloads um and you start to think about what could be in that exabyte of data we've talked about how you extract that value we've talked about sort of um how you start but if the scale is big not everybody's going to start at a petabyte or an exabyte to garrett's point the ability to start small and grow into these products or excuse me these projects i think a is a really um fundamental concept here because you're not going to just go by i'm going to kick off a five petabyte project whether you do that on disk or flash it's going to be expensive right but if you could start at a couple hundred terabytes not just as a proof of concept but as something that you know you could get predictable value out of that then you could say hey this either scales linearly or non-linearly in a way that i can then go map my investments to how i can go dig deeper into this that's how all of these things are gonna that's how these successful projects are going to start because the people that are starting with these very large you know sort of um expansive you know greenfield projects at multi-petabyte scale it's gonna be hard to realize near-term value excellent we gotta wrap but but garrett i wonder if you could close when you look forward you talk to customers do you see this unification of of file and object is it is this an evolutionary trend is it something that is that that is that is that is going to be a lever that customers use how do you see it evolving over the next two three years and beyond yeah i mean i think from our perspective i mean just from what we're seeing from the numbers within the market the amount of growth that's happening with unstructured data is really just starting to finally really kind of hit this data deluge or whatever you want to call it that we've been talking about for so many years it really does seem to now be becoming true as we start to see things scale out and really folks settle into okay i'm going to use the cloud to to start and maybe train my models but now i'm going to get it back on prem because of latency or security or whatever the the um decision points are there this is something that is not going to slow down and i think you know folks like pure having the ability to have the tools that they give us um to use and bring to market with our customers are really key and critical for us so i see it as a huge growth area and a big focus for us moving forward guys great job unpacking a topic that you know it's covered a little bit but i think we we covered some ground that is uh that is new and so thank you so much for those insights and that data really appreciate your time thanks steve thanks yeah thanks dave okay and thank you for watching the convergence of file and object keep it right there right back after this short break innovation impact influence welcome to the cube disruptors developers and practitioners learn from the voices of leaders who share their personal insights from the hottest digital events around the globe enjoy the best this community has to offer on the cube your global leader in high-tech digital coverage [Music] okay now we're going to get the customer perspective on object and we'll talk about the convergence of file and object but really focusing on the object piece this is a content program that's being made possible by pure storage and it's co-created with the cube christopher cb bond is here he's a lead architect for microfocus the enterprise data warehouse and principal data engineer at microfocus cb welcome good to see you thanks dave good to be here so tell us more about your role at microfocus it's a pan microfocus role of course we know the company is a multinational software firm and acquired the software assets of hp of course including vertica tell us where you fit yeah so microfocus is uh you know it's like i said wide worldwide uh company that uh sells a lot of software products all over the place to governments and so forth and um it also grows often by acquiring other companies so there is the problem of of integrating new companies and their data and so what's happened over the years is that they've had a a number of different discrete data systems so you've got this data spread all over the place and they've never been able to get a full complete introspection on the entire business because of that so my role was come in design a central data repository an enterprise data warehouse that all reporting could be generated against and so that's what we're doing and we selected vertica as the edw system and pure storage flashblade as the communal repository okay so you obviously had experience with with vertica in your in your previous role so it's not like you were starting from scratch but but paint a picture of what life was like before you embarked on this sort of consolidated a approach to your your data warehouse what was it just disparate data all over the place a lot of m a going on where did the data live right so again the data was all over the place including under people's desks in just dedicated you know their their own private uh sql servers it a lot of data in in um microfocus is run on sql server which has pros and cons because that's a great uh transactional database but it's not really good for analytics in my opinion so uh but a lot of stuff was running on that they had one vertica instance that was doing some select uh reporting wasn't a very uh powerful system and it was what they call vertica enterprise mode where had dedicated nodes which um had the compute and storage um in the same locus on each uh server okay so vertica eon mode is a whole new world because it separates compute from storage you mentioned eon mode uh and the ability to to to scale storage and compute independently we wanted to have the uh analytics olap stuff close to the oltp stuff right so that's why they're co-located very close to each other and so uh we could what's nice about this situation is that these s3 objects it's an s3 object store on the pure flash plate we could copy those over if we needed to uh aws and we could spin up um a version of vertica there and keep going it's it's like a tertiary dr strategy because we actually have a we're setting up a second flashblade vertica system geo-located elsewhere for backup and we can get into it if you want to talk about how the latest version of the pure software for the flashblade allows synchronization across network boundaries of those flash plays which is really nice because if uh you know there's a giant sinkhole opens up under our colo facility and we lose that thing then we just have to switch the dns and we were back in business off the dr and then if that one was to go we could copy those objects over to aws and be up and running there so we're feeling pretty confident about being able to weather whatever comes along so you're using the the pure flash blade as an object store um most people think oh object simple but slow uh not the case for you is that right not the case at all it's ripping um well you have to understand about vertica and the way it stores data it stores data in what they call storage containers and those are immutable okay on disk whether it's on aws or if you had a enterprise mode vertica if you do an update or delete it actually has to go and retrieve that object container from disk and it destroys it and rebuilds it okay which is why you don't you want to avoid updates and deletes with vertica because the way it gets its speed is by sorting and ordering and encoding the data on disk so it can read it really fast but if you do an operation where you're deleting or updating a record in the middle of that then you've got to rebuild that entire thing so that actually matches up really well with s3 object storage because it's kind of the same way uh it gets destroyed and rebuilt too okay so that matches up very well with vertica and we were able to design this system so that it's append only now we had some reports that were running in sql server okay uh which were taking seven days so we moved that to uh to vertica from sql server and uh we rewrote the queries which were which had been written in t sql with a bunch of loops and so forth and we were to get this is amazing it went from seven days to two seconds to generate this report which has tremendous value uh to the company because it would have to have this long cycle of seven days to get a new introspection in what they call their knowledge base and now all of a sudden it's almost on demand two seconds to generate it that's great and that's because of the way the data is stored and uh the s3 you asked about oh you know is it slow well not in that context because what happens really with vertica eon mode is that it can they have um when you set up your compute nodes they have local storage also which is called the depot it's kind of a cache okay so the data will be drawn from the flash and cached locally uh and that was it was thought when they designed that oh you know it's that'll cut down on the latency okay but it turns out that if you have your compute nodes close meaning minimal hops to the flashblade that you can actually uh tell vertica you know don't even bother caching that stuff just read it directly on the fly from the from the flashblade and the performance is still really good it depends on your situation but i know for example a major telecom company that uh uses the same topology as we're talking about here they did the same thing they just they just dropped the cache because the flash player was able to to deliver the the data fast enough so that's you're talking about that that's speed of light issues and just the overhead of of of switching infrastructure is that that gets eliminated and so as a result you can go directly to the storage array that's correct yeah it's it's like it's fast enough that it's it's almost as if it's local to the compute node uh but every situation is different depending on your uh your knees if you've got like a few tables that are heavily used uh then yeah put them um put them in the cash because that'll be probably a little bit faster but if you have a lot of ad hoc queries that are going on you know you may exceed the storage of the local cache and then you're better off having it uh just read directly from the uh from the flash blade got it look it pure's a fit i mean i sound like a fanboy but pure is all about simplicity so is object so that means you don't have to you know worry about wrangling storage and worrying about luns and all that other you know nonsense and and file i've been burned by hardware in the past you know where oh okay they're building to a price and so they cheap out on stuff like fans or other things and these these components fail and the whole thing goes down but this hardware is super super good quality and uh so i'm i'm happy with the quality that we're getting so cb last question what's next for you where do you want to take this uh this this initiative well we are in the process now of we um when so i i designed this system to combine the best of the kimball approach to data warehousing and the inland approach okay and what we do is we bring over all the data we've got and we put it into a pristine staging layer okay like i said it's uh because it's append only it's essentially a log of all the transactions that are happening in this company just they appear okay and then from the the kimball side of things we're designing the data marts now so that that's what the end users actually interact with and so we're we're taking uh the we're examining the transactional systems to say how are these business objects created what's what's the logic there and we're recreating those logical models in uh in vertica so we've done a handful of them so far and it's working out really well so going forward we've got a lot of work to do to uh create just about every object that that the company needs cb you're an awesome guest to really always a pleasure talking to you and uh thank you congratulations and and good luck going forward stay safe thank you [Music] okay let's summarize the convergence of file and object first i want to thank our guests matt burr scott sinclair garrett belsener and c.b bohn i'm your host dave vellante and please allow me to briefly share some of the key takeaways from today's program so first as scott sinclair of esg stated surprise surprise data's growing and matt burr he helped us understand the growth of unstructured data i mean estimates indicate that the vast majority of data will be considered unstructured by mid-decade 80 or so and obviously unstructured data is growing very very rapidly now of course your definition of unstructured data and that may vary across across a wide spectrum i mean there's video there's audio there's documents there's spreadsheets there's chat i mean these are generally considered unstructured data but of course they all have some type of structure to them you know perhaps it's not as strict as a relational database but there's certainly metadata and certain structure to these types of use cases that i just mentioned now the key to what pure is promoting is this idea of unified fast file and object uffo look object is great it's inexpensive it's simple but historically it's been less performant so good for archiving or cheap and deep types of examples organizations often use file for higher performance workloads and let's face it most of the world's data lives in file formats what pure is doing is bringing together file and object by for example supporting multiple protocols ie nfs smb and s3 s3 of course has really given new life to object over the past decade now the key here is to essentially enable customers to have the best of both worlds not having to trade off performance for object simplicity and a key discussion point that we've had on the program has been the impact of flash on the long slow death of spinning disk look hard disk drives they had a great run but hdd volumes they peaked in 2010 and flash as you well know has seen tremendous volume growth thanks to the consumption of flash in mobile devices and then of course its application into the enterprise and that's volume is just going to keep growing and growing and growing the price declines of flash are coming down faster than those of hdd so it's the writing's on the wall it's just a matter of time so flash is riding down that cost curve very very aggressively and hdd has essentially become you know a managed decline business now by bringing flash to object as part of the flashblade portfolio and allowing for multiple protocols pure hopes to eliminate the dissonance between file and object and simplify the choice in other words let the workload decide if you have data in a file format no problem pure can still bring the benefits of simplicity of object at scale to the table so again let the workload inform what the right strategy is not the technical infrastructure now pure course is not alone there are others supporting this multi-protocol strategy and so we asked matt burr why pure or what's so special about you and not surprisingly in addition to the product innovation he went right to pure's business model advantages i mean for example with its evergreen support model which was very disruptive in the marketplace you know frankly pure's entire business disrupted the traditional disk array model which was fundamentally was flawed pure forced the industry to respond and when it achieved escape velocity velocity and pure went public the entire industry had to react and a big part of the pure value prop in addition to this business model innovation that we just discussed is simplicity pure's keep its simple approach coincided perfectly with the ascendancy of cloud where technology organizations needed cloud-like simplicity for certain workloads that were never going to move into the cloud they're going to stay on-prem now i'm going to come back to this but allow me to bring in another concept that garrett and cb really highlighted and that is the complexity of the data pipeline and what do you mean what do i mean by that and why is this important so scott sinclair articulated he implied that the big challenge is organizations their data full but insights are scarce scarce a lot of data not as much insights it takes time too much time to get to those insights so we heard from our guests that the complexity of the data pipeline was a barrier to getting to faster insights now cb bonds shared how he streamlined his data architecture using vertica's eon mode which allowed him to scale compute independently of storage so that brought critical flexibility and improved economics at scale and flashblade of course was the back-end storage for his data warehouse efforts now the reason i think this is so important is that organizations are struggling to get insights from data and the complexity associated with the data pipeline and data life cycles let's face it it's overwhelming organizations and there the answer to this problem is a much longer and different discussion than unifying object and file that's you know i can spend all day talking about that but let's focus narrowly on the part of the issue that is related to file and object so the situation here is that technology has not been serving the business the way it should rather the formula is twisted in the world of data and big data and data architectures the data team is mired in complex technical issues that impact the time to insights now part of the answer is to abstract the underlying infrastructure complexity and create a layer with which the business can interact that accelerates instead of impedes innovation and unifying file and object is a simple example of this where the business team is not blocked by infrastructure nuance like does this data reside in a file or object format can i get to it quickly and inexpensively in a logical way or is the infrastructure in a stovepipe and blocking me so if you think about the prevailing sentiment of how the cloud is evolving to incorporate on premises workloads that are hybrid and configurations that are working across clouds and now out to the edge this idea of an abstraction layer that essentially hides the underlying infrastructure is a trend we're going to see evolve this decade now is uffo the be all end-all answer to solving all of our data pipeline challenges no no of course not but by bringing the simplicity and economics of object together with the ubiquity and performance of file uffo makes it a lot easier it simplifies life organizations that are evolving into digital businesses which by the way is every business so we see this as an evolutionary trend that further simplifies the underlying technology infrastructure and does a better job supporting the data flows for organizations so they don't have to spend so much time worrying about the technology details that add a little value to the business okay so thanks for watching the convergence of file and object and thanks to pure storage for making this program possible this is dave vellante for the cube we'll see you next time [Music] you
SUMMARY :
on the nfs side um but you know we
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
garrett belsner | PERSON | 0.99+ |
matt burr | PERSON | 0.99+ |
2010 | DATE | 0.99+ |
2050 | DATE | 0.99+ |
270 terabytes | QUANTITY | 0.99+ |
seven days | QUANTITY | 0.99+ |
2021 | DATE | 0.99+ |
scott sinclair | PERSON | 0.99+ |
2035 | DATE | 0.99+ |
2019 | DATE | 0.99+ |
four | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
two seconds | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
matt burr | PERSON | 0.99+ |
first phase | QUANTITY | 0.99+ |
dave | PERSON | 0.99+ |
dave vellante | PERSON | 0.99+ |
scott sinclair | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
250 terabytes | QUANTITY | 0.99+ |
10 terabyte | QUANTITY | 0.99+ |
zero percent | QUANTITY | 0.99+ |
100 | QUANTITY | 0.99+ |
steve | PERSON | 0.99+ |
gary | PERSON | 0.99+ |
two billion dollar | QUANTITY | 0.99+ |
garrett | PERSON | 0.99+ |
two minutes | QUANTITY | 0.99+ |
two weeks later | DATE | 0.99+ |
three topics | QUANTITY | 0.99+ |
two sides | QUANTITY | 0.99+ |
two weeks ago | DATE | 0.99+ |
billion dollars | QUANTITY | 0.99+ |
mid-decade 80 | DATE | 0.99+ |
today | DATE | 0.99+ |
cdw | PERSON | 0.98+ |
three phases | QUANTITY | 0.98+ |
80 | QUANTITY | 0.98+ |
billions of objects | QUANTITY | 0.98+ |
10 month | QUANTITY | 0.98+ |
one device | QUANTITY | 0.98+ |
an hour | QUANTITY | 0.98+ |
one platform | QUANTITY | 0.98+ |
scott | ORGANIZATION | 0.97+ |
last year | DATE | 0.97+ |
five petabyte | QUANTITY | 0.97+ |
scott | PERSON | 0.97+ |
cassandra | PERSON | 0.97+ |
one | QUANTITY | 0.97+ |
single block | QUANTITY | 0.97+ |
one system | QUANTITY | 0.97+ |
next decade | DATE | 0.96+ |
tons of places | QUANTITY | 0.96+ |
both worlds | QUANTITY | 0.96+ |
vertica | TITLE | 0.96+ |
matt | PERSON | 0.96+ |
both | QUANTITY | 0.96+ |
69 of organizations | QUANTITY | 0.96+ |
billion dollars | QUANTITY | 0.95+ |
pandemic | EVENT | 0.95+ |
first | QUANTITY | 0.95+ |
three great guests | QUANTITY | 0.95+ |
next year | DATE | 0.95+ |
DV Pure Storage 208
>> Thank you, sir. All right, you ready to roll? >> Ready. >> All right, we'll go ahead and go in five, four, three, two. >> Okay, let's summarize the convergence of file and object. First, I want to thank our guests, Matt Burr, Scott Sinclair, Garrett Belsner, and CB Bonne. I'm your host, Dave Vellante, and please allow me to briefly share some of the key takeaways from today's program. So first, as Scott Sinclair of ESG stated surprise, surprise, data's growing. And Matt Burr, he helped us understand the growth of unstructured data. I mean, estimates indicate that the vast majority of data will be considered unstructured by mid decade, 80% or so. And obviously, unstructured data is growing very, very rapidly. Now, of course, your definition of unstructured data, now that may vary across a wide spectrum. I mean, there's video, there's audio, there's documents, there's spreadsheets, there's chat. I mean, these are generally considered unstructured data but of course they all have some type of structure to them. You know, perhaps it's not as strict as a relational database, but there's certainly metadata and certain structure to these types of use cases that I just mentioned. Now, the key to what Pure is promoting is this idea of unified fast file and object, U-F-F-O. Look, object is great, it's inexpensive, it's simple, but historically, it's been less performant, so good for archiving, or cheap and deep types of examples. Organizations often use file for higher performance workloads and let's face it, most of the world's data lives in file formats. What Pure is doing is bringing together file and object by, for example, supporting multiple protocols, ie, NFS, SMB, and S3. S3, of course, has really given a new life to object over the past decade. Now, the key here is to essentially enable customers to have the best of both worlds, not having to trade off performance for object simplicity. And a key discussion point that we've had in the program has been the impact of Flash on the long, slow, death of spinning disk. Look, hard disk drives, they had a great run, but HDD volumes, they peaked in 2010, and Flash, as you well know, has seen tremendous volume growth thanks to the consumption of Flash in mobile devices and then of course, its application into the enterprise. And as volume is just going to keep growing and growing, and growing. the price declines of Flash are coming down faster than those of HDD. So it's, the writing's on the wall. It's just a matter of time. So Flash is riding down that cost curve very, very aggressively and HDD has essentially become a managed decline business. Now, by bringing Flash to object as part of the FlashBlade portfolio and allowing for multiple protocols, Pure hopes to eliminate the dissonance between file and object and simplify the choice. In other words, let the workload decide. If you have data in a file format, no problem. Pure can still bring the benefits of simplicity of object at scale to the table. So again, let the workload inform what the right strategy is not the technical infrastructure. Now Pure, of course, is not alone. There are others supporting this multi-protocol strategy. And so we asked Matt Burr why Pure, what's so special about you? And not surprisingly, in addition to the product innovation, he went right to Pure's business model advantages. I mean, for example, with its Evergreen support model which was very disruptive in the marketplace. You know, frankly, Pure's entire business disrupted the traditional disk array model which was, fundamentally, it was flawed. Pure forced the industry to respond. And when it achieved escape velocity and Pure went public, the entire industry had to react. And a big part of the Pure value prop in addition to this business model innovation that we just discussed is simplicity. Pure's keep it simple approach coincided perfectly with the ascendancy of cloud where technology organizations needed cloud-like simplicity for certain workloads that were never going to move into the cloud. They were going to stay on-prem. Now I'm going to come back to this but allow me to bring in another concept that Garrett and CB really highlighted, and that is the complexity of the data pipeline. And what do I mean, what do I mean by that, and why is this important? So Scott Sinclair articulated or he implied that the big challenge is organizations, they're data full, but insights are scarce; a lot of data, not as much insights, and it takes time, too much time to get to those insights. So we heard from our guests that the complexity of the data pipeline was a barrier to getting to faster insights. Now, CB Bonne shared how he streamlined his data architecture using Vertica's Eon Mode which allowed him to scale, compute, independently of storage, so that brought critical flexibility and improved economics at scale. And FlashBlade, of course, was the backend storage for his data warehouse efforts. Now, the reason I think this is so important is that organizations are struggling to get insights from data and the complexity associated with the data pipeline and data lifecycles, let's face it, it's overwhelming organizations. And there, the answer to this problem is a much longer and different discussion than unifying object and file. That's, you know, I could spend all day talking about that, but let's focus narrowly on the part of the issue that is related to file and object. So the situation here is the technology has not been serving the business the way it should. Rather, the formula is twisted in the world of data and big data, and data architectures. The data team is mired in complex technical issues that impact the time to insights. Now, part of the answer is to abstract the underlying infrastructure complexity and create a layer with which the business can interact that accelerates instead of impedes innovation. And unifying file and object is a simple example of this where the business team is not blocked by infrastructure nuance, like does this data reside in the file or object format? Can I get to it quickly and inexpensively in a logical way or is the infrastructure in a stovepipe and blocking me? So if you think about the prevailing sentiment of how the cloud is evolving to incorporate on premises, workloads that are hybrid, and configurations that are working across clouds, and now out to the edge, this idea of an abstraction layer that essentially hides the underlying infrastructure is a trend we're going to see evolve this decade. Now, is UFFO the be-all end-all answer to solving all of our data pipeline challenges? No, no, of course not. But by bringing the simplicity and economics of object together with the ubiquity and performance of file, UFFO makes it a lot easier. It simplifies a life organizations that are evolving into digital businesses, which by the way, is every business. So, we see this as an evolutionary trend that further simplifies the underlying technology infrastructure and does a better job supporting the data flows for organizations so they didn't have to spend so much time worrying about the technology details that add little value to the business. Okay, so thanks for watching the convergence of file and object and thanks to Pure Storage for making this program possible. This is Dave Vellante for theCUBE. We'll see you next time.
SUMMARY :
All right, you ready to roll? in five, four, three, two. that impact the time to insights.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Matt Burr | PERSON | 0.99+ |
Scott Sinclair | PERSON | 0.99+ |
Garrett Belsner | PERSON | 0.99+ |
ESG | ORGANIZATION | 0.99+ |
80% | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
CB Bonne | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
2010 | DATE | 0.99+ |
First | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
first | QUANTITY | 0.98+ |
four | QUANTITY | 0.98+ |
three | QUANTITY | 0.98+ |
both worlds | QUANTITY | 0.98+ |
Flash | TITLE | 0.97+ |
CB | PERSON | 0.97+ |
Vertica | ORGANIZATION | 0.97+ |
Pure Storage | ORGANIZATION | 0.96+ |
Pure | ORGANIZATION | 0.96+ |
Garrett | PERSON | 0.96+ |
Evergreen | ORGANIZATION | 0.86+ |
past decade | DATE | 0.59+ |
UFFO | ORGANIZATION | 0.59+ |
Pure Storage 208 | COMMERCIAL_ITEM | 0.59+ |
Pure | PERSON | 0.58+ |
this decade | DATE | 0.5+ |
FlashBlade | ORGANIZATION | 0.43+ |
FlashBlade | TITLE | 0.37+ |
Matt Burr, General Manager, FlashBlade, Pure Storage | The Convergence of File and Object
from around the globe it's thecube presenting the convergence of file and object brought to you by pure storage we're back with the convergence of file and object a special program made possible by pure storage and co-created with the cube so in this series we're exploring that convergence between file and object storage we're digging into the trends the architectures and some of the use cases for unified fast file and object storage uffo with me is matt burr who's the vice president general manager of flashblade at pure storage hello matt how you doing i'm doing great morning dave how are you good thank you hey let's start with a little 101 you know kind of the basics what is unified fast file and object yeah so look i mean i think you got to start with first principles talking about the rise of unstructured data so when we think about unstructured data you sort of think about the projections 80 of data by 2025 is going to be unstructured data whether that's machine generated data or you know ai and ml type workloads you start to sort of see this i don't want to say it's a boom uh but it's sort of a renaissance for unstructured data if you will where we move away from you know what we've traditionally thought of as general purpose nas and and file shares to you know really things that focus on uh fast object taking advantage of s3 cloud native applications that need to integrate with applications on site um you know ai workloads ml workloads tend to look to share data across uh you know multiple data sets and you really need to have a platform that can deliver both highly performant and scalable fast file and object from one system so talk a little bit more about some of the drivers that you know bring forth that need to unify file an object yeah i mean look you know there's a there's there's a real challenge um in managing you know bespoke uh bespoke infrastructure or architectures around general purpose nas and daz etc so um if you think about how a an architect sort of looks at an application they might say well okay i need to have um you know fast daz storage proximal to the application um but that's gonna require a tremendous amount of dabs which is a tremendous amount of drives right hard drives are you know historically pretty pretty pretty unwieldy to manage because you're replacing them relatively consistently at multi-petabyte scale so you start to look at things like the complexity of das you start to look at the complexity of general purpose nas and you start to just look at quite frankly something that a lot of people don't really want to talk about anymore but actual data center space right like consolidation matters the ability to take you know something that's the size of a microwave like a modern flash blade or a modern um you know uffo device replaces something that might be you know the size of three or four or five refrigerators so matt why is is now the right time for this i mean for years nobody really paid much attention to object s3 already obviously changed you know that course most of the world's data is still stored in file formats and you get there with nfs or smb why is now the time to think about unifying object and and file well because we're moving to things like a contactless society um you know the the things that we're going to do are going to just require a tremendous amount more compute power network and quite frankly storage throughput and you know i can give you two sort of real primary examples here right um you know warehouses are being you know taken over by robots if you will um it's not a war it's a it's a it's sort of a friendly advancement in you know how do i how do i store a box in a warehouse and you know we have we have a customer who focuses on large sort of big box distribution warehousing and you know a box that carried a an object uh two weeks ago might have a different box size two weeks later well that robot needs to know where the space is in the data center in order to put it but also needs to be able to process hey i don't want to put the thing that i'm going to access the most in the back of the warehouse i'm going to put that thing in the front of the warehouse all of those types of data you know sort of real time you can think of the robot as almost an edge device uh is processing in real time unstructured data and its object right so it's sort of the emergence of these new types of workloads and i give you the opposite example the other end of the spectrum is ransomware right you know today you know we'll talk to customers and they'll say quite commonly hey if you know anybody can sell me a backup device i need something that can restore quickly if you had the ability to restore something in 270 terabytes an hour or 250 terabytes an hour that's much faster when you're dealing with a ransomware attack you want to get your data back quickly you know so i want to actually i was going to ask you about that later but since you brought it up what is the right i guess call it architecture for for for ransomware i mean how and explain like how unified object and file would support me i get the fast recovery but how would you recommend a customer uh go about architecting a ransomware proof you know system yeah well you know with with flashblade and and with flasharray there's an actual feature called called safe mode and that safe mode actually protects uh the snapshots and and the data from uh sort of being is a part of the of the ransomware event and so if you're in a type of ransomware situation like this you're able to leverage safe mode and you say okay what happens in a ransomware attack is you can't get access to your data and so you know the bad guy the perpetrator is basically saying hey i'm not going to give you access to your data until you pay me you know x in bitcoin or whatever it might be right um with with safe mode those snapshots are actually protected outside of the ransomware blast zone and you can bring back those snapshots because what's your alternative if you're not doing something like that your alternative is either to pay and unlock your data or you have to start retouring restoring excuse me from tape or slow disk that could take you days or weeks to get your data back so leveraging safe mode um you know in either the flash for the flash blade product is a great way to go about uh architecting against ransomware i got to put my i'm thinking like a customer now so safe mode so that's an immutable mode right can't change the data um is it can can an administrator go in and change that mode can he turn it off do i still need an air gap for example what would you recommend there yeah so there there are still um uh you know sort of our back or rollback role-based access control policies uh around who can access that safe mode and who can right okay so uh anyway subject for a different day i want to i want to actually bring up uh if you don't object a topic that i think used to be really front and center and it now be is becoming front and center again i mean wikibon just produced a research note forecasting the future of flash and hard drives and those of you who follow us know we've done this for quite some time and you can if you could bring up the chart here you you could see and we see this happening again it was originally we forecast the the death of of quote unquote high spin speed disk drives which is kind of an oxymoron but you can see on here on this chart this hard disk had a magnificent journey but they peaked in volume in manufacturing volume in 2010 and the reason why that is is so important is that volumes now are steadily dropping you can see that and we use wright's law to explain why this is a problem and wright's law essentially says that as you your cumulative manufacturing volume doubles your cost to manufacture decline by a constant percentage now i won't go too much detail on that but suffice it to say that flash volumes are growing very rapidly hdd volumes aren't and so flash because of consumer volumes can take advantage of wright's law and that constant reduction and that's what's really important for the next generation which is always more expensive to build and so this kind of marks the beginning of the end matt what do you think what what's the future hold for spinning disc in your view uh well i can give you the answer on two levels on a personal level uh it's why i come to work every day uh you know the the eradication or or extinction of an inefficient thing um you know i like to say that inefficiency is the bane of my existence uh and i think hard drives are largely inefficient and i'm willing to accept the sort of long-standing argument that um you know we've seen this transition in block right and we're starting to see it repeat itself in in unstructured data um and i'm willing to accept the argument that cost is a vector here and it most certainly is right hdds have been considerably cheaper uh than than than flash storage um you know even to this day uh you know up to this point right but we're starting to approach the point where you sort of reach a 3x sort of you know differentiator between the cost of an hdd and an sdd and you know that really is that point in time when uh you begin to pick up a lot of volume and velocity and so you know that tends to map directly to you know what you're seeing here which is you know a slow decline uh which i think is going to become even more rapid kind of probably starting around next year where you start to see sds excuse me ssds uh you know really replacing hdds uh at a much more rapid clip particularly on the unstructured data side and it's largely around cost the the workloads that we talked about robots and warehouses or you know other types of advanced machine learning and artificial intelligence type applications and workflows you know they require a degree of performance that a hard drive just can't deliver we are we are seeing sort of the um creative innovative uh disruption of an entire industry right before our eyes it's a fun thing to live through yeah and and we would agree i mean it doesn't the premise there is it doesn't have to be less expensive we think it will be by you know the second half or early second half of this decade but even if it's a we think around a 3x delta the value of of ssd relative to spinning disk is going to overwhelm just like with your laptop you know it got to the point where you said why would i ever have a spinning disc in my laptop we see the same thing happening here um and and so and we're talking about you know raw capacity you know put in compression and dedupe and everything else that you really can't do with spinning discs because of the performance issues you can do with flash okay let's come back to uffo can we dig into the challenges specifically that that this solves for customers give me give us some examples yeah so you know i mean if we if we think about the examples um you know the the robotic one um i think is is is the one that i think is the marker for you know kind of of of the the modern side of of of what we see here um but what we're you know what we're what we're seeing from a trend perspective which you know not everybody's deploying robots right um you know there's there's many companies that are you know that aren't going to be in either the robotic business uh or or even thinking about you know sort of future type oriented type things but what they are doing is greenfield applications are being built on object um generally not on not on file and and not on block and so you know the rise of of object as sort of the the sort of let's call it the the next great protocol for um you know for uh for for modern workloads right this is this is that that modern application coming to the forefront and that could be anything from you know financial institutions you know right down through um you know we've even see it and seen it in oil and gas uh we're also seeing it across across healthcare uh so you know as as as companies take the opportunity as industries to take this opportunity to modernize you know they're modernizing not on things that are are leveraging you know um you know sort of archaic disk technology they're they're they're really focusing on on object but they still have file workflows that they need to that they need to be able to support and so having the ability to be able to deliver those things from one device in a capacity orientation or a performance orientation while at the same time dramatically simplifying the overall administration of your environment both physically and non-physically is a key driver so the great thing about object is it's simple it's a kind of a get put metaphor um it's it scales out you know because it's got metadata associated with the data uh and and it's cheap the drawback is you don't necessarily associate it with high performance and and as well most applications don't you know speak in that language they speak in the language of file you know or as you mentioned block so i i see real opportunities here if i have some some data that's not necessarily frequently accessed you know every day but yet i want to then whether end of quarter or whatever it is i want to i want to or machine learning i want to apply some ai to that data i want to bring it in and then apply a file format uh because for performance reasons is that right maybe you could unpack that a little bit yeah so um you know we see i mean i think you described it well right um but i don't think object necessarily has to be slow um and nor does it have to be um you know because when you think about you brought up a good point with metadata right being able to scale to a billions of objects being able to scale to billions of objects excuse me is of value right um and i think people do traditionally associate object with slow but it's not necessarily slow anymore right we we did a sort of unofficial survey of of of our of our customers and our employee base and when people described object they thought of it as like law firms and storing a word doc if you will um and that that's just you know i think that there's a lack of understanding or a misnomer around what modern what modern object has become and perform an object particularly at scale when we're talking about billions of objects you know that's the next frontier right um is it at pace performance wise with you know the other protocols no but it's making leaps and grounds so you talked a little bit more about some of the verticals that you see i mean i think when i think of financial services i think transaction processing but of course they have a lot of tons of unstructured data are there any patterns you're seeing by by vertical market um we're you know we're not that's the interesting thing um and you know um as a as a as a as a company with a with a block heritage or a block dna those patterns were pretty easy to spot right there were a certain number of databases that you really needed to support oracle sql some postgres work etc then kind of the modern databases around cassandra and things like that you knew that there were going to be vmware environments you know you could you could sort of see the trends and where things were going unstructured data is such a broader horizontal um thing right so you know inside of oil and gas for example you have you know um you have specific applications and bespoke infrastructures for those applications um you know inside of media entertainment you know the same thing the the trend that we're seeing the commonality that we're seeing is the modernization of you know object as a starting point for all the all of the net new workloads within within those industry verticals right that's the most common request we see is what's your object roadmap what's your you know what's your what's your object strategy you know where do you think where do you think object is going so um there isn't any um you know sort of uh there's no there's no path uh it's really just kind of a wide open field in front of us with common requests across all industries so the amazing thing about pure just as a kind of a little you know quasi you know armchair historian the industry is pure was really the only company in many many years to be able to achieve escape velocity break through a billion dollars i mean three part couldn't do it isilon couldn't do it compellent couldn't do it i could go on but pure was able to achieve that as an independent company uh and so you become a leader you look at the gartner magic quadrant you're a leader in there i mean if you've made it this far you've got to have some chops and so of course it's very competitive there are a number of other storage suppliers that have announced products that unify object and file so i'm interested in how pure differentiates why pure um it's a great question um and it's one that uh you know having been a long time puritan uh you know i take pride in answering um and it's actually a really simple answer um it's it's business model innovation and technology right the the technology that goes behind how we do what we do right and i don't mean the product right innovation is product but having a better support model for example um or having on the business model side you know evergreen storage right where we sort of look at your relationship to us as a subscription right um you know we're gonna sort of take the thing that that you've had and we're gonna modernize that thing in place over time such that you're not rebuying that same you know terabyte or you know petabyte of storage that you've that you that you've paid for over time so um you know sort of three legs of the stool uh that that have made you know pure clearly differentiated i think the market has has recognized that um you're right it's it's hard to break through to a billion dollars um but i look forward to the day that you know we we have two billion dollar products and i think with uh you know that rise in in unstructured data growing to 80 by 2025 and you know the massive transition that you know you guys have noted in in in your hdd slide i think it's a huge opportunity for us on you know the other unstructured data side of the house you know the other thing i'd add matt and i've talked to cause about this is is it's simplicity first i've asked them why don't you do this why don't you do it and the answer is always the same is that adds complexity and we we put simplicity for the customer ahead of everything else and i think that served you very very well what about the economics of of unified file and object i mean if you bringing additional value presumably there's a there there's a cost to that but there's got to be also a business case behind it what kind of impact have you seen with customers yeah i mean look i'll i'll go back to something i mentioned earlier which is just the reclamation of floor space and power and cooling right um you know there's a you know there's people people people want to search for kind of the the sexier element if you will when it comes to looking at how we how you derive value from something but the reality is if you're reducing your power consumption by you know by by a material percentage um power bills matter in big in big data centers you know customers typically are are facing you know a paradigm of well i i want to go to the cloud but you know the clouds are not being more expensive than i thought it was going to be or you know i've figured out what i can use in the cloud i thought it was going to be everything but it's not going to be everything so hybrid's where we're landing but i want to be out of the data center business and i don't want to have a team of 20 storage people to match you know to administer my storage um you know so there's sort of this this very tangible value around you know hey if i could manage um you know multiple petabytes with one full-time engineer uh because the system uh to your and kaza's point was radically simpler to administer didn't require someone to be running around swapping drives all the time would that be a value the answer is yes 100 of the time right and then you start to look at okay all right well on the uffo side from a product perspective hey if i have to manage a you know bespoke environment for this application if i have to manage a bespoke environment for this application and a spoke environment for this application and this focus environment for this application i'm managing four different things and can i actually share data across those four different things there's ways to share data but most customers it just gets too complex how do you even know what your what your gold.master copy is of data if you have it in four different places or you try to have it in four different places and it's four different siloed infrastructures so when you get to the sort of the side of you know how do we how do you measure value in uffo it's actually being able to have all of that data concentrated in one place so that you can share it from application to application got it i'm interested we use a couple minutes left i'm interested in the the update on flashblade you know generally but also i have a specific question i mean look getting file right is hard enough uh you just announced smb support for flashblade i'm interested in you know how that fits in i think it's kind of obvious with file and object converging but give us the update on on flashblade and maybe you could address that specific question yeah so um look i mean we're we're um you know tremendously excited about the growth of flashblade uh you know we we we found workloads we never expected to find um you know the rapid restore workload was one that was actually brought to us from from a customer actually um and has become you know one of our one of our top two three four you know workloads so um you know we're really happy with the trend we've seen in it um and you know mapping back to you know thinking about hdds and ssds you know we're well on a path to building a billion dollar business here so you know we're very excited about that but to your point you know you don't just snap your fingers and get there right um you know we've learned that doing file and object uh is is harder than block um because there's more things that you have to go do for one you're basically focused on three protocols s b nfs and s3 not necessarily in that order um but to your point about s b uh you know we we are on the path through to releasing um you know smb full full native smb support in in the system that will allow us to uh service customers we have a limitation with some customers today where they'll have an smb portion of their nfs workflow um and we do great on the nfs side um but you know we didn't we didn't have the ability to plug into the s p component of their workflow so that's going to open up a lot of opportunity for us um on on that front um and you know we continue to you know invest significantly across the board in in areas like security which is you know become more than just a hot button you know today security's always been there but it feels like it's blazing hot today and so you know going through the next couple years we'll be looking at uh you know developing some some uh you know pretty material security elements of the product as well so uh well on a path to a billion dollars is the net on that and uh you know we're we're fortunate to have have smb here and we're looking forward to introducing that to to those customers that have you know nfs workloads today with an s b component yeah nice tailwind good tam expansion strategy matt thanks so much we're out of time but really appreciate you coming on the program we appreciate you having us and uh thanks much dave good to see you all right good to see you and you're watching the convergence of file and object keep it right there we'll be back with more right after this short break [Music]
SUMMARY :
i need to have um you know fast daz
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
2010 | DATE | 0.99+ |
Matt Burr | PERSON | 0.99+ |
250 terabytes | QUANTITY | 0.99+ |
270 terabytes | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
matt burr | PERSON | 0.99+ |
today | DATE | 0.99+ |
billion dollars | QUANTITY | 0.98+ |
two levels | QUANTITY | 0.98+ |
billions of objects | QUANTITY | 0.98+ |
two weeks later | DATE | 0.98+ |
80 | QUANTITY | 0.98+ |
two weeks ago | DATE | 0.98+ |
one system | QUANTITY | 0.98+ |
an hour | QUANTITY | 0.97+ |
cassandra | PERSON | 0.97+ |
matt | PERSON | 0.97+ |
next year | DATE | 0.96+ |
billions of objects | QUANTITY | 0.96+ |
dave | PERSON | 0.96+ |
one device | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
first principles | QUANTITY | 0.93+ |
second half | QUANTITY | 0.93+ |
billion dollar | QUANTITY | 0.91+ |
petabyte | QUANTITY | 0.9+ |
four different siloed infrastructures | QUANTITY | 0.89+ |
two billion dollar | QUANTITY | 0.89+ |
one place | QUANTITY | 0.89+ |
next couple years | DATE | 0.88+ |
80 of data | QUANTITY | 0.88+ |
early second half of this decade | DATE | 0.87+ |
20 storage people | QUANTITY | 0.86+ |
four different things | QUANTITY | 0.86+ |
five refrigerators | QUANTITY | 0.86+ |
one | QUANTITY | 0.84+ |
oracle sql | TITLE | 0.81+ |
one full-time | QUANTITY | 0.8+ |
wikibon | ORGANIZATION | 0.79+ |
four different places | QUANTITY | 0.79+ |
first | QUANTITY | 0.79+ |
3x | QUANTITY | 0.78+ |
a lot of people | QUANTITY | 0.78+ |
FlashBlade | ORGANIZATION | 0.78+ |
end of quarter | DATE | 0.77+ |
a couple minutes | QUANTITY | 0.77+ |
two sort | QUANTITY | 0.75+ |
isilon | ORGANIZATION | 0.74+ |
s3 | TITLE | 0.74+ |
three part | QUANTITY | 0.72+ |
100 of | QUANTITY | 0.7+ |
terabyte | QUANTITY | 0.7+ |
three legs | QUANTITY | 0.68+ |
two | QUANTITY | 0.68+ |
multiple petabytes | QUANTITY | 0.68+ |
vice president | PERSON | 0.65+ |
many years | QUANTITY | 0.61+ |
flashblade | ORGANIZATION | 0.57+ |
many companies | QUANTITY | 0.56+ |
tons | QUANTITY | 0.55+ |
gartner | ORGANIZATION | 0.53+ |
General Manager | PERSON | 0.53+ |
multi | QUANTITY | 0.51+ |
general manager | PERSON | 0.45+ |
Pure | ORGANIZATION | 0.34+ |
Matt Kixmoeller, Pure Storage & Michael Ferranti, Portworx | Kubecon + CloudNativeCon NA 2020
>> Narrator: From around the globe, it's theCUBE. With coverage of KubeCon and CloudNativeCon North America 2020, virtual. Brought to you by Red Hat, the Cloud Native Computing Foundation and ecosystem partners. >> Hi, I'm Joep Piscaer. Welcome to theCUBEs coverage of KubeCon, CloudNativeCon 2020. So I'm joined today by Matt Kixmoeller, he's VP of strategy at Pure Storage, as well as Michael Ferranti, he's the senior director of product marketing at Portworx now acquired by Pure Storage. Fellows, welcome to the show. >> Thanks here. >> I want to start out with you know , how about the lay of the land of storage in the Cloud Native space in the Kubernetes space. You know, what's hard? what's happening? What are the trends that you see going on? Matt, if you could shed some light on that for me? >> Yeah, I think you know, from a Pure point of view obviously we just told customers will they maturing their comprehensive deployments and particularly leaning towards persistant, you know applications and so you know we noticed within our customer base that there was quite a lot of deployments of a Portworx on Pure Storage. And that inspired us to start talking to one another you know, almost six plus months ago that eventually ended in us bringing the two companies together. So it's been a great journey from the Pure point of view, bringing Portworx into the Pure family. And, you know, we're working through now with our joint customers, integration strategies and how to really broaden the use of the technology. So that's quite exciting times for us. >> And of course, it's good to hear that the match goes beyond just the marketing color, like the brand color. >> Absolutely. Yeah. I mean, the fact that both companies were orange and you know, their logo looked like kind of a folded up version of ours, just started things off on the right foot >> A match made in heaven, right? So I want to talk a little bit about you know, the acquisition, what's happened there and especially, you know looking at Portworx as a company, and as a product set, it's fairly popular in the cloud community. A lot of traction with customers. So I want to zoom in on the acquisition itself and kind of the roadmap going forward merging the two companies and adding Portworx to that Pure portfolio. Matt, if you could shed some light on that as well. >> Yeah. Why don't I start and then Michael can jump in as well? So, you know, we at Pure had been really working for years now to outfit our all flash storage arrays for the container use case and shipped a piece of software that we call PSO. That was really a super CSI driver that allowed us to do intelligent placement of you know, persistent volumes on Pure arrays. But the more time we spent in the market, the more we just started to engage with customers and realized that there were a whole number of use cases that didn't really want a hardware based solution, you know. They either wanted to run completely in the cloud, hybrid between on-prem and cloud and leverage bare metal hardware. And so you know, we came to the conclusion that you know, first off, although positioning arrays for the market was the right thing to do, we wouldn't really be able to serve the broader needs restoratively for containers, if you did that. And then, you know, the second thing I think was that we heard from customers that they wanted a much richer data management stack. You know, it's not just about providing the business versus the volume for the container, but you know, all the capabilities around snapshoting and replication and mobilization and mobility between on-prem and cloud were necessary. And so, you know, Portworx we bought to bear not only a software based solution into our portfolio, but really that full data management stack platform in addition to just storage. And so as we look to integrate our product lines you know, we're looking to deliver a consistent experience for data management, for Kubernetes whatever infrastructure customer would like to, whether they want to run on all flash arrays, white box servers, bare metal, VMs or on cloud storage as well. You know, all of that can have a consistent experience with the Portworx platform. >> Yeah, and because you know, data management especially in this world of containers is you know, it's a little more difficult it's definitely more fragmented across you know, multiple clouds, multiple cloud vendors, multiple cloud services, multiple instances of a service. So the fragmentation has you know, given IT departments quite the headache in operationally managing all that. So Michael you know, what's kind of the use case for Portworx in this fragmented cloud storage space. >> Yeah. It's a great question. You know, the used cases are many and varied, you know to put it in a little bit of historical perspective you know, I've been attending coupons either (indistinct) for about five or six years now, kind of losing count. And we really started seeing Kubernetes as kind of an agile way to run CI/CD environments and other test dev environments. And there were just a handful of customers that were really running production workloads at the very, very beginning. If you fast forward to today, Kubernetes is being used to tackle some of the biggest central board level problems that enterprises face, because they need that scale and they need that agility. So you know, COVID's accelerated that. So we see customers say in the retail space, who are having to cope with a massive increase in traffic on their website. People searching for kind of you know, the products that they can't find anywhere else. Are they available? Can I buy them online? And so they're re-architecting those web services to use often open source databases in this case Elasticsearch, in order to create a great user experiences. And they're managing that across clouds and across environments using Kubernetes. Another customer that I would say kind of a very different use case but also one that matches that scale would be Esri which unfortunately the circumstances of becoming a household name are a lot of the covert tracking ArcGIS system to keep track of, tracing and outbreaks. They're running that service in the cloud using Portworx. And again, it's all about how do we reliably and agilely deploy applications that are always available and create that experience that our customers need. And so we see kind of you know, financial services doing similar things healthcare, pharmaceutical, doing similar things. Again, the theme is it's the biggest business problems that we're using now, not just the kind of the low hanging fruit as we used to talk about. >> Yeah exactly. Because you know storage, is it a lot of the times it's kind of a boiler plate functionality you know, it's there it works. And if it doesn't, you know, the problem with storage in a cloud data space is that fragmentation right? Is that enormous you know, on the one hand that you don't have a scale on the other hand, the tons of different services that can hold data that need protecting as well as data management. So I want to zoom in on a recent development in the Portworx portfolio where the PX backup product has spun out its own little product. You know, what's the strategy there, Michael? >> Yeah, so I think, you know fundamentally data protection needs to change in a Kubernetes context. The way in which we protected applications in the past was very closely related to the way in which we protected servers. Because we would run one app per server. So if we protected the server our application was protected. Kubernetes breaks that model now an individual application is made up of dozens or hundreds of components that are spread across multiple servers. And you have container images, you have configuration I mean you have data, and it's very difficult for any one person to understand where any of that is in the cluster at any given moment. And so you need to leverage automation and the ability for Kubernetes to understand where a particular set of components is deployed and use that Kubernetes native functionality to take what we call application aware backups. So what PX backup provides is data protection engineered from the ground up for this new application delivery model that we see within Kubernetes. So unlike traditional backup and recovery solutions that were very machine focused, we can allow a team to back up a single application within their Kubernetes cluster, all of the applications in a namespace or the entire cluster all at once, and do so in a self-service manner where integrated with your corporate identity systems individuals can be responsible for protecting their own applications. So we marry kind of a couple of really important concepts. One is kind of the application specific nature of Kubernetes the self service desire of DevOps teams, as well as with the page you go model, where you can have this flexible consumption model, where as you grow, you can pay more. You don't have to do an upfront payment in order to protect your Kubernetes applications. >> Yeah. I think one key thing that Michael hit on was just how this obligation is designed to fit like a glove with the Kubernetes admin. I see a lot of parallels to what happened over a decade ago in the VMware space when you know, VMware came about they needed to be backed up differently. And a little company called Veem built a tool that was purpose-built for it. And it just had a really warm embrace by the VMware community because it really felt like it was built for them, not some legacy enterprise backup application that was forced to fit into this new use case. And you know, we think that the opportunity is very similar on Kubernetes backup and perhaps the difference of the environment is even more profound than on the VMware side where you know, the Kubernetes admin really wants something that fits in their operational model, deploys within the cluster itself, backs up to object storage. Is just perfect purpose-built for this use case. And so we see a huge opportunity for that, and we believe that for a lot of customers, this might be the easiest place for them to start trying to Portworx portfolio. You know, you've got an existing competitors cluster download this, give it a shot, it'll work on any instructions you've got going with Kubernetes today. >> And especially because, you know, looking at the kind of breakdown of Kubernetes in a way data is, you know, infrastructure is provisioned. Data is placing in cloud services. It's no longer the cluster admin necessarily, that gets to decide where data goes, what application has access to it, you know, that's in the hands of the developers. And that's a pretty big shift you know, it used to be the VI admin the virtualization admin that did that, had control over where data was living, where data was accessed out, how it was accessed. But now we see developers kind of taking control over their infrastructure resources. They get to decide where it runs, how it runs what services to use, what applications to tie it into. So I'm curious, you know, how our Portworx and PX backup kind of help the developer stay in control and still have that freedom of choice. >> Yeah, we think of it in terms of data services. So I have a database and I needed to be highly available. I needed to be encrypted, backed up. I might need a DR. An off site DR schedule. And with Portworx, you can think about adding these services HA, security, backup, capacity management as really just I want to check a box and now I have this service available. My database is now highly available. It's backed up, it's encrypted. I can migrate it. I can attach a backup schedule to it. So 'cause within a Kubernetes cluster some apps are going to need that entire menu of services. And some apps might not need any of those services because we're only in Testa phage, everything is multiplexed into a single cluster. And so being able to turn off and turn on these various data services is how we empower a developer, a DevOps team to take an application all the way from test dev, into production, without having to really change anything about their Kubernetes deployments besides, you know, a flag within their YAML file. It makes it really, really easy to get the performance and the security and the availability that we were used to with VM based applications via that admin now within Kubernetes. >> So Matt, I want to spend the last couple of minutes talking about the bigger picture, right? We've talked about Portworx, PX backup. I want to take a look at the broader storage picture of cloud native and kind of look at the Pure angle on the trends on what you see happening in this space. >> Yeah absolutely. You know, a couple of high-level things I would, you know, kind of talk about, you know, the first buzz that I think, you know hybrid cloud deployments are the de facto now. And so when people are picking storage, whether they be you know, a storage for a traditional database application or next gen application, cloud native application, the thought from the beginning is how do I architect for hybrid? And so you know, within the Pure portfolio, we've really thought about how we build solutions that work with cloud native apps like Portworx, but also traditional applications. And our cloud block store allows you know, those to be mobilized to the cloud without, with minimal re-architecture. Another big trend that we see is the growth of object storage. And, you know if you look at the first generation of object storage, object storage is what? 15 plus years old and many of the first deployments were characterized by really low costs low performance, kind of the last retention layer if you will, for unimportant content. But then this web application thing happens and people started to build web apps that used object storage as their primary storage. And so now, as people try to bring those cloud native applications on-prem and build them in a multicloud way there's a real growth in the need for you know, high-performance kind of applications object storage. And so we see this real change to the needs and requirements on the object storage landscape. And it's one that in particular, we're trying to serve with our FlashBlade product that provides a unified file and object access, because many of those applications are kind of graduating from file or moving towards object, but they can't do that overnight. And so being able to provide a high-performance way to deliver unstructured data (indistinct) object files solve is very strategic right now. >> Well, that's insightful. Thanks. So I want to thank you both for being here. And, you know, I look forward to hearing about Portworx and Pure in the future as is acquisition. You know, it integrates and new products and new developments come out from the Pure side. So thanks both for being here and thank you at home for watching. I'm Joep Piscaer, thanks for watching the theCUBE's coverage of KubeCon CloudNativeCon 2020. Thanks. >> Yeah. Thanks too. >> Yeah. Thank you. (gentle music)
SUMMARY :
Brought to you by Red Hat, he's the senior director What are the trends that you see going on? Yeah, I think you know, beyond just the marketing and you know, their logo looked like and kind of the roadmap going forward And so you know, we came So the fragmentation has you know, And so we see kind of you know, And if it doesn't, you know, One is kind of the application And you know, we think and PX backup kind of help the developer and the availability that we were used to and kind of look at the the need for you know, And, you know, I look forward to hearing
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Michael Ferranti | PERSON | 0.99+ |
Joep Piscaer | PERSON | 0.99+ |
Matt Kixmoeller | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
Cloud Native Computing Foundation | ORGANIZATION | 0.99+ |
Portworx | ORGANIZATION | 0.99+ |
two companies | QUANTITY | 0.99+ |
Matt | PERSON | 0.99+ |
both companies | QUANTITY | 0.99+ |
KubeCon | EVENT | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
dozens | QUANTITY | 0.99+ |
both | QUANTITY | 0.98+ |
Veem | ORGANIZATION | 0.98+ |
Kubernetes | TITLE | 0.98+ |
VMware | ORGANIZATION | 0.98+ |
second thing | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
one app | QUANTITY | 0.98+ |
ArcGIS | TITLE | 0.97+ |
first deployments | QUANTITY | 0.97+ |
One | QUANTITY | 0.97+ |
Portworx | TITLE | 0.97+ |
first generation | QUANTITY | 0.97+ |
Elasticsearch | TITLE | 0.97+ |
CloudNativeCon | EVENT | 0.97+ |
today | DATE | 0.96+ |
Pure | ORGANIZATION | 0.96+ |
CloudNativeCon North America 2020 | EVENT | 0.96+ |
six years | QUANTITY | 0.95+ |
about five | QUANTITY | 0.94+ |
one person | QUANTITY | 0.94+ |
PX | ORGANIZATION | 0.94+ |
one key thing | QUANTITY | 0.94+ |
single application | QUANTITY | 0.93+ |
15 plus years old | QUANTITY | 0.93+ |
CloudNativeCon 2020 | EVENT | 0.93+ |
six plus months ago | DATE | 0.89+ |
single cluster | QUANTITY | 0.87+ |
theCUBE | ORGANIZATION | 0.85+ |
Kubecon | ORGANIZATION | 0.8+ |
Esri | TITLE | 0.8+ |
COVID | TITLE | 0.79+ |
a decade ago | DATE | 0.77+ |
Charlie Giancarlo, Pure Storage | CUBE Conversation, June 2020
>> From theCUBE Studios in Palo Alto and Boston, connecting with thought leaders all around the world, this is a CUBE Conversation. (intense music) >> Hi, everybody, this is Dave Vellante in theCUBE, and as you know, I've been doing a CEO series, and welcome to the isolation economy. We're here at theCUBE's remote studio, and really pleased to have Charlie Giancarlo, who is the CEO of PureStorage. Charlie, I wish we were face-to-face at Pure Accelerate, but this'll have to do. Thanks for coming on. >> You know, Dave, it's always fun to be face-to-face with you. At Pure Accelerate when we do it in person is great fun, but we do what we have to do, and actually, this has been a great event for us, so appreciate you coming on air with me. >> Yeah, and we're going to chat about that, but I want to start off with this meme that's been going around the internet. I was going to use the wrecking ball. I don't know if you've seen that. It's got the people, the executives in the office building saying, "Eh, digital transformation; "not in my lifetime," complacency, and then this big wrecking ball, the COVID-19. You've probably seen it, but as you can see here, somebody created a survey, Who's leading the digital transformation at your company? The CEO, the CTO, or of course circled is COVID-19, and so we've seen that, right? You had no choice but to be a digital company. >> Well, there's that, and there's also the fact that the CEOs who've been wanting to push a digital transformation against a team that wants to stick with the status quo, it gives the CEO now, and even within our own company in Pure, to drive towards that digital transformation when people didn't really take up the mantle. So no, it's a great opportunity for digital transformation, and of course, the companies that have been doing it all along have been getting ahead during this crisis, and the ones that haven't are having some real trouble. And you and I have had some really interesting conversations. Again, that's, I think, the thing I miss most, not only having you in theCUBE, but the side conversations at the cocktail parties, et cetera. And we've talked about IP, and China, and the history of the US, and all kinds of interesting things there, but one of the things I want to put forth, and I know you guys, Kix especially, has done a lot of work on Tech For Good, but the narrative pre-COVID, PC I guess we'd call it, was really a lot of vitriol toward big tech especially, but you know what? That tech lash... Without tech, where would we be right now? >> Well, just think about it, right? Where would we be without videoconferencing, without the internet, right? We'd be sheltered in place with literally nothing to do, and all business would stop, and of course many businesses that require in-person have, but thank God you can still get goods at your home. You can still get food, you can still get all these things that today is enabled by technology. We've seen this ourselves, in terms of having to make emergency shipments during our first quarter to critical infrastructure to keep things going. It's been quite a quarter. I was saying to my team recently that we had just gotten everyone together in February for our sales kickoff for the year, and it felt like a full year since I had seen them all. >> Well, I had interviewed, I think, is it Mike Fitzgerald, your head of supply chain. >> Yes. >> In March, and he was saying, "No. "We have no disruptions. "We're delivering for clients," and we certainly saw that in your results in the quarter. >> Yeah, no, we're very fortunate, but we had been planning for doing our normal business continuity disaster planning, and actually, once we saw COVID in Asia in January we started exercising all those muscles, including pre-shipping product around to depos around the world in case transportation got clogged, which it in fact did. So we were well-prepared, but we're also, I think, very fortunate in terms of the fact that we had a very distributed supply chain. >> Yeah, I mean you guys obviously did a good job. You saw in Dell's earnings they held pretty firm. HPE, on the other hand, really saw some disruption, so congratulations to you and the team on that. So as we think about exiting this isolation economy, we've done work that shows about 44% of CIOs see a U-shaped recovery, but it's very fragmented. It varies by industry. It varies by how digital the organizations are. Are they able to provide physical distancing? How essential are these organizations? And so I'm sure you're seeing that in your customer base as well. How are you thinking about exiting this isolation economy? >> Well, I've certainly resisted trying to predict a U- or a V-shape, because I think there are many more unknowns than there are knowns, and in particular, we don't know if there's a second wave. If there is a second wave, is it going to be more or less lethal than the first wave? And as you know, maybe some of your audience knows, I contracted COVID in March. So I've done a lot of reading on not just COVID, but also on the Spanish flu of 1918-1919. It's going to take a while before this settles down, and we don't know what it's going to look like the rest of the year or next year. So a lot of the recovery is going to depend on that. What we can do, however, is make sure that we're prepared to work from home, work in the office, that we make sure that our team out in the field is well-placed to be able to support our customers in the environment, and the way that we're incenting our overall team now has less to do with the macro than it does with our specific segment, and what I mean by that is we're incenting our team to continue to build market share, and to continue to outperform our competition as we go forward, and also on our customer satisfaction figure, which you know is our Net Promoter Score, which is the highest in the industry. So that's how we're incenting our team. >> Yeah, and we're going to talk about that, and by the way, yes, I did know, and it's great to see you healthy, and I'd be remiss if I didn't also express my condolences, Matt, the loss of Matt Danziger, your head of IR, terrible tragedy. Of course Matt had some roots in Boston, went to school in Maine. >> Yeah. >> Loved Cape Cod, and so really sad loss, I'm sure, for all of the Puritans. >> It's affected us all very personally, because Matt was just an incredible team member, a great friend, and so young and vital. When someone that young dies for almost unexplainable reasons. It turned out to be a congenital heart condition that nobody knew about, but it just breaks... It just breaks everyone's heart, so thank you for your condolences. I appreciate it. >> You're welcome. Okay, so let's get into the earnings a little bit. I want to just pull up one of the charts that shows roughly, I have approximately Q1 because some companies like NetApp, Dell, HPE, are sort of staggered, but the latest results you saw IBM growing at 19%. Now we know that was mainframe-driven in a very easy compare. Pure plus 12, and then everybody else in the negative. Dell, minus five, so actually doing pretty well relative to NetApp and HPE, who, as I said, had some challenges with deliveries. But let's talk about your quarter. You continue to be the one sort of shining star in the storage business. Let's get into it. What are your big takeaways that you want us to know about? >> Well, of course I'd rather see everybody in the black, right, everybody in the positive, but we continue to take market share and continue to grow 20 to 30% faster than the rest of the industry combined, and it's quarter after quarter. It's not just a peak in one quarter and then behind in another quarter. Every quarter we're ahead of the rest of the industry, and I think the reasoning is really quite straightforward. We're the one company that invests in storage as if it's high technology. You do hear quite often, and even among some customers, that storage is commoditized, and all of our competitors invest in it, or don't invest in it, as if it's a commoditized market. Our view is quite straightforward. The science and the engineering of computing and data centers continues to evolve, continues to advance, has to advance if we continue down this path of becoming more of a digital economy. As we all know, processors advance in speed and capability. Networking advances in terms of speed and capability. Well, data storage is a third of data center spend, and if it doesn't continue to advance at the same pace or faster than everything else, it becomes a major bottleneck. We've been the innovator. If you look at a number of different studies, year after year, now over six or seven years, we are the leader in innovation in the data storage market, and we're being rewarded for that by penetrating more and more of the customer base. >> All right, let's talk about that. And you mentioned in your keynote at Accelerate that you guys spend more on R&D as a percentage of revenue than anybody, and so I want to throw out some stats. I'm sorry, folks, I don't have a slide on this. HPE spends about 1.8 billion a year on R&D, about 6% of revenues. IBM, I've reported on IBM and how it's spending the last 10 years, spent a huge amount on dividends and stock buybacks, and they spent six billion perpetually on R&D, which is now 8% of revenue. Dell at five billion. Of course Dell used to spend well under a billion before the EMC acquisition. That's about 6% of revenue. And NetApp, 800 million, much higher. They're a pure play, about 13%. Pure spends 430 million last year on R&D, which is over 30% of revenue on R&D, to your point. >> Yeah, yeah, well, as I said, we treat it like it's high technology, which it is, right? If you're not spending at an appropriate level you're going to fall behind, and so we continue to advance. I will say that you mentioned big numbers by the other players, but I was part of a big organization as well with a huge R&D budget, but what matters is what percent of the revenue of a specific area are you spending, right? You mentioned Dell and VMware. A very large fraction of their spend is on VMware. Great product and great company, but very little is being spent in the area of storage. >> Well, and the same thing's true for IBM, and I've made this point. In fact, I made this point about Snowflake last week in my breaking analysis. How is Snowflake able to compete with all these big whales? And the same thing for you guys. Every dime you spend on R&D goes to making your storage products better for your customers. Your go-to-market, same thing. Your partner ecosystem, same thing, and so you're the much more focused play. >> Right, well I think it boils down to one very simple thing, right? Most of our competitors are, you might call them one-stop shops, so the shopping mall of IT gear, right? The Best Buy, if you will, of information technology. We're really the sole best of breed player in data storage, right, and if you're a company that wants two vendors, you might choose one that's a one-stop shop. If you have the one-stop shop, the next one you want is a best of breed player, right? And we fill that role for our customers. >> Look it, this business is a technology business, and technology and innovation is driven by research and development, period, the end. But I want to ask you, so the storage business generally, look, you're kind of the one-eyed man in the land of the blind here. I mean the storage business has been somewhat on the back burner. In part it's your fault because you put so much flash into the data center, gave so much headroom that organizations didn't have to buy spindles anymore to get to performance, the cloud has also been a factor. But look, last decade was a better decade for storage than the previous decade when you look at the exits that you guys had and escape velocity, Nutanix, if you can kind of put them in there, too. Much larger than say the Compellents or 3PARs. They didn't make it to a billion. So my question is storage businesses, is it going to come back as a growth business? Like you said, you wish everybody were in the black here. >> Right, well a lot of what's being measured, of course, is enterprise on-prem storage, right? If we add on-prem and cloud, it actually continues to be a big growth business, because data is not shrinking. In fact, data is still growing faster than the price reduction of the media underneath, right, so it's still growing. And as you know, more recently we've introduced what we call Pure as-a-Service and Cloud Block Store. So now we have our same software, which we call Purity, that runs on our on-prem arrays, also running on AWS, and currently in beta on Azure. So from our point of view this is a... First of all, it's a big market, about $30 to $40 billion total. If you add in cloud, it's another $10 to $15 billion, which is a new opportunity for us. Last year we were about 1.65 billion. We're still less than, as you know, less than 10% of the overall market. So the opportunity for us to grow is just tremendous out there, and whether or not total storage grows, for us it's less important right now than the market share that we pick up. >> Right, okay, so I want to stay on that for a minute and talk about... I love talking about the competition. So what I'm showing here with this kind of wheel slide is data from our data partner ETR, and they go out every quarter. They have a very simple methodology. It's like Net Promoter Score, and it's very consistent. They say relative to last year, are you adopting the platform, that's the lime green, and so this is Pure's data. Are you increasing spend by 6% or more? That's the 32%, the forest green. Is spending going to be flat? Is it going to decrease by more than 6%? That's the 9%. And then are you replacing the platform, 2%. Now this was taken at the height of the US lockdown. This last survey. >> Wow. >> So you can see the vast majority of customers are either keeping spending the same, or they're spending more. >> Yeah. >> So that's very, very strong. And I want to just bring up another data point, which is we like to plot that Net Score here on the vertical axis, and then what we call market share. It's not like IDC market share, but it's pervasiveness in the survey. And you can see here, to your point, Pure is really the only, and I've cited the other vendors on the right hand, that box there, you're the only company in the green with a 40% Net Score, and you can see everybody else is well below the line in the red, but to your point, you got a long way to go in terms of gaining market share. >> Exactly, right, and the reason... I think the reason why you're seeing that is really our fundamental and basic value is that our product and our company is easy to do business with and easy to operate, and it's such a pleasure to use versus the competition that customers really appreciate the product and the company. We do have a Net Promoter Score of over 80, which I think you'd be hard-pressed to find another company in any industry with Net Promoter Scores that high. >> Yeah, so I want to stay on the R&D thing for a minute, because you guys bet the company from day one on simplicity, and that's really where you put a lot of effort. So the cloud is vital here, and I want to get your perspective on it. You mentioned your Cloud Block Store, which I like that, it's native to AWS. I think you're adding other platforms. I think you're adding Azure as well, and I'm sure you'll do Google. >> Azure, Azure's in beta, yes. >> Yeah, Google's just a matter of time. Alibaba, you'll get them all, but the key here is that you're taking advantage of the native services, and let's take AWS as an example. You're using EC2, and high priority instances of EC2, as an example, to essentially improve block storage on Amazon. Amazon loves it because it sells Compute. Maybe the storage guys in Amazon don't love it so much, but it's all about the customer, and so the native cloud services are critical. I'm sure you're going to do the same thing for Azure and other clouds, and that takes a lot of investment, but I heard George Kurian today addressing some analysts, talking about they're the only company doing kind of that cloud native approach. Where are you placing your bets? How much of it is cloud versus kind of on-prem, if you will? >> Yeah, well... So first of all, an increasing fraction is cloud, as you might imagine, right? We started off with a few dozen developers, and now we're at many more than that. Of course the majority of our revenue still comes from on-prem, but the value is the following in our case, which is that we literally have the same software operating, from a customer and from a application standpoint. It is the same software operating on-prem as in the cloud, which means that the customer doesn't have to refactor their application to move it into the cloud, and we're the one vendor that's focused on block. What NetApp is doing is great, but it's a file-based system. It's really designed for smaller workloads and low performance workloads. Our system's designed for high performance enterprise workloads, Tier 1 workloads in the cloud. To say that they're both cloud sort of washes over the fact that they're almost going after two completely separate markets. >> Well, I think it's interesting that you're both really emphasizing cloud native, which I think is very important. I think that some of the others have some catching up to do in that regard, and again, that takes a big investment in not just wrapping your stack, and shoving it in the cloud, and hosting it in the cloud. You're actually taking advantage of the local services. >> Well, I mean one thing I'll mention was Amazon gave us an award, which they give to very few vendors. It's called the Well-Architected AWS Award, because we've designed it not to operate, let's say, in a virtualized environment on AWS. We really make use of the native AWS EC2 services. It is designed like a web service on EC2. >> And the reason why this is so important is just, again, to share with our audience is because when you start talking about multi-cloud and hybrid cloud, you want the same exact experience on-prem as you do in the cloud, whether it's hybrid or across clouds, and the key is if you're using cloud native services, you have the most efficient, the highest performance, lowest latency, and lowest cost solution. That is going to be... That's going to be a determinate of the winner. >> Yes, I believe so. Customers don't want to be doing... Be working with software that is going to change, fundamentally change and cause them to have to refactor their applications. If it's not designed natively to the cloud, then when Amazon upgrades it may cause a real problem with the software or with the environment, and so customers don't want that. They want to know they're cloud native. >> Well, your task over the next 10 years is something. Look it, it's very challenging to grow a company the size of Pure, period, but let's face it, you guys caught EMC off-guard. You were driving a truck through the Symmetrics base and the VNX base. Not that that was easy. (chuckling) And they certainly didn't make it easy for ya. But now we've got this sort of next chapter, and I want to talk a little bit about this. You guys call it the Modern Data Experience. You laid it out last Accelerate, kind of your vision. You talked about it more at this year's Accelerate. I wonder if you could tell us the key takeaways from your conference this year. >> Right, the key takeaway... So let me talk about both. I'll start with Modern Data Experience and then key takeaways from this Accelerate. So Modern Data Experience, for those that are not yet familiar with it, is the idea that an on-prem experience would look very similar, if not identical, to a cloud experience. That is to say that applications and orchestrators just use APIs to be able to call upon and have delivered the storage environment that they want to see instantaneously over a high speed network. The amazing thing about storage, even today, is that it's highly mechanical, it's highly hardware-oriented to where if you have a new application and you want storage, you actually have to buy an array and connect it. It's physical. Where we want to be is just like in the cloud. If you have a new application and you want storage or you want data services, you just write a few APIs in your application and it's delivered immediately and automatically, and that's what we're delivering on-prem with the Modern Data Experience. What we're also doing, though, is extending that to the cloud, and with Cloud Block Store as part of this, with that set of interfaces and management system exactly the same as on-prem, you now have that cloud experience across all the clouds without having to refactor applications in one or the other. So that's our Modern Data Experience. That's the vision that drives us. We've delivered more and more against it starting at the last Accelerate, but even more now. Part of this is being able to deliver storage that is flexible and able to be delivered by API. On this Accelerate we delivered our Purity 6.0 for Flash Array, which adds not only greater resiliency characteristics, but now file for the first time in a Flash Array environment, and so now the same Flash Array can deliver both file and block. Which is a unified experience, but all delivered by API and simple to operate. We've also delivered, more recently, Flash Array 3.0... I'm sorry, Purity 3.0 on FlashBlade that delivers the ability for FlashBlade now to have very high resiliency characteristics, and to be able to even better deliver the ability to restore applications when there's been a failure of their data systems very, very rapidly, something that we call Rapid Restore. So these are huge benefits. And the last one I'll mention, Pure as-a-Service allows a customer today to be able to contract for storage as a service on-prem and in the cloud with one unified subscription. So they only pay for what they use. They only pay for what they use when they use it, and they only pay for it, regardless of where it's used, on-prem or in the cloud, and it's a true subscription model. It's owned and operated by Pure, but the customer gets the benefit of only paying for what they use, regardless of where they use it. >> Awesome, thanks for that run through. And a couple other notes that I had, I mean you obviously talked about the support for the work from home and remote capabilities. Automation came up a lot. >> Yep. >> You and I, I said, we have these great conversations, and one of the ones I would have with you if we were having a drink somewhere would be if you look at productivity stats in US and Europe, they're declining-- >> Yes. >> Pretty dramatically. And if you think about the grand challenges we have, the global challenges, whether it's pandemics, or healthcare, or feeding people, et cetera, we're not going to be able to meet those challenges without automation. I mean people, for years, have been afraid of automation. "Oh, we're going to lose jobs." We don't have enough people to solve all these problems, and so I think that's behind us, right-- >> Yeah, I agree. >> The fear of automation. So that came up. Yeah, go ahead, please. >> I once met with Alan Greenspan. You may remember him. >> Of course. >> This is after he was the chairman, and he said, "Look, I've studied the economies now "for the last 100 years, "and the fact of the matter is "that wealth follows productivity." The more productive you are as a society, that means the greater the wealth that exists for every individual, right? The standard of living follows productivity, and without productivity there's no wealth creation for society. So to your point, yeah, if we don't become more productive, more efficient, people don't live better, right? >> Yeah, I knew you'd have some good thoughts on that, and of course, speaking of Greenspan, we're seeing a little bit of rational exuberance maybe in the market. (chuckling) Pretty amazing. But you also talked about containers, and persisting containers, and Kubernetes, the importance of Kubernetes. That seems to be a big trend that you guys are hopping on as well. >> You bet. It is the wave of the future. Now, like all waves of the future, it's going to take time. Containers work entirely differently from VMs and from machines in terms of how they utilize resources inside a data center environment, and they are extraordinarily dynamic. They require the ability to build up, tear down connections to storage, and create storage, and spin it down at very, very rapid rates, and again, it's all API-driven. It's all responsive, not to human operators, but it's got to be responsive to the application itself and to the orchestration environment. And again, I'll go back to what we talked about with our Modern Data Experience. It's exactly the kind of experience that our customers want to be able to be that responsive to this new environment. >> My last question is from John Furrier. He asked me, "Hey, Charlie knows a lot about networking." We were talking about multi-cloud. Obviously cross-cloud networks are going to become increasingly important. People are trying to get rid of their MPLS networks, really moving to an SD-WAN environment. Your thoughts on the evolution of networking over the next decade. >> Well, I'll tell you. I'm a big believer that even SD-WANs, over time, are going to become obsolete. Another way to phrase it is the new private network is the internet. I mean look at it now. What does SD-WAN mean when nobody's in the local office, right? No one's in the remote office; they're all at home. And so now we need to think about the fact... Sometimes it's called Zero Trust. I don't like that term. Nobody wants to talk about zero anything. What it really is about is that there is no internal network anymore. The fact of the matter is even for... Let's say I'm inside my own company's network. Well, do they trust my machine? Maybe not. They may trust me but not my machine, and so what we need to have is going to a cloud model where all communication to all servers goes through a giant, call it a firewall or a proxy service, where everything is cleaned before it's delivered. People, individuals only get, and applications, only get access to the applications that they're authorized to use, not to a network, because once they're in the network they can get anywhere. So they should only get access to the applications they're able to use. So my personal opinion is the internet is the future private network, and that requires a very different methodology for authentication for security and so forth, and if we think that we protect ourselves now by firewalls, we have to rethink that. >> Great perspectives. And by the way, you're seeing more than glimpses of that. You look at Zscaler's results recently, and that's kind of the security cloud, and I'm glad you mentioned that you don't like that sort of Zero Trust. You guys, even today, talked about near zero RPO. That's an honest statement-- >> Right. >> Because there's no such thing as zero RPO. (chuckling) >> Right, yeah. >> Charlie, great to have you on. Thanks so much for coming back in theCUBE. Great to see you again. >> Dave, always a pleasure. Thank you so much, and hopefully next time in person. >> I hope so. All right, and thank you for watching, everybody. This is Dave Vellante for theCUBE, and we'll see you next time. (smooth music)
SUMMARY :
leaders all around the world, and really pleased to it's always fun to be executives in the office building and of course, the companies for our sales kickoff for the year, your head of supply chain. and we certainly saw that in and actually, once we saw HPE, on the other hand, and the way that we're incenting our overall team and it's great to see you healthy, I'm sure, for all of the Puritans. so thank you for your condolences. but the latest results you and continue to grow 20 to 30% faster and how it's spending the last 10 years, and so we continue to advance. Well, and the same the next one you want is a and development, period, the end. than the market share that we pick up. height of the US lockdown. are either keeping spending the same, the red, but to your point, and it's such a pleasure to So the cloud is vital here, and so the native cloud It is the same software operating and hosting it in the cloud. It's called the and the key is if you're and cause them to have to You guys call it the and in the cloud with for the work from home and so I think that's behind us, right-- So that came up. I once met with Alan Greenspan. that means the greater the wealth That seems to be a big trend that you guys They require the ability to build up, over the next decade. The fact of the matter is even for... and that's kind of the security cloud, such thing as zero RPO. Charlie, great to have you on. Thank you so much, and and we'll see you next time.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Mike Fitzgerald | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
20 | QUANTITY | 0.99+ |
June 2020 | DATE | 0.99+ |
six billion | QUANTITY | 0.99+ |
George Kurian | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Matt | PERSON | 0.99+ |
February | DATE | 0.99+ |
Pure Accelerate | ORGANIZATION | 0.99+ |
Maine | LOCATION | 0.99+ |
March | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Matt Danziger | PERSON | 0.99+ |
Charlie Giancarlo | PERSON | 0.99+ |
$10 | QUANTITY | 0.99+ |
Europe | LOCATION | 0.99+ |
Asia | LOCATION | 0.99+ |
Charlie | PERSON | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
8% | QUANTITY | 0.99+ |
US | LOCATION | 0.99+ |
Last year | DATE | 0.99+ |
January | DATE | 0.99+ |
five billion | QUANTITY | 0.99+ |
Alan Greenspan | PERSON | 0.99+ |
last week | DATE | 0.99+ |
19% | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
6% | QUANTITY | 0.99+ |
NetApp | ORGANIZATION | 0.99+ |
9% | QUANTITY | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
$15 billion | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
theCUBE | ORGANIZATION | 0.99+ |
less than 10% | QUANTITY | 0.99+ |
PureStorage | ORGANIZATION | 0.99+ |
800 million | QUANTITY | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
430 million | QUANTITY | 0.99+ |
32% | QUANTITY | 0.99+ |
zero RPO | QUANTITY | 0.99+ |
$40 billion | QUANTITY | 0.99+ |
2% | QUANTITY | 0.99+ |
EC2 | TITLE | 0.99+ |
EMC | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Cloud Block Store | TITLE | 0.99+ |
last year | DATE | 0.99+ |
John Furrier | PERSON | 0.99+ |
first time | QUANTITY | 0.99+ |
COVID-19 | OTHER | 0.99+ |
both | QUANTITY | 0.99+ |
one quarter | QUANTITY | 0.99+ |
Accelerate | ORGANIZATION | 0.99+ |
next year | DATE | 0.98+ |
about $30 | QUANTITY | 0.98+ |
Flash Array 3.0 | TITLE | 0.98+ |
ETR | ORGANIZATION | 0.98+ |
Cape Cod | LOCATION | 0.98+ |
Ben White, Domo | Virtual Vertica BDC 2020
>> Announcer: It's theCUBE covering the Virtual Vertica Big Data Conference 2020, brought to you by Vertica. >> Hi, everybody. Welcome to this digital coverage of the Vertica Big Data Conference. You're watching theCUBE and my name is Dave Volante. It's my pleasure to invite in Ben White, who's the Senior Database Engineer at Domo. Ben, great to see you, man. Thanks for coming on. >> Great to be here and here. >> You know, as I said, you know, earlier when we were off-camera, I really was hoping I could meet you face-to-face in Boston this year, but hey, I'll take it, and, you know, our community really wants to hear from experts like yourself. But let's start with Domo as the company. Share with us what Domo does and what your role is there. >> Well, if I can go straight to the official what Domo does is we provide, we process data at BI scale, we-we-we provide BI leverage at cloud scale in record time. And so what that means is, you know, we are a business-operating system where we provide a number of analytical abilities to companies of all sizes. But we do that at cloud scale and so I think that differentiates us quite a bit. >> So a lot of your work, if I understand it, and just in terms of understanding what Domo does, there's a lot of pressure in terms of being real-time. It's not, like, you sometimes don't know what's coming at you, so it's ad-hoc. I wonder if you could sort of talk about that, confirm that, maybe add a little color to it. >> Yeah, absolutely, absolutely. That's probably the biggest challenge it is to being, to operating Domo is that it is an ad hoc environment. And certainly what that means, is that you've got analysts and executives that are able to submit their own queries with out very... With very few limitations. So from an engineering standpoint, that challenge in that of course is that you don't have this predictable dashboard to plan for, when it comes to performance planning. So it definitely presents some challenges for us that we've done some pretty unique things, I think, to address those. >> So it sounds like your background fits well with that. I understand your people have called you a database whisperer and an envelope pusher. What does that mean to a DBA in this day and age? >> The whisperer part is probably a lost art, in the sense that it's not really sustainable, right? The idea that, you know, whatever it is I'm able to do with the database, it has to be repeatable. And so that's really where analytics comes in, right? That's where pushing the envelope comes in. And in a lot of ways that's where Vertica comes in with this open architecture. And so as a person who has a reputation for saying, "I understand this is what our limitations should be, but I think we can do more." Having a platform like Vertica, with such an open architecture, kind of lets you push those limits quite a bit. >> I mean I've always felt like, you know, Vertica, when I first saw the stone breaker architecture and talked to some of the early founders, I always felt like it was the Ferrari of databases, certainly at the time. And it sounds like you guys use it in that regard. But talk a little bit more about how you use Vertica, why, you know, why MPP, why Vertica? You know, why-why can't you do this with RDBMS? Educate us, a little bit, on, sort of, the basics. >> For us it was, part of what I mentioned when we started, when we talked about the very nature of the Domo platform, where there's an incredible amount of resiliency required. And so Vertica, the MPP platform, of course, allows us to build individual database clusters that can perform best for the workload that might be assigned to them. So the open, the expandable, the... The-the ability to grow Vertica, right, as your base grows, those are all important factors, when you're choosing early on, right? Without a real idea of how growth would be or what it will look like. If you were kind of, throwing up something to the dark, you look at the Vertica platform and you can see, well, as I grow, I can, kind of, build with this, right? I can do some unique things with the platform in terms of this open architecture that will allow me to not have to make all my decisions today, right? (mutters) >> So, you're using Vertica, I know, at least in part, you're working with AWS as well, can you describe sort of your environment? Do you give anything on-prem, is everything in cloud? What's your set up look like? >> Sure, we have a hybrid cloud environment where we have a significant presence in public files in our own private cloud. And so, yeah, having said that, we certainly have a really an extensive presence, I would say, in AWS. So, they're definitely the partner of our when it comes to providing the databases and the server power that we need to operate on. >> From a standpoint of engineering and architecting a database, what were some of the challenges that you faced when you had to create that hybrid architecture? What did you face and how did you overcome that? >> Well, you know, some of the... There were some things we faced in terms of, one, it made it easy that Vertica and AWS have their own... They play well together, we'll say that. And so, Vertica was designed to work on AWS. So that part of it took care of it's self. Now our own private cloud and being able to connect that to our public cloud has been a part of our own engineering abilities. And again, I don't want to make little, make light of it, it certainly not impossible. And so we... Some of the challenges that pertain to the database really were in the early days, that you mentioned, when we talked a little bit earlier about Vertica's most recent eon mode. And I'm sure you'll get to that. But when I think of early challenges, some of the early challenges were the architecture of enterprise mode. When I talk about all of these, this idea that we can have unique databases or database clusters of different sizes, or this elasticity, because really, if you know the enterprise architecture, that's not necessarily the enterprise architecture. So we had to do some unique things, I think, to overcome that, right, early. To get around the rigidness of enterprise. >> Yeah, I mean, I hear you. Right? Enterprise is complex and you like when things are hardened and fossilized but, in your ad hoc environment, that's not what you needed. So talk more about eon mode. What is eon mode for you and how do you apply it? What are some of the challenges and opportunities there, that you've found? >> So, the opportunities were certainly in this elastic architecture and the ability to separate in the storage, immediately meant that for some of the unique data paths that we wanted to take, right? We could do that fairly quickly. Certainly we could expand databases, right, quickly. More importantly, now you can reduce. Because previously, in the past, right, when I mentioned the enterprise architecture, the idea of growing a database in itself has it's pain. As far as the time it takes to (mumbles) the data, and that. Then think about taking that database back down and (telephone interference). All of a sudden, with eon, right, we had this elasticity, where you could, kind of, start to think about auto scaling, where you can go up and down and maybe you could save some money or maybe you could improve performance or maybe you could meet demand, At a time where customers need it most, in a real way, right? So it's definitely a game changer in that regard. >> I always love to talk to the customers because I get to, you know, I hear from the vendor, what they say, and then I like to, sort of, validate it. So, you know, Vertica talks a lot about separating compute and storage, and they're not the only one, from an architectural standpoint who do that. But Vertica stresses it. They're the only one that does that with a hybrid architecture. They can do it on-prem, they can do it in the cloud. From your experience, well first of all, is that true? You may or may not know, but is that advantageous to you, and if so, why? >> Well, first of all, it's certainly true. Earlier in some of the original beta testing for the on-prem eon modes that we... I was able to participate in it and be aware of it. So it certainly a realty, they, it's actually supported on Pure storage with FlashBlade and it's quite impressive. You know, for who, who will that be for, tough one. It's probably Vertica's question that they're probably still answering, but I think, obviously, some enterprise users that probably have some hybrid cloud, right? They have some architecture, they have some hardware, that they themselves, want to make use of. We certainly would probably fit into one of their, you know, their market segments. That they would say that we might be the ones to look at on-prem eon mode. Again, the beauty of it is, the elasticity, right? The idea that you could have this... So a lot of times... So I want to go back real quick to separating compute. >> Sure. Great. >> You know, we start by separating it. And I like to think of it, maybe more of, like, the up link. Because in a true way, it's not necessarily separated because ultimately, you're bringing the compute and the storage back together. But to be able to decouple it quickly, replace nodes, bring in nodes, that certainly fits, I think, what we were trying to do in building this kind of ecosystem that could respond to unknown of a customer query or of a customer demand. >> I see, thank you for that clarification because you're right, it's really not separating, it's decoupling. And that's important because you can scale them independently, but you still need compute and you still need storage to run your work load. But from a cost standpoint, you don't have to buy it in chunks. You can buy in granular segments for whatever your workload requires. Is that, is that the correct understanding? >> Yeah, and to, the ability to able to reuse compute. So in the scenario of AWS or even in the scenario of your on-prem solution, you've got this data that's safe and secure in (mumbles) computer storage, but the compute that you have, you can reuse that, right? You could have a scenario that you have some query that needs more analytic, more-more fire power, more memory, more what have you that you have. And so you can kind of move between, and that's important, right? That's maybe more important than can I grow them separately. Can I, can I borrow it. Can I borrow that compute you're using for my (cuts out) and give it back? And you can do that, when you're so easily able to decouple the compute and put it where you want, right? And likewise, if you have a down period where customers aren't using it, you'd like to be able to not use that, if you no longer require it, you're not going to get it back. 'Cause it-it opened the door to a lot of those things that allowed performance and process department to meet up. >> I wonder if I can ask you a question, you mentioned Pure a couple of times, are you using Pure FlashBlade on-prem, is that correct? >> That is the solution that is supported, that is supported by Vertica for the on-prem. (cuts out) So at this point, we have been discussing with them about some our own POCs for that. Before, again, we're back to the idea of how do we see ourselves using it? And so we certainly discuss the feasibility of bringing it in and giving it the (mumbles). But that's not something we're... Heavily on right now. >> And what is Domo for Domo? Tell us about that. >> Well it really started as this idea, even in the company, where we say, we should be using Domo in our everyday business. From the sales folk to the marketing folk, right. Everybody is going to use Domo, it's a business platform. For us in engineering team, it was kind of like, well if we use Domo, say for instance, to be better at the database engineers, now we've pointed Domo at itself, right? Vertica's running Domo in the background to some degree and then we turn around and say, "Hey Domo, how can we better at running you?" So it became this kind of cool thing we'd play with. We're now able to put some, some methods together where we can actually do that, right. Where we can monitor using our platform, that's really good at processing large amounts of data and spitting out useful analytics, right. We take those analytics down, make recommendation changes at the-- For now, you've got Domo for Domo happening and it allows us to sit at home and work. Now, even when we have to, even before we had to. >> Well, you know, look. Look at us here. Right? We couldn't meet in Boston physically, we're now meeting remote. You're on a hot spot because you've got some weather in your satellite internet in Atlanta and we're having a great conversation. So-so, we're here with Ben White, who's a senior database engineer at Domo. I want to ask you about some of the envelope pushing that you've done around autonomous. You hear that word thrown around a lot. Means a lot of things to a lot of different people. How do you look at autonomous? And how does it fit with eon and some of the other things you're doing? >> You know, I... Autonomous and the idea idea of autonomy is something that I don't even know if that I have already, ready to define. And so, even in my discussion, I often mention it as a road to it. Because exactly where it is, it's hard to pin down, because there's always this idea of how much trust do you give, right, to the system or how much, how much is truly autonomous? How much already is being intervened by us, the engineers. So I do hedge on using that. But on this road towards autonomy, when we look at, what we're, how we're using Domo. And even what that really means for Vertica, because in a lot of my examples and a lot of the things that we've engineered at Domo, were designed to maybe overcome something that I thought was a limitation thing. And so many times as we've done that, Vertica has kind of met us. Like right after we've kind of engineered our architecture stuff, that we thought that could help on our side, Vertica has a release that kind of addresses it. So, the autonomy idea and the idea that we could analyze metadata, make recommendations, and then execute those recommendations without innervation, is that road to autonomy. Once the database is properly able to do that, you could see in our ad hoc environment how that would be pretty useful, where with literally millions of queries every hour, trying to figure out what's the best, you know, profile. >> You know for- >> (overlapping) probably do a better job in that, than we could. >> For years I felt like IT folks sometimes were really, did not want that automation, they wanted the knobs to turn. But I wonder if you can comment. I feel as though the level of complexity now, with cloud, with on-prem, with, you know, hybrid, multicloud, the scale, the speed, the real time, it just gets, the pace is just too much for humans. And so, it's almost like the industry is going to have to capitulate to the machine. And then, really trust the machine. But I'm still sensing, from you, a little bit of hesitation there, but light at the end of the tunnel. I wonder if you can comment? >> Sure. I think the light at the end of the tunnel is even in the recent months and recent... We've really begin to incorporate more machine learning and artificial intelligence into the model, right. And back to what we're saying. So I do feel that we're getting closer to finding conditions that we don't know about. Because right now our system is kind of a rule, rules based system, where we've said, "Well these are the things we should be looking for, these are the things that we think are a problem." To mature to the point where the database is recognizing anomalies and taking on pattern (mutters). These are problems you didn't know happen. And that's kind of the next step, right. Identifying the things you didn't know. And that's the path we're on now. And it's probably more exciting even than, kind of, nailing down all the things you think you know. We figure out what we don't know yet. >> So I want to close with, I know you're a prominent member of the, a respected member of the Vertica Customer Advisory Board, and you know, without divulging anything confidential, what are the kinds of things that you want Vertica to do going forward? >> Oh, I think, some of the in dated base for autonomy. The ability to take some of the recommendations that we know can derive from the metadata that already exists in the platform and start to execute some of the recommendations. And another thing we've talked about, and I've been pretty open about talking to it, talking about it, is the, a new version of the database designer, I think, is something that I'm sure they're working on. Lightweight, something that can give us that database design without the overhead. Those are two things, I think, as they nail or basically the database designer, as they respect that, they'll really have all the components in play to do in based autonomy. And I think that's, to some degree, where they're heading. >> Nice. Well Ben, listen, I really appreciate you coming on. You're a thought leader, you're very open, open minded, Vertica is, you know, a really open community. I mean, they've always been quite transparent in terms of where they're going. It's just awesome to have guys like you on theCUBE to-to share with our community. So thank you so much and hopefully we can meet face-to-face shortly. >> Absolutely. Well you stay safe in Boston, one of my favorite towns and so no doubt, when the doors get back open, I'll be coming down. Or coming up as it were. >> Take care. All right, and thank you for watching everybody. Dave Volante with theCUBE, we're here covering the Virtual Vertica Big Data Conference. (electronic music)
SUMMARY :
brought to you by Vertica. of the Vertica Big Data Conference. I really was hoping I could meet you face-to-face And so what that means is, you know, I wonder if you could sort of talk about that, confirm that, is that you don't have this predictable dashboard What does that mean to a DBA in this day and age? The idea that, you know, And it sounds like you guys use it in that regard. that can perform best for the workload that we need to operate on. Some of the challenges that pertain to the database and you like when things are hardened and fossilized and the ability to separate in the storage, but is that advantageous to you, and if so, why? The idea that you could have this... And I like to think of it, maybe more of, like, the up link. And that's important because you can scale them the compute and put it where you want, right? that is supported by Vertica for the on-prem. And what is Domo for Domo? From the sales folk to the marketing folk, right. I want to ask you about some of the envelope pushing and a lot of the things that we've engineered at Domo, than we could. But I wonder if you can comment. nailing down all the things you think you know. And I think that's, to some degree, where they're heading. It's just awesome to have guys like you on theCUBE Well you stay safe in Boston, All right, and thank you for watching everybody.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
AWS | ORGANIZATION | 0.99+ |
Dave Volante | PERSON | 0.99+ |
Ben White | PERSON | 0.99+ |
Boston | LOCATION | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
Atlanta | LOCATION | 0.99+ |
Ferrari | ORGANIZATION | 0.99+ |
Domo | ORGANIZATION | 0.99+ |
Vertica Customer Advisory Board | ORGANIZATION | 0.99+ |
Ben | PERSON | 0.99+ |
two things | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
Vertica | TITLE | 0.98+ |
theCUBE | ORGANIZATION | 0.97+ |
Vertica Big Data Conference | EVENT | 0.97+ |
Domo | TITLE | 0.97+ |
Domo | PERSON | 0.96+ |
Virtual Vertica Big Data Conference | EVENT | 0.96+ |
Virtual Vertica Big Data Conference 2020 | EVENT | 0.96+ |
first | QUANTITY | 0.95+ |
eon | TITLE | 0.92+ |
one | QUANTITY | 0.87+ |
today | DATE | 0.87+ |
millions of queries | QUANTITY | 0.84+ |
FlashBlade | TITLE | 0.82+ |
Virtual Vertica | EVENT | 0.75+ |
couple | QUANTITY | 0.7+ |
Pure FlashBlade | COMMERCIAL_ITEM | 0.58+ |
BDC 2020 | EVENT | 0.56+ |
MPP | TITLE | 0.55+ |
times | QUANTITY | 0.51+ |
RDBMS | TITLE | 0.48+ |
UNLIST TILL 4/2 - Vertica in Eon Mode: Past, Present, and Future
>> Paige: Hello everybody and thank you for joining us today for the virtual Vertica BDC 2020. Today's breakout session is entitled Vertica in Eon Mode past, present and future. I'm Paige Roberts, open source relations manager at Vertica and I'll be your host for this session. Joining me is Vertica engineer, Yuanzhe Bei and Vertica Product Manager, David Sprogis. Before we begin, I encourage you to submit questions or comments during the virtual session. You don't have to wait till the end. Just type your question or comment as you think of it in the question box, below the slides and click Submit. Q&A session at the end of the presentation. We'll answer as many of your questions as we're able to during that time, and any questions that we don't address, we'll do our best to answer offline. If you wish after the presentation, you can visit the Vertica forums to post your questions there and our engineering team is planning to join the forums to keep the conversation going, just like a Dev Lounge at a normal in person, BDC. So, as a reminder, you can maximize your screen by clicking the double arrow button in the lower right corner of the slides, if you want to see them bigger. And yes, before you ask, this virtual session is being recorded and will be available to view on demand this week. We are supposed to send you a notification as soon as it's ready. All right, let's get started. Over to you, Dave. >> David: Thanks, Paige. Hey, everybody. Let's start with a timeline of the life of Eon Mode. About two years ago, a little bit less than two years ago, we introduced Eon Mode on AWS. Pretty specifically for the purpose of rapid scaling to meet the cloud economics promise. It wasn't long after that we realized that workload isolation, a byproduct of the architecture was very important to our users and going to the third tick, you can see that the importance of that workload isolation was manifest in Eon Mode being made available on-premise using Pure Storage FlashBlade. Moving to the fourth tick mark, we took steps to improve workload isolation, with a new type of subcluster which Yuanzhe will go through and to the fifth tick mark, the introduction of secondary subclusters for faster scaling and other improvements which we will cover in the slides to come. Getting started with, why we created Eon Mode in the first place. Let's imagine that your database is this pie, the pecan pie and we're loading pecan data in through the ETL cutting board in the upper left hand corner. We have a couple of free floating pecans, which we might imagine to be data supporting external tables. As you know, the Vertica has a query engine capability as well which we call external tables. And so if we imagine this pie, we want to serve it with a number of servers. Well, let's say we wanted to serve it with three servers, three nodes, we would need to slice that pie into three segments and we would serve each one of those segments from one of our nodes. Now because the data is important to us and we don't want to lose it, we're going to be saving that data on some kind of raid storage or redundant storage. In case one of the drives goes bad, the data remains available because of the durability of raid. Imagine also, that we care about the availability of the overall database. Imagine that a node goes down, perhaps the second node goes down, we still want to be able to query our data and through nodes one and three, we still have all three shards covered and we can do this because of buddy projections. Each neighbor, each nodes neighbor contains a copy of the data from the node next to it. And so in this case, node one is sharing its segment with node two. So node two can cover node one, node three can cover node two and node one back to node three. Adding a little bit more complexity, we might store the data in different copies, each copy sorted for a different kind of query. We call this projections in Vertica and for each projection, we have another copy of the data sorted differently. Now it gets complex. What happens when we want to add a node? Well, if we wanted to add a fourth node here, what we would have to do, is figure out how to re-slice all of the data in all of the copies that we have. In effect, what we want to do is take our three slices and slice it into four, which means taking a portion of each of our existing thirds and re-segmenting into quarters. Now that looks simple in the graphic here, but when it comes to moving data around, it becomes quite complex because for each copy of each segment we need to replace it and move that data on to the new node. What's more, the fourth node can't have a copy of itself that would be problematic in case it went down. Instead, what we need is we need that buddy to be sitting on another node, a neighboring node. So we need to re-orient the buddies as well. All of this takes a lot of time, it can take 12, 24 or even 36 hours in a period when you do not want your database under high demand. In fact, you may want to stop loading data altogether in order to speed it up. This is a planned event and your applications should probably be down during this period, which makes it difficult. With the advent of cloud computing, we saw that services were coming up and down faster and we determined to re-architect Vertica in a way to accommodate that rapid scaling. Let's see how we did it. So let's start with four nodes now and we've got our four nodes database. Let's add communal storage and move each of the segments of data into communal storage. Now that's the separation that we're talking about. What happens if we run queries against it? Well, it turns out that the communal storage is not necessarily performing and so the IO would be slow, which would make the overall queries slow. In order to compensate for the low performance of communal storage, we need to add back local storage, now it doesn't have to be raid because this is just an ephemeral copy but with the data files, local to the node, the queries will run much faster. In AWS, communal storage really does mean an S3 bucket and here's a simplified version of the diagram. Now, do we need to store all of the data from the segment in the depot? The answer is no and the graphics inside the bucket has changed to reflect that. It looks more like a bullseye, showing just a segment of the data being copied to the cache or to the depot, as we call it on each one of the nodes. How much data do you store on the node? Well, it would be the active data set, the last 30 days, the last 30 minutes or the last. Whatever period of time you're working with. The active working set is the hot data and that's how large you want to size your depot. By architecting this way, when you scale up, you're not re-segmenting the database. What you're doing, is you're adding more compute and more subscriptions to the existing shards of the existing database. So in this case, we've added a complete set of four nodes. So we've doubled our capacity and we've doubled our subscriptions, which means that now, the two nodes can serve the yellow shard, two nodes can serve the red shard and so on. In this way, we're able to run twice as many queries in the same amount of time. So you're doubling the concurrency. How high can you scale? Well, can you scale to 3X, 5X? We tested this in the graphics on the right, which shows concurrent users in the X axis by the number of queries executed in a minute along the Y axis. We've grouped execution in runs of 10 users, 30 users, 50, 70 up to 150 users. Now focusing on any one of these groups, particularly up around 150. You can see through the three bars, starting with the bright purple bar, three nodes and three segments. That as you add nodes to the middle purple bar, six nodes and three segments, you've almost doubled your throughput up to the dark purple bar which is nine nodes and three segments and our tests show that you can go to 5X with pretty linear performance increase. Beyond that, you do continue to get an increase in performance but your incremental performance begins to fall off. Eon architecture does something else for us and that is it provides high availability because each of the nodes can be thought of as ephemeral and in fact, each node has a buddy subscription in a way similar to the prior architecture. So if we lose node four, we're losing the node responsible for the red shard and now node one has to pick up responsibility for the red shard while that node is down. When a query comes in, and let's say it comes into one and one is the initiator then one will look for participants, it'll find a blue shard and a green shard but when it's looking for the red, it finds itself and so the node number one will be doing double duty. This means that your performance will be cut in half approximately, for the query. This is acceptable until you are able to restore the node. Once you restore it and once the depot becomes rehydrated, then your performance goes back to normal. So this is a much simpler way to recover nodes in the event of node failure. By comparison, Enterprise Mode the older architecture. When we lose the fourth node, node one takes over responsibility for the first shard and the yellow shard and the red shard. But it also is responsible for rehydrating the entire data segment of the red shard to node four, this can be very time consuming and imposes even more stress on the first node. So performance will go down even further. Eon Mode has another feature and that is you can scale down completely to zero. We call this hibernation, you shut down your database and your database will maintain full consistency in a rest state in your S3 bucket and then when you need access to your database again, you simply recreate your cluster and revive your database and you can access your database once again. That concludes the rapid scaling portion of, why we created Eon Mode. To take us through workload isolation is Yuanzhe Bei, Yuanzhe. >> Yuanzhe: Thanks Dave, for presenting how Eon works in general. In the next section, I will show you another important capability of Vertica Eon Mode, the workload isolation. Dave used a pecan pie as an example of database. Now let's say it's time for the main course. Does anyone still have a problem with food touching on their plates. Parents know that it's a common problem for kids. Well, we have a similar problem in database as well. So there could be multiple different workloads accessing your database at the same time. Say you have ETL jobs running regularly. While at the same time, there are dashboards running short queries against your data. You may also have the end of month report running and their can be ad hoc data scientists, connect to the database and do whatever the data analysis they want to do and so on. How to make these mixed workload requests not interfere with each other is a real challenge for many DBAs. Vertica Eon Mode provides you the solution. I'm very excited here to introduce to you to the important concept in Eon Mode called subclusters. In Eon Mode, nodes they belong to the predefined subclusters rather than the whole cluster. DBAs can define different subcluster for different kinds of workloads and it redirects those workloads to the specific subclusters. For example, you can have an ETL subcluster, dashboard subcluster, report subcluster and the analytic machine learning subcluster. Vertica Eon subcluster is designed to achieve the three main goals. First of all, strong workload isolation. That means any operation in one subcluster should not affect or be affected by other subclusters. For example, say the subcluster running the report is quite overloaded and already there can be, the data scienctists running crazy analytic jobs, machine learning jobs on the analytics subcluster and making it very slow, even stuck or crash or whatever. In such scenario, your ETL and dashboards subcluster should not be or at least very minimum be impacted by this crisis and which means your ETL job which should not lag behind and dashboard should respond timely. We have done a lot of improvements as of 10.0 release and will continue to deliver improvements in this category. Secondly, fully customized subcluster settings. That means any subcluster can be set up and tuned for very different workloads without affecting other subclusters. Users should be able to tune up, tune down, certain parameters based on the actual needs of the individual subcluster workload requirements. As of today, Vertica already supports few settings that can be done at the subcluster level for example, the depot pinning policy and then we will continue extending more that is like resource pools (mumbles) in the near future. Lastly, Vertica subclusters should be easy to operate and cost efficient. What it means is that the subcluster should be able to turn on, turn off, add or remove or should be available for use according to rapid changing workloads. Let's say in this case, you want to spin up more dashboard subclusters because we need higher scores report, we can do that. You might need to run several report subclusters because you might want to run multiple reports at the same time. While on the other hand, you can shut down your analytic machine learning subcluster because no data scientists need to use it at this moment. So we made automate a lot of change, the improvements in this category, which I'll explain in detail later and one of the ultimate goal is to support auto scaling To sum up, what we really want to deliver for subcluster is very simple. You just need to remember that accessing subclusters should be just like accessing individual clusters. Well, these subclusters do share the same catalog. So you don't have to work out the stale data and don't need to worry about data synchronization. That'd be a nice goal, Vertica upcoming 10.0 release is certainly a milestone towards that goal, which will deliver a large part of the capability in this direction and then we will continue to improve it after 10.0 release. In the next couple of slides, I will highlight some issues about workload isolation in the initial Eon release and show you how we resolve these issues. First issue when we initially released our first or so called subcluster mode, it was implemented using fault groups. Well, fault groups and the subcluster have something in common. Yes, they are both defined as a set of nodes. However, they are very different in all the other ways. So, that was very confusing in the first place, when we implement this. As of 9.3.0 version, we decided to detach subcluster definition from the fault groups, which enabled us to further extend the capability of subclusters. Fault groups in the pre 9.3.0 versions will be converted into subclusters during the upgrade and this was a very important step that enabled us to provide all the amazing, following improvements on subclusters. The second issue in the past was that it's hard to control the execution groups for different types of workloads. There are two types of problems here and I will use some example to explain. The first issue is about control group size. There you allocate six nodes for your dashboard subcluster and what you really want is on the left, the three pairs of nodes as three execution groups, and each pair of nodes will need to subscribe to all the four shards. However, that's not really what you get. What you really get is there on the right side that the first four nodes subscribed to one shard each and the rest two nodes subscribed to two dangling shards. So you won't really get three execusion groups but instead only get one and two extra nodes have no value at all. The solution is to use subclusters. So instead of having a subcluster with six nodes, you can split it up into three smaller ones. Each subcluster will guarantee to subscribe to all the shards and you can further handle this three subcluster using load balancer across them. In this way you achieve the three real exclusion groups. The second issue is that the session participation is non-deterministic. Any session will just pick four random nodes from the subcluster as long as this covers one shard each. In other words, you don't really know which set of nodes will make up your execution group. What's the problem? So in this case, the fourth node will be doubled booked by two concurrent sessions. And you can imagine that the resource usage will be imbalanced and both queries performance will suffer. What is even worse is that these queries of the two concurrent sessions target different table They will cause the issue, that depot efficiency will be reduced, because both session will try to fetch the files on to two tables into the same depot and if your depot is not large enough, they will evict each other, which will be very bad. To solve this the same way, you can solve this by declaring subclusters, in this case, two subclusters and a load balancer group across them. The reason it solved the problem is because the session participation would not go across the boundary. So there won't be a case that any node is double booked and in terms of the depot and if you use the subcluster and avoid using a load balancer group, and carefully send the first workload to the first subcluster and the second to the second subcluster and then the result is that depot isolation is achieved. The first subcluster will maintain the data files for the first query and you don't need to worry about the file being evicted by the second kind of session. Here comes the next issue, it's the scaling down. In the old way of defining subclusters, you may have several execution groups in the subcluster. You want to shut it down, one or two execution groups to save cost. Well, here comes the pain, because you don't know which nodes may be used by which session at any point, it is hard to find the right timing to hit the shutdown button of any of the instances. And if you do and get unlucky, say in this case, you pull the first four nodes, one of the session will fail because it's participating in the node two and node four at that point. User of that session will notice because their query fails and we know that for many business this is critical problem and not acceptable. Again, with subclusters this problem is resolved. Same reason, session cannot go across the subcluster boundary. So all you need to do is just first prevent query sent to the first subcluster and then you can shut down the instances in that subcluster. You are guaranteed to not break any running sessions. Now, you're happy and you want to shut down more subclusters then you hit the issue four, the whole cluster will go down, why? Because the cluster loses quorum. As a distributed system, you need to have at least more than half of a node to be up in order to commit and keep the cluster up. This is to prevent the catalog diversion from happening, which is important. But do you still want to shut down those nodes? Because what's the point of keeping those nodes up and if you are not using them and let them cost you money right. So Vertica has a solution, you can define a subcluster as secondary to allow them to shut down without worrying about quorum. In this case, you can define the first three subclusters as secondary and the fourth one as primary. By doing so, this secondary subclusters will not be counted towards the quorum because we changed the rule. Now instead of requiring more than half of node to be up, it only require more than half of the primary node to be up. Now you can shut down your second subcluster and even shut down your third subcluster as well and keep the remaining primary subcluster to be still running healthily. There are actually more benefits by defining secondary subcluster in addition to the quorum concern, because the secondary subclusters no longer have the voting power, they don't need to persist catalog anymore. This means those nodes are faster to deploy, and can be dropped and re-added. Without the worry about the catalog persistency. For the most the subcluster that only need to read only query, it's the best practice to define them as secondary. The commit will be faster on this secondary subcluster as well, so running this query on the secondary subcluster will have less spikes. Primary subcluster as usual handle everything is responsible for consistency, the background tasks will be running. So DBAs should make sure that the primary subcluster is stable and assume is running all the time. Of course, you need to at least one primary subcluster in your database. Now with the secondary subcluster, user can start and stop as they need, which is very convenient and this further brings up another issue is that if there's an ETL transaction running and in the middle, a subcluster starting and it become up. In older versions, there is no catalog resync mechanism to keep the new subcluster up to date. So Vertica rolls back to ETL session to keep the data consistency. This is actually quite disruptive because real world ETL workloads can sometimes take hours and rolling back at the end means, a large waste of resources. We resolved this issue in 9.3.1 version by introducing a catalog resync mechanism when such situation happens. ETL transactions will not roll back anymore, but instead will take some time to resync the catalog and commit and the problem is resolved. And last issue I would like to talk about is the subscription. Especially for large subcluster when you start it, the startup time is quite long, because the subscription commit used to be serialized. In one of the in our internal testing with large catalogs committing a subscription, you can imagine it takes five minutes. Secondary subcluster is better, because it doesn't need to persist the catalog during the commit but still take about two seconds to commit. So what's the problem here? Let's do the math and look at this chart. The X axis is the time in the minutes and the Y axis is the number of nodes to be subscribed. The dark blues represents your primary subcluster and light blue represents the secondary subcluster. Let's say the subcluster have 16 nodes in total and if you start a secondary subcluster, it will spend about 30 seconds in total, because the 2 seconds times 16 is 32. It's not actually that long time. but if you imagine that starting secondary subcluster, you expect it to be super fast to react to the fast changing workload and 30 seconds is no longer trivial anymore and what is even worse is on the primary subcluster side. Because the commit is much longer than five minutes let's assume, then at the point, you are committing to six nodes subscription all other nodes already waited for 30 minutes for GCLX or we know the global catalog lock, and the Vertica will crash the nodes, if any node cannot get the GCLX for 30 minutes. So the end result is that your whole database crashed. That's a serious problem and we know that and that's why we are already planning for the fix, for the 10.0, so that all the subscription will be batched up and all the nodes will commit at the same time concurrently. And by doing that, you can imagine the primary subcluster can finish commiting in five minutes instead of crashing and the secondary subcluster can be finished even in seconds. That summarizes the highlights for the improvements we have done as of 10.0, and I hope you already get excited about Emerging Eon Deployment Pattern that's shown here. A primary subcluster that handles data loading, ETL jobs and tuple mover jobs is the backbone of the database and you keep it running all the time. At the same time defining different secondary subcluster for different workloads and provision them when the workload requirement arrives and then de-provision them when the workload is done to save the operational cost. So can't wait to play with the subcluster. Here as are some Admin Tools command you can start using. And for more details, check out our Eon subcluster documentation for more details. And thanks everyone for listening and I'll head back to Dave to talk about the Eon on-prem. >> David: Thanks Yuanzhe. At the same time that Yuanzhe and the rest of the dev team were working on the improvements that Yuanzhe described in and other improvements. This guy, John Yovanovich, stood on stage and told us about his deployment at at&t where he was running Eon Mode on-prem. Now this was only six months after we had launched Eon Mode on AWS. So when he told us that he was putting it into production on-prem, we nearly fell out of our chairs. How is this possible? We took a look back at Eon and determined that the workload isolation and the improvement to the operations for restoring nodes and other things had sufficient value that John wanted to run it on-prem. And he was running it on the Pure Storage FlashBlade. Taking a second look at the FlashBlade we thought alright well, does it have the performance? Yes, it does. The FlashBlade is a collection of individual blades, each one of them with NVMe storage on it, which is not only performance but it's scalable and so, we then asked is it durable? The answer is yes. The data safety is implemented with the N+2 redundancy which means that up to two blades can fail and the data remains available. And so with this we realized DBAs can sleep well at night, knowing that their data is safe, after all Eon Mode outsources the durability to the communal storage data store. Does FlashBlade have the capacity for growth? Well, yes it does. You can start as low as 120 terabytes and grow as high as about eight petabytes. So it certainly covers the range for most enterprise usages. And operationally, it couldn't be easier to use. When you want to grow your database. You can simply pop new blades into the FlashBlade unit, and you can do that hot. If one goes bad, you can pull it out and replace it hot. So you don't have to take your data store down and therefore you don't have to take Vertica down. Knowing all of these things we got behind Pure Storage and partnered with them to implement the first version of Eon on-premise. That changed our roadmap a little bit. We were imagining it would start with Amazon and then go to Google and then to Azure and at some point to Alibaba cloud, but as you can see from the left column, we started with Amazon and went to Pure Storage. And then from Pure Storage, we went to Minio and we launched Eon Mode on Minio at the end of last year. Minio is a little bit different than Pure Storage. It's software only, so you can run it on pretty much any x86 servers and you can cluster them with storage to serve up an S3 bucket. It's a great solution for up to about 120 terabytes Beyond that, we're not sure about performance implications cause we haven't tested it but for your dev environments or small production environments, we think it's great. With Vertica 10, we're introducing Eon Mode on Google Cloud. This means not only running Eon Mode in the cloud, but also being able to launch it from the marketplace. We're also offering Eon Mode on HDFS with version 10. If you have a Hadoop environment, and you want to breathe new fresh life into it with the high performance of Vertica, you can do that starting with version 10. Looking forward we'll be moving Eon mode to Microsoft Azure. We expect to have something breathing in the fall and offering it to select customers for beta testing and then we expect to release it sometime in 2021 Following that, further on horizon is Alibaba cloud. Now, to be clear we will be putting, Vertica in Enterprise Mode on Alibaba cloud in 2020 but Eon Mode is going to trail behind whether it lands in 2021 or not, we're not quite sure at this point. Our goal is to deliver Eon Mode anywhere you want to run it, on-prem or in the cloud, or both because that is one of the great value propositions of Vertica is the hybrid capability, the ability to run in both your on prem environment and in the cloud. What's next, I've got three priority and roadmap slides. This is the first of the three. We're going to start with improvements to the core of Vertica. Starting with query crunching, which allows you to run long running queries faster by getting nodes to collaborate, you'll see that coming very soon. We'll be making improvements to large clusters and specifically large cluster mode. The management of large clusters over 60 nodes can be tedious. We intend to improve that. In part, by creating a third network channel to offload some of the communication that we're now loading onto our spread or agreement protocol. We'll be improving depot efficiency. We'll be pushing down more controls to the subcluster level, allowing you to control your resource pools at the subcluster level and we'll be pairing tuple moving with data loading. From an operational flexibility perspective, we want to make it very easy to shut down and revive primaries and secondaries on-prem and in the cloud. Right now, it's a little bit tedious, very doable. We want to make it as easy as a walk in the park. We also want to allow you to be able to revive into a different size subcluster and last but not least, in fact, probably the most important, the ability to change shard count. This has been a sticking point for a lot of people and it puts a lot of pressure on the early decision of how many shards should my database be? Whether it's in 2020 or 2021. We know it's important to you so it's important to us. Ease of use is also important to us and we're making big investments in the management console, to improve managing subclusters, as well as to help you manage your load balancer groups. We also intend to grow and extend Eon Mode to new environments. Now we'll take questions and answers
SUMMARY :
and our engineering team is planning to join the forums and going to the third tick, you can see that and the second to the second subcluster and the improvement to the
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David Sprogis | PERSON | 0.99+ |
David | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
John Yovanovich | PERSON | 0.99+ |
10 users | QUANTITY | 0.99+ |
Paige Roberts | PERSON | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
Yuanzhe Bei | PERSON | 0.99+ |
John | PERSON | 0.99+ |
five minutes | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
30 seconds | QUANTITY | 0.99+ |
50 | QUANTITY | 0.99+ |
second issue | QUANTITY | 0.99+ |
12 | QUANTITY | 0.99+ |
Yuanzhe | PERSON | 0.99+ |
120 terabytes | QUANTITY | 0.99+ |
30 users | QUANTITY | 0.99+ |
two types | QUANTITY | 0.99+ |
2021 | DATE | 0.99+ |
Paige | PERSON | 0.99+ |
30 minutes | QUANTITY | 0.99+ |
three pairs | QUANTITY | 0.99+ |
second | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
nine nodes | QUANTITY | 0.99+ |
first subcluster | QUANTITY | 0.99+ |
two tables | QUANTITY | 0.99+ |
two nodes | QUANTITY | 0.99+ |
first issue | QUANTITY | 0.99+ |
each copy | QUANTITY | 0.99+ |
2 seconds | QUANTITY | 0.99+ |
36 hours | QUANTITY | 0.99+ |
second subcluster | QUANTITY | 0.99+ |
fourth node | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
six nodes | QUANTITY | 0.99+ |
third subcluster | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
twice | QUANTITY | 0.99+ |
First issue | QUANTITY | 0.99+ |
three segments | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
three bars | QUANTITY | 0.99+ |
24 | QUANTITY | 0.99+ |
5X | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
16 nodes | QUANTITY | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
each segment | QUANTITY | 0.99+ |
first node | QUANTITY | 0.99+ |
three slices | QUANTITY | 0.99+ |
Each subcluster | QUANTITY | 0.99+ |
each nodes | QUANTITY | 0.99+ |
three nodes | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
two subclusters | QUANTITY | 0.98+ |
three servers | QUANTITY | 0.98+ |
four shards | QUANTITY | 0.98+ |
3X | QUANTITY | 0.98+ |
three | QUANTITY | 0.98+ |
two concurrent sessions | QUANTITY | 0.98+ |
UNLIST TILL 4/2 - Vertica Big Data Conference Keynote
>> Joy: Welcome to the Virtual Big Data Conference. Vertica is so excited to host this event. I'm Joy King, and I'll be your host for today's Big Data Conference Keynote Session. It's my honor and my genuine pleasure to lead Vertica's product and go-to-market strategy. And I'm so lucky to have a passionate and committed team who turned our Vertica BDC event, into a virtual event in a very short amount of time. I want to thank the thousands of people, and yes, that's our true number who have registered to attend this virtual event. We were determined to balance your health, safety and your peace of mind with the excitement of the Vertica BDC. This is a very unique event. Because as I hope you all know, we focus on engineering and architecture, best practice sharing and customer stories that will educate and inspire everyone. I also want to thank our top sponsors for the virtual BDC, Arrow, and Pure Storage. Our partnerships are so important to us and to everyone in the audience. Because together, we get things done faster and better. Now for today's keynote, you'll hear from three very important and energizing speakers. First, Colin Mahony, our SVP and General Manager for Vertica, will talk about the market trends that Vertica is betting on to win for our customers. And he'll share the exciting news about our Vertica 10 announcement and how this will benefit our customers. Then you'll hear from Amy Fowler, VP of strategy and solutions for FlashBlade at Pure Storage. Our partnership with Pure Storage is truly unique in the industry, because together modern infrastructure from Pure powers modern analytics from Vertica. And then you'll hear from John Yovanovich, Director of IT at AT&T, who will tell you about the Pure Vertica Symphony that plays live every day at AT&T. Here we go, Colin, over to you. >> Colin: Well, thanks a lot joy. And, I want to echo Joy's thanks to our sponsors, and so many of you who have helped make this happen. This is not an easy time for anyone. We were certainly looking forward to getting together in person in Boston during the Vertica Big Data Conference and Winning with Data. But I think all of you and our team have done a great job, scrambling and putting together a terrific virtual event. So really appreciate your time. I also want to remind people that we will make both the slides and the full recording available after this. So for any of those who weren't able to join live, that is still going to be available. Well, things have been pretty exciting here. And in the analytic space in general, certainly for Vertica, there's a lot happening. There are a lot of problems to solve, a lot of opportunities to make things better, and a lot of data that can really make every business stronger, more efficient, and frankly, more differentiated. For Vertica, though, we know that focusing on the challenges that we can directly address with our platform, and our people, and where we can actually make the biggest difference is where we ought to be putting our energy and our resources. I think one of the things that has made Vertica so strong over the years is our ability to focus on those areas where we can make a great difference. So for us as we look at the market, and we look at where we play, there are really three recent and some not so recent, but certainly picking up a lot of the market trends that have become critical for every industry that wants to Win Big With Data. We've heard this loud and clear from our customers and from the analysts that cover the market. If I were to summarize these three areas, this really is the core focus for us right now. We know that there's massive data growth. And if we can unify the data silos so that people can really take advantage of that data, we can make a huge difference. We know that public clouds offer tremendous advantages, but we also know that balance and flexibility is critical. And we all need the benefit that machine learning for all the types up to the end data science. We all need the benefits that they can bring to every single use case, but only if it can really be operationalized at scale, accurate and in real time. And the power of Vertica is, of course, how we're able to bring so many of these things together. Let me talk a little bit more about some of these trends. So one of the first industry trends that we've all been following probably now for over the last decade, is Hadoop and specifically HDFS. So many companies have invested, time, money, more importantly, people in leveraging the opportunity that HDFS brought to the market. HDFS is really part of a much broader storage disruption that we'll talk a little bit more about, more broadly than HDFS. But HDFS itself was really designed for petabytes of data, leveraging low cost commodity hardware and the ability to capture a wide variety of data formats, from a wide variety of data sources and applications. And I think what people really wanted, was to store that data before having to define exactly what structures they should go into. So over the last decade or so, the focus for most organizations is figuring out how to capture, store and frankly manage that data. And as a platform to do that, I think, Hadoop was pretty good. It certainly changed the way that a lot of enterprises think about their data and where it's locked up. In parallel with Hadoop, particularly over the last five years, Cloud Object Storage has also given every organization another option for collecting, storing and managing even more data. That has led to a huge growth in data storage, obviously, up on public clouds like Amazon and their S3, Google Cloud Storage and Azure Blob Storage just to name a few. And then when you consider regional and local object storage offered by cloud vendors all over the world, the explosion of that data, in leveraging this type of object storage is very real. And I think, as I mentioned, it's just part of this broader storage disruption that's been going on. But with all this growth in the data, in all these new places to put this data, every organization we talk to is facing even more challenges now around the data silo. Sure the data silos certainly getting bigger. And hopefully they're getting cheaper per bit. But as I said, the focus has really been on collecting, storing and managing the data. But between the new data lakes and many different cloud object storage combined with all sorts of data types from the complexity of managing all this, getting that business value has been very limited. This actually takes me to big bet number one for Team Vertica, which is to unify the data. Our goal, and some of the announcements we have made today plus roadmap announcements I'll share with you throughout this presentation. Our goal is to ensure that all the time, money and effort that has gone into storing that data, all the data turns into business value. So how are we going to do that? With a unified analytics platform that analyzes the data wherever it is HDFS, Cloud Object Storage, External tables in an any format ORC, Parquet, JSON, and of course, our own Native Roth Vertica format. Analyze the data in the right place in the right format, using a single unified tool. This is something that Vertica has always been committed to, and you'll see in some of our announcements today, we're just doubling down on that commitment. Let's talk a little bit more about the public cloud. This is certainly the second trend. It's the second wave maybe of data disruption with object storage. And there's a lot of advantages when it comes to public cloud. There's no question that the public clouds give rapid access to compute storage with the added benefit of eliminating data center maintenance that so many companies, want to get out of themselves. But maybe the biggest advantage that I see is the architectural innovation. The public clouds have introduced so many methodologies around how to provision quickly, separating compute and storage and really dialing-in the exact needs on demand, as you change workloads. When public clouds began, it made a lot of sense for the cloud providers and their customers to charge and pay for compute and storage in the ratio that each use case demanded. And I think you're seeing that trend, proliferate all over the place, not just up in public cloud. That architecture itself is really becoming the next generation architecture for on-premise data centers, as well. But there are a lot of concerns. I think we're all aware of them. They're out there many times for different workloads, there are higher costs. Especially if some of the workloads that are being run through analytics, which tend to run all the time. Just like some of the silo challenges that companies are facing with HDFS, data lakes and cloud storage, the public clouds have similar types of siloed challenges as well. Initially, there was a belief that they were cheaper than data centers, and when you added in all the costs, it looked that way. And again, for certain elastic workloads, that is the case. I don't think that's true across the board overall. Even to the point where a lot of the cloud vendors aren't just charging lower costs anymore. We hear from a lot of customers that they don't really want to tether themselves to any one cloud because of some of those uncertainties. Of course, security and privacy are a concern. We hear a lot of concerns with regards to cloud and even some SaaS vendors around shared data catalogs, across all the customers and not enough separation. But security concerns are out there, you can read about them. I'm not going to jump into that bandwagon. But we hear about them. And then, of course, I think one of the things we hear the most from our customers, is that each cloud stack is starting to feel even a lot more locked in than the traditional data warehouse appliance. And as everybody knows, the industry has been running away from appliances as fast as it can. And so they're not eager to get locked into another, quote, unquote, virtual appliance, if you will, up in the cloud. They really want to make sure they have flexibility in which clouds, they're going to today, tomorrow and in the future. And frankly, we hear from a lot of our customers that they're very interested in eventually mixing and matching, compute from one cloud with, say storage from another cloud, which I think is something that we'll hear a lot more about. And so for us, that's why we've got our big bet number two. we love the cloud. We love the public cloud. We love the private clouds on-premise, and other hosting providers. But our passion and commitment is for Vertica to be able to run in any of the clouds that our customers choose, and make it portable across those clouds. We have supported on-premises and all public clouds for years. And today, we have announced even more support for Vertica in Eon Mode, the deployment option that leverages the separation of compute from storage, with even more deployment choices, which I'm going to also touch more on as we go. So super excited about our big bet number two. And finally as I mentioned, for all the hype that there is around machine learning, I actually think that most importantly, this third trend that team Vertica is determined to address is the need to bring business critical, analytics, machine learning, data science projects into production. For so many years, there just wasn't enough data available to justify the investment in machine learning. Also, processing power was expensive, and storage was prohibitively expensive. But to train and score and evaluate all the different models to unlock the full power of predictive analytics was tough. Today you have those massive data volumes. You have the relatively cheap processing power and storage to make that dream a reality. And if you think about this, I mean with all the data that's available to every company, the real need is to operationalize the speed and the scale of machine learning so that these organizations can actually take advantage of it where they need to. I mean, we've seen this for years with Vertica, going back to some of the most advanced gaming companies in the early days, they were incorporating this with live data directly into their gaming experiences. Well, every organization wants to do that now. And the accuracy for clickability and real time actions are all key to separating the leaders from the rest of the pack in every industry when it comes to machine learning. But if you look at a lot of these projects, the reality is that there's a ton of buzz, there's a ton of hype spanning every acronym that you can imagine. But most companies are struggling, do the separate teams, different tools, silos and the limitation that many platforms are facing, driving, down sampling to get a small subset of the data, to try to create a model that then doesn't apply, or compromising accuracy and making it virtually impossible to replicate models, and understand decisions. And if there's one thing that we've learned when it comes to data, prescriptive data at the atomic level, being able to show end of one as we refer to it, meaning individually tailored data. No matter what it is healthcare, entertainment experiences, like gaming or other, being able to get at the granular data and make these decisions, make that scoring applies to machine learning just as much as it applies to giving somebody a next-best-offer. But the opportunity has never been greater. The need to integrate this end-to-end workflow and support the right tools without compromising on that accuracy. Think about it as no downsampling, using all the data, it really is key to machine learning success. Which should be no surprise then why the third big bet from Vertica is one that we've actually been working on for years. And we're so proud to be where we are today, helping the data disruptors across the world operationalize machine learning. This big bet has the potential to truly unlock, really the potential of machine learning. And today, we're announcing some very important new capabilities specifically focused on unifying the work being done by the data science community, with their preferred tools and platforms, and the volume of data and performance at scale, available in Vertica. Our strategy has been very consistent over the last several years. As I said in the beginning, we haven't deviated from our strategy. Of course, there's always things that we add. Most of the time, it's customer driven, it's based on what our customers are asking us to do. But I think we've also done a great job, not trying to be all things to all people. Especially as these hype cycles flare up around us, we absolutely love participating in these different areas without getting completely distracted. I mean, there's a variety of query tools and data warehouses and analytics platforms in the market. We all know that. There are tools and platforms that are offered by the public cloud vendors, by other vendors that support one or two specific clouds. There are appliance vendors, who I was referring to earlier who can deliver package data warehouse offerings for private data centers. And there's a ton of popular machine learning tools, languages and other kits. But Vertica is the only advanced analytic platform that can do all this, that can bring it together. We can analyze the data wherever it is, in HDFS, S3 Object Storage, or Vertica itself. Natively we support multiple clouds on-premise deployments, And maybe most importantly, we offer that choice of deployment modes to allow our customers to choose the architecture that works for them right now. It still also gives them the option to change move, evolve over time. And Vertica is the only analytics database with end-to-end machine learning that can truly operationalize ML at scale. And I know it's a mouthful. But it is not easy to do all these things. It is one of the things that highly differentiates Vertica from the rest of the pack. It is also why our customers, all of you continue to bet on us and see the value that we are delivering and we will continue to deliver. Here's a couple of examples of some of our customers who are powered by Vertica. It's the scale of data. It's the millisecond response times. Performance and scale have always been a huge part of what we have been about, not the only thing. I think the functionality all the capabilities that we add to the platform, the ease of use, the flexibility, obviously with the deployment. But if you look at some of the numbers they are under these customers on this slide. And I've shared a lot of different stories about these customers. Which, by the way, it still amaze me every time I talk to one and I get the updates, you can see the power and the difference that Vertica is making. Equally important, if you look at a lot of these customers, they are the epitome of being able to deploy Vertica in a lot of different environments. Many of the customers on this slide are not using Vertica just on-premise or just in the cloud. They're using it in a hybrid way. They're using it in multiple different clouds. And again, we've been with them on that journey throughout, which is what has made this product and frankly, our roadmap and our vision exactly what it is. It's been quite a journey. And that journey continues now with the Vertica 10 release. The Vertica 10 release is obviously a massive release for us. But if you look back, you can see that building on that native columnar architecture that started a long time ago, obviously, with the C-Store paper. We built it to leverage that commodity hardware, because it was an architecture that was never tightly integrated with any specific underlying infrastructure. I still remember hearing the initial pitch from Mike Stonebreaker, about the vision of Vertica as a software only solution and the importance of separating the company from hardware innovation. And at the time, Mike basically said to me, "there's so much R&D in innovation that's going to happen in hardware, we shouldn't bake hardware into our solution. We should do it in software, and we'll be able to take advantage of that hardware." And that is exactly what has happened. But one of the most recent innovations that we embraced with hardware is certainly that separation of compute and storage. As I said previously, the public cloud providers offered this next generation architecture, really to ensure that they can provide the customers exactly what they needed, more compute or more storage and charge for each, respectively. The separation of compute and storage, compute from storage is a major milestone in data center architectures. If you think about it, it's really not only a public cloud innovation, though. It fundamentally redefines the next generation data architecture for on-premise and for pretty much every way people are thinking about computing today. And that goes for software too. Object storage is an example of the cost effective means for storing data. And even more importantly, separating compute from storage for analytic workloads has a lot of advantages. Including the opportunity to manage much more dynamic, flexible workloads. And more importantly, truly isolate those workloads from others. And by the way, once you start having something that can truly isolate workloads, then you can have the conversations around autonomic computing, around setting up some nodes, some compute resources on the data that won't affect any of the other data to do some things on their own, maybe some self analytics, by the system, etc. A lot of things that many of you know we've already been exploring in terms of our own system data in the product. But it was May 2018, believe it or not, it seems like a long time ago where we first announced Eon Mode and I want to make something very clear, actually about Eon mode. It's a mode, it's a deployment option for Vertica customers. And I think this is another huge benefit that we don't talk about enough. But unlike a lot of vendors in the market who will dig you and charge you for every single add-on like hit-buy, you name it. You get this with the Vertica product. If you continue to pay support and maintenance, this comes with the upgrade. This comes as part of the new release. So any customer who owns or buys Vertica has the ability to set up either an Enterprise Mode or Eon Mode, which is a question I know that comes up sometimes. Our first announcement of Eon was obviously AWS customers, including the trade desk, AT&T. Most of whom will be speaking here later at the Virtual Big Data Conference. They saw a huge opportunity. Eon Mode, not only allowed Vertica to scale elastically with that specific compute and storage that was needed, but it really dramatically simplified database operations including things like workload balancing, node recovery, compute provisioning, etc. So one of the most popular functions is that ability to isolate the workloads and really allocate those resources without negatively affecting others. And even though traditional data warehouses, including Vertica Enterprise Mode have been able to do lots of different workload isolation, it's never been as strong as Eon Mode. Well, it certainly didn't take long for our customers to see that value across the board with Eon Mode. Not just up in the cloud, in partnership with one of our most valued partners and a platinum sponsor here. Joy mentioned at the beginning. We announced Vertica Eon Mode for Pure Storage FlashBlade in September 2019. And again, just to be clear, this is not a new product, it's one Vertica with yet more deployment options. With Pure Storage, Vertica in Eon mode is not limited in any way by variable cloud, network latency. The performance is actually amazing when you take the benefits of separate and compute from storage and you run it with a Pure environment on-premise. Vertica in Eon Mode has a super smart cache layer that we call the depot. It's a big part of our secret sauce around Eon mode. And combined with the power and performance of Pure's FlashBlade, Vertica became the industry's first advanced analytics platform that actually separates compute and storage for on-premises data centers. Something that a lot of our customers are already benefiting from, and we're super excited about it. But as I said, this is a journey. We don't stop, we're not going to stop. Our customers need the flexibility of multiple public clouds. So today with Vertica 10, we're super proud and excited to announce support for Vertica in Eon Mode on Google Cloud. This gives our customers the ability to use their Vertica licenses on Amazon AWS, on-premise with Pure Storage and on Google Cloud. Now, we were talking about HDFS and a lot of our customers who have invested quite a bit in HDFS as a place, especially to store data have been pushing us to support Eon Mode with HDFS. So as part of Vertica 10, we are also announcing support for Vertica in Eon Mode using HDFS as the communal storage. Vertica's own Roth format data can be stored in HDFS, and actually the full functionality of Vertica is complete analytics, geospatial pattern matching, time series, machine learning, everything that we have in there can be applied to this data. And on the same HDFS nodes, Vertica can actually also analyze data in ORC or Parquet format, using External tables. We can also execute joins between the Roth data the External table holds, which powers a much more comprehensive view. So again, it's that flexibility to be able to support our customers, wherever they need us to support them on whatever platform, they have. Vertica 10 gives us a lot more ways that we can deploy Eon Mode in various environments for our customers. It allows them to take advantage of Vertica in Eon Mode and the power that it brings with that separation, with that workload isolation, to whichever platform they are most comfortable with. Now, there's a lot that has come in Vertica 10. I'm definitely not going to be able to cover everything. But we also introduced complex types as an example. And complex data types fit very well into Eon as well in this separation. They significantly reduce the data pipeline, the cost of moving data between those, a much better support for unstructured data, which a lot of our customers have mixed with structured data, of course, and they leverage a lot of columnar execution that Vertica provides. So you get complex data types in Vertica now, a lot more data, stronger performance. It goes great with the announcement that we made with the broader Eon Mode. Let's talk a little bit more about machine learning. We've been actually doing work in and around machine learning with various extra regressions and a whole bunch of other algorithms for several years. We saw the huge advantage that MPP offered, not just as a sequel engine as a database, but for ML as well. Didn't take as long to realize that there's a lot more to operationalizing machine learning than just those algorithms. It's data preparation, it's that model trade training. It's the scoring, the shaping, the evaluation. That is so much of what machine learning and frankly, data science is about. You do know, everybody always wants to jump to the sexy algorithm and we handle those tasks very, very well. It makes Vertica a terrific platform to do that. A lot of work in data science and machine learning is done in other tools. I had mentioned that there's just so many tools out there. We want people to be able to take advantage of all that. We never believed we were going to be the best algorithm company or come up with the best models for people to use. So with Vertica 10, we support PMML. We can import now and export PMML models. It's a huge step for us around that operationalizing machine learning projects for our customers. Allowing the models to get built outside of Vertica yet be imported in and then applying to that full scale of data with all the performance that you would expect from Vertica. We also are more tightly integrating with Python. As many of you know, we've been doing a lot of open source projects with the community driven by many of our customers, like Uber. And so now with Python we've integrated with TensorFlow, allowing data scientists to build models in their preferred language, to take advantage of TensorFlow. But again, to store and deploy those models at scale with Vertica. I think both these announcements are proof of our big bet number three, and really our commitment to supporting innovation throughout the community by operationalizing ML with that accuracy, performance and scale of Vertica for our customers. Again, there's a lot of steps when it comes to the workflow of machine learning. These are some of them that you can see on the slide, and it's definitely not linear either. We see this as a circle. And companies that do it, well just continue to learn, they continue to rescore, they continue to redeploy and they want to operationalize all that within a single platform that can take advantage of all those capabilities. And that is the platform, with a very robust ecosystem that Vertica has always been committed to as an organization and will continue to be. This graphic, many of you have seen it evolve over the years. Frankly, if we put everything and everyone on here wouldn't fit on a slide. But it will absolutely continue to evolve and grow as we support our customers, where they need the support most. So, again, being able to deploy everywhere, being able to take advantage of Vertica, not just as a business analyst or a business user, but as a data scientists or as an operational or BI person. We want Vertica to be leveraged and used by the broader organization. So I think it's fair to say and I encourage everybody to learn more about Vertica 10, because I'm just highlighting some of the bigger aspects of it. But we talked about those three market trends. The need to unify the silos, the need for hybrid multiple cloud deployment options, the need to operationalize business critical machine learning projects. Vertica 10 has absolutely delivered on those. But again, we are not going to stop. It is our job not to, and this is how Team Vertica thrives. I always joke that the next release is the best release. And, of course, even after Vertica 10, that is also true, although Vertica 10 is pretty awesome. But, you know, from the first line of code, we've always been focused on performance and scale, right. And like any really strong data platform, the execution engine, the optimizer and the execution engine are the two core pieces of that. Beyond Vertica 10, some of the big things that we're already working on, next generation execution engine. We're already actually seeing incredible early performance from this. And this is just one example, of how important it is for an organization like Vertica to constantly go back and re-innovate. Every single release, we do the sit ups and crunches, our performance and scale. How do we improve? And there's so many parts of the core server, there's so many parts of our broader ecosystem. We are constantly looking at coverages of how we can go back to all the code lines that we have, and make them better in the current environment. And it's not an easy thing to do when you're doing that, and you're also expanding in the environment that we are expanding into to take advantage of the different deployments, which is a great segue to this slide. Because if you think about today, we're obviously already available with Eon Mode and Amazon, AWS and Pure and actually MinIO as well. As I talked about in Vertica 10 we're adding Google and HDFS. And coming next, obviously, Microsoft Azure, Alibaba cloud. So being able to expand into more of these environments is really important for the Vertica team and how we go forward. And it's not just running in these clouds, for us, we want it to be a SaaS like experience in all these clouds. We want you to be able to deploy Vertica in 15 minutes or less on these clouds. You can also consume Vertica, in a lot of different ways, on these clouds. As an example, in Amazon Vertica by the Hour. So for us, it's not just about running, it's about taking advantage of the ecosystems that all these cloud providers offer, and really optimizing the Vertica experience as part of them. Optimization, around automation, around self service capabilities, extending our management console, we now have products that like the Vertica Advisor Tool that our Customer Success Team has created to actually use our own smarts in Vertica. To take data from customers that give it to us and help them tune automatically their environment. You can imagine that we're taking that to the next level, in a lot of different endeavors that we're doing around how Vertica as a product can actually be smarter because we all know that simplicity is key. There just aren't enough people in the world who are good at managing data and taking it to the next level. And of course, other things that we all hear about, whether it's Kubernetes and containerization. You can imagine that that probably works very well with the Eon Mode and separating compute and storage. But innovation happens everywhere. We innovate around our community documentation. Many of you have taken advantage of the Vertica Academy. The numbers there are through the roof in terms of the number of people coming in and certifying on it. So there's a lot of things that are within the core products. There's a lot of activity and action beyond the core products that we're taking advantage of. And let's not forget why we're here, right? It's easy to talk about a platform, a data platform, it's easy to jump into all the functionality, the analytics, the flexibility, how we can offer it. But at the end of the day, somebody, a person, she's got to take advantage of this data, she's got to be able to take this data and use this information to make a critical business decision. And that doesn't happen unless we explore lots of different and frankly, new ways to get that predictive analytics UI and interface beyond just the standard BI tools in front of her at the right time. And so there's a lot of activity, I'll tease you with that going on in this organization right now about how we can do that and deliver that for our customers. We're in a great position to be able to see exactly how this data is consumed and used and start with this core platform that we have to go out. Look, I know, the plan wasn't to do this as a virtual BDC. But I really appreciate you tuning in. Really appreciate your support. I think if there's any silver lining to us, maybe not being able to do this in person, it's the fact that the reach has actually gone significantly higher than what we would have been able to do in person in Boston. We're certainly looking forward to doing a Big Data Conference in the future. But if I could leave you with anything, know this, since that first release for Vertica, and our very first customers, we have been very consistent. We respect all the innovation around us, whether it's open source or not. We understand the market trends. We embrace those new ideas and technologies and for us true north, and the most important thing is what does our customer need to do? What problem are they trying to solve? And how do we use the advantages that we have without disrupting our customers? But knowing that you depend on us to deliver that unified analytics strategy, it will deliver that performance of scale, not only today, but tomorrow and for years to come. We've added a lot of great features to Vertica. I think we've said no to a lot of things, frankly, that we just knew we wouldn't be the best company to deliver. When we say we're going to do things we do them. Vertica 10 is a perfect example of so many of those things that we from you, our customers have heard loud and clear, and we have delivered. I am incredibly proud of this team across the board. I think the culture of Vertica, a customer first culture, jumping in to help our customers win no matter what is also something that sets us massively apart. I hear horror stories about support experiences with other organizations. And people always seem to be amazed at Team Vertica's willingness to jump in or their aptitude for certain technical capabilities or understanding the business. And I think sometimes we take that for granted. But that is the team that we have as Team Vertica. We are incredibly excited about Vertica 10. I think you're going to love the Virtual Big Data Conference this year. I encourage you to tune in. Maybe one other benefit is I know some people were worried about not being able to see different sessions because they were going to overlap with each other well now, even if you can't do it live, you'll be able to do those sessions on demand. Please enjoy the Vertica Big Data Conference here in 2020. Please you and your families and your co-workers be safe during these times. I know we will get through it. And analytics is probably going to help with a lot of that and we already know it is helping in many different ways. So believe in the data, believe in data's ability to change the world for the better. And thank you for your time. And with that, I am delighted to now introduce Micro Focus CEO Stephen Murdoch to the Vertica Big Data Virtual Conference. Thank you Stephen. >> Stephen: Hi, everyone, my name is Stephen Murdoch. I have the pleasure and privilege of being the Chief Executive Officer here at Micro Focus. Please let me add my welcome to the Big Data Conference. And also my thanks for your support, as we've had to pivot to this being virtual rather than a physical conference. Its amazing how quickly we all reset to a new normal. I certainly didn't expect to be addressing you from my study. Vertica is an incredibly important part of Micro Focus family. Is key to our goal of trying to enable and help customers become much more data driven across all of their IT operations. Vertica 10 is a huge step forward, we believe. It allows for multi-cloud innovation, genuinely hybrid deployments, begin to leverage machine learning properly in the enterprise, and also allows the opportunity to unify currently siloed lakes of information. We operate in a very noisy, very competitive market, and there are people, who are in that market who can do some of those things. The reason we are so excited about Vertica is we genuinely believe that we are the best at doing all of those things. And that's why we've announced publicly, you're under executing internally, incremental investment into Vertica. That investments targeted at accelerating the roadmaps that already exist. And getting that innovation into your hands faster. This idea is speed is key. It's not a question of if companies have to become data driven organizations, it's a question of when. So that speed now is really important. And that's why we believe that the Big Data Conference gives a great opportunity for you to accelerate your own plans. You will have the opportunity to talk to some of our best architects, some of the best development brains that we have. But more importantly, you'll also get to hear from some of our phenomenal Roth Data customers. You'll hear from Uber, from the Trade Desk, from Philips, and from AT&T, as well as many many others. And just hearing how those customers are using the power of Vertica to accelerate their own, I think is the highlight. And I encourage you to use this opportunity to its full. Let me close by, again saying thank you, we genuinely hope that you get as much from this virtual conference as you could have from a physical conference. And we look forward to your engagement, and we look forward to hearing your feedback. With that, thank you very much. >> Joy: Thank you so much, Stephen, for joining us for the Vertica Big Data Conference. Your support and enthusiasm for Vertica is so clear, and it makes a big difference. Now, I'm delighted to introduce Amy Fowler, the VP of Strategy and Solutions for FlashBlade at Pure Storage, who was one of our BDC Platinum Sponsors, and one of our most valued partners. It was a proud moment for me, when we announced Vertica in Eon mode for Pure Storage FlashBlade and we became the first analytics data warehouse that separates compute from storage for on-premise data centers. Thank you so much, Amy, for joining us. Let's get started. >> Amy: Well, thank you, Joy so much for having us. And thank you all for joining us today, virtually, as we may all be. So, as we just heard from Colin Mahony, there are some really interesting trends that are happening right now in the big data analytics market. From the end of the Hadoop hype cycle, to the new cloud reality, and even the opportunity to help the many data science and machine learning projects move from labs to production. So let's talk about these trends in the context of infrastructure. And in particular, look at why a modern storage platform is relevant as organizations take on the challenges and opportunities associated with these trends. The answer is the Hadoop hype cycles left a lot of data in HDFS data lakes, or reservoirs or swamps depending upon the level of the data hygiene. But without the ability to get the value that was promised from Hadoop as a platform rather than a distributed file store. And when we combine that data with the massive volume of data in Cloud Object Storage, we find ourselves with a lot of data and a lot of silos, but without a way to unify that data and find value in it. Now when you look at the infrastructure data lakes are traditionally built on, it is often direct attached storage or data. The approach that Hadoop took when it entered the market was primarily bound by the limits of networking and storage technologies. One gig ethernet and slower spinning disk. But today, those barriers do not exist. And all FlashStorage has fundamentally transformed how data is accessed, managed and leveraged. The need for local data storage for significant volumes of data has been largely mitigated by the performance increases afforded by all Flash. At the same time, organizations can achieve superior economies of scale with that segregation of compute and storage. With compute and storage, you don't always scale in lockstep. Would you want to add an engine to the train every time you add another boxcar? Probably not. But from a Pure Storage perspective, FlashBlade is uniquely architected to allow customers to achieve better resource utilization for compute and storage, while at the same time, reducing complexity that has arisen from the siloed nature of the original big data solutions. The second and equally important recent trend we see is something I'll call cloud reality. The public clouds made a lot of promises and some of those promises were delivered. But cloud economics, especially usage based and elastic scaling, without the control that many companies need to manage the financial impact is causing a lot of issues. In addition, the risk of vendor lock-in from data egress, charges, to integrated software stacks that can't be moved or deployed on-premise is causing a lot of organizations to back off the all the way non-cloud strategy, and move toward hybrid deployments. Which is kind of funny in a way because it wasn't that long ago that there was a lot of talk about no more data centers. And for example, one large retailer, I won't name them, but I'll admit they are my favorites. They several years ago told us they were completely done with on-prem storage infrastructure, because they were going 100% to the cloud. But they just deployed FlashBlade for their data pipelines, because they need predictable performance at scale. And the all cloud TCO just didn't add up. Now, that being said, well, there are certainly challenges with the public cloud. It has also brought some things to the table that we see most organizations wanting. First of all, in a lot of cases applications have been built to leverage object storage platforms like S3. So they need that object protocol, but they may also need it to be fast. And the said object may be oxymoron only a few years ago, and this is an area of the market where Pure and FlashBlade have really taken a leadership position. Second, regardless of where the data is physically stored, organizations want the best elements of a cloud experience. And for us, that means two main things. Number one is simplicity and ease of use. If you need a bunch of storage experts to run the system, that should be considered a bug. The other big one is the consumption model. The ability to pay for what you need when you need it, and seamlessly grow your environment over time totally nondestructively. This is actually pretty huge and something that a lot of vendors try to solve for with finance programs. But no finance program can address the pain of a forklift upgrade, when you need to move to next gen hardware. To scale nondestructively over long periods of time, five to 10 years plus is a crucial architectural decisions need to be made at the outset. Plus, you need the ability to pay as you use it. And we offer something for FlashBlade called Pure as a Service, which delivers exactly that. The third cloud characteristic that many organizations want is the option for hybrid. Even if that is just a DR site in the cloud. In our case, that means supporting appplication of S3, at the AWS. And the final trend, which to me represents the biggest opportunity for all of us, is the need to help the many data science and machine learning projects move from labs to production. This means bringing all the machine learning functions and model training to the data, rather than moving samples or segments of data to separate platforms. As we all know, machine learning needs a ton of data for accuracy. And there is just too much data to retrieve from the cloud for every training job. At the same time, predictive analytics without accuracy is not going to deliver the business advantage that everyone is seeking. You can kind of visualize data analytics as it is traditionally deployed as being on a continuum. With that thing, we've been doing the longest, data warehousing on one end, and AI on the other end. But the way this manifests in most environments is a series of silos that get built up. So data is duplicated across all kinds of bespoke analytics and AI, environments and infrastructure. This creates an expensive and complex environment. So historically, there was no other way to do it because some level of performance is always table stakes. And each of these parts of the data pipeline has a different workload profile. A single platform to deliver on the multi dimensional performances, diverse set of applications required, that didn't exist three years ago. And that's why the application vendors pointed you towards bespoke things like DAS environments that we talked about earlier. And the fact that better options exists today is why we're seeing them move towards supporting this disaggregation of compute and storage. And when it comes to a platform that is a better option, one with a modern architecture that can address the diverse performance requirements of this continuum, and allow organizations to bring a model to the data instead of creating separate silos. That's exactly what FlashBlade is built for. Small files, large files, high throughput, low latency and scale to petabytes in a single namespace. And this is importantly a single rapid space is what we're focused on delivering for our customers. At Pure, we talk about it in the context of modern data experience because at the end of the day, that's what it's really all about. The experience for your teams in your organization. And together Pure Storage and Vertica have delivered that experience to a wide range of customers. From a SaaS analytics company, which uses Vertica on FlashBlade to authenticate the quality of digital media in real time, to a multinational car company, which uses Vertica on FlashBlade to make thousands of decisions per second for autonomous cars, or a healthcare organization, which uses Vertica on FlashBlade to enable healthcare providers to make real time decisions that impact lives. And I'm sure you're all looking forward to hearing from John Yavanovich from AT&T. To hear how he's been doing this with Vertica and FlashBlade as well. He's coming up soon. We have been really excited to build this partnership with Vertica. And we're proud to provide the only on-premise storage platform validated with Vertica Eon Mode. And deliver this modern data experience to our customers together. Thank you all so much for joining us today. >> Joy: Amy, thank you so much for your time and your insights. Modern infrastructure is key to modern analytics, especially as organizations leverage next generation data center architectures, and object storage for their on-premise data centers. Now, I'm delighted to introduce our last speaker in our Vertica Big Data Conference Keynote, John Yovanovich, Director of IT for AT&T. Vertica is so proud to serve AT&T, and especially proud of the harmonious impact we are having in partnership with Pure Storage. John, welcome to the Virtual Vertica BDC. >> John: Thank you joy. It's a pleasure to be here. And I'm excited to go through this presentation today. And in a unique fashion today 'cause as I was thinking through how I wanted to present the partnership that we have formed together between Pure Storage, Vertica and AT&T, I want to emphasize how well we all work together and how these three components have really driven home, my desire for a harmonious to use your word relationship. So, I'm going to move forward here and with. So here, what I'm going to do the theme of today's presentation is the Pure Vertica Symphony live at AT&T. And if anybody is a Westworld fan, you can appreciate the sheet music on the right hand side. What we're going to what I'm going to highlight here is in a musical fashion, is how we at AT&T leverage these technologies to save money to deliver a more efficient platform, and to actually just to make our customers happier overall. So as we look back, and back as early as just maybe a few years ago here at AT&T, I realized that we had many musicians to help the company. Or maybe you might want to call them data scientists, or data analysts. For the theme we'll stay with musicians. None of them were singing or playing from the same hymn book or sheet music. And so what we had was many organizations chasing a similar dream, but not exactly the same dream. And, best way to describe that is and I think with a lot of people this might resonate in your organizations. How many organizations are chasing a customer 360 view in your company? Well, I can tell you that I have at least four in my company. And I'm sure there are many that I don't know of. That is our problem because what we see is a repetitive sourcing of data. We see a repetitive copying of data. And there's just so much money to be spent. This is where I asked Pure Storage and Vertica to help me solve that problem with their technologies. What I also noticed was that there was no coordination between these departments. In fact, if you look here, nobody really wants to play with finance. Sales, marketing and care, sure that you all copied each other's data. But they actually didn't communicate with each other as they were copying the data. So the data became replicated and out of sync. This is a challenge throughout, not just my company, but all companies across the world. And that is, the more we replicate the data, the more problems we have at chasing or conquering the goal of single version of truth. In fact, I kid that I think that AT&T, we actually have adopted the multiple versions of truth, techno theory, which is not where we want to be, but this is where we are. But we are conquering that with the synergies between Pure Storage and Vertica. This is what it leaves us with. And this is where we are challenged and that if each one of our siloed business units had their own stories, their own dedicated stories, and some of them had more money than others so they bought more storage. Some of them anticipating storing more data, and then they really did. Others are running out of space, but can't put anymore because their bodies aren't been replenished. So if you look at it from this side view here, we have a limited amount of compute or fixed compute dedicated to each one of these silos. And that's because of the, wanting to own your own. And the other part is that you are limited or wasting space, depending on where you are in the organization. So there were the synergies aren't just about the data, but actually the compute and the storage. And I wanted to tackle that challenge as well. So I was tackling the data. I was tackling the storage, and I was tackling the compute all at the same time. So my ask across the company was can we just please play together okay. And to do that, I knew that I wasn't going to tackle this by getting everybody in the same room and getting them to agree that we needed one account table, because they will argue about whose account table is the best account table. But I knew that if I brought the account tables together, they would soon see that they had so much redundancy that I can now start retiring data sources. I also knew that if I brought all the compute together, that they would all be happy. But I didn't want them to tackle across tackle each other. And in fact that was one of the things that all business units really enjoy. Is they enjoy the silo of having their own compute, and more or less being able to control their own destiny. Well, Vertica's subclustering allows just that. And this is exactly what I was hoping for, and I'm glad they've brought through. And finally, how did I solve the problem of the single account table? Well when you don't have dedicated storage, and you can separate compute and storage as Vertica in Eon Mode does. And we store the data on FlashBlades, which you see on the left and right hand side, of our container, which I can describe in a moment. Okay, so what we have here, is we have a container full of compute with all the Vertica nodes sitting in the middle. Two loader, we'll call them loader subclusters, sitting on the sides, which are dedicated to just putting data onto the FlashBlades, which is sitting on both ends of the container. Now today, I have two dedicated storage or common dedicated might not be the right word, but two storage racks one on the left one on the right. And I treat them as separate storage racks. They could be one, but i created them separately for disaster recovery purposes, lashing work in case that rack were to go down. But that being said, there's no reason why I'm probably going to add a couple of them here in the future. So I can just have a, say five to 10, petabyte storage, setup, and I'll have my DR in another 'cause the DR shouldn't be in the same container. Okay, but I'll DR outside of this container. So I got them all together, I leveraged subclustering, I leveraged separate and compute. I was able to convince many of my clients that they didn't need their own account table, that they were better off having one. I eliminated, I reduced latency, I reduced our ticketing I reduce our data quality issues AKA ticketing okay. I was able to expand. What is this? As work. I was able to leverage elasticity within this cluster. As you can see, there are racks and racks of compute. We set up what we'll call the fixed capacity that each of the business units needed. And then I'm able to ramp up and release the compute that's necessary for each one of my clients based on their workloads throughout the day. And so while they compute to the right before you see that the instruments have already like, more or less, dedicated themselves towards all those are free for anybody to use. So in essence, what I have, is I have a concert hall with a lot of seats available. So if I want to run a 10 chair Symphony or 80, chairs, Symphony, I'm able to do that. And all the while, I can also do the same with my loader nodes. I can expand my loader nodes, to actually have their own Symphony or write all to themselves and not compete with any other workloads of the other clusters. What does that change for our organization? Well, it really changes the way our database administrators actually do their jobs. This has been a big transformation for them. They have actually become data conductors. Maybe you might even call them composers, which is interesting, because what I've asked them to do is morph into less technology and more workload analysis. And in doing so we're able to write auto-detect scripts, that watch the queues, watch the workloads so that we can help ramp up and trim down the cluster and subclusters as necessary. There has been an exciting transformation for our DBAs, who I need to now classify as something maybe like DCAs. I don't know, I have to work with HR on that. But I think it's an exciting future for their careers. And if we bring it all together, If we bring it all together, and then our clusters, start looking like this. Where everything is moving in harmonious, we have lots of seats open for extra musicians. And we are able to emulate a cloud experience on-prem. And so, I want you to sit back and enjoy the Pure Vertica Symphony live at AT&T. (soft music) >> Joy: Thank you so much, John, for an informative and very creative look at the benefits that AT&T is getting from its Pure Vertica symphony. I do really like the idea of engaging HR to change the title to Data Conductor. That's fantastic. I've always believed that music brings people together. And now it's clear that analytics at AT&T is part of that musical advantage. So, now it's time for a short break. And we'll be back for our breakout sessions, beginning at 12 pm Eastern Daylight Time. We have some really exciting sessions planned later today. And then again, as you can see on Wednesday. Now because all of you are already logged in and listening to this keynote, you already know the steps to continue to participate in the sessions that are listed here and on the previous slide. In addition, everyone received an email yesterday, today, and you'll get another one tomorrow, outlining the simple steps to register, login and choose your session. If you have any questions, check out the emails or go to www.vertica.com/bdc2020 for the logistics information. There are a lot of choices and that's always a good thing. Don't worry if you want to attend one or more or can't listen to these live sessions due to your timezone. All the sessions, including the Q&A sections will be available on demand and everyone will have access to the recordings as well as even more pre-recorded sessions that we'll post to the BDC website. Now I do want to leave you with two other important sites. First, our Vertica Academy. Vertica Academy is available to everyone. And there's a variety of very technical, self-paced, on-demand training, virtual instructor-led workshops, and Vertica Essentials Certification. And it's all free. Because we believe that Vertica expertise, helps everyone accelerate their Vertica projects and the advantage that those projects deliver. Now, if you have questions or want to engage with our Vertica engineering team now, we're waiting for you on the Vertica forum. We'll answer any questions or discuss any ideas that you might have. Thank you again for joining the Vertica Big Data Conference Keynote Session. Enjoy the rest of the BDC because there's a lot more to come
SUMMARY :
And he'll share the exciting news And that is the platform, with a very robust ecosystem some of the best development brains that we have. the VP of Strategy and Solutions is causing a lot of organizations to back off the and especially proud of the harmonious impact And that is, the more we replicate the data, Enjoy the rest of the BDC because there's a lot more to come
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Stephen | PERSON | 0.99+ |
Amy Fowler | PERSON | 0.99+ |
Mike | PERSON | 0.99+ |
John Yavanovich | PERSON | 0.99+ |
Amy | PERSON | 0.99+ |
Colin Mahony | PERSON | 0.99+ |
AT&T | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
John Yovanovich | PERSON | 0.99+ |
Vertica | ORGANIZATION | 0.99+ |
Joy King | PERSON | 0.99+ |
Mike Stonebreaker | PERSON | 0.99+ |
John | PERSON | 0.99+ |
May 2018 | DATE | 0.99+ |
100% | QUANTITY | 0.99+ |
Wednesday | DATE | 0.99+ |
Colin | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Vertica Academy | ORGANIZATION | 0.99+ |
five | QUANTITY | 0.99+ |
Joy | PERSON | 0.99+ |
2020 | DATE | 0.99+ |
two | QUANTITY | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
Stephen Murdoch | PERSON | 0.99+ |
Vertica 10 | TITLE | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Philips | ORGANIZATION | 0.99+ |
tomorrow | DATE | 0.99+ |
AT&T. | ORGANIZATION | 0.99+ |
September 2019 | DATE | 0.99+ |
Python | TITLE | 0.99+ |
www.vertica.com/bdc2020 | OTHER | 0.99+ |
One gig | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Second | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
15 minutes | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
Patrick Smith, Pure Storage & Eric Greffier, Cisco | Cisco Live EU Barcelona 2020
>> Announcer: Live from Barcelona, Spain, it's theCUBE! Covering Cisco Live 2020. Brought to you by Cisco and its ecosystem partners. >> Welcome back, this is theCUBE's live coverage of Cisco Live 2020, here in Barcelona. Our third year of the show, over 17,000 in attendance between the Cisco people, their large partner ecosystem, and the customers, I'm Stu Miniman, my cohost for this segment is Dave Vellante. John Furrier's scouring the show for all of the news at the event, and joining us, we have two first time guests on the program, first, sitting to my left is Patrick Smith, who is the field CTO for EMEA with Pure Storage. Sitting to his left is Eric Greffier, who is the managing director of EMEAR specialists with Cisco, so you have a slightly larger region than Patrick does, gentlemen, thanks so much for joining us. >> Patrick: Great to be here. >> All right, so, we know this show, we were talking that broad ecosystem, and of course Cisco in the data center group has very strong storage partnerships, highlighted by their converged infrastructure stacks. I wrote my research many many years ago, Cisco's brilliant job was when they entered the server market, they made sure that that fragmented storage ecosystem, they made partnerships across the board. And of course, when Pure's ascendancy with the flash era made the stack, so helping to paint those data centers orange with your Cisco partnership, so Patrick, give us the update here, 2020, what's interesting and important to know about Pure Storage and Cisco customer base? >> You know, we continue to see significant adoption of FlashStack, our converged infrastructure with Cisco. Driving just great interest and great growth, both for Pure and for Cisco with the UCS platform, and the value that the customers see in FlashStack, bringing together storage, networking and compute together with overall automation of the stack, and that really gives customers fantastic time to value. And that's what they're looking for in this day and age. >> All right, and Eric, what differentiates the partnership with Pure, versus, as you said, you do work with many of the storage companies out there. >> Well, we had a baby together, it was called FlashStack, and it was couple of years ago now, and as you said, I think the key element for us is really to have those CVDs, those Cisco Validated Designs together, and FlashStack was a great addition to our existing partnership at that time, talking about a couple of years ago. And of course, with the flash technology of Pure, we've seen the demand that we'd say going and going, and it has been amazing, amazing trajectory together. >> But talk a little bit more about the CVDs, the different use cases that you're seeing. You don't have to go through all 20, but maybe pick a couple of your favorite children. >> Well, just to make sure that people understand what CVD means, it's Cisco Validated Design, and this is kind of an outcome in the form of a document, which is available for customers and partners, which is the outcome of the partnership from R&D to R&D, which is just telling customers and partners what they need to order and have in it to fit all of this together for a specific business outcome. And the reason why we have multiple CVDs, is we have one CVD per use case. So the more use cases we have together, the more the CVD's precise, and you just have to follow the CVD design principles. Of course, the later swarms, and maybe Patrick can say a word, but we've been of course doing things regarding analytics and AI, because this is a big demand right now, so maybe Patrick, you want to say a word on this. >> Yeah, you guys were first with the AI and bringing AI and storage together with your partnership with Nvidia, so maybe double down on that. >> The FlashBlade was our move into building a storage platform for AI and model analytics, and we've seen tremendous success with that in lots of different verticals. And so with Cisco we launched FlashStack for AI, which brings together FlashBlade networking, and Cisco's fantastic compute platform with capability for considerable scale of Nvidia GPUs. So an in-a-box capability to really deliver fast time to market solutions for the growing world of analytics and modern AI, people want quick insight into the vast amounts of data we have, and so FlashStack for AI is really important for us being able to deliver as part of the Cisco ecosystem, and provide customers with a platform for success. >> What's happening with modernization, generally, but specifically in Europe, obviously Cisco, long history in Europe, Pure, you've got a presence here, good presence, but obviously much newer. Larger proportion, far larger proportion is in North America, so it's a real opportunity for you guys. What are you seeing in terms of modernization of infrastructure, and apps in the European community? >> Modernization I think is particularly important, and it's more and more seen under the guise of digital transformation, because investing in infrastructure just doesn't get the credit that sometimes it deserves. But the big push there is really all around simpler infrastructure, easier management, and the push for automation. Organizations don't want to have large infrastructure support teams who are either installing or managing in a heavy touch way, their environments, and so the push towards automation, not just at the infrastructure layer, but all the way up the stack, is really key. And you know, we were talking earlier, behind us we have the DevNet sessions here, all about how customers of Cisco and by correlation Pure, can really optimize the management to their environment, use technology like Intersight, like Ansible and others, to really minimize the overhead of managing technology, deliver services faster to customers and be more agile, in this always-on world that we live in, there's no time to really add a human to the cycle of managing infrastructure. >> I think we've been very proud over the years because this notion of converged infrastructure, which was, the promise was to simplify and modernize the data centers, before it was like, "Everything needs to get connected to anything," and coming was this notion of a pod, everything converged, "We've done the job for you, mister customer, "just think about adding some pod." This has been the promise for the last 10 years, and we've been very proud, almost to have created this market, but it wouldn't have been possible without the partnership with the storage players, and with Pure, we've been one step further in terms of simplifying things for customers. >> I love the extension you're talking about, because absolutely converged infrastructure was supposed to deliver on that simplicity, and it was, let's think of the entire rack as a unit of how we manage it, but with today's applications, with the speed of change happening in the environment, we've gone beyond human speed, and so therefore if we don't have the automation that you were talking about, we can't keep up with what the business needs to be able to do there. >> Yeah, that's what it's all about, it's the rapid rate of change. Whether it's business services, whether it's supporting developers in the developer environment, more and more our customers are becoming software development organizations, their developers are a key resource, and making them as efficient as possible is really important, so being able to quickly spin up development environments, new environments for developers, using snapshot technology, giving them the latest sets of data to test their applications on, is really central to enabling and empowering the developer. >> You know, you talk about Cisco's play and kind of creation of the converged infrastructure, Mark, and I think that's fair, by the way. Others may claim it, but I think the mantle goes to you. But there were two friction points, or headwinds, that we pointed out early in the day, the first was organizational, the servers team, the storage team, the network team didn't speak together, then the practitioner told us one day, "Look, you want to solve that problem, "put it in and watch what happens." 'Cause if you try to figure out the organization you'll never get there, and that sort of took care of itself. The other was the channel. The channel likes things separate, they can add value, they have this sort of box selling mentality, so I wonder if you could update us on what the mindset is in the channel, and how that's evolved. >> Yeah, it's a great question. I think the channel actually really likes the simplicity of a converged infrastructure to sell, it's a very simple message, and it really empowers the channel to take, to your point about organization, they have the full stack, all in one sellable item, and so they don't have to fight for the different components, it's one consistent unit that they sell as a whole, and so I think it simplifies the channel, and actually, we find that customers are actively seeking out, it's shown by our growth with FlashStack that customers are actually seeking out the channel partners who are selling FlashStack. >> Yeah, and do you think the channel realizes, "Wow, we really do have to go up the stack, "add more value, do things like partner with"? >> Well for most of the partners, they were heavily specialized on storage or compute or network, so for most of them, supporting the converged infrastructure was to be able to put a foot into another market, which was an expansion for them, which was part number one. Part number two, maybe the things that we've been missing, because since the beginning we had APIs around all those platforms. I don't believe in the early days, I'm talking about five years from now, that they got, that they could really really build something upon the converged infrastructure. Now, if you go through the DevNet area here at Cisco Live, you will see that I think this is the time now for them to understand, and really build new services on top of it, so I believe the value for the channel is pretty obvious now, more than ever. >> Well yeah, it's a great point, you don't usually hear converged infrastructure and infrastructure as code in the same conversation, but the maturation of the platforms underneath are bringing things together. >> They really are, in the same way that IT organizations are freeing up more time to focus up the stack on automation and added value, the same is true of the partners. It's interesting the corollary between the two. >> So I have a question on your act two, so what got us here the last 10 years, both firms were disruptors. Cisco came in and disrupted the compute space, it was misunderstood, "Cisco getting into servers, "that'll never work!" "Well, really not getting into servers, "we're changing the game." "Ah, okay," 10 years later. Pure, all-flash, really created some havoc in the industry, injected a ton of flash into the data center, practically drove a truck through the legacy business. Okay, so very successful. What's act two for you guys, what do you envision, disruptors, are you more incrementalists, I'd love to hear your thoughts on that. >> I start, Patrick. Probably for us, phase two is what you heard yesterday morning, I think Liz Anthony did a great speech regarding Cisco Intersight Workload Optimizer, sorry for the name, this is a bit long, but what it means is now we truly connect the infrastructure to the application performance, and the fact that we can place and discuss about converged infrastructure but in the context of what truly matters for customers, which is application, this is the first time ever you're going to see such amount of R&D put into bringing the two worlds together. So this is just the beginning, but I think this was probably for me yesterday one of the most important announcement ever. And by the way, Pure is coming with this announcement, so if you as a customer buy Cisco Intersight Workload Optimizer, you'll get everything you need to know about Pure and if you have to move things around the storage area, you know the tool will be doing it for you. So we are really the two of us in this announcement, so Patrick, if you want to? >> No, I mean as Eric mentioned, Intersight's important for Cisco, it's important for us, we're very proud to be early integrators as a third party into Intersight to allow that simple management, but you know, as you talk about the future, we were viewed as disruptors when we first came to market with flash array, and we consider still ourselves to be disruptors and innovators, and the amount of our revenue that we invest in innovation, in what is a really focused product portfolio, I think is showing benefits, and you've seen the announcements over the last six months or so with FlashArray//C, bringing all the benefits of flash to tier two applications, and just the interest that that has generated is huge. In the world of networking with NVMe, we have a fabric in RoCEv2, just increasing the performance for business applications that will have fantastic implications for things like SAP, time and performance-critical databases, and then what we announced with direct memory with adding SCM as a read cache onto flash array as well. Really giving customers investment protection for what they bought from us already, because they can, as you well know, Evergreen gives customers an asset that continues to appreciate in value, which is completely the opposite. >> And you're both sort of embracing that service consumption model, I mean Cisco's becoming a very large proportion of your business, you guys have announced some actual straight cloud plays, you've built an aray inside of AWS, which is pretty innovative, so. >> Yes, and as well as the cloud play with Cloud Block Store in AWS, there's Pure as a service, which takes that cloud-like consumption model and allows a customer to run it in their own data center without owning the assets, and that's really interesting, because customers have got used to the cloud-like consumption model, and paying as an OpEx rather than CapEx, and so bringing that into their own facility, and only paying for the data you have written, really does change the game in terms of how they consume and think about their storage environments. >> Patrick, we'd just love to get your viewpoint, you've been talking to a lot of customers this week, you said you've been checking out the DevNet zone, for people that didn't make it to the show here, what have they been missing, what would their peers be telling them in the hallway conversations? >> There's a huge amount as we've been talking about, there's a huge amount on automation, and actually we see it as we go into customers, the number of people we're now talking to who are developers but not developers developing business applications but developers developing code for managing infrastructure is key, and you see it all around the DevNet zone. And then, the focus on containers, I've been talking about it for a long time, and containers is so important for enterprises going forward. We have a great play in that space, and I think as we roll forward, the next three to five years, containers is just going to be the important technology that will be prevalent across enterprises large and small. >> Dave: Yeah, we agree. >> Eric and Patrick, thank you so much for giving us the update, congratulations on all the progress and definitely look forward to keeping an eye on your progress. >> Thanks very much. >> All right, Dave Vellante and I will be back with much more here from Cisco Live 2020 in Barcelona, thanks for watching theCUBE. (techno music)
SUMMARY :
Brought to you by Cisco and its ecosystem partners. and the customers, I'm Stu Miniman, and of course Cisco in the data center group and the value that the customers see in FlashStack, with Pure, versus, as you said, and as you said, I think the key element for us the different use cases that you're seeing. the more the CVD's precise, and you just have to follow and bringing AI and storage together and we've seen tremendous success with that and apps in the European community? and so the push towards automation, the data centers, before it was like, the automation that you were talking about, in the developer environment, and kind of creation of the converged infrastructure, the channel to take, to your point about organization, because since the beginning we had APIs and infrastructure as code in the same conversation, They really are, in the same way Cisco came in and disrupted the compute space, and the fact that we can place and discuss and just the interest that that has generated is huge. you guys have announced some actual straight cloud plays, and only paying for the data you have written, the next three to five years, Eric and Patrick, thank you so much with much more here from Cisco Live 2020 in Barcelona,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Eric Greffier | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Patrick | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Europe | LOCATION | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
Eric | PERSON | 0.99+ |
Barcelona | LOCATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Liz Anthony | PERSON | 0.99+ |
Patrick Smith | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
North America | LOCATION | 0.99+ |
two | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
Evergreen | ORGANIZATION | 0.99+ |
Intersight | ORGANIZATION | 0.99+ |
2020 | DATE | 0.99+ |
yesterday | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
yesterday morning | DATE | 0.99+ |
third year | QUANTITY | 0.99+ |
FlashStack | TITLE | 0.99+ |
first | QUANTITY | 0.99+ |
10 years later | DATE | 0.99+ |
Barcelona, Spain | LOCATION | 0.99+ |
EMEAR | ORGANIZATION | 0.98+ |
20 | QUANTITY | 0.98+ |
both firms | QUANTITY | 0.98+ |
couple of years ago | DATE | 0.98+ |
Ansible | ORGANIZATION | 0.98+ |
Cloud Block Store | TITLE | 0.98+ |
first time | QUANTITY | 0.98+ |
Pure | ORGANIZATION | 0.98+ |
DevNet | TITLE | 0.97+ |
this week | DATE | 0.97+ |
today | DATE | 0.96+ |
two friction points | QUANTITY | 0.96+ |
two worlds | QUANTITY | 0.96+ |
Mark | PERSON | 0.96+ |
EMEA | ORGANIZATION | 0.96+ |
Pure Storage | ORGANIZATION | 0.96+ |
both | QUANTITY | 0.95+ |
over 17,000 | QUANTITY | 0.95+ |
theCUBE | ORGANIZATION | 0.94+ |
one | QUANTITY | 0.92+ |
Rob Lee & Rob Walters, Pure Storage | AWS re:Invent 2019
>> Voiceover: Live, from Las Vegas it's theCUBE Covering AWS re:Invent 2019. Brought to you by Amazon Web Services and Intel, along with its ecosystem partners. >> We're back at AWS re:Invent, this is theCUBE, the leader in live tech coverage. I'm Dave Vellante with my co-host, Justin Warren. This is day one of AWS re:Invent. Rob Lee is here, he's the Vice President and Chief Architect at Pure Storage. And he's joined by Rob Walters, who is the Vice President, General Manager of Storage as a Service at Pure. Robs, welcome to theCUBE. >> Thanks for having us back. >> Yep, thank you. >> Dave: You're welcome. Rob, we'll start with, Rob Lee we'll start with you. So re:Invent, this is the eighth re:Invent, I think the seventh for theCUBE, what's happened at the show, any key takeaways? >> Yeah, absolutely it's great to be back. We were here last year obviously big launch of cloud data services, so it's great to be back a year in. And just kind of reflect back on how the year's gone for uptick at cloud data services, our native US. And it's been a banner year. So we saw over the last year CloudSnap go GA Cloud Block Store go GA and you know just really good customer uptake, adoption and kind of interest out of the gate. So it's kind of great to be back. Great to kind of share what we've down over the last year as well as just get some feedback and more interest from future customers and prospects as well. >> So Rob W, with your background in the cloud what's you take on this notion of storage as a service? How do you guys think about that and how do you look at that? >> Sure, well this is an ever more increasingly important way to consume storage. I mean we're seeing customers who've been you know got used to the model, the economic model, the as a service model in the cloud, now looking to get those benefits on-prem and in the hybrid cloud too. Which if you know, you look at our portfolio we have both there, as part of the Pure as a service. >> Right okay, and then so Pure Accelerate you guys announced Cloud Block Store. >> Yeah, that's when we took it GA. Right so we've been working with customers in a protracted beta process over the last year to really refine the fit and use cases for tier one block workloads and so we took that GA in Accelerate. >> So this is an interesting, you're a partner obviously with Amazon I would think many parts of Amazon love Cloud Block Store 'cause you're using EC2, you're front-ending S3 like you're helping Amazon sell services and you're delivering a higher level of availability and performance in certain workloads, relative to EVS. So there's probably certain guys at Amazon that aren't so friendly with you. So that's an interesting dynamic, but talk about the positioning of Cloud Block Store. Any sort of updates on uptake? What are customers excited about? What can you share? >> Yeah, no absolutely You know I'd say primarily we're most pleased with the variety of workloads and use cases that customers are bringing us into. I think when we started out on this journey we saw tremendous promise for the technology to really improve the AWS Echo system and customer experience for people that wanted to consume block storage in the cloud. What we learned as we started working with customers is that because of the way we've architected the product brought a lot of the same capabilities we deliver on our flash arrays today into AWS, it's allowed customers to take us into all the same types of workloads that they put flash arrays into. So that's their tier one mission critical environments, their VMware workloads, their Oracle workloads, their SAP workloads. They're also looking at us from everything from to do lift and shift, test and dev in the cloud, as well as DR right, and that again I think speaks to a couple things. It speaks to the durability, the higher level of service that we're able to deliver in AWS, but also the compatibility with which we're able to deliver the same sets of features and have it operate in exactly the same way on-prem and in the cloud. 'Cause look, if you're going to DR the last time, the last point in time you want to discover that there's a caveat, hey this feature doesn't quite work the way you expect is when you have a DR failover. And so the fact that we set out with this mission in mind to create that exact level of sameness, you know it's really paying dividends in the types of use cases that customers are bringing us into. >> So you guys obviously a big partner of VMware, you're done very well in that community. So VMware cloud on AWS, is that a tailwind for you guys or can you take advantage of that at this point? >> Yeah no, so I think the way I look at it is both VMware, Pure, AWS, I think we're all responding to the same market demands and customer needs. Which at the end of the day is, look if I'm an enterprise customer the reality is, I'm going to have some of my workloads running on-premise, I'm going to have some of my workloads running in the cloud, I expect you the vendors to help me manage this diverse, hybrid environment. And what I'd say is, there are puts and takes how the different vendors are going about it but at the end of the day that's the customer need. And so you know we're going about this through a very targeted storage-centric approach because that's where we provide service today. You know and you see VMware going after it from the kind of application, hypervisor kind of virtualization end of things. Over time we've had a great partnership with VMware on-premise, and as both Cloud Block Store and VMware Cloud mature, we'd look to replicate the same motion with them in that offering. >> Yeah, I mean to to extent I mean you think about VMware moving workloads with their customers into the cloud, more mission critical stuff comes into the cloud, it's been hard to get a lot of those workloads in to date and that's maybe the next wave of cloud. Rob W., I have a question for you. You know Amazon's been kind of sleepy in storage over the, S3, EBS, okay great. They dropped a bunch of announcements this year and so it seems like there's more action now in the cloud. What's your sort of point of view as to how you make that an opportunity for Pure? >> The way I've always looked at it is, there's been a way of getting your storage done and delivered on AWS and there's been the way that enterprises have done things on-premise. And I think that was a sort of a longer term bet from AWS that that was the way things will tend to fall towards into the public cloud. And now we see, all of the hyperscalers quite honestly with on-prem, hybrid opportunities. With the like Outpost today, et cetera. The hybrid is a real things, it's not just something people said that couldn't get to the cloud, you know it's a real thing. So I think that actually opens up opportunity from both sides. True enterprise class features that our enterprise class customers are looking for in the cloud through something like CBS are now available. But I think you know at Amazon and other hyperscale are reaching back down into the on-prem environments to help with the onboarding of enterprises up into the cloud >> So the as a service side of things makes life a little bit interesting from my perspective, because that's kind of new for Pure to provide that storage as a service, but also for enterprises as you say, they're used to running things in a particular way so as they move to cloud they're kind of having to adapt and change and yet they don't fully want to. Hybrid is a real thing, there are real workloads that need to perform in a hybrid fashion. So what does that mean for you providing storage as a service, and still to Rob Lee's point, still providing that consistency of experience across the entire product portfolio. 'Cause that's quite an achievement and many other as storage providers haven't actually been able to pull that off. So how do you keep all of those components working coherently together and still provide what customers are actually looking for? >> I think you have to go back to what the basics of what customers are actually looking for. You know they're looking to make smart use of their finances capex potentially moving towards opex, that kind of consumption model is growing in popularity. And I think a lot of enterprises are seeing less and less value in the sort of nuts and bolts storage management of old. And we can provide a lot of that through the as a service offering. So had to look past the management and monitoring. We've always done the Evergreen service subscription, so with software and hardware upgrades. So we're letting their sort of shrinking capex budget and perhaps their limited resources work on the more strategically important elements of their IT strategies, including hybrid-cloud. >> Rob Lee, one of the things we've talked about in the past is AI. I'm interested in sort of the update on the AI workloads . We heard a lot obviously today on the main stage about machine learning, machine intelligence, AI, transformations, how is that going, the whole AI push? You guys were first, really the first storage company to sort of partner up and deliver solutions in that area. Give us the update there. Wow's it going, what are you learning? >> Yeah, so it's going really well. So it continues to be a very strong driver of our flash play business, and again it's really driven by it's a workload that succeeds with very large sums of data, it succeeds when you can push those large sums of data at high speed into modern compute, and rinse and repeat very frequently. And the fourth piece which I think is really helping to propel some of the business there, is you know, as enterprises, as customers get further on into the AI deployment journeys what they're finding is the application space evolves very quickly there. And the ability for infrastructure in general, but storage in particular, because that's where so much data gravity exists to be flexible to adapt to different applications and changing application requirements really helps speed them up. So said another way, if the application set that your data scientists are using today are going to change in six months, you can't really be building your storage infrastructure around a thesis of what that application looks like and then go an replace it in six months. And so that message as customers have been through now the first, first and a half iterations of that and really sort of internalize, hey AI is a space that's rapidly evolving we need infrastructure that can evolve and grow with us, that's helping drive a lot of second looks and a lot of business back to us. And I would actually tie this back to your previous question which is the direction that Amazon have taken with some of their new storage offerings and how that ties into storage as a service. If I step back as a whole, what I'd say is both Amazon and Pure, what we see is there's now a demand really for multiple classes of service for storage, right. Fast is important, it's going to continue to get more and more important, whether it's AI, whether it's low latency transactional databases, or some other workload. So fast always matters, cost always matters. And so you're going to have this stratification, whether it's in the cloud, whether its on flash with SCM, TLC, QLC, you want the benefits of all of those. What you don't want is to have to manage the complexity of tying and stitching all those pieces together yourself, and what you certainly don't want is a procurement model that locks you out or in to one of these tiers, or in one of these locations. And so if you think about it in the long term, and not to put words in the other Rob's mouth, where I think you see us going with Pure as a service is moving to a model that really shifts the conversation with customers to say, look the way you should be transacting with storage vendors, and we're going to lead the charge is class of service, maybe protocol, and that's about it. It's like where do you want this data to exist? How fast do you want it? Where on the price performance curve do you want to be? How do you want it to be protected? And give us room to take care of it from there. >> That's right, that's right. This isn't about the storage array anymore. You know you look at the modern data experience message this is about what do you need from your storage, from a storage attribute perspective rather than a physical hardware perspective and let us worry about the rest. >> Yeah you have to abstract that complexity. You guys have, I mean simple is the reason why you were able to achieve escape velocity along with obviously great product and pretty good management as well. But you'll never sub optimize simplicity to try to turn some knobs. I mean I've learned that following you guys over the years. I mean that's your philosophy. >> No absolutely, and what I'd say is as technology evolves, as the components evolve into this world of multis, multi-protocol, multi-tier, multi-class of service, you know the focus on that simplicity and taking even more if it on becomes ever more important. And that's a place where, getting to your question about AI we help customers implement AI, we also do a lot of AI within our own products in our fleet. That's a place where our AI driven ops really have a place to shine in delivering that kind of best optimization of price, performance, tiers of service, so on, so forth, within the product lines. >> What are you guys seeing at the macro? I mean that to say, you've achieved escape velocity, check. Now you're sort of entering the next chapter of Pure. You're the big share gainer, but obviously growing slower than you had in previous years. Part of that we think is this, part of your fault. You put so much flash into the marketplace. It's given people a lot of headroom. Obviously NaN pricing has been an issue, you guys have addressed that on your calls, but still gaining share much, much more quickly than most. Most folks are shrinking. So what are you seeing at the macro, what are customers telling you in terms of their long term strategy with regard to storage? >> Well, so I'll start, I'll let Rob add in. What I'd say is we see in the macro a shift, a clear shift to flash. We've called the shots since day one, but what I'd say is that's accelerating. And that's accelerating with pricing dynamics, with and you know we talked about a lot of the NaN pricing and all that kind of stuff, but in the macro I think there's a clear realization now that customers want to be on flash. It's just a matter of what's the sensible rate? What's the price kind of curve to get there? And we see a couple meaningful steps. We saw it originally with our flash array line taking out 15K spinning drives, 10K's really falling. With QLC coming online and what we're doing in FlashArray//C the 7200 RPM drive kind of in the enterprise, you know those days are numbered, right. And I think for many customers at this point it's really a matter of, okay how quickly can we get there and when does it make sense to move, as opposed to, does it make sense. In many ways it's really exciting. Because if you think about it, the focus for so long has been in those tier one environments, but in many ways the tier two environments are the ones that could most benefit from a move to flash because a couple things happen there. Because they're considered lower tier, lower cost they tend to spread like bunnies, they tend to be kind of more neglected parts of the environment and so having customers now be able to take a second look at modernizing, consolidating those environments is both helpful from a operational point of view, it's also helpful from the point of view of getting them to be able to make that data useful again. >> I would also say that those exact use cases are perfect candidates for an as a service consumption model because we can actually raise the utilization, actually helping customers manage to a much more utilized set of arrays than the over consumption, under consumption game they're trying to play right now with their annual capex cycles. >> And so how aggressive do you see customers wanting to take advantage of that as a service consumption model? Is it mixed or is it like, we want this? >> There's a lot of customers who are just like we want this and we want it now. We've seen a very good traction and adoption so yeah, it's a surprisingly large, complex enterprise customer adoption as well. >> A lot of enterprise, they've gotten used to the idea of cloud from AWS. They like that model of dealing with things and they want to bring that model of operating on site, because they want cloud everywhere. They don't actually want to transform the cloud into enterprise. >> No, exactly, I mean if I go back 20 plus years to when I was doing hands on IT, the idea that we as a team would let go of any of the widgetry that we are responsible for, never would have happened. But then you've had this parallel path of public cloud experience, and people are like well I don't even need to be doing that anymore. And we get better results. Oh and it's secure as well? And that list just goes on. And so now as you say, the enterprise wants to bring it back on-prem for all of those benefits. >> One of the other things that we've been tracking, and maybe it falls in the category of cloud 2.0 is the sort of new workload forming. And I'll preface it this way, you know the early days, the past decade of cloud infrastructures of service have been about, yeah I'm going to spin up some EC2, I'm going to need some S3, whatever, I need some storage, but today it seems like, there's all this data now and then you're seeing new workloads driven by platforms like Snowflake, Redshift, you know clearly throw in some ML tools like Databricks and it's driving a lot of compute now but it's also driving insights. People are really pulling insights out of that data. I just gave you cloud examples, are you seeing on-prem examples as well, or hybrid examples, and how do you guys fit into that? >> Yeah, no absolutely. I think this is a secular trend that was kicked off by open source and the public cloud. But it certainly affects, I would say, the entire tech landscape. You know a lot of it is just about how applications are built. If you about, think back to the late '80s, early '90s you had large monoliths, you had Oracle, and it did everything, soup to nuts. Your transactional system, your data warehouse, ERP, cool, we got it all. That's not how applications are built anymore. They're built with multiple applications working together. You've got, whether it's Kafka connecting into some scale out analytics database, connected into Cassandra, connected right. It's just the modern way of how applications are built. And so whether that's connecting data between SaaS services in the cloud, whether it's connecting data between multiple different application sets that are running on-prem, we definitely see that trend. And so when you peel back the covers of that, what we see, what we hear from customers as they make that shift, as they try to stand up infrastructure to meet those need, is again the need for flexibility. As multiple applications are sharing data, are handing off data as part of a pipeline or as part of a workflow, it becomes ever more important for the underlying infrastructure, the storage array if you will, to be able to deliver high performance to multiple applications. And so the era of saying, hey look I'm going to design a storage array to be super optimized for Oracle and nothing else like you're only going to solve part of the problem now. And so this is why you see us taking, within Pure the approach that we do with how we optimize performance, whether it's across FlashArray, FlashBlade, or Cloud Block Store. >> Excellent, well guys we got to leave it there. Thanks so much for coming on theCUBE and sharing your thoughts with us. And have a good rest of re:Invent. >> Thanks for having us back >> Dave: All right, pleasure >> Thank you >> All right, keep it right there everybody. We'll be back to wrap day one. Dave Vellante for Justin Warren. You're watching theCUBE from AWS re:Invent 2019. Right back (electronic music)
SUMMARY :
Brought to you by Amazon Web Services and Intel, Rob Lee is here, he's the Vice President So re:Invent, this is the eighth re:Invent, and kind of interest out of the gate. and in the hybrid cloud too. you guys announced Cloud Block Store. and so we took that GA in Accelerate. but talk about the positioning of Cloud Block Store. And so the fact that we set out with this mission in mind So VMware cloud on AWS, is that a tailwind for you guys And so you know we're going about this as to how you make that an opportunity for Pure? that couldn't get to the cloud, you know it's a real thing. So what does that mean for you I think you have to go back to what the basics Wow's it going, what are you learning? Where on the price performance curve do you want to be? this is about what do you need from your storage, I mean I've learned that following you guys over the years. you know the focus on that simplicity So what are you seeing at the macro, are the ones that could most benefit from a move to flash than the over consumption, under consumption game There's a lot of customers who are just like They like that model of dealing with things And so now as you say, the enterprise wants to and maybe it falls in the category of cloud 2.0 And so this is why you see us taking, within Pure and sharing your thoughts with us. We'll be back to wrap day one.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Justin Warren | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Rob Lee | PERSON | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
Rob Walters | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Rob | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Rob W. | PERSON | 0.99+ |
last year | DATE | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Rob W | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Cloud Block Store | TITLE | 0.99+ |
Echo | COMMERCIAL_ITEM | 0.99+ |
Intel | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
15K | QUANTITY | 0.99+ |
US | LOCATION | 0.99+ |
10K | QUANTITY | 0.99+ |
both sides | QUANTITY | 0.99+ |
20 plus years | QUANTITY | 0.99+ |
fourth piece | QUANTITY | 0.99+ |
EBS | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
late '80s | DATE | 0.98+ |
EC2 | TITLE | 0.98+ |
early '90s | DATE | 0.98+ |
CBS | ORGANIZATION | 0.98+ |
GA | LOCATION | 0.97+ |
six months | QUANTITY | 0.97+ |
Robs | PERSON | 0.97+ |
one | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
second look | QUANTITY | 0.96+ |
first and a half | QUANTITY | 0.95+ |
Pure Accelerate | ORGANIZATION | 0.95+ |
S3 | TITLE | 0.95+ |
day one | QUANTITY | 0.95+ |
seventh | QUANTITY | 0.94+ |
FlashBlade | TITLE | 0.94+ |
Outpost | ORGANIZATION | 0.94+ |
Evergreen | ORGANIZATION | 0.93+ |
One | QUANTITY | 0.92+ |
Pure | ORGANIZATION | 0.92+ |
tier two | QUANTITY | 0.92+ |
Matt Kixmoeller, Pure Storage | CUBEConversation, November 2019
(jazzy music) >> From our studios in the heart of Silicon Valley, Palo Alto, California, this is a CUBE conversation. >> Hello and welcome to theCUBE studios in Palo Alto, California for another CUBE conversation, where we go in-depth with thought leaders driving innovation across the tech industry. I'm your host, Peter Burris. Digital business is forcing companies to rethink what data means to them, and that means we have to rethink how we're going to manage, use, and take care of our data. A lot of companies are still thinking that we can use old data practices to solve new data requirements, and that disconnect is causing tension in a lot of businesses. So how do they overcome that gap? How do they modernize their data practices and approaches to ensure that they have the options and the flexibility and the capabilities they need as they drive their businesses forward? To have that conversation, we're joined by Matt Kixmoeller, who's the vice president of strategy at Pure Storage. Matt, welcome back to theCUBE. >> Thanks so much. Glad to be here. >> All right, let's start with the obvious. Give us a quick update on Pure. >> Oh, it's a super fun time at Pure right now. We just rounded our 10th birthday, so a lot of celebration going around at the company, and we're just back from our Accelerate conference, where we launched some new products and had quite a good time in Austin. >> Well, tell us a little bit about what was the big story from the Accelerate conference in Austin? >> Well a couple big things. First off, we announced the GA of our cloud block store product. You know, this is where we really take the core Pure value proposition and bring it to run natively on the public cloud. We GAed on the AWS platform, and we actually also just announced a tech preview on Azure. So that was a big part of it. You know, that product's all about helping customers take their tier one applications and transparently move them to the cloud. >> So I mentioned upfront this notion of an impedance mismatch, a disconnect between the requirements or the drive to use data differently, and that's a major feature in digital business transformation. And traditional practices of how data storage and management is conducted, as you talk to customers, how is that challenge manifesting itself in practical as well as strategic ways? >> Yeah, I mean, if you look at our average customer at Pure, they're in the journey of understanding digital transformation, and it's obvious to say, perhaps, but data's at the core of that. And so let's look at, you know, we do a lot of work, for example, in the audio industry. And you might think, okay, the auto industry, kind of a traditional space. They've been around forever manufacturing big, expensive things. But if you look at a modern car company, number one, they're a software developer. There's an amazing amount of software inside cars. And this is similar with everybody that's in digital business. They're now having to build their own software, get it to market quickly. That's a key part of their differentiation. Number two, they're increasingly IoT companies, and so they're having to learn how to harvest all this data that's coming off of their cars, bring it back to the core, analyze it, use it in real-time, and use it in much post-real time to design the next car and get smarter about how they do their work. And then number three, they're operating huge technology environments to run these platforms, and so they need to drive down their cost of data, their cost of goods, if you will, to be able to operate successfully and have an edge and be able to develop more. >> So build software faster, manage storage more efficiently, and move more rapidly and quickly. >> Absolutely. And then mine all that for insights and do more with that data to build the next product every year, every cycle. >> So what is it about the old practices that don't lend themselves to being able to be more efficient, faster, and more productive in how they deliver systems? >> Well, the problem with storage today, if you look at storage just as a layer within the data center, it's probably the least cloud-like of any part of IT. You know, the cloud model, I don't mean cloud the destination, I mean the operating model, has really been taken well to the virtualization and servers and networking layer, but storage, you still have a land of lots of bespoke infrastructure, dedicated silos for each chunk of data, and a lot of manual management. And so when we talk to our typical customer, they're not doing exciting things with data. They're in the drudgery of running the factory of data down there, spending all their time just keeping it working, and they're horribly inefficient in terms of infrastructure they have to use, because it's so bespoke. You know, the term snowflakes is often used in the cloud world. We've just got a million snowflakes in storage. >> So I've always thought that, well, it's not just what I think, but there's a general recognition that every business organizes itself, institutionalizes its work, establishes value propositions around what it thinks are its core differentiating assets. A digital business, increasingly that's data. But I think what you just said sounds like that in the storage world, the assets remain the devices. They remain the LUNs. They remain the physical things. They remain the administrative practices. And we have to find a way to make more of that go away so we can focus more on the data that's being delivered out of the storage. Have I got that right? >> Absolutely. I mean, I think it's just putting data at the core of the strategy and having people actually build an architecture around it. Today what we see is a lot of people build their data strategy piecemeal by project, not having an opportunity to step back and just really think about it from the core. And, you know, at Pure, one of the things we talked about at Accelerate was our vision that we call the modern data experience, and this is just really rethinking the entire experience of storage, hitting the reset button, and trying to bring the lessons of the cloud to how you manage data in enterprise. >> So let's talk about it. If we think about the modern data experience, give us a couple of kind of highlights of what are the two or three things that you absolutely must do differently? >> Well, the first thing is just cloud everywhere. And again, this is cloud the model, not cloud the place. And so the first thing we do is talk about the lessons of cloud with customers. Standardization of services. Not having bespoke infrastructure. You know, designing a set of tiers of storage and delivering that, and then really working on automation. Standardization and automation so that customers can be self-serviced. It's easy to say, but when we go into most storage environments today, they just don't operate like this, right? It's still very bespoke. And so giving customers the tools to be able to design their storage layers, their tiers, if you will, and deliver those services in an on-demand fashion. >> So one of the things that we've uncovered when we talk to customers is as they try to do more exciting and advanced types of workloads in clouds, and discovering that the range of data services provided by the cloud are not as robust, they're not as numerous, they're not as usable as some of the data services that you historically were able to get out of on-premise technology. Now, you mentioned that you are bringing your core management infrastructure into the cloud. Are you able therefore to provide a more rich and complementary range of data services without undermining or compromising that cloud experience? >> I think the key is that cloud experience, that increasingly you need that cloud experience, and it's not either/or, it's both. And so folks have realized that the cloud isn't a panacea. They can absolutely do their work on-prem with data at a lower cost and larger scale and higher performance. They can leverage the cloud for agility. And what's strategic is to have that bridge that allows them to go back and forth depending on the needs of the project. And so when we say cloud everywhere, that assumes that you're going to want to use things on the cloud, in the cloud, and on-prem, and you need a strategic layer of technology for data that can bridge both sides. That's a key part of what we try to deliver. >> So as you talk to your customers, are they utilizing Pure as a way of, or basically the Pure approach to the modern data experience as a way of getting other elements of IT and other elements of the business to think differently and to use data as the foundation for thinking about IT and digital business differently? >> Absolutely. I mean, I'll give you an example. One of our customers is a manufacturing customer. They run a large SAP instance. They wanted to have more agility in how they develop their SAP application. And so they use Pure on-prem to host that application, but they leverage our cloud block store offering to be able to do test dev in the cloud. And this allows them to easily spin up instances, copy production data to the cloud to be able to do test dev around it. And so it's brought new levels of cloud agility to what was a traditional kind of on-prem app. >> That's a great example. Are there any other types of things beyond just test dev that you can think about where the ability to have the certainty associated with Pure and the flexibility associated with the cloud is changing the way IT's thinking? >> I think another big one is DR. You know, if you look at DR investments, folks don't necessarily want to have a second data center. And so being able to leverage the cloud as the DR site not only reduces the cost of DR, but that data's already there, so it then unlocks test dev and other use cases around the cloud. And so that's a big one we see people interested in around cloud block store. >> Now, Pure, even before the modern data experience, was one of the early talkers or early storage companies to talk about how important multi-cloud is going to be. >> Absolutely. >> How does Pure as a target facilitate the practical reality, the pragmatic reality that large enterprises are going to source cloud services from multiple different providers? >> Yeah, I mean, I think, you know, customers are earlier in their journey right now around cloud, so for them, it's more about hybrid cloud than multi-cloud. Multi-cloud is a place they want to get to eventually. But incumbent upon that means a standardized set of services so that storage can speak and be the same, whether it be on this cloud, on that cloud, or on-prem. And look, there's work to do on both sides of the equation, right? If you look at on-prem storage, tier one block storage, we saw that as a gap in the public cloud, so that's why we brought to market cloud block store. If you look at what most people use in the public cloud, it's object storage. Well, most enterprises don't have an object store on-prem. It's one of the reasons we added an object interface to our FlashBlade product. And so this isn't just about bringing things to the public cloud. It's also about bringing some of the public cloud storage services on-prem and making sure they can connect. >> Obviously Pure is associated with storage devices even though you, modern data experience, and what you did at Accelerate is introducing new service classes into how you're going to engage your customers and how customers can source your expertise in their business. But how is that changing Pure? >> I think you picked up on a really interesting thing there around service classes, because one of the things, you know, from the earliest days of Pure, one of our goals was to deliver on the all-flash data center. You know, we obviously brought out tier one flash products to go after the highest end. But we realized that if we wanted to be able to go after all data across the data center, you needed to be able to serve more than one class of data. And so another big push that we announced at Accelerate this year was a QLC-based flash device, the FlashArray//C. And this allows us to really go after that second tier of larger scale and tier two application data in enterprise, to be able to bring that same all-flash cloud experience to this tier two data. >> So what's next? >> I think a big piece of that is we just announced that, so going after that is a large piece of it. The other thing we're really working on is driving up the level of automation and intelligence within the product line. If you look at the first generation of Pure, it was all about simple, right? You know, we have a SaaS-based management experience with Pure1, and we delivered consumer simplicity to this enterprise storage landscape, which was remarkably refreshing to folks. But if you look at this next generation, customers are looking for more intelligence and automation, and so the way you deliver simple to a more sophisticated customer today is open APIs, smart automation, plugin with the orchestration frameworks they're using. And so we're doing a lot of work not only in our API level and our automation level, but also the behind the scenes with our meta AI engine to understand workload and to make intelligent decisions for the customer without them having to deal with it. >> Matt, well, thank you once again for being on theCUBE. >> Likewise. Thanks, Peter. >> And thanks for joining us for another CUBE conversation. I'm Peter Burris. See you next time. (jazzy music)
SUMMARY :
From our studios in the heart and the capabilities they need Glad to be here. All right, let's start with the obvious. so a lot of celebration going around at the company, and bring it to run natively on the public cloud. or the drive to use data differently, and so they need to drive down their cost of data, and move more rapidly and quickly. and do more with that data to build the next product They're in the drudgery of running that in the storage world, one of the things we talked about at Accelerate that you absolutely must do differently? And so the first thing we do is and discovering that the range of data services that the cloud isn't a panacea. And this allows them to easily spin up instances, and the flexibility associated with the cloud And so being able to leverage the cloud as the DR site Now, Pure, even before the modern data experience, so that storage can speak and be the same, and what you did at Accelerate because one of the things, you know, and so the way you deliver simple See you next time.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Peter Burris | PERSON | 0.99+ |
Matt Kixmoeller | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Peter | PERSON | 0.99+ |
November 2019 | DATE | 0.99+ |
Accelerate | ORGANIZATION | 0.99+ |
Austin | LOCATION | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Matt | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
One | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
both sides | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
second tier | QUANTITY | 0.98+ |
10th birthday | QUANTITY | 0.98+ |
three things | QUANTITY | 0.98+ |
first generation | QUANTITY | 0.98+ |
Accelerate | EVENT | 0.98+ |
second data center | QUANTITY | 0.98+ |
first thing | QUANTITY | 0.97+ |
First | QUANTITY | 0.97+ |
each chunk | QUANTITY | 0.97+ |
Pure | ORGANIZATION | 0.96+ |
more than one class | QUANTITY | 0.96+ |
tier one | QUANTITY | 0.96+ |
CUBE | ORGANIZATION | 0.96+ |
today | DATE | 0.95+ |
tier two | QUANTITY | 0.95+ |
Pure Storage | ORGANIZATION | 0.93+ |
this year | DATE | 0.91+ |
Azure | TITLE | 0.9+ |
Silicon Valley, Palo Alto, California | LOCATION | 0.89+ |
a million snowflakes | QUANTITY | 0.83+ |
tier two application | QUANTITY | 0.82+ |
FlashBlade | TITLE | 0.81+ |
theCUBE | ORGANIZATION | 0.8+ |
Number two | QUANTITY | 0.79+ |
Pure | COMMERCIAL_ITEM | 0.76+ |
Accelerate conference | EVENT | 0.74+ |
number one | QUANTITY | 0.73+ |
number three | QUANTITY | 0.67+ |
Pure | TITLE | 0.64+ |
Pure1 | COMMERCIAL_ITEM | 0.61+ |
block | TITLE | 0.61+ |
FlashArray | ORGANIZATION | 0.56+ |
goals | QUANTITY | 0.49+ |
SAP | TITLE | 0.42+ |
Matt “Kix” Kixmoeller, Pure Storage | Pure Accelerate 2019
>> Announcer: From Austin, Texas, it's theCUBE, covering Pure Storage Accelerate 2019, brought to you by Pure Storage. (air whooshes) >> Welcome to theCUBE's day two coverage of Pure Accelerate 2019 from Austin, Texas. I am Lisa Martin, Dave Vellante is my co-host, and we're pleased to welcome back to theCUBE, here is VP of Strategy Matt Kixmoeller. Kix, welcome back! >> Thank you very much, happy to be here. >> This has been a, being shot out of a cannon. Yesterday and today, lots of news. First of all, happy 10th anniversary to you and Pure. >> Thank you very much, yeah. >> Tremendous amount of innovation, as Tara Lee said yesterday, overnight in 10 years. (laughs) >> It's a really fun time at Pure. Just something about the nostalgia of 10 years gets people, naturally, to start thinking about what the next 10 years are about. And so, there's just a lot of that spirit right now at the company, so it's almost like people are really charging into the second chapter with a lot of energy, so that's cool. >> A lot of energy, I think, all fueled by this massive sea of orange that has descended on Austin. >> Absolutely. >> So, four announcements yesterday. Let's start with Cloud Block Store, what you guys are doing with AWS, and kind of this vision of Pure's cloud strategy. >> Yeah, look, the cloud discussions I've had with customers here at the show have been awesome. And I think more than anything, people have realized that we've really built something very unique with Cloud Block Store, something that doesn't exist anywhere else in the industry right now. And, you know, if you look at kind of other storage vendors over the time, people have certainly taken their storage OSes and put them in the cloud kind of as a test-dev experiment, a way to try things out, but never really thinking, "I want to build something "that runs tier-one applications." And that was our goal from day one. We looked at the Amazon platform and said, they really built EBS, their block offering, as kind of a way to beat boot VMs, but it was really never meant for a way to run mission-critical applications. So they've been very open in partnering with us to say, look, let's bring this capability onto the platform. And we really rearchitected our Purity Operating Environment, and so, the whole lower half of that is really optimized for the AWS services to help customers move tier-one apps to the cloud. >> Was that joint engineering, or was it really mostly Pure doing that work? >> You know, it was Pure engineering in the sense that we wrote the code, but there was a lot of co-architecture work with AWS so we could fundamentally understand the basics of all of their services and how to optimize for it. And one of the big realizations and choices that came out of that was not to base the storage layer of this on EBS, but instead to base it on S3. And if you look at your average cloud customer, they really use S3 as the storage basis for the apps they build on Amazon, and so, S3 is the 11-nines durable storage platform there. And so our whole goal here was, how do you use S3, but still deliver the level of performance you'd expect out of a tier-one block environment? >> Well, when you read the sort of cloud storage press release du jour, you can't really get into the nuance, but if I understand it correctly, you guys essentially have architected, using AWS services, a new class of block storage that runs on AWS, but looks like Pure. >> That's exactly it. >> So you're essentially front-ending cheap S3 storage with high-priority EC2s, you've got some mirroring for rights to give it high availability, and again, it looks like Pure. >> Kix: Yep. >> So you win, 'cause you're making money on the software, (laughs) AWS is selling services, and the customer has a Pure experience. Did we get that right? >> Yeah, and I think the combination, the one-two punch, that's been very interesting for customers is not only what we're doing with Cloud Block Store, but the new Pure as-a-Service offering. And so, Pure as-a-Service is our as-a-service consumption mechanism that allows you to essentially subscribe to or rent Pure arrays from Pure in your data center, but it's a license that can go between on-prem and cloud. And so, imagine you're a customer that is mostly on-prem today, but you have that mandate, "I've got to get to the cloud." You might need more storage, but the last thing you want to do is commit to another three- or five-year purchase of a storage array that just puts off that cloud journey that much longer. So a customer can subscribe to Pure as-a-Service, they'll maybe subscribe to 100 terabytes, and we put an array in their data center right now, but a year from now, they decide they're going to move 50 terabytes to Cloud Block Store in Amazon, that's just a transparent movement, they're already licensed for it. And so that-- >> And there's already, oh, sorry, sorry. >> Kix: No, go ahead. >> There's already customers that are in beta with Cloud Block Store, is that correct? >> Correct, yeah. >> Lisa: Any interesting insights that you can share without giving away secret sauce? >> Oh, absolutely. You know, I think the thing that pleased us the most about the beta was really the divergence of use cases. You know, we created this, but there's always, you create something, and you don't know what people are going to do with it, right? And so, we have this goal of going after tier-one apps. Obviously, there's a lot of people that are just focused on migration, "How do I get the tier-one app from on-prem to cloud?" And so that was what I would say would be the dominant use case. But there were a lot of interested in test-dev type use cases. And really interesting, I think we saw it in both directions. So we saw some customers who wanted to develop their app in the cloud, but then deploy on-prem. We saw the opposite, we saw people that wanted to develop on-prem but then deploy in the scalable infrastructure in the cloud. And so I thought that was quite interesting. >> How much of the impetus to do that offering was hardcore customer demand, "We need this," versus, "Hey, we need to embrace the cloud "and make it a tailwind and not be defensive about it"? >> You know, I think when we looked at what was going to be the buy-in criteria for the storage array of tomorrow, fundamentally, this is it, right? People want on-prem infrastructure that's connected to the cloud and provides them a roadmap or a bridge to the cloud. And I think we've seen a big change in mindset over even the last couple years. I'd say two or three years ago, the mindset from customers was, "I'm all in on cloud." I think we've seen that soften, where they've realized that the cloud is not a panacea, it's usually actually not cheaper or faster, but it is more agile, it is more flexible, and so, a combination of on-prem and cloud is the right answer. And so, what does that mean from a storage platform? Storage is the hard part. And so, I then need a storage architecture that can support both on-prem and cloud and drive commonality, as opposed to having it be totally different architecture. >> Was Outposts at all a catalyst in your thinking on this, or was this happening way before you even saw that? >> No, we started this effort before that, but I think Outposts is a good example, I believe, of how Amazon is just getting serious about saying, look, we can't ask everybody to rearchitect every application for web scale. There are certain apps that it won't make sense to rearchitect. How do we bring those to the cloud in an efficient way? And those are really the types of applications and the first-generation Cloud Block Store is perfect for. You connect your existing on-prem app, move it to the cloud without changing it, and then maybe slowly you rearchitect parts of the application, you evolve it over time, but that's not a gate to going to the cloud anymore. >> I like the way you said it, you thought about what storage is going to look like in the next 10 years. And we've said this a lot, it's the cloud experience, bringing that cloud experience to your data is what storage is going to look like, you know, wherever it lives, is going to look like in the next 10 years. >> Absolutely, and I think the other real mindset shift I think we've seen is how people are thinking about truly running their on-prem environment more like a service. You know, if you look at, the key message that we had at the show here was really the Modern Data Experience, and defining for customers what that meant. And in a lot of ways, I've been in the storage industry for a little while, I think back, 20 years ago, the buzzword was utility storage. I think one of our competitors had that as their slogan sometime in the '90s. >> Yeah, right. >> And the reality, though, is when you talk to most storage teams, they just never did that. They still ran a bunch of arrays on a project-by-project basis, and it didn't look at all like the cloud. And so, now people have learned the lessons from the public cloud and said, "We really need to apply those on-prem "to truly bring our infrastructure together "into much more of a virtual pool, "truly deliver it on demand, abstract consumption "from the back-end infrastructure to give flexibility." And so, that's really what we're trying to deliver with the Modern Storage Experience, is to say, look, let's get out of the world of array-by-array management. If a customer buys 50 or 100 of our arrays, how do they take that pool of arrays and turn it into a block service, turn it into a file service, turn it into an object service for their customers, with real abstractions and real APIs for those services that have nothing to do with the back-end infrastructure? >> Dave: Mm-hm. >> When Charlie talked yesterday, Kix, about the Modern Data Experience, the three S's pop up. >> Kix: Yeah. (clears throat) >> Simple, seamless, sustainable. But as IT is getting more and more complex, and customers are in a multi-cloud environment, not necessarily from a strategic perspective, right, acquisition, et cetera, how does Pure actually take that word, simple, from a marketing concept into reality for your customers? >> Yeah, you know, I think simple is the most underappreciated but biggest differentiator (coughs) that Pure has. I was recalling for someone, you talked to Coz earlier today. I had a conversation about three weeks into the existence of Pure, (coughs) excuse me, with Coz, and we were just debating, I mean, this is before we wrote any code at all, about, what would be Pure's long-term differentiator? And I was kind of like, "Ah, we'll be the flash people, or high-performance, or whatever," and he's like, "No, no, no, we're going to be simple. "We are going to deliver a culture that drives "simplicity into our products, "and that'll be game-changing." And I thought he was a little crazy at the time, but he's absolutely turned out to be right. And if you look over the years, that started with just an appliance experience, a 10-card install, just a really easy environment. But that's manifested itself into every product we create. And it's really hard to reverse-engineer that. It's an engineering discipline thing that you have to build into the DNA of the company. >> Yeah, he kind of shared that with us, Lisa. He was basically, my words, saying, you don't ever want to suboptimize simple to get a little knob turn on performance, because you'll be turning knobs your entire career. There's a lot of storage arrays out there that, it's all about turning the knobs. >> Kix: Yeah, well-- >> If you can't fix it, you feature it. >> Oh, and if you think about really trying to automate something, it's really hard to automate complex stuff. If something's simple, if it's consistent, it plugs into an automation framework. >> You talked about "get your 10X"-- >> Kix: Yeah. >> I think, is that what you said? And an entrepreneur who was very successful once told me, "I look for two things, a large market and a 10X impact." >> Yep. >> So, what is your 10X? >> You know, we have two 10Xs at the show this year. So first was really kind of a 10-year jump in performance. When we first entered, people were used to 10-millisecond latency from disk, and we introduced them to one-millisecond latency. Now, with the shipping in direct memory and bringing SCM into the architecture, we can do 100 microseconds. That's another 10X. And so, it's hard to ignore that. >> Lisa: That's game-changing, as you said yesterday. >> (coughs) Exactly. The other is really around our next product, FlashArray C, which brings flash to tier-two data. And there, it's all about consolidation. Most people have not used flash to fix tier one, but their biggest problem now is tier two. They have less-important applications, but because they haven't optimized that, it's taking up way too much of IT time. And so, FlashArray C is, "How do I go "and basically consolidate 10X consolidation "at that tier-two level to really bring "sanity to tier-two storage?" >> And you've got NAM pricing, we talked to Charlie about this, that it ultimately should be a tailwind for you guys as NAM pricing comes down, as NOR fab capacity's coming online in China to go after the thumb drives, right, so that's going to leave the enterprise for all the traditional flash guys that we know and love. So that should open up new markets for you. Today, if you look at pricing for flash C class storage, if I got it right, I'm guessing $1, $1.50 a gigabyte. You see hybrid still at probably half that, 65, 70 cents. Do you see that compressing over the next, let's call it 18, 24 months? >> Absolutely, I mean, what we can do with this product is really bring out flash at disk prices. And so, if you think about the difference, I mean, what we now have in the product line is two platforms, FlashArray X, optimized for performance, at hundreds of microseconds of latency, but C, at a little bit slower performance, still in the millisecond range, can really get down now to those disk prices you just mentioned. And so, it fundamentally gives customers the chance to ask, "Can I really now eliminate disk from the data center?" You know, as I said in my keynote, that the slogan from Pure from day one has been "the all-flash data center." And 10 years ago, people didn't believe it. We were maybe leaning over our skis a little bit in doing that. It now really feels possible to go and have the all-flash data center. >> Well, I'll tell you, we believed it. David Floyer picked up on it early on, and he was-- >> Kix: Yeah. >> He was actually probably too aggressive with (laughs) his forecast. We missed the NAND supply constraints. >> Kix: Yeah. >> But now that seems to be loosening up. >> Well, and, look, one of the things that really helps us build the perfect product around QLC is the work we've done to integrate with raw flash. We cannot just use QLC, but we can use it really efficiently, and the challenge there is to make it reliable. It's inherently a less-reliable flash. And so, that's what we're good at, taking things that are less reliable and making them enterprise-grade. >> And your custom flash modules allow that? >> Yeah. >> Can you add some color to that? >> Basically, what we do is we source raw NANDs, put it in our system, but then do all the work in software to manage the flash. And so, when you have a less-reliable flash medium like QLC, generally, what you have to do is add more flash to overprovision and be careful writing to it. And so, when do it globally, we don't do it inside every SSD, we can do it across the whole system, which makes the whole thing more efficient, thus allowing us to drive costs down even more. >> Hm. >> One of the things that we have heard over the last day and a half from customers, even those that were onstage yesterday, those that were on theCUBE yesterday and those that will come on today, is, they talk about the customer experience. They don't talk about FlashBlade, FlashArray, they're not talking about product names. They're talking about maybe workloads that they're running on there. But the interesting thing is, when we go to some other shows, you hear a lot of names of boxes. >> Kix: Yep. >> We haven't heard that. Talk to me a little bit about how Pure has evolved and really maybe even created this customer experience that's focused on simplicity, on outcomes, that is, in your perspective, why people aren't talking about the specific technologies-- >> Kix: Yeah. >> But rather, this single pane of glass that they have. >> Look, when we started the company, I obviously talked to a lot of customers, and I found, in general, there was frustration with products, but they also just generally didn't like their storage company. And so, from day one, we said, how do we reinvent the experience? Of course, we have to build a better product, and we can use flash as kind of an excuse to do that, but we also want to work on the business model of storage, and we also want to work on the customer experience, the support experience, the just 360 view of how you deal with a vendor. And so, from day one, we've been very disciplined about all of that. Going all-flash was a key part of the product. Evergreen has probably been our quintessential investment in just, how do you change that buying cycle? And so, you can buy into an experience and nondisrupt the way they evolve, versus replace your storage array every three to five years. And then, I think the overall customer experience just comes from the culture of the company, right? Everybody at Pure is centered on making customers happy, doing the right thing, being a vendor that you actually want to work with. And that's not something you can really legislate, that's not something you can put rules around, it's just the culture at Pure. >> When we talked about Evergreen yesterday with a number of customers, including Formula 1. I said, "You know, as a marketer, "how much of that nondisruptive operations, "take me from marketing to reality," and all of them articulated the exact value prop that you guys talk about. It was really remarkable. And another customer that we talked to, I think from a legal firm here in the U.S., didn't even do a POC, talked to a peer of his at another company that was a Pure fan-- >> Kix: Yep. >> And (snaps fingers) bought it right on the spot. So the validation that you're getting from the voice of the customer is pretty remarkable. >> Yeah, this is our number one asset, right? And I mean, so when we think about, how do we spread the religion of Pure, it's just all about giving voice to our customers, so they can share their stories. 'Cause that's so much more credible than anything we say, obviously, as a vendor. >> You're one of only two billion-dollar independent storage companies, which, we love independent storage companies, 'cause, you know, the competition's great. How far out do you look and do you think about being an independent storage company? You've seen, as a "somewhat" historian of the industry, you've seen TAM expansion, you guys are working hard on TAM expansion now, new workloads. You got backup stuff goin' on. You got the cloud as an opportunity, multi-cloud as an opportunity. So you got some runway there. >> Yeah. >> Beyond that, you've seen companies try to vertically integrate, buy backup software companies, you know, a converged infrastructure, whatever it is. How far out do you think about it from a business model standpoint? Or do you not worry about that? >> You know, look, to put it in context a little bit, you look at the latest IDC numbers, we're maybe one-third in to the transition to flash, right? The world still buys two-thirds disk, one-third flash. That's a huge opportunity. We're now five or six globally in storage. That's a few spots that we have to go, right? And so, we're not at all market-share limited, or opportunity limited, even within the storage industry, so we could make a much, much larger company. And so, that's mission number one at Pure. But when we think beyond that, that's just a launching point. And so, you've seen us do some stuff here at the show where we're getting into different types of storage. The first obvious expansion is, let's make sure anything that is a storage product comes from Pure, and there's obvious categories we don't play in today. You saw us introduce a new product around VM Analytics Pro, where we're reaching up the stack and adding real value at the VM tier, taking our Meta AI technology and using to give VM-level optimization recommendations. And so, yeah, I think we increasingly understand that IT's a full-stack game, and so storage is maybe the hardest part of the stack, and that gives us a great base to work from, but we don't constrain our engineers to say, you can only solve storage problems. >> Geography's another upside for you. I mean, most of your business, the vast majority of your business, is in the U.S., whereas you take a company like some of these other ones around here, more than half their business is outside the U.S, so. >> Yeah, no, our international businesses, we've been international five or six years now, and it felt like the first couple years are investment years, and it took time. But we're really starting to see them grow and take hold, and so, it's great to see the international business grow. And I think Pure as a company is also learning to really think internationally, not just because we want the opportunity, but the largest customers in the world that we now deal with have international operations, and they want to deal with one Pure globally. >> So when you're talking, and maybe this has even happened the last day and a half, with a prospective customer who is still investing a lot on-prem, still not yet gone the route of flash, as you were saying, those numbers speak for themselves. What do you say to them? >> If they're not on flash yet? >> Lisa: Yeah, yeah, to show them the benefits. I mean, what's that conversation like? >> It's rare, to be honest, now to find customers who haven't started with flash. But I think the biggest thing I try to encourage folks is that flash is not just about performance. And when I look at the history of people who have embraced Pure, they usually start with some performance need, but very quickly, they realize it's all about simplicity, it's all about efficiency. And if they can make storage fundamentally simpler and more efficient, they free up dollars to put towards innovation. And we unlock the ability to drive dollars towards innovation, and then we drive storage to the new innovation projects, like analytics, like AI, et cetera. And so, we just try to talk about that broader opportunity. And I think that's the hardest thing for people to grasp, because the IT history has always been lots of ROI pitches that say, "Hey, this thing costs a lot, but trust me, "you'll make it up in all these other benefits," that no one believes. And so, you just have to get them to taste it to begin with, and when they see it for themselves, that's when it clicks and they start to really understand the ROI around that. >> Well, congratulations on 10 years of Pure unlocking innovation, not just internally, but externally across the globe. We appreciate your time, Kix. >> Thank you, we're looking forward to the next 10 years. >> All right, to the next 10! For Dave Vellante, I'm Lisa Martin. You're watching theCUBE from Pure Accelerate 2019. (upbeat music)
SUMMARY :
brought to you by Pure Storage. Welcome to theCUBE's to you and Pure. Tremendous amount of innovation, And so, there's just a lot of that spirit sea of orange that has descended what you guys are doing with AWS, of that is really optimized for the AWS services And if you look at your average cloud customer, but if I understand it correctly, you guys essentially front-ending cheap S3 storage with high-priority EC2s, and the customer has a Pure experience. consumption mechanism that allows you to essentially And there's already, And so that was what I would say And I think we've seen a big change in mindset parts of the application, you evolve it over time, I like the way you said it, you thought about at the show here was really the Modern Data Experience, And the reality, though, is when you talk to most about the Modern Data Experience, the three S's and customers are in a multi-cloud environment, And if you look over the years, Yeah, he kind of shared that with us, Lisa. If you can't fix it, Oh, and if you think about really trying is that what you said? And so, it's hard to ignore that. as you said yesterday. "at that tier-two level to really bring for all the traditional flash guys that we know and love. And so, it fundamentally gives customers the chance to ask, and he was-- We missed the NAND supply constraints. to be loosening up. And so, that's what we're good at, And so, when you have a less-reliable flash medium like QLC, that we have heard over the last day and a half talking about the specific technologies-- But rather, And so, you can buy into an experience And another customer that we talked to, So the validation that you're getting And I mean, so when we think about, You got the cloud as an opportunity, How far out do you think about it and so storage is maybe the hardest part of the stack, the vast majority of your business, is in the U.S., and so, it's great to see the international business grow. the last day and a half, with a prospective customer to show them the benefits. And I think that's the hardest thing for people to grasp, but externally across the globe. All right, to the next 10!
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Tara Lee | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Evergreen | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
David Floyer | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Charlie | PERSON | 0.99+ |
50 terabytes | QUANTITY | 0.99+ |
China | LOCATION | 0.99+ |
three | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
50 | QUANTITY | 0.99+ |
one-millisecond | QUANTITY | 0.99+ |
U.S. | LOCATION | 0.99+ |
S3 | TITLE | 0.99+ |
100 microseconds | QUANTITY | 0.99+ |
10-millisecond | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
Cloud Block Store | TITLE | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
second chapter | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
two platforms | QUANTITY | 0.99+ |
six years | QUANTITY | 0.99+ |
10-year | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
Pure | ORGANIZATION | 0.99+ |
Matt Kixmoeller | PERSON | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
six | QUANTITY | 0.99+ |
10-card | QUANTITY | 0.99+ |
five-year | QUANTITY | 0.99+ |
Kix | PERSON | 0.99+ |
100 terabytes | QUANTITY | 0.99+ |
five years | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Austin | LOCATION | 0.99+ |
Yesterday | DATE | 0.99+ |
two | QUANTITY | 0.99+ |
first couple years | QUANTITY | 0.99+ |
two things | QUANTITY | 0.99+ |
TAM | ORGANIZATION | 0.99+ |
U.S | LOCATION | 0.98+ |
first | QUANTITY | 0.98+ |
11-nines | QUANTITY | 0.98+ |
10th anniversary | QUANTITY | 0.98+ |
20 years ago | DATE | 0.98+ |
first-generation | QUANTITY | 0.98+ |
more than half | QUANTITY | 0.98+ |
EBS | ORGANIZATION | 0.98+ |
one-third | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
two | DATE | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
John Colgrove, Pure Storage | Pure Accelerate 2019
(upbeat music) >> Narrator: From Austin, Texas it's theCUBE, covering Pure Storage Accelerate 2019. Brought to you by Pure Storage. >> Welcome back to theCUBE. Lisa Martin, Dave Vellante is my co-host. I'm at Pure accelerate 2019 in Austin, Texas and Dave and I are really pleased to welcome to theCUBE, for the first time, John Colgrove, Coz, CTO and founder of Pure Storage, Coz, welcome to theCUBE. >> Ah, I'm glad to be here thanks for having me. >> And happy 10th anniversary. So, 10 years ago I'm sure you couldn't have envisioned 3, 000 people, Austin being taken over by a sea of orange. But let's go back 10 years, Why did you found Pure Storage? >> Why did I found it? Well, I wasn't really ready to be retired yet. Flash, I started have seen from when I worked at Amdahl many years ago all the way through Veritas, I saw disks continuing to get bigger and bigger and effectively slower and slower. Cause when they don't get any faster and they get bigger, they get slower from by their data. And flash was a catalyst that was going to to change that. But it was the catalyst. What we really wanted to do was to completely change the storage industry. Everything that had annoyed me about the storage industries through all the years in Veritas, All the complexity, all the bad customer practices that the industry forced on people, I wanted to change all that. think of what you demand from your personal tech from your iphone or your laptop or your tablet. Customers should demand that kind of quality, service ability, ease of use from their enterprise IT gear. >> When I started my career in the early '80s I was at IDC and they didn't have a storage analyst. And I started following mainframes and I learned a lot about channel command words and IO subsystems and I came to the conclusion that this is a really hard thing, hard problem to solve. And, so, I got interested in it. You obviously did as well. I'm interested in when you went from Amdahl to Veritas, you had to do some unnatural acts with software to make IO better, 'cause of the spinning disk and understanding the latencies and the scatty chatty protocols and everything else. When you went and thought about Pure and when you think about great architects and I've obviously put you in that category, you chose flash, others like another great architect, Moshe have said you know what I can even squeeze more out of spinning disks. What led you to flash versus trying to squeeze more blood from the spinning disk stone? If I can phrase it that way. >> I think I tend to be more of an extremist on things like that. And I think that's been the key to Pure's success. We were not the first all flash startup. We were the first to focus on affordable flash. Right, if you're going to change the world you have to make something for everyone not for an elite few. But the other thing was we were all flash. There were a lot of other startups that were hybrids that were squeezing more out of the disk and we just went all flash from the beginning. Everything about us is all flash. So, as the future goes more and more towards all flash, we're in a stronger and stronger position. >> And you think that was the game changer that led Pure to be that unicorn that IPO'd four years ago versus those other startups who are trying to do similar things with flash? >> So, that focus helped us a lot with that. The biggest thing that, as I said before flash was a catalyst. The biggest thing we brought to the industry is the simplicity and the evergreen business model. And it's really cool to see all the big companies that we've competed against all these years mimicking a lot of that, but that's the differentiator. Flash was the catalyst that lets you do that. >> Well, so, I'm interested as a little bit of an industry historian and some of the factors that led to your ability to achieve escape velocity which used to be defined as an IPO. I mean, I would argue the 3PAR achieved its escape velocity, I was a $250 million company before it got acquired for 2.5 billion or whatever it was, never reached a billion never even came close. You were the first storage company since NetApp to achieve billion dollar revenue. And you're well on your way to 2 billion, you'll do probably 1.7 this year. In addition to what you've said are there other factors that we should consider in our B school case study on Pure? >> I think one of the things we've tried to do is we've tried to build a company that's going to be in it for that long term. So, we never wanted to settle for an acquisition. We want to build a long term enduring great brand and part of that you have to build more of a partnership with your customers. You have to be a good partner to your partners. Right, if you are short-term focused if you try to squeeze every dollar you can out of people, they don't like you, they don't want to come back. If you build something great and you partner well with the environment around you you can build something long lasting. And we wanted to do that from the beginning, we focused a lot on culture and things like that to help us do that. >> Well it's impressive, congratulations are in order, 'cause 3PAR couldn't do it, Compelling couldn't do it, Isilon, on and on and on. And and EMC at the time was really about EMC that's how you went after. They were able to do virtualization and freeze the market on 3PAR . They were able to do a low cost call it the compellant killer. They were never able to figure out, now maybe they got distracted with elliot management and everything else, but they were never able to figure out how to squash you guys. And that's impressive that you're able to live through that. >> Well, thanks. I mean one of the things we've always tried to do is be supremely disruptive, and that does make it harder for them. >> So, I got to ask I got to challenge you on a couple of things that have come out largely from your competitors but I want to get your take on it. The first one is scale out how come Pure doesn't scale out? I'll leave it there. I have my own thoughts that I've shared with Lisa but. Two controller design. >> Yeah one thing I'd point out is well, FLashBlade, one of our products, is scale out. Flash array, our first product, is not scale out. Scale out isn't a capability for a customer, it's an architecture in how you build the product. When I scale out I have more complicated software. I have more components. More components lead to more failures. Right, if I have a piece of memory and it's going to fail at a certain annual failure rate and I have 10 pieces of memory, I'm going to fail it 10 times that same rate. So, scale out introduces complexity, it introduces more components. And then you have to say what do you get from it. So, if our customers needed a lot more performance than we're delivering, if they needed a lot more scale than we're delivering in the flash array product, we'd then react to that and go build scale out. Where the flash array sells, we don't see that as a major market need, it's more of a niche. Where FlashBlade sells, then there is much more of a need for that and that's why FlashBlade was scale out from day one. >> Well my correct to that the other thing you get from scale out is non disruptive controller swaps but you've solved that in other ways right? >> You say you get non disruptive controller swaps, I will point out that if you look at these scale out architectures out there there's a set of them that do provide that, but actually the larger set of them don't provide it. Because what they're doing is they're making what they view and what the customer views as one monolithic array built from a set of scale out components. So, in those architectures you can't swap out one part of the scale out, you have to swap out the whole thing. >> The other thing I heard, I love this analogy is you don't really see planes anymore. You see them but you really don't want to fly 'em cause they're old with four engines versus two engines 'cause the two engine planes are so, much more reliable. All right the other question is on proprietary flash modules. You guys have chosen your philosophies, do things that you can't do with just off the shelf components. So, you've gone proprietary and this history there, I mean 3PAR with Custom ASICs but I'd like you to share with us your philosophy on what you're doing there. >> So, kind of, there's a couple dimensions to that. Number one, we have gone with proprietary flash modules but in our flash array, we could plug in off-the-shelf drives any time we want. And in fact today our XR2 line, the lower end models use off the shelf flash and the higher end models use the proprietary. What we get with the proprietary is our own firmware on there. Right, it's the same nanochips, the same nanocontrollers, it's all the same components but it's our firmware. And our firmware only has to support one application, our purity operating system. However the customer reads and writes data into the array, we write it the same way down to the flash. We read it back the same way from the flash. So, by making simpler firmware that only has to solve that one problem, we get better performance out of the flash. We get longer life out of the flash and we get order less that one third of the failures of flash drives. Now the flash drives we were using were already failing, a lot less than disk drives. But we've gotten better than three times the reliability by going to our own flash modules. >> Tiering, your philosophy on tiering. Five, 10 years ago there was a big thing on automated tiering, we're going to put the hot data on the high performance either disk or flash and the slow data on the cheap stuff. Your philosophy on tiering, I think I infer you don't believe in tiering. Why not? Or maybe I don't want to put words in our mouth. >> Well so, tiering is another thing that it adds complexity. So, why do you tier? You tier because you say oh I can't afford all of the better things so, I'm going to layer it in with something that's a little cheaper. If you can get by without tiering that's a better solution it's a simpler solution. >> Simplicity is a theme here. The copy of your acquisition your a file system guru to my knowledge what I've read about them, strong file system. What do you intend to do with that? it's concerned about it forking your existing products. How do you respond? >> So, the compuverde file system, we're going to put that on top of our flash array line and make that a unified architecture where you can support block in file. Compuverde is a very complete file protocol stack. And file protocols are a lot more complex than block protocols. Implementing all of the SMB protocol is not an easy thing it takes a bunch of time. So, it's a way to accelerate that and get a very complete protocol stack for that product. Flash blade will continue on with its own scale out file protocols, file and object protocols independent of that. >> Last question I had is on, there's some criticism that's been laid on you guys on the evergreen. The controller, performance of controller upgrades which I we have not heard, we didn't hear that from customers, we've asked some customers that, but I'd love to get your take on, why is there no guarantee of performance improvements as you go to subsequent controller swap outs. Your thoughts? >> So, what we guarantee is you'll get the like or better. So, you might get a new set of controllers that are perform about the same, you might get one a little better. Generally speaking every time we've done it so far it's moved to better. It doesn't move to radically better, but it moves to the better. So, we are guaranteeing that, it's just a question of how much do you chose to deliver with that. What you're doing is you're keeping the array new. It's not so much about making huge strides in the performance it's about keeping the array new. >> But there's another nuance there that I want to test I mean, just conceptionally it seems to me, because the way you ship software constantly that you're making incremental improvements throughout that three year period. First of all is that an accurate assertion? >> it's actually very accurate. The first time we started really looking at how much better we realized that we had moved the needle on the old gear about, I think it was about 60% up during the time period so, yeah there was sort of a little less gains. >> Okay, so, the proper measurement is okay from what's the performance from day one delta to the controller upgrade? That is more significant versus the controller swap day, whatever and plus one if that makes sense. >> Well, I think both are valid ways to look at it. The biggest thing is the customer doesn't have to migrate and the migrations are the most horrible event in storage. Right it's like moving your house for everyone who has moved, you got to pack everything up. Things could get broken things could get lost, it's just a mess. You don't have to do that and the array just gets bigger, denser, more power efficient it gets better and better over time. And you're on that forever, we are happy to do controller swaps after three years, six years, nine years, 12 years. We will continue to do that as long as customers are paying for that it's our job to keep improving it and to keep making it better. >> We've done a lot of research on array migrations. At a minimum, your anti to do a array migration is $50,000. That's what our data shows. We talk to a very large practitioner last night he said, "When I'm doing an array migration I start six "to eight months ahead of time because it takes that long "to do an array migration, array migrations are horrendous "and anything you can do to avoid those is worth it." So, that's all I had that awesome. Thank you for addressing those questions. >> So, the acceleration, pun intended, that Pure has achieved in its first 10 years we talk about customers all the time we've had a number on yesterday from law firms to utilities to F1, we'll have more on today. But in order to achieve what Pure has, you have had to build a culture that's pretty unique. One, this vibrant orange color that just screams energy, boldness too, we're in Austin, Texas, Dell Technology's backyard. Give us a little bit as we wrap here about how you and your co-founders have developed and really fostered this culture of passion that is delivering more than your competitors would like to see. >> Well, so, one of the things that was a key part of the culture is we didn't just hire a bunch of storage people. We had a few early on cause you need some experience in the history but an awful lot of the people we hired came from other backgrounds. Other engineers, marketing people, et cetera, they did not come from storage. And what we challenge people to do when they come in the door is we're hiring them because of their brain power, right. We don't own minimal rights somewhere, we don't have buildings we don't have a lot of assets. Our asset is our people and what they can produce. And obviously if you think back, well, when I was the only employee, right, I was doing every job. Ideally everyone we've hired since can do whatever we've hired them to do a lot better than I could do it. And that's a philosophy you want to keep going. Every person in Pure should be focused on using their brains, using their creativity to deliver the most value possible to disrupt things where they can, to always look for how we do things better, and to always be looking to hire better than them. >> So, it kind of gets into the next 10 years. Don't hate me for saying this but in retrospect the first 10 years you had it kind of easy. You caught EMC off guard, you drove a truck through their install base, NetApp miss flash. You guys executed obviously, we talked about that billion dollar company. Next 10 years, a little different. Where's the TAM expansion come from in the next 10 years? It's Multicloud, it's new AI workloads, it's lower cost solutions that get you more of the market, it's partnering with backup. But you got cloud, you got competitors that are starting to figure it out. How do you see the next 10 years to go from beyond where you are and that next pike. >> Well, so, I'll start by saying when you start a company, you dream of success and the first 10 years have been as good as you could possibly have dreamt. So, A, hopefully the next 10 years will continue that way. I think you touched upon one thing is the cloud. People have been through the hype cycle of saying the entire world is going to be cloud, there's only going to be three data centers in the world and it's going to be Amazon, Microsoft and Google. They now understand the cloud is a tool and you need to use it properly. So, one of the focuses we're going to be working on over the next several years is making sure that someone can have their data, their application on prem. They can decide I want to put it in the cloud. Move there seamlessly. Move there as easily as you move from one of your cell phones to the next model. Move from one cloud to another cloud. Move from that cloud back on prem. Whether you want to move the data, the applications, both and get the same kind of service, the same kind of experience. That's going to be a big thing. >> You got a lot of work to do there, but yeah. But there's an opportunity isn't there? >> It's the way everybody wants to run, it's the way everybody should run. Running an IT service to deliver value to your company, value to your organization should not be rocket science. And our job at Pure is to make that accessible to everybody so, everybody can deliver that kind of quality experience to their organization. >> And it's an obvious question but you see that as technically feasible over the next five to 10 years? >> Yeah it is technically feasible. This goes back to one of the things that I was mentioning before with flash as a catalyst. One of the thing flash helps do to make this simpler is it frees you from the geometry constraints of disk. You don't have to care as much. Another thing that's making it possible, is faster networking, right. And better networking. And then again you have all the compute and GPUs and co-processors and things pushing things. As you get to where resources are more plentiful, then you have the ability to trade off some of the I've got to get like every microsecond out of this thing for the simplicity, for that ease of use. And that lets you deliver something better in the long run. Right, if I perfectly tune something I might be able to do a little bit better but I'm not going to be able to keep it in tune and I'm going to spend my whole life retuning it and retuning it and finding it out of sync. Simplicity, that drives so much efficiency. Agility, that drives so, much value. >> Well, Coz, thank you so, much for joining Dave and me on theCUBE this morning from Accelerate day two. You talked about flash being a catalyst that sounds to me like Coz has been one of the major catalysts of Pure's success. Happy 10th anniversary, we look forward to the next 10. >> Thanks a lot and thanks for having me. >> For Coz and Dave Vellante, I am Lisa Martin, you're watching theCUBE from Pure Accelerate, 2019. (techno music)
SUMMARY :
Brought to you by Pure Storage. and Dave and I are really pleased to welcome So, 10 years ago I'm sure you couldn't have envisioned Everything that had annoyed me about the storage industries to Veritas, you had to do some unnatural acts But the other thing was we were all flash. And it's really cool to see all the big companies and some of the factors that led to your ability and part of that you have to build more of a partnership And and EMC at the time I mean one of the things we've always tried to do So, I got to ask I got to challenge you And then you have to say what do you get from it. that if you look at these scale out architectures out there but I'd like you to share with us your philosophy Now the flash drives we were using were already failing, I think I infer you don't believe in tiering. all of the better things so, I'm going to What do you intend to do with that? Implementing all of the SMB protocol is not an easy thing as you go to subsequent controller swap outs. of how much do you chose to deliver with that. because the way you ship software constantly on the old gear about, I think it was about 60% up Okay, so, the proper measurement is okay from and the migrations are the most horrible event in storage. "and anything you can do to avoid those is worth it." about how you and your co-founders have developed of the culture is we didn't just hire a bunch the first 10 years you had it kind of easy. and you need to use it properly. You got a lot of work to do there, but yeah. And our job at Pure is to make that accessible to everybody to make this simpler is it frees you of the major catalysts of Pure's success. For Coz and Dave Vellante, I am Lisa Martin,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amazon | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Lisa Martin | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
10 times | QUANTITY | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
John Colgrove | PERSON | 0.99+ |
$250 million | QUANTITY | 0.99+ |
10 pieces | QUANTITY | 0.99+ |
Amdahl | ORGANIZATION | 0.99+ |
2.5 billion | QUANTITY | 0.99+ |
$50,00 | QUANTITY | 0.99+ |
eight months | QUANTITY | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
12 years | QUANTITY | 0.99+ |
2 billion | QUANTITY | 0.99+ |
nine years | QUANTITY | 0.99+ |
3, 000 people | QUANTITY | 0.99+ |
six years | QUANTITY | 0.99+ |
two engines | QUANTITY | 0.99+ |
Dell Technology | ORGANIZATION | 0.99+ |
first product | QUANTITY | 0.99+ |
six | QUANTITY | 0.99+ |
two engine | QUANTITY | 0.99+ |
EMC | ORGANIZATION | 0.99+ |
last night | DATE | 0.99+ |
Veritas | ORGANIZATION | 0.99+ |
1.7 | QUANTITY | 0.99+ |
first 10 years | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
yesterday | DATE | 0.99+ |
iphone | COMMERCIAL_ITEM | 0.99+ |
2019 | DATE | 0.99+ |
this year | DATE | 0.99+ |
Lisa | PERSON | 0.99+ |
10 years ago | DATE | 0.98+ |
three times | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
billion dollar | QUANTITY | 0.98+ |
3PAR | ORGANIZATION | 0.98+ |
one problem | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
first time | QUANTITY | 0.98+ |
four years ago | DATE | 0.98+ |
one application | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
many years ago | DATE | 0.98+ |
TAM | ORGANIZATION | 0.98+ |
Austin | LOCATION | 0.97+ |
Five | DATE | 0.97+ |
one thing | QUANTITY | 0.97+ |
10th anniversary | QUANTITY | 0.97+ |
First | QUANTITY | 0.97+ |
Coz | PERSON | 0.97+ |
theCUBE | ORGANIZATION | 0.97+ |
NetApp | ORGANIZATION | 0.96+ |
three year | QUANTITY | 0.96+ |
early '80s | DATE | 0.96+ |
three data centers | QUANTITY | 0.96+ |
four engines | QUANTITY | 0.96+ |
about 60% | QUANTITY | 0.95+ |
Two controller | QUANTITY | 0.95+ |
One | QUANTITY | 0.94+ |
10 years | QUANTITY | 0.94+ |
Pure | ORGANIZATION | 0.94+ |
first storage company | QUANTITY | 0.93+ |
Moshe | PERSON | 0.93+ |
three years | QUANTITY | 0.93+ |
CTO | PERSON | 0.93+ |
Tom Koppelman, Cisco & Mike Bundy, Pure Storage | Cisco Live US 2019
>> Live from San Diego, California, it's theCUBE, covering Cisco Live US 2019. Brought to you by Cisco and its ecosystem partners. >> Welcome back to theCUBE. Our coverage of Cisco Live day three is in full effect. I am Lisa Martin with Dave Vellante and we have a couple of guests joining us. We've got Mike Bundy, head of Cisco Strategic Alliances from, guess where? The jacket should give it away, Pure Storage. And Tom Koppelman, the VP of Architecture Sales America for Cisco, hi guys! >> Hi. >> Hi. >> How ya doing? >> Thanks for bringing more brightness to our set. >> Yeah I forgot my sunglasses. >> I know, we're in the buzzy, bright DevNet Zone. We've been here all week. Great event, massive event, my goodness. 28,000 folks or so, Mike let's start with you. Give us a status of the Pure-Cisco relationship, the evolution of that, where you guys are now. What is exciting? >> Sure, so the relationship, it's unbelievable in terms of the amount of synergies and energy we have together. In fact, Tom at Cisco was really involved in the early genesis of this relationship, prior to me joining the company. And, in the last couple years, we've probably doubled in terms of our go-to-market and sell to customers together. So, tremendous growth. Partnership brings a value to us because of the strong heritage that we have from a DevNet tie-in, in terms of all the automation that we have on the platform, so. It's just a tremendous, tremendous, great partnership. >> And Tom, Cisco has a massive partner ecosystem, a lot of choice. What is it about Pure Storage that is providing advantages to Cisco? Where it's helping customers really kind of bridge this gap between hyper-converged, multi-cloud hybrid, all that jazz? >> Right, so Pure was a first mover in terms of flash storage, right. We saw demand from our customers wanting that technology to improve their data center environments. And when we partnered up early, we were able to kind of capture that momentum, right. And, when I think about our go-to-market with Pure, which is really where I kind of focus, there's very little friction in that relationship, right. There's not competitive overlap. There's not things like that. It's technology that the customers want, that they ask for, and a good field go-to-market in leadership on both sides that are willing to invest and get engaged and move the relationship forward. >> So what else are you guys doing besides just the go-to-market partnership because I got a hold of this timeline of Cisco Validated Designs that Pure and Cisco have put out over the last five years, four years. >> Right. >> And there's like 13 milestones on there. So that's roughly three a year. Of course, it started with Pure's IPO. So that's when Cisco said, all right, these guys are real. Start working with them. And in the early days, of course, you started with FlashStack. That was the flagship product. And then VDI, everybody does VDI, analysts are like, yeah, yeah, everybody does VDI. But then it started really accelerating the cadence. So it's more than just go-to-market. What's beneath that go-to-market? >> Yeah, good question. >> You want to? >> You hit the highlights of the CVD's and whatnot. >> I would say that Pure, this is our number one partnership that we have from an alliance perspective. The investment is far exceeding other partnerships we have. So, the amount of product integration that we're doing is tremendous, as you see there. We've focused on ACI and multi data centers the last couple years. We've focusing on AI and machine learning, most recently. And beyond that, we just signed an agreement and have released resell of Cisco SAN switches in the marketplace. It's the resell agreement we've ever done as a company and it just further shows the commitment in resources that we're willing to put into making sure the partnership is successful and continues to grow and evolve. >> And on top of that the investment in Cisco Intersight, in integrating with Cisco Intersight, the management platform, which is very important to us, it just shows commitment of the partnership. >> Let's talk more about that. So, how does that work? What problems is that solving for customers? >> Well, Cisco Intersight is our cloud based management offering for compute and Pure has integrated their storage platform as part of that solution. So allowing customers, whether it's a converged solution, just raw compute, a hyper-converged solution, but allowing them to manage those pools and deliver that via a cloud solution. >> So Pure plugs into the Cisco API. Now you're part of that stack, essentially. So it's transparent to the customer. And, Cisco's management plane takes care of all that. >> That's exactly right, correct, yes. >> Its' a big deal for us because it was the first integration with Intersight from any storage partner that Cisco has, right. So first to market. We want to embrace hyper-convergence, which is a big important priority for Cisco, and also bridging that gap. So as we compete against single vendor stacks, we have the right solution that customers are looking for. And ultimately, that's why it's so important for us. >> Yeah, Pure is big on firsts. First to flash, you just mentioned another first, you were the first with NVMe, before that you were with the evergreen. I mean, you like being first. >> First orange sport coat. >> That's definitely first there. (laughing) >> Let's talk about customer value though. Obviously, that's what it's all about. As we look at, not just the tremendous amount of choice that customer have when it comes to technology partners, but also the amount of data that's being generated, that's growing astronomically. Yet, organizations are getting so little value out of that because they can't extract the insights. What are you guys doing together leveraging the superpowers of AI and machine learning to help customers in any industry search a really, not just monetize that data, but really accelerate their businesses. Tom you're smiling so let's start with you. >> Yeah, so we came out with an AI server, right, our ML 480, and we've integrated that. Pure has invested, we've both invested and done an integration between FlashBlade, and I'll let Mike talk a little about FlashBlade and the value proposition of FlashBlade, but integrated that with our AI server. And our AI server is an Nvidia powered server, so it essentially gives you scale of processing and capabilities to allow you capitalize on all that data so the customers can get the information they need out of that. If you want to take a second on FlashBlade. >> And you know, AI is the buzz. It's the hot two letter acronym in the industry these days. $13 billion infrastructure opportunity, et cetera, et cetera. So, what Pure is really focused on is, data is the new oil of commodities for customers and clients. What we've built is a platform called FlashBlade, an architecture called the Data Hub, that allows you to not have to copy data and move it around and create silos in data warehouses. So, you can much easier execute a data strategy with the Data Hub architecture, using FlashBlade. When you look at machine learning in terms of how you build a data pipeline so that you can then get to quicker results from a business application standpoint with AI. That's what we've built together with Cisco. We're uber, uber, super excited a number of customers already in the last couple months. >> So I'm going to push a little on that, AI server, AI storage, people don't associate storage and server guys with AI. But if I hear you correctly, there's a $13 billion opportunity for workloads. To manage workloads running on your servers and your storage. >> Correct. >> And so you're optimizing them for AI workloads. >> Absolutely, exactly right. >> So you're not necessarily inventing AI. You're providing infrastructure so that people can leverage AI, is that right? >> Yes. >> Yeah, and the same way that we've built APIs together to work with Intersight, we do that in a way that allows our customers to leverage Cafe, other applications that can help build that data pipeline. We build the platform from the infrastructure level, it makes the management easy and we partner with all of the applications at the top end, but also the middleware and that software prepackage layer that connects via APIs to us. So, it's easy, it's agile, it's manageable, it's a cloud-like experience for the customers, right. >> Easy, agile, all awesome but security. Absolutely critical today. What are you guys doing, Tom I'll start with you, how are you guys working together infuse and integrate security into the technology so that from a customer's perspective, those risks dial down. >> So, Cisco is integrating security across all of our product portfolio, right. And, that includes our data center portfolio, all the way through our campus, our WAN, all those portfolios. We continue to look for opportunities to integrate, whether it's dual-factor authentication or things like secure data center where the highly scalable, multi-instance firewall in front of a data center, things like that. So we're definitely looking for areas and angles and opportunities for us to, not only integrate it from a product standpoint, but also ensure that we are talking that story with our customers so that they know they can leverage Cisco for the full architecture from a security standpoint. >> And the same thing on the storage of the data from an encryption perspective, and as the data gets moved or is mobile, that level of security and policy follows it wherever the data is moved. >> So, what should we expect, what's next in the time? What's 14 going to look like? You don't have top give us specifics but are we going to see blockchain CVDs? What should observers think about the partnership going forward? What could we look forward to? >> Yeah, I mean, the adoption of Container capability is tremendous in our customers environment. Cisco has a cloud Container platform available today. We're integrating that into FlashStack in the very near future. Embracing the cloud. Disaster recovery and data protection it's very hot for customers. Improving that experience so that you have faster restoration times, you're able to look at multi-tier strategy that's very easy to manage from a storage perspective, leveraging S3 with Amazon, Azure, et cetera. So, that's a couple things that are on the short term building block together. >> Yeah, I was going to comment on certainly multi cloud and Containers, those would be two of the big ones that I'd hit on, right. And, in the event of multi cloud leveraging, converged and hyper-converged together to better solve a customer's problems. >> So I was going to ask you. So hyper-converged now becomes a bridge to the cloud if, in fact, that's where customers want to go. >> Yes, it can be. >> Absolutely. >> Yeah, it can be, yes. >> Absolutely. >> Well guys thank you so much for joining Dave and me on the program, sharing with us the momentum that the Pure-Cisco relationship has and what excites you for the future. We appreciate your time. >> Thank you. >> Thank guys. >> For Dave Vellante, I'm Lisa Martin, you're watching theCUBE live from Cisco Live San Diego. Thanks for watching. (electronic music)
SUMMARY :
Brought to you by Cisco and its ecosystem partners. And Tom Koppelman, the VP of Architecture Sales more brightness to our set. the evolution of that, where you guys are now. of the amount of synergies and energy we have together. What is it about Pure Storage that is It's technology that the customers want, that they ask for, that Pure and Cisco have put out over the last And in the early days, of course, and it just further shows the commitment in resources it just shows commitment of the partnership. So, how does that work? and deliver that via a cloud solution. So Pure plugs into the Cisco API. the first integration with Intersight from any storage before that you were with the evergreen. That's definitely first there. but also the amount of data that's being generated, about FlashBlade and the value proposition so that you can then get to quicker results So I'm going to push a little on that, You're providing infrastructure so that and the same way that we've built APIs together to work and integrate security into the technology that we are talking that story with our customers And the same thing on the storage of the data Yeah, I mean, the adoption of Container capability is And, in the event of multi cloud leveraging, So hyper-converged now becomes a bridge to the cloud and me on the program, sharing with us the momentum you're watching theCUBE live from Cisco Live San Diego.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Tom Koppelman | PERSON | 0.99+ |
Mike Bundy | PERSON | 0.99+ |
Tom | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
$13 billion | QUANTITY | 0.99+ |
Mike | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
San Diego, California | LOCATION | 0.99+ |
13 milestones | QUANTITY | 0.99+ |
uber | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
Pure | ORGANIZATION | 0.99+ |
28,000 folks | QUANTITY | 0.99+ |
Nvidia | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
First | QUANTITY | 0.99+ |
FlashBlade | TITLE | 0.99+ |
both sides | QUANTITY | 0.98+ |
Intersight | ORGANIZATION | 0.98+ |
San Diego | LOCATION | 0.98+ |
both | QUANTITY | 0.97+ |
two letter | QUANTITY | 0.97+ |
America | ORGANIZATION | 0.97+ |
three a year | QUANTITY | 0.96+ |
first integration | QUANTITY | 0.96+ |
Azure | ORGANIZATION | 0.96+ |
four years | QUANTITY | 0.96+ |
S3 | TITLE | 0.95+ |
ML 480 | COMMERCIAL_ITEM | 0.95+ |
CVD | ORGANIZATION | 0.95+ |
Data Hub | TITLE | 0.95+ |
ACI | ORGANIZATION | 0.93+ |
Cisco Strategic Alliances | ORGANIZATION | 0.93+ |
Pure Storage | ORGANIZATION | 0.9+ |
Cisco Intersight | ORGANIZATION | 0.9+ |
2019 | DATE | 0.9+ |
FlashStack | TITLE | 0.89+ |
firsts | QUANTITY | 0.87+ |
agile | TITLE | 0.87+ |
first mover | QUANTITY | 0.86+ |
last couple years | DATE | 0.83+ |
Live | EVENT | 0.82+ |
DevNet | ORGANIZATION | 0.82+ |
14 | QUANTITY | 0.8+ |
theCUBE | ORGANIZATION | 0.8+ |
FlashStack | ORGANIZATION | 0.79+ |
Katie Colbert, Pure Storage & Kaustubh Das, Cisco | Cisco Live EU 2019
>> Live from Barcelona, Spain, it's theCUBE, covering Cisco Live Europe. Brought to you by Cisco and its ecosystem partners. >> Welcome back to Barcelona, everybody. You're watching theCUBE, the leader in live tech coverage. My name is Dave Vellante. I'm here with my cohost, Stu Miniman. This is day one of Cisco Live Barcelona. Katie Colbert is here. She's the vice president of alliances at Pure Storage, and she's joined by Kaustubh Das, otherwise known as KD, who's the vice president of computing systems at Cisco. Katie and KD, welcome to theCUBE, good to see you. >> Thank you. >> Thank you. >> Alright, so let's start off, KD2, if you could just tell us about the partnership. Where did it start, how did it evolve? We'll get into it. >> We just had a terrific partnership, and the reason it's so great is it's really based on some foundational things that are super compatible. Pure Storage, Cisco, both super technology-driven companies, innovating. They're both also super programmatic companies. They'll do everything via API. It's very modern in that sense, the frameworks that we work on. And then from a business perspective, it's very compatible. We're chasing common markets, very few conflicts. So it's been rooted in solid foundations. And then, we've actually invested over the years to build more and more solutions for our customers jointly. So it's been terrific. >> So, Katie, I hate to admit how long we talk about partnering with Cisco >> It's going to age us. >> So you and I won't admit how many decades it's been partnering with Cisco, but here we are, 2019, Cisco's a very different company than it was a decade or two ago. >> Absolutely. >> Tell me what it's like working with them, especially as a company that's primarily in storage and data at Pure, what it means to partner with them. >> Absolutely, you're right. So, worked with Cisco as a partner for many years at the beginning of my career, then went away for, I'd say, a good 10 years, and joined Pure in June, and I will tell you one of the most exciting reasons why I joined Pure was the Pure and Cisco relationship. When I worked with them at the beginning of my career, it was great and I would tell you it's even better now. I will say that the momentum that these two companies have in the market is very phenomenal. A lot of differentiation from our products separately, but both together, I think that it's absolutely been very successful, and to KD's point, the investment that both companies are making really is just astronomical, and I see that our customers are the beneficiaries of that. It makes it so much easier for them to deploy and use the technologies together, which is exciting. >> So we always joke about Barney deals, I love you, you love me, I mean, it's clear you guys go much much deeper than that. So I want to probe at that a little bit. Particularly from an engineering standpoint, whether it's validated designs or other innovations that you guys are working on together, can we peel the onion on that one a little bit? Talk about what you guys are doing below that line. >> I'll start there then I'll hand it over to the engineering leader from Cisco. But if you think about the pace of this, the partnership, I think, is roughly 3 or so years old. We've 16 Cisco-validated designs for our FlashStack infrastructure. So that is just unbelievable. So, huge amount of investment from engineers, product managers, on both sides of the fence. >> Yeah, totally second that. We start out with the... Cisco-validated designs are like blueprints, so we start out with the blueprints for the standard workloads: Oracle, SAP. And we keep those fresh as new versions come out. But then I think we've taken it further into new spaces of late. ACI, we saw in the keynote this morning, it's going everywhere, it's going multi-site. We've done some work on marrying that with the clustering service of Pure Storage. On top of that, we're doing some work in AI and ML, which is super exciting, so we got some CBDs around that that's just coming out. We're doing some work on automation, coupling Intersight, which is Cisco's cloud-based automation suite, with Pure Storage and Pure Storage's ability to integrate into the Intersight APIs. We talked about it, in fact, I talked about it in my session at the Cisco Live in the summer last year, and now we've got that out as a product. So tremendous amount of work, both in traditional areas as well as some of these new spaces. >> Maybe we can unpack that Intersight piece a bit, because people might look at it initially and say, "Okay, multi-cloud, on-prem, all these environments, "but is this just a networking tool?" And working we're working with someone with Pure, maybe explain a little bit the scope and how, if I'm a Pure administrator, how I live into this world. >> Absolutely, so let's start with what is Intersight, just for a foundational thing. Intersight is our software management tool driven from the cloud. So everything from the personality of the server, the bios settings, the WLAN settings, the networking and the compute pieces of it, that gets administered from the cloud, but it does more. What it does is it can deliver playbooks from the cloud that give the server a certain kind of personality for the workload that it's supporting. So then the next question that anyone asks is, "Now that we have this partnership, "well can it do the same thing for storage? "Can it actually provision that storage, "get that up and running?" And the answer is yes, it can, but it's better because what it can not only do is, not only can it do that, getting that done is super simple. All Pure Storage needed to do was to write some of those Intersight APIs and deliver that playbook from the cloud, from a remote location potentially, into whatever your infrastructure is, provisioning compute, provisioning networking, provisioning storage, in a truly modern cloud-driven environment, right? So I think that's phenomenal what it does for our customers. >> Yeah, I'd agree with that. And I think it'll even become more important as the companies are partnering around our multi-cloud solutions. So, as you probably saw earlier this year in February, sorry, the end of 2018, Pure announced our first leaning into hybrid cloud, so that's Pure Cloud Data Services. That enables us to have Purity, which is our operating system on our storage, running in AWS to begin with. So you can pretty easily start to think about where this partnership is going to go, especially as it pertains to Intersight integration. >> And just to bounce on that, strategically, you can see the alignment there as well. I mean, Cisco's been talking about multi-cloud for a bit now, we've done work to enable similar development environments, whether we're doing something on-prem or in the cloud, so that you can move workloads from one to the other, or actually you can make workloads on both sides talk to each other, and, again, combined with what Katie just said, it makes it a really really compelling solution. >> Like you said, you've got pretty clear swimming lanes for the two companies. There's very little overlap here. You can't have too many of these types of partnerships, right, I mean, you got 25 thousand engineers almost, but still, you still have limited resources. So what makes this one so special, and why are you able to spend so much time and effort, each of you? >> I could start, so from a Pure perspective, I think the cultures are aligned, you called it out there, there's inherently not a lot of overlap in terms of where core competencies are. Pure is not looking at all to become a networking company. And just a lot of synergies in the market make it one that our engineers want to invest in. We have really picked Cisco as our lean-in partner, truthfully, I run all of the alliances at Pure, and a lion's share of my resources really are focused at that partnership. >> Yeah, and if you look at both these companies, Pure is a relative youngster among the storage companies, a new, modern, in a good way, a new, modern company built on modern software practices and so forth. Cisco, although a pretty veteran company, but Cisco compute is relatively new as well as a compute provider. So we are very similar in how our design philosophies work and how modern our infrastructures are, and that gets us to delivering results, delivering solutions to our customers with relatively less effort from our engineers. And that pace of innovation that we can do with Pure is not something we can do with every other company. >> We had a session earlier today, and we went pretty deep into AI, but it's probably worth touching on that. I guess my question here is, what are the customers asking you guys for in terms of AI infrastructure? What's that infrastructure look like that's powering the machine and intelligence era? >> You want to start? >> You want to go, I'll go first. This is a really exciting space, and not only is it exciting because AI is exciting, it's actually exciting because we've got some unique ingredients across Pure and Cisco to make this happen. What does AI feed on? AI feeds on data. The model requires that volume of data to actually train itself We've got an infrastructure, so we just released the C4ATML, the UCC4ATML, highly powered infrastructure, eight GPUs, interconnected, 180 terabytes on board, high network bandwidth, but it needs something to feed it the data, and what Pure's got with their FlashBlade is that ability to actually feed data to this AI infrastructure so that we can train bigger models or train these models faster. Makes for a fantastic solution because these ingredients are just custom made for each other. >> Anything you can add? >> Absolutely I'd agree with that. Really, if you look at AI and what it needs to be successful, and, first of all, all of our customers, if they're not thinking about it, they should be, and I will tell you most of them are, is, how do you ingest that amount of data? If you can't ingest that quickly, it's not going to be of use. So that's a big piece of it, and that's really what the new Cisco platform, I mean, the folks over at Pure are just thrilled about the new Cisco product, and then you take a look at the FlashBlade and how it's able to really scale out unstructured data, object it and file, really to make that useful, so when you have to scrub that data to be able to use it and correlate it, FlashBlade is the perfect solution. So really, this is two companies coming together with the best of breed technologies. >> And the tooling in that world is exploding, open source innovation, it needs a place to run all the Kafkas and the Caffes and the TensorFlows and the Pythons. It's not just confined to data scientists anymore. It's really starting to seep throughout the organization, are you seeing that? >> Yeah. >> What's happening is you've got the buzzwords going around, and that leads to businesses and the leaders of businesses saying, "We've got to have an AI strategy. "We've got to hire these data scientists." But at the same time, the data scientists can get started on the laptop, they can get started on the cloud. When they want to deploy this, they need an enterprise class, resilient, automated infrastructure that fits into the way they do their work. You've got to have something that's built on these components, so what we provide together is that infrastructure for the ITTs so that the data scientists, when they build their beautiful models, have a place to deploy them, have a place to put that into production, and can actually have that life cycle running in a much more smooth production-grade environment. >> Okay, so you guys are three years in, roughly. Where do you want to take this thing, what's the vision? Give us a little road map for the future as to what this partnership looks like down the road. >> Yeah, so I can start. So I think there's a few different vectors. We're going to continue driving the infrastructure for the traditional workloads. That's it, that's a big piece that we do, we continue doing that. We're going to drive a lot more on the automation side, I think there's such a lot of potential with what we've got on Intersight, with the automation that Pure supports, bring those together and really make it simple for our customers to get this up and running and manage that life cycle. And third vector's going to be imparting those new use cases, whether it be AI or more data analytics type use cases. There's a lot of potential that it unleashes for our customers and there's a lot of potential of bringing these technologies together to partner. So you'll see a lot more of that from us. I don't know, will you add something? >> Yeah, no, I absolutely agree. And I would say more FlashStack, look for more FlashStack CVDs, and AI, I think, is one to watch. We believe Cisco, really, this step that Cisco's made, is going to take AI infrastructure to the next level. So we're going to be investing much more heavily into that. And then cloud, from a hybrid cloud, how do these two companies leverage FlashStack and all the innovation we've done on prem together to really enable the multi-cloud. >> Great, alright, well Katie and KD, thanks so much for coming to theCUBE. It was great to have you. >> Great. Thanks for having us. >> Thank you very much. >> You're welcome, alright. Keep it right there everybody. Stu and I will be back with our next guest right after this short break. You're watching theCUBE Live from Cisco Live Barcelona. We'll be right back. (techy music)
SUMMARY :
Brought to you by Cisco and its ecosystem partners. Welcome back to Barcelona, everybody. if you could just tell us about the partnership. and the reason it's so great is it's really based So you and I won't admit how many at Pure, what it means to partner with them. and I see that our customers are the beneficiaries of that. or other innovations that you guys are working on together, I'll start there then I'll hand it over to so we start out with the blueprints maybe explain a little bit the scope and how, and deliver that playbook from the cloud, So you can pretty easily start to think so that you can move workloads from one to the other, and why are you able to spend And just a lot of synergies in the market And that pace of innovation that we can do with Pure what are the customers asking you guys for is that ability to actually feed data and how it's able to really scale out unstructured data, and the TensorFlows and the Pythons. and that leads to businesses and the leaders of businesses as to what this partnership looks like down the road. for our customers to get this up and running and AI, I think, is one to watch. thanks so much for coming to theCUBE. Thanks for having us. Stu and I will be back with our next guest
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Katie | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Katie Colbert | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Kaustubh Das | PERSON | 0.99+ |
KD | PERSON | 0.99+ |
Pure | ORGANIZATION | 0.99+ |
Stu | PERSON | 0.99+ |
June | DATE | 0.99+ |
two companies | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
16 | QUANTITY | 0.99+ |
both companies | QUANTITY | 0.99+ |
three years | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
KD2 | PERSON | 0.99+ |
10 years | QUANTITY | 0.99+ |
180 terabytes | QUANTITY | 0.99+ |
Barcelona | LOCATION | 0.99+ |
ACI | ORGANIZATION | 0.99+ |
C4ATML | COMMERCIAL_ITEM | 0.99+ |
one | QUANTITY | 0.99+ |
25 thousand engineers | QUANTITY | 0.98+ |
both sides | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
Barcelona, Spain | LOCATION | 0.98+ |
FlashBlade | TITLE | 0.98+ |
SAP | ORGANIZATION | 0.98+ |
Barney | ORGANIZATION | 0.98+ |
FlashBlade | COMMERCIAL_ITEM | 0.98+ |
Pure Storage | ORGANIZATION | 0.97+ |
Intersight | ORGANIZATION | 0.96+ |
earlier this year | DATE | 0.96+ |
UCC4ATML | COMMERCIAL_ITEM | 0.95+ |
end of 2018 | DATE | 0.95+ |
Pythons | TITLE | 0.94+ |
eight GPUs | QUANTITY | 0.93+ |
each | QUANTITY | 0.93+ |
second | QUANTITY | 0.92+ |
TensorFlows | TITLE | 0.9+ |
two ago | DATE | 0.9+ |
February | DATE | 0.88+ |
Purity | ORGANIZATION | 0.87+ |
theCUBE | ORGANIZATION | 0.86+ |
Intersight | TITLE | 0.86+ |
Zongjie Diao, Cisco and Mike Bundy, Pure Storage | Cisco Live EU 2019
(bouncy music) >> Live, from Barcelona, Spain, it's theCUBE, covering Cisco Live Europe. Brought to you by Cisco and its ecosystem partners. >> Welcome back everyone. Live here in Barcelona it's theCUBE's exclusive coverage of Cisco Live 2019. I'm John Furrier. Dave Vellante, my co-host for the week, and Stu Miniman, who's also here doing interviews. Our next two guests is Mike Bundy, Senior Director of Global Cisco Alliance with Pure Storage, and Z, who's in charge of product strategy for Cisco. Welcome to theCUBE. Thanks for joining us. >> Thank you for having us here. >> You're welcome. >> Thank you. >> We're in the DevNet zone. It's packed with people learning real use cases, rolling up their sleeves. Talk about the Cisco Pure relationship. How do you guys fit into all this? What's the alliance? >> You want to start? >> Sure. So, we have a partnership with Cisco, primarily around a solution called Flashstack in the converged infrastructure space. And most recently, we've evolved a new use-case and application together for artificial intelligence that Z's business unit have just released a new platform that works with Cisco and NVIDEA to accomplish customer application needs mainly in machine learning but all aspects of artificial intelligence, so. >> So AI is obviously a hot trend in machine learning but today at Cisco, the big story was not about the data center as much anymore as it's the data at the center of the value proposition which spans the on-premises, IoT edge, and multiple clouds so data now is everywhere. You've got to store it. It's going to be stored in the cloud, it's on-premise. So data at the center means a lot of things. You can program with it. It's got to be addressable. It has to be smart and aware and take advantage of the networking. So with all of that as the background, backdrop, what is the AI approach? How should people think about AI in context to storing data, using data? Not just moving packets from point A to point B, but you're storing it, you're pulling it out, you're integrating it into applications. A lot of moving parts there. What's the-- >> Yeah, you got a really good point here. When people think about machine learning, traditionally they just think about training. But we look at it as more than just training. It's the whole data pack line that starts with collecting the data, store the data, analyze the data, train the data, and then deploy it. And then put the data back. So it's really a very, it's a cycle there. It's where you need to consider how you actually collect the data from edge, how you store them, in the speed that you can, and give the data to the training side. So I believe when we work with Pure, we try to create this as a whole data pack line and think about the entire data movement and the storage need that we look at here. >> So we're in the DevNet zone and I'm looking at the machine learning with Python, ML Library, (mumbles) Flow, Appache Spark, a lot of this data science type stuff. >> Yup. >> But increasingly, AI is a workload that's going mainstream. But what are the trends that you guys are seeing in terms of traditional IT's involvement? Is it still sort of AI off on an island? What are you seeing there? >> So I'll take a guess, a stab at it. So really, every major company industry that we work with have AI initiatives. It's the core of the future for their business. What we're trying to do is partner with IT to get ahead of the large infrastructure demands that will come from those smaller, innovative projects that are in pilot mode so that they are a partner to the business and the data scientists rather than a laggard in the business, the way that sometimes the reputation that IT gets. We want to be the infrastructure, solid, like a cloud-like experience for the data scientists so they can worry more about the applications, the data, what it means to the business, and less about the infrastructure. >> Okay. And so you guys are trying to simplify that infrastructure, whether it's converged infrastructure, and other unifying approaches. Are you seeing the shift of that heavy lifting, of people now shifting resources to new workloads like AI? Maybe you could discuss what the trends are there? >> Yeah, absolutely. So I think AI started with more like a data science experiment. You see a couple of data scientists experimenting. Now it's really getting into mainstream. More and more people are into that. And as, I apologize. >> Mike. >> Mike. >> Mike, can we restart that question? (all laughing) My deep apology. I need a GPU or something in my brain. I need to store that data better. >> You're on Fortnite. Go ahead. >> Yes, so as Mike has said earlier on, it's not just the data scientists. It's actually an IT challenge as well and I think with Cisco, what we're trying to do with Pure here is, you know that Cisco thing, we're saying, "We're a bridge." We want to bridge the gap between the data scientists and the IT and make it not just AI as an experiment but AI at scale, at production level, and be ready to actually create real impact with the technology infrastructure that we can enable. >> Mike, talk about Pure's position. You guys have announced Pure in the cloud? >> Yes. >> You're seeing that software focus. Software is the key here. >> Absolutely. >> You're getting into a software model. AI and machine learning, all this we're talking about is software. Data is now available to be addressed and managed in that software life cycle. How is the role of the software for you guys with converged infrastructure at the center of all the Cisco announcements. You were out on stage today with converged infrastructure to the edge. >> Yes, so, if you look at the platform that we built, it's referenced back, being called the Data Hub. The Data Hub has a very tight synergy with all the applications you're referring to: Spark, Tensor Flow, et cetera, et cetera, Cafe. So, we look at it as the next generation analytics and the platform has a super layer on top of all those applications because that's going to really make the integration possible for the data scientists so they can go quicker and faster. What we're trying to do underneath that is use the Data Hub that no matter what the size, whether it's small data, large data, transaction based or more bulk data warehouse type applications, the Data Hub and the FlashBlade solution underneath handle all of that very, very different and probably more optimized and easier than traditional legacy infrastructures. Even traditional, even Flash, from some of our competitors, because we built this purpose-built application for that. Not trying to go backwards in terms of technology. >> So I want to put both you guys on the spot for a question. We hear infrastructure as code going on many, many years since theCUBE started nine years ago. Infrastructure as code, now it's here. The network is programmable, the infrastructure is programmable, storage is programmable. When a customer or someone asks you, how is infrastructure, networks, and storage programmable and what do I do? I used to provision storage, I've got servers. I'm going to the cloud. What do I do? How do I become AI enabled so that I could program the infrastructure? How do you guys answer that question? >> So a lot of that comes to the infrastructure management layer. How do you actually, using policy and using the right infrastructure management to make the right configuration you want. And I think one thing from programmability is also flexibility. Instead of having just a fixed configuration, what we're doing with Pure here is really having that flexibility where you can put Pure storage, different kind of storage with different kind of compute that we have. No matter we're talking about two hour use, four hour use, that kind of compute power is different and can max with different storage, depending on what the customer use case is. So that flexibility driven to the programmability that is managed by the infrastructure management layer. And we're extending that. So Pure and Cisco's infrastructure management actually tying together. It's really single pane of glass within the side that we can actually manage both Pure and Cisco. That's the programmability that we're talking about. >> Your customers get Pure storage, end-to-end manageability? >> With the Cisco compute, it's a single pane of glass. >> Okay. >> So where do I buy? I want to get started. What do you got for me? (laughing) >> It's pretty simple. It's three basic components. Cisco Compute and a platform for machine learning that's powered by NVIDEA GPUs; Cisco FlashBlade, which is the Data Hub and storage component; and then network connectivity from the number one network provider in the world, from Cisco. It's very simple. >> And it's a SKU, it's a solution? >> Yup, it's very simple. It's data-driven. It's not tied to a specific SKU. It's more flexible than that so you have better optimization of the network. You don't buy a 1000 series X and then only use 50% of it. It's very customizable. >> Okay, do I can customize it for my, whatever, data science team or my IT workloads? >> Yes, and provision it for multi-purpose, same way a service provider would if you're a large IT organization. >> Trend around breaking silos has been discussed heavily. Can you talk about multiple clouds, on-premise in cloud and edge all coming together? How should companies think about their data architecture because silos are good for certain things, but to make multi-cloud work and all this end-to-end and intent-based networking and all the power of AI's around the corner, you got to have the data out there and it's got to be horizontally scalable, if you will. How do you break down those silos? What's your advice, is there a use case for an architecture? >> I think it's a classic example of how IT has evolved to not think just silos and be multi-cloud. So what we advocate is to have a data platform that transpires the entire community, whether it's development, test, engineering, production applications, and that runs holistically across the entire organization. That would include on-prem, it would include integration with the cloud because most companies now require that. So you can have different levels of high availability or lower cost if your data needs to be archived. So it's really building and thinking about the data as a platform across the company and not just silos for various applications. >> So replication never goes away. >> Never goes away. (laughing) >> It's going to be around for a long, long time. >> Dev Test never goes away either. >> Your thoughts on this? >> Yeah, so adding on top of that, we believe where your infrastructure should go is where the data goes. You want to follow where the data is and that's exactly why we want to partner with Pure here because we see a lot of the data are sitting today in the very important infrastructure which is built by Pure Storage and we want to make sure that we're not just building a silo box sitting there where you have to pour the data in there all the time, but actually connect to our server with Pure Storage in the most manageable way. And for IT, it's the same kind of manual layer. You're not thinking about, oh, I have to manage all this silo box, or the shadow IT that some data scientists would have under their desk. That's the least thing you want. >> And the other thing that came up in the key note today, which we've been saying on theCUBE, and all the experts reaffirm, is that moving data costs money. You've got latency costs and also just cost to move traffic around. So moving compute to the edge or moving compute to the data has been a big, hot trend. How has the compute equation changed? Because I've got storage. I'm not just moving packets around. I'm storing it, I'm moving it around. How does that change the compute? Does that put more emphasis on the compute? >> It's definitely putting a lot more emphasis on compute. I think it's where you want compute to happen. You can pull all the data and want it to happen in the center place. That's fine if that's the way you want to manage it. If you have already simplified the data, you want to put it in that's the way. If you want to do it at the edge, near where the data source is, you can also do the cleaning there. So we want to make sure that, no matter how you want to manage it, we have the portfolio that can actually help you to manage that. >> And it's alternative processors. You mentioned NVIDEA. >> Exactly. >> You guys are the first to do a deal with them. >> And other ways, too. You've got to take advantage of technology like Kubernetes, as an example. So you can move the containers where they need to be and have policy managers for the compute requirements and also storage, so that you don't have contention or data integrity issues. So embracing those technologies in a multi-cloud world is very, very essential. >> Mike, I want to ask you a question around customer trends. What are you seeing as a pattern from a customer standpoint, as they prepare for AI, and start re-factoring some of their IT and/or resources, is there a certain use-case that they set up with Pure in terms of how they set up their storage? Is it different by customer? Is there a common trend that you see? >> Yeah, there are some commonalities. Take financial services, quant-trading as an example. We have a number of customers that leverage our platform for that because it's very time-sensitive, high-availability data. So really, I think that the trend overall of that would be: step back, take a look at your data, and focus on, how can I correlate and organize that? And really get it ready so that whatever platform you use from a storage standpoint, you're thinking about all aspects of data and get it in a format, in a form, where you can manage and catalog, because that's kind of essential to the entire thing. >> It really highlights the key things that we've been saying in storage for a long time. High-availability, integrity of the data, and now you've got application developers programming with data. With APIs, you're slinging APIs around like it's-- >> The way it should be. >> That's the way it should be. This is like Nirvana finally got here. How far along are we in the progress? How far? Are we early? Are we moving the needle? Where are the customers? >> You mean in terms of a partnership? >> Partnership, customer AI, in general. You guys, you've got storage, you've got networking and compute all working together. It has to be flexible, elastic, like the cloud. >> My feeling, Mike can correct me, or you can disagree with me. (laughing) I think right now, if we look at what all the analysts are saying, and what we're saying, I think most of the companies, more than 50% of companies either have deployed AI MO or are considering a plan of deploying that. But having said that, we do see that we're still at a relatively early stage because the challenges of making AI deployment at scale, where data scientists and IT are really working together. You need that level of security and that level of skill of infrastructure and software and evolving DevNet. So my feeling is we're still at a relatively early stage. >> Yeah, I think we are in the early adopter phase. We've had customers for the last two years that have really been driving this. We work with about seven of the automated car-driving companies. But if you look at the data from Morgan Stanley and other analysts, there's about a $13 billion infrastructure that's required for AI over the next three years, from 2019-2021, so that is probably 6X, 7X what it is today, so we haven't quite hit that bell curve yet. >> So people are doing their homework right now, setting up their architecture? >> It's the leaders. It's leaders in the industry, not the mainstream. >> Got it. >> And everybody else is going to close that gap, and that's where you guys come in, is helping them do that. >> That's scale. (talking over one another) >> That's what we built this platform with Cisco on, is really, the Flashstack for AI is around scale, for tens and twenties of petabytes of data that will be required for these applications. >> And it's a targeted solution for AI with all the integration pieces with Cisco built in? >> Yes. >> Great, awesome. We'll keep track of it. It's exciting. >> Awesome. >> It's cliche to say future-proof but in this case, it literally is preparing for the future. The bridge to the future, as the new saying at Cisco goes. >> Yes, absolutely. >> This is theCube coverage live in Barcelona. We'll be back with more live coverage after this short break. Thanks for watching. I'm John Furrier with Dave Vallente. Stay with us. (upbeat electronic music)
SUMMARY :
Brought to you by Cisco and its ecosystem partners. Dave Vellante, my co-host for the week, We're in the DevNet zone. in the converged infrastructure space. So data at the center means a lot of things. the data to the training side. at the machine learning with Python, ML Library, But what are the trends that you guys are seeing and less about the infrastructure. And so you guys are trying to simplify So I think AI started with I need to store that data better. You're on Fortnite. and the IT and make it not just AI as an experiment You guys have announced Pure in the cloud? Software is the key here. How is the role of the software and the platform has a super layer on top So I want to put both you guys on the spot So a lot of that comes to the What do you got for me? network provider in the world, from Cisco. It's more flexible than that so you have Yes, and provision it for multi-purpose, and it's got to be horizontally scalable, if you will. and that runs holistically across the entire organization. (laughing) That's the least thing you want. How does that change the compute? That's fine if that's the way you want to manage it. And it's alternative processors. and also storage, so that you don't have Mike, I want to ask you a where you can manage and catalog, High-availability, integrity of the data, That's the way it should be. It has to be flexible, elastic, like the cloud. and that level of skill of infrastructure that's required for AI over the next three years, It's leaders in the industry, not the mainstream. and that's where you guys come in, is helping them do that. That's scale. is really, the Flashstack for AI is around scale, It's exciting. it literally is preparing for the future. I'm John Furrier with Dave Vallente.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Mike | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Dave Vallente | PERSON | 0.99+ |
Mike Bundy | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Barcelona | LOCATION | 0.99+ |
four hour | QUANTITY | 0.99+ |
50% | QUANTITY | 0.99+ |
Pure Storage | ORGANIZATION | 0.99+ |
Zongjie Diao | PERSON | 0.99+ |
Morgan Stanley | ORGANIZATION | 0.99+ |
more than 50% | QUANTITY | 0.99+ |
Python | TITLE | 0.99+ |
1000 series X | COMMERCIAL_ITEM | 0.99+ |
today | DATE | 0.99+ |
Pure | ORGANIZATION | 0.98+ |
7X | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
Barcelona, Spain | LOCATION | 0.98+ |
both | QUANTITY | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
one thing | QUANTITY | 0.98+ |
6X | QUANTITY | 0.98+ |
nine years ago | DATE | 0.98+ |
NVIDEA | ORGANIZATION | 0.97+ |
Global Cisco Alliance | ORGANIZATION | 0.97+ |
Flash | TITLE | 0.97+ |
two guests | QUANTITY | 0.96+ |
Appache Spark | TITLE | 0.96+ |
2019-2021 | DATE | 0.96+ |
Nirvana | ORGANIZATION | 0.96+ |
Flow | TITLE | 0.93+ |
$13 billion | QUANTITY | 0.93+ |
FlashBlade | COMMERCIAL_ITEM | 0.91+ |
Fortnite | TITLE | 0.91+ |
Z | PERSON | 0.9+ |
Data Hub | TITLE | 0.9+ |
Europe | LOCATION | 0.9+ |
Spark | TITLE | 0.89+ |
three basic components | QUANTITY | 0.88+ |
ML Library | TITLE | 0.88+ |
tens and twenties of petabytes of data | QUANTITY | 0.88+ |
about seven of the automated car-driving companies | QUANTITY | 0.84+ |
last two years | DATE | 0.83+ |
Cisco Live 2019 | EVENT | 0.82+ |
two hour | QUANTITY | 0.81+ |
Cisco | EVENT | 0.8+ |
Flashstack | TITLE | 0.79+ |
single pane of | QUANTITY | 0.78+ |
single pane of glass | QUANTITY | 0.77+ |
Dev Test | TITLE | 0.77+ |
about | QUANTITY | 0.74+ |
Cisco Pure | ORGANIZATION | 0.73+ |
next three years | DATE | 0.72+ |
Kubernetes | TITLE | 0.69+ |
FlashBlade | TITLE | 0.65+ |
DevNet | TITLE | 0.65+ |
Katie Colbert & Kaustubh Das | Cisco Live EU 2019
>> Live from Barcelona, Spain, it's The Cube, covering Cisco Live Europe. Brought to you by Cisco and its ecosystem partners. >> Welcome back to Barcelona, everybody. You're watching The Cube, the leader in live tech coverage. My name is Dave Vellante. I'm here with my cohost, Stu Miniman. This is day one of Cisco Live Barcelona. Katie Colbert is here. She's the vice president of alliances at Pure Storage, and she's joined by Kaustubh Das, otherwise known as KD, who's the vice president of computing systems at Cisco. Katie and KD, welcome to The Cube, good to see you. >> Thank you. >> Thank you. >> Alright, so let's start off, KD2, if you could just tell us about the partnership. Where did it start, how did it evolve? We'll get into it. >> We just had a terrific partnership, and the reason it's so great is it's really based on some foundational things that are super compatible. Pure Storage, Cisco, both super technology-driven companies, innovating. They're both also super programmatic companies. They'll do everything via API. It's very modern in that sense, the frameworks that we work on. And then from a business perspective, it's very compatible. We're chasing common markets, very few conflicts. So it's been rooted in solid foundations. And then, we've actually invested over the years to build more and more solutions for our customers jointly. So it's been terrific. >> So, Katie, I hate to admit how long we talk about partnering with Cisco >> It's going to age us. >> So you and I won't admit how many decades it's been partnering with Cisco, but here we are, 2019, Cisco's a very different company than it was a decade or two ago. >> Absolutely. >> Tell me what it's like working with them, especially as a company that's primarily in storage and data at Pure, what it means to partner with them. >> Absolutely, you're right. So, worked with Cisco as a partner for many years at the beginning of my career, then went away for, I'd say, a good 10 years, and joined Pure in June, and I will tell you one of the most exciting reasons why I joined Pure was the Pure and Cisco relationship. When I worked with them at the beginning of my career, it was great and I would tell you it's even better now. I will say that the momentum that these two companies have in the market is very phenomenal. A lot of differentiation from our products separately, but both together, I think that it's absolutely been very successful, and to KD's point, the investment that both companies are making really is just astronomical, and I see that our customers are the beneficiaries of that. It makes it so much easier for them to deploy and use the technologies together, which is exciting. >> So we always joke about Barney deals, I love you, you love me, I mean, it's clear you guys go much much deeper than that. So I want to probe at that a little bit. Particularly from an engineering standpoint, whether it's validated designs or other innovations that you guys are working on together, can we peel the onion on that one a little bit? Talk about what you guys are doing below that line. >> I'll start there then I'll hand it over to the engineering leader from Cisco. But if you think about the pace of this, the partnership, I think, is roughly 3 or so years old. We've 16 Cisco-validated designs for our FlashStack infrastructure. So that is just unbelievable. So, huge amount of investment from engineers, product managers, on both sides of the fence. >> Yeah, totally second that. We start out with the... Cisco-validated designs are like blueprints, so we start out with the blueprints for the standard workloads: Oracle, SAP. And we keep those fresh as new versions come out. But then I think we've taken it further into new spaces of late. ACI, we saw in the keynote this morning, it's going everywhere, it's going multi-site. We've done some work on marrying that with the clustering service of Pure Storage. On top of that, we're doing some work in AI and ML, which is super exciting, so we got some CBDs around that that's just coming out. We're doing some work on automation, coupling Intersight, which is Cisco's cloud-based automation suite, with Pure Storage and Pure Storage's ability to integrate into the Intersight APIs. We talked about it, in fact, I talked about it in my session at the Cisco Live in the summer last year, and now we've got that out as a product. So tremendous amount of work, both in traditional areas as well as some of these new spaces. >> Maybe we can unpack that Intersight piece a bit, because people might look at it initially and say, "Okay, multi-cloud, on-prem, all these environments, "but is this just a networking tool?" And working we're working with someone with Pure, maybe explain a little bit the scope and how, if I'm a Pure administrator, how I live into this world. >> Absolutely, so let's start with what is Intersight, just for a foundational thing. Intersight is our software management tool driven from the cloud. So everything from the personality of the server, the bios settings, the WLAN settings, the networking and the compute pieces of it, that gets administered from the cloud, but it does more. What it does is it can deliver playbooks from the cloud that give the server a certain kind of personality for the workload that it's supporting. So then the next question that anyone asks is, "Now that we have this partnership, "well can it do the same thing for storage? "Can it actually provision that storage, "get that up and running?" And the answer is yes, it can, but it's better because what it can not only do is, not only can it do that, getting that done is super simple. All Pure Storage needed to do was to write some of those Intersight APIs and deliver that playbook from the cloud, from a remote location potentially, into whatever your infrastructure is, provisioning compute, provisioning networking, provisioning storage, in a truly modern cloud-driven environment, right? So I think that's phenomenal what it does for our customers. >> Yeah, I'd agree with that. And I think it'll even become more important as the companies are partnering around our multi-cloud solutions. So, as you probably saw earlier this year in February, sorry, the end of 2018, Pure announced our first leaning into hybrid cloud, so that's Pure Cloud Data Services. That enables us to have Purity, which is our operating system on our storage, running in AWS to begin with. So you can pretty easily start to think about where this partnership is going to go, especially as it pertains to Intersight integration. >> And just to bounce on that, strategically, you can see the alignment there as well. I mean, Cisco's been talking about multi-cloud for a bit now, we've done work to enable similar development environments, whether we're doing something on-prem or in the cloud, so that you can move workloads from one to the other, or actually you can make workloads on both sides talk to each other, and, again, combined with what Katie just said, it makes it a really really compelling solution. >> Like you said, you've got pretty clear swimming lanes for the two companies. There's very little overlap here. You can't have too many of these types of partnerships, right, I mean, you got 25 thousand engineers almost, but still, you still have limited resources. So what makes this one so special, and why are you able to spend so much time and effort, each of you? >> I could start, so from a Pure perspective, I think the cultures are aligned, you called it out there, there's inherently not a lot of overlap in terms of where core competencies are. Pure is not looking at all to become a networking company. And just a lot of synergies in the market make it one that our engineers want to invest in. We have really picked Cisco as our lean-in partner, truthfully, I run all of the alliances at Pure, and a lion's share of my resources really are focused at that partnership. >> Yeah, and if you look at both these companies, Pure is a relative youngster among the storage companies, a new, modern, in a good way, a new, modern company built on modern software practices and so forth. Cisco, although a pretty veteran company, but Cisco compute is relatively new as well as a compute provider. So we are very similar in how our design philosophies work and how modern our infrastructures are, and that gets us to delivering results, delivering solutions to our customers with relatively less effort from our engineers. And that pace of innovation that we can do with Pure is not something we can do with every other company. >> We had a session earlier today, and we went pretty deep into AI, but it's probably worth touching on that. I guess my question here is, what are the customers asking you guys for in terms of AI infrastructure? What's that infrastructure look like that's powering the machine and intelligence era? >> You want to start? >> You want to go, I'll go first. This is a really exciting space, and not only is it exciting because AI is exciting, it's actually exciting because we've got some unique ingredients across Pure and Cisco to make this happen. What does AI feed on? AI feeds on data. The model requires that volume of data to actually train itself We've got an infrastructure, so we just released the C4ATML, the UCC4ATML, highly powered infrastructure, eight GPUs, interconnected, 180 terabytes on board, high network bandwidth, but it needs something to feed it the data, and what Pure's got with their FlashBlade is that ability to actually feed data to this AI infrastructure so that we can train bigger models or train these models faster. Makes for a fantastic solution because these ingredients are just custom made for each other. >> Anything you can add? >> Absolutely I'd agree with that. Really, if you look at AI and what it needs to be successful, and, first of all, all of our customers, if they're not thinking about it, they should be, and I will tell you most of them are, is, how do you ingest that amount of data? If you can't ingest that quickly, it's not going to be of use. So that's a big piece of it, and that's really what the new Cisco platform, I mean, the folks over at Pure are just thrilled about the new Cisco product, and then you take a look at the FlashBlade and how it's able to really scale out unstructured data, object it and file, really to make that useful, so when you have to scrub that data to be able to use it and correlate it, FlashBlade is the perfect solution. So really, this is two companies coming together with the best of breed technologies. >> And the tooling in that world is exploding, open source innovation, it needs a place to run all the Kafkas and the Caffes and the TensorFlows and the Pythons. It's not just confined to data scientists anymore. It's really starting to seep throughout the organization, are you seeing that? >> Yeah. >> What's happening is you've got the buzzwords going around, and that leads to businesses and the leaders of businesses saying, "We've got to have an AI strategy. "We've got to hire these data scientists." But at the same time, the data scientists can get started on the laptop, they can get started on the cloud. When they want to deploy this, they need an enterprise class, resilient, automated infrastructure that fits into the way they do their work. You've got to have something that's built on these components, so what we provide together is that infrastructure for the ITTs so that the data scientists, when they build their beautiful models, have a place to deploy them, have a place to put that into production, and can actually have that life cycle running in a much more smooth production-grade environment. >> Okay, so you guys are three years in, roughly. Where do you want to take this thing, what's the vision? Give us a little road map for the future as to what this partnership looks like down the road. >> Yeah, so I can start. So I think there's a few different vectors. We're going to continue driving the infrastructure for the traditional workloads. That's it, that's a big piece that we do, we continue doing that. We're going to drive a lot more on the automation side, I think there's such a lot of potential with what we've got on Intersight, with the automation that Pure supports, bring those together and really make it simple for our customers to get this up and running and manage that life cycle. And third vector's going to be imparting those new use cases, whether it be AI or more data analytics type use cases. There's a lot of potential that it unleashes for our customers and there's a lot of potential of bringing these technologies together to partner. So you'll see a lot more of that from us. I don't know, will you add something? >> Yeah, no, I absolutely agree. And I would say more FlashStack, look for more FlashStack CVDs, and AI, I think, is one to watch. We believe Cisco, really, this step that Cisco's made, is going to take AI infrastructure to the next level. So we're going to be investing much more heavily into that. And then cloud, from a hybrid cloud, how do these two companies leverage FlashStack and all the innovation we've done on prem together to really enable the multi-cloud. >> Great, alright, well Katie and KD, thanks so much for coming to The Cube. It was great to have you. >> Great. Thanks for having us. >> Thank you very much. >> You're welcome, alright. Keep it right there everybody. Stu and I will be back with our next guest right after this short break. You're watching The Cube Live from Cisco Live Barcelona. We'll be right back. (techy music)
SUMMARY :
Brought to you by Cisco and its ecosystem partners. Welcome back to Barcelona, everybody. if you could just tell us about the partnership. and the reason it's so great is it's really based So you and I won't admit how many at Pure, what it means to partner with them. and I see that our customers are the beneficiaries of that. or other innovations that you guys are working on together, I'll start there then I'll hand it over to so we start out with the blueprints maybe explain a little bit the scope and how, and deliver that playbook from the cloud, So you can pretty easily start to think so that you can move workloads from one to the other, and why are you able to spend And just a lot of synergies in the market And that pace of innovation that we can do with Pure what are the customers asking you guys for is that ability to actually feed data and how it's able to really scale out unstructured data, and the TensorFlows and the Pythons. and that leads to businesses and the leaders of businesses as to what this partnership looks like down the road. for our customers to get this up and running and AI, I think, is one to watch. thanks so much for coming to The Cube. Thanks for having us. Stu and I will be back with our next guest
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Katie | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Katie Colbert | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
KD | PERSON | 0.99+ |
Kaustubh Das | PERSON | 0.99+ |
Pure | ORGANIZATION | 0.99+ |
Stu | PERSON | 0.99+ |
June | DATE | 0.99+ |
two companies | QUANTITY | 0.99+ |
KD2 | PERSON | 0.99+ |
both | QUANTITY | 0.99+ |
16 | QUANTITY | 0.99+ |
both companies | QUANTITY | 0.99+ |
three years | QUANTITY | 0.99+ |
2019 | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
10 years | QUANTITY | 0.99+ |
Barcelona | LOCATION | 0.99+ |
ACI | ORGANIZATION | 0.99+ |
180 terabytes | QUANTITY | 0.99+ |
C4ATML | COMMERCIAL_ITEM | 0.99+ |
25 thousand engineers | QUANTITY | 0.99+ |
both sides | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
Barcelona, Spain | LOCATION | 0.98+ |
SAP | ORGANIZATION | 0.98+ |
FlashBlade | TITLE | 0.98+ |
The Cube Live | TITLE | 0.98+ |
Barney | ORGANIZATION | 0.98+ |
earlier this year | DATE | 0.97+ |
Pure Storage | ORGANIZATION | 0.97+ |
FlashBlade | COMMERCIAL_ITEM | 0.97+ |
Intersight | ORGANIZATION | 0.97+ |
end of 2018 | DATE | 0.96+ |
UCC4ATML | COMMERCIAL_ITEM | 0.95+ |
each | QUANTITY | 0.95+ |
Pythons | TITLE | 0.94+ |
second | QUANTITY | 0.92+ |
eight GPUs | QUANTITY | 0.91+ |
TensorFlows | TITLE | 0.9+ |
two ago | DATE | 0.9+ |
Purity | ORGANIZATION | 0.89+ |
Pure Cloud Data Services | ORGANIZATION | 0.89+ |
February | DATE | 0.89+ |
Chadd Kenney, PureStorage | CUBEConversation, November 2018
(bright instrumental music) >> Hi everyone, I'm John Furrier. Here in the Cube Studios in Palo Alto, for a special Cube conversation on some big news from PureStorage. We're here with Chadd Kenney, who's the Vice President of Product and Solutions at PureStorage. Big Cloud news. A historic announcement from PureStorage. One of the fastest growing startups in the storage business. Went public, I've been following these guys since creation. Great success story in Silicon Valley and certainly innovative products. Now announcing a Cloud product. Cloud data services, now in market. Chadd, this is huge. >> It's exciting time. Thank you so much for having us. >> So you guys, obviously storage success story, but now the reality is changed. You know we've been saying in the Cube, nothing changes, you get storage computer networking, old way, new way in the Cloud. Game is still the same. Storage isn't going away. You got to store the data somewhere and the data tsunami is coming. Still coming with Edge and a bunch of other things. Cloud more important than ever. To get it right is super important. So, what is the announcement of Cloud Data Service. Explain what the product is, why you guys built it, why now. >> Awesome. So, a couple different innovations that are part of this launch to start with. We have Cloud Block Store which is taking Purity, which is our operating system found on-prem and actually moving it to AWS. And we spent a bunch of time optimizing these solutions to make it so that, we could actually take on tier one, mission critical applications. A key differentiator is that most folks were really chasing after test-dev and leveraging the Cloud for that type of use case. Whereas Cloud Block Storage, really kind of industry strength and ready for mission critical applications. We also took protection mechanisms from FlashArray on-premises and actually made it so that you could use CloudSnap and move and protect data into the public Cloud via portable snapshot technology. Which we can dig into a little bit later. And then the last part is, we thought it was really ripe to change data protection, just as a whole. Most people are doing kind of disc to disc, to tape, and then moving tape offsite. We believe the world has shifted. There's a big problem in data protection. Restoring data is not happening in the time frame that its needed, and SLAs aren't being met, and users are not happy with the overall solution as a whole. We believe that restorations from Flash are incredibly important to the business, but in order to get there you have to offset the economics. So what we're building is a Flash to Flash to Cloud solution which enables folks to be able to take the advantages of the economics of Cloud and be able to then have a caching mechanism of Flash on-premises. So that they can restore things relatively quickly for the predominant set of data that they have out there. >> And just so I get everything right here. You guys only been on-premises only, this is now a cloud solution. It's software. >> Correct. >> Why now? Why wait 'til now, is the timing right? What's the internal conversation? And why should customers know, is this the right time. >> So, the evolution of cloud has been pretty interesting as we've gone through it. Most customers went from kind of a 100% on-premises, the Cloud came out and said, hey, I'm going to move everything to the Cloud. They found that didn't work great for enterprise applications. And so they kind of moved back and realized that hybrid cloud was going to be a world with they wanted to leverage both. We're seeing a lot of other shifts in the market. VMware already having RDS in platform. Now it's true hybrid cloud kind of playing out there. Amazon running an AWS. It's a good mixture just to showcase where people really want to be able to leverage the capabilities of both. >> So it's a good time because the customers are re-architecting as well. >> It's all about- >> Hybrid applications are definitely what people want. >> 100% and the application stack, I think was the core focus that really shifted over time. Instead of just focusing on hybrid cloud infrastructure, it was really about how applications could leverage multiple types of clouds to be able to leverage the innovation and services that they could provide. >> You know, I've always been following the IT business for 30 years or so and it's always been an interesting trend. You buy something from a vendor and there's a trade-offs. And there's always the payback periods, but now I think with this announcement that's interesting is you got the product mix that allows customers to have choice and pick what they want. There's no more trade-offs. If they like cloud, they go to cloud. If they like on-premise, you go on-premises. >> It sounds like an easy concept, but the crazy part to this is the Cloud divide is real. They are very different environments. As we've talked to customers, they were very lost on how it was going to take and enterprise application and actually leverage the innovations within the Cloud. They wanted it, they needed it, but at the same time, they weren't able to deliver up on it. And so, we realized that the data layer, fundamentally was the area that could give them that bridge between those two environments. And we could add some core values to the Cloud for even the next generation developer who's developing in the Cloud to bring in, better overall resiliency. Management and all sorts of new features that they weren't able to take advantage of in traditional public cloud. >> You know Chugg wants to do minimal about the serviceless trend and how awesome that is. It's just, look at the resource pool as a serviceless pool of resource. So is this storageless? >> So it's still backed by storage, obviously. >> No, I was just making a joke. No wait, that you're looking at it as what serviceless is to the user. You guys are providing that same kind of storage pool, addressable through the application of, >> Correct >> as if it's storageless. And what's great about taking 100% software platform and moving it to the Cloud is, customer can spin this up in like minutes. And what's great about it is, they can spend many, many, many instances of these things for various different use cases that they have out there, and get true utility out of it. So they're getting the agility that they really want while not having to offset the values that they really come to love about PureStorage on-premises. Now they can actually get it all on the public cloud as well. >> I want to dig into the products a little bit. Before we get there, I want you to answer the question that's probably on people's minds. I know you've been at Pure, really from the beginning. So, you've seen the history. Most people look at you guys and say, well you're a hardware vendor. I have Pure boxes everywhere, you guys doing a great job. You've pioneered the Flash, speed game on storage. People want, kill latency as they say. You guys have done a great job. But wait a minute, this is software. Explain how you guys did this, why it's important. People might not know that this is a software solution. They might be know you for hardware. What's the difference? Is there a difference? Why should they care and what's the impact? >> So, great question. Since we sell hardware products, most people see us as a hardware company. But at the end of the day, the majority of vinge and dev is software. We're building software to make, originally, off the shelf components to be enterprise worthy. Over time we decided to optimize the hardware too, and that pairing between the software and hardware gets them inherently great values. And this is why we didn't just take our software and just kind of throw it into every cloud and say have it, to customers. Like a lot of folks did. We spent a lot of time, just like we did on our hardware platform, optimizing for AWS to start with. So that we could truly be able to leverage the inherent technologies that they have, but build software to make it even better. >> It's interesting, I interviewed Andy Bechtolsheim at VMworld, and he's a chairman of Arista. He's called, Les Peckgesem calls him the Rembrandt of motherboards. And he goes, "John, we're in the software business." And he goes, "Let me tell ya, hardware's easy. Software's hard." >> I agree. >> So everyone's pretty much in the software business. This is not a change for Pure. >> No, this is the same game we've been in. >> Great. Alright, let's get into the products. The first one is Cloud Block Store for AWS. Which is the way Amazon does the branch. So it's on Amazon, or for Amazon as they say. They use different words. So this is Pure software in the Cloud. Your company, technically Pure software. >> Yup. >> In the Cloud as software, no hardware. >> Yup. >> A 100% storage, API support, always encrypted, seamless management and orchestration, DR backup migration between clouds. >> Yup. >> That's kind of the core premise. So what does the product do, what's the purpose of the product. On the Amazon piece, if I'm a customer of Pure or a prospect for Pure, what does the product give me? What's the capabilities? >> Great. I would say that the biggest thing that customers get is just leverage for their application stack to be able to utilize the Cloud. And let me give you a couple of examples 'cause they're kind of fun. So first off, Cloud Block Storage is just software that sits in the Cloud that has many of the same utilities that run on-premises. Any by doing so, you get the ability to be able to do stuff like I want to replicate, as a DR target. So maybe I don't have a secondary site out there, and I want to have a DR target that spin up in the event of a disaster. You can easily set up bi-directional replication to the instance that you have running in the Cloud. It's the exact same experience. The exact same APIs and you get our cloud data management with Pure1 to be able to see both sites. One single pane of glass, and make sure everything is up and running and doing well. You could also though, leverage a test-dev environment. So let's saying I'm running production on-premises, I can then go ahead and replicate to the Cloud, spin up an instance for test-dev, and running reporting, run analytics. Run anything else that I wanted on top of that. And spin up compute relatively quickly. Maybe I don't have it on-prem. Next, we could focus on replicating for protection. Let's say for compliance, I want to have many instances to be able to restore back in the event of a disaster or in the event that I just want to look back during a period of time. The last part is, not just on-prem to the Cloud, but leveraging the Cloud for even better resiliency to take enterprise applications and actually move them without having to do massive re-architecture. If you look at what happens, Amazon recommends typically, that you have data in two different availability zones. So that when you put an application on top of it, it can be resilient to any sort of failures within an AZ. What we've done is we've taken our active cluster technology which is active-active replication between two instances, and made it so that you can actually replicate between two availability zones. And your application now doesn't need to be re-architected whatsoever. >> So you basically, if I get this right, you had core software that made all that Flash, on the box which is on-premise, which is a hardware solution. Which sounds like it was commodity boxes so this, components. >> Just like the Cloud. >> You take it to the Cloud as an amazing amount of boxes out there. They have tons of data centers. So you treat the Cloud as if it's a virtual device, so to speak. >> Correct. I mean the Cloud functionally is just compute and storage, and networking on the back end has been abstracted by some sort of layer in front of it. We're leveraging compute resources for our controllers and we're leveraging persistent storage media for our storage. But what we've done in software is optimize a bunch of things. An example just as one is, in the Cloud when you, procure storage, you pay for all of it, whether you leverage it or not. We incorporate de-dupe, compression, thin provisioning, AES 256 encryption on all data arrest. These are data services that are just embedded in that aren't traditionally found in a traditional cloud. >> This makes so much sense. If you're an application developer, you focus on building the app. Not worrying about where the storage is and how it's all managed. 'Cause you want persistent data and uni-managed state, and all this stuff going on. And I just need a dashboard, I just need to know where the storage is. Is it available and bring it to the table. >> And make it easy with the same APIs that you were potentially running on, on-premises. And that last part that I would say is that, the layered services that are built into Purity, like our snapshot technology and being able to refresh test-dev environments or create 10 sandboxes for 10 developers in the Cloud and add compute instances to them, is not only instantaneous, but it's space saving as you actually do it. Where as in the normal cloud offerings, you're paying for each one of those instances. >> And the agility is off the charts, it's amazing. Okay, final question on this one is, how much is it's going to cost? How does a customer consume it? Is it in the marketplace? Do I just click a button, spin up things? How's the interface? What's the customer interaction and engagement with the product? How they buy it, how much it costs? Can you share the interaction with the customer? >> So we're just jumping into beta, so a lot of this is still being worked out. But what I will tell you is it's the exact same experience that customers have come to love with Pure. You can go download the Cloud formation template into your catalog with an AWS. So you can spin up instances. The same kind of consumption models that we've built on-prem will be applied to cloud. So it will be a very similar consumption model, which has been super consumer friendly that customers have loved from us over the years. And it will be available in the mid part of next year, and so people will be able to beta it today, test it out, see how it works, and then put it into full production in mid part of next year. >> And operationally, in the work flows, the customers don't skip a beat. It's the same kind of format, languages and the words, the word flow. It feels like Pure all the way through. >> Correct. And not only are we a 100% built on a rest API, but all of the things we've built in with, Python libraries that automate this for developers, to PowerShell toolkits, to Ansible playbooks. All the stuff we've built on codeupyourstorage.com are all applicable to both sites and you get Pure1, our Cloud based management system to be able to see all of it in one single pane of glass. >> Okay, let's move on. So the next piece I think is interesting. I'll get your thoughts on this is that the whole protection piece. On-premises, really kind of held back from the Cloud, mainly to protect the data. So you guys got CloudSnap for AWS, what does this product do? Is this the protection piece? How does this work? What is the product? What's the features and what's the value? >> So, StorReduce was a recent acquisition that we did that enables de-duplication on top of an S3 target. And so it allows you to store an S3 de-duplicated into a smaller form factor and we're pairing that with both an on-premises addition which will have a flash plate behind it for super fast restores. So think of that as a caching tier for your backups, but then also be able to replicate that out to the public cloud and leverage store reduce natively in the public cloud as well. >> So that's the store reduce product. So store reduce on it is that piece. As an object store? >> It is, yes. And we pair that with CloudSnap which is natively integrated within FlashArray, so you can also do snapshots to a FlashBlade for fast restores for both NFS, and you can send it also to S3 in the public cloud. And so you get the inherent abilities to even do, VM level granularity or volume level granularity as well from a FlashArray directly, without needing to have any additional hardware. >> Okay so the data services are the; Block Storage, Store Reduce and CloudSnap on a four AWS. >> Correct. >> How would you encapsulate this from a product and solution standpoint? How would you describe that to a customer in an elevator or just a quick value statement? What's in it for them? >> Sure. So Pure's been seen by customers as innovation engine that optimized applications and allowed them to do, I would say, amazing things into the enterprise. What we're doing now, is we're evolving that solution out of just an on-premises solution and making it available in a very agile Cloud world. We know this world is evolving dramatically. We know people really want to be able to take advantage of the innovations within the Cloud, and so what we're doing is we're finally bridging the gap between on-premises and the Cloud. Giving them the same user experience that they've come to love with Pure and all of the Clouds that they potentially need to develop in. >> Okay so from the announcement standpoint, you guys got Cloud Block Storage limited public beta, right out of the gate. GA in mid 2019. CloudSnap is GA at announcement and Store Reduce is going into beta, first half of 2019. >> Correct, we're excited about it. >> So for the skeptics out there who are- Hey you know, Chadd, I got to tell ya. I love the Cloud, but I'm a little bit nervous. How do I test and get a feeling for- this is going to be simple, if I'm going to jump in and look at this. What should I look at first? What sequence, should I try this? Do you guys have a playbook, for them to either kick the tires or how should they explore to get proficient in the new solution. >> Good question. Right, so for one if you're a FlashArray customer, CloudSnap gives you the ability to be able to take this new entity, called a portable Snapshot. Which is data paired with metadata, and allow you to be able to move data off of a FlashArray. You can put it to an NFS target or you can send it to the Cloud. And so that's the most logical one that folks will probably leverage first because it's super exciting for them to be able to leverage the Cloud and spin up instances, if they'd like to. Or protect back to their own prem. Also, Cloud Block Storage, great because you can spin it up relatively quickly and test out applications between the two. One area that I think customers are going to be really excited about is you could run an analytics environment in the Cloud and spin up a bunch of compute from your production instance by just replicating it up into the Cloud. The last part is, I think backup is not super sexy. Nobody like to talk about it, but it's a significant pain point that's out there, and I think we can make some major in-roads in helping businesses get better SLAs. We're very, very interested to see the great solutions people bring with- >> So, I'm going to put you on the spot here and ask you, there's always the, love the cliche, is it a vitamin or is it an Asprin. Is there a pain point? So obviously backup, I would agree. Backup and recovery, certainly with the disaster, you see the wildfires going on here in California. You can't stop thinking about what the, disaster recovery plan and then you got top line growth with application developers. The kind of the vitamin, if you will. What are the use cases, low hanging fruit for someone to like test this out from a pain point standpoint. Is it backup and what's the growth angle? I wanted to test out this new solution, what should I look at first? What would you recommend? >> It's a very tough question. So, CloudSnap is obviously the easy one. I'd say Cloud Block Store is one that I think, people will. I look at my biggest, customers biggest challenges out there it's how do I get application portable. So I think Cloud Block Store really gives you the application portability. So I think it's finally achieving that whole, hybrid cloud world. But at the end of the day, backup is really big pain point that the enterprise deals with, like right this second. So there's areas where we believe we can add inherent values to them with being able to do fast restores from Flash. That meets SLA's very quickly and is an easy fix. >> And you guys feel good about the data protection aspect of this? >> Yes, very much so. >> Awesome. I want to get your personal take on this. You were early on in Pure. What's the vibe inside the company? This is Cloud and people love Cloud. There's benefits for Cloud, as well as on-premises. What's the mood like inside PureStorage? You've seen from the beginning, now you're a public company and growing up really, really fast. What's the vibe like inside PureStorage? >> It's funny, it hasn't really changed all that much, in the cultural side of the thing, of the business. I love where I work because of the people. The people bring so much fun to the business, so much innovation and we have a mindset that's heavily focused on customer first. And that's one of the things. I always tell this kind of story is, when we first started, we sat in a room on a whiteboard and wrote up, what is everything that sucks about storage. And instead of trying to figure out how we make a 2.0 version of some storage array, we actually figured out what are all the customer pain points that we needed to satisfy and then we built innovations to go do that. Not go chase the competition, but actually go alleviate customer challenges. And we just continue to kind of focus on customer first and so the whole company kind of, rallies around that. And I think you see a very different motion that what you do in most companies because we love hearing about customer results of our products. Engineering just will rally around when a customer shows up just to hear exactly their experience associated to it. And so with this, I think what they see is a continued evolution of the things we've been doing and they love seeing and providing customer solutions in areas that they were challenged to deal with in the past. >> What was some of the customer feedback when you guys started going, hey, you've got a new product, you're doing all of that early work. And you got to go talk to some people and knock on the, hey, what do you think, would you like the Cloud, a little bit of the Cloud. How would you like the Cloud to be implemented? What was some of the things you heard from customers? >> A lot of them said, if you can take your core tenets, which was simplicity, efficiency, reliability, and customer focus around consumption, and if you could give that to me in the Cloud, that would be the Nirvana. So, when we looked at this model, that's exactly what we did. We said, let's take what people love about us on-prem, and give 'em the exact same experience in the Cloud. >> That's great and that's what you guys have done. Congratulations. >> Thanks so much. >> Great to hear the Cloud story here Chadd Kenney, Vice President of Products and Solutions at PureStorage. Taking the formula of success on-premises with Flash and the success there, and bringing it to the Cloud. That's the big deal in this announcement. I'm John Furrier here in the Palo Alto studios, thanks for watching. (upbeat instrumental music)
SUMMARY :
One of the fastest growing startups in the storage business. Thank you so much for having us. and the data tsunami is coming. of the economics of Cloud and be able to then have And just so I get everything right here. What's the internal conversation? So, the evolution of cloud has been So it's a good time because the customers 100% and the application stack, You know, I've always been following the IT business for but the crazy part to this is the Cloud divide is real. It's just, look at the resource pool You guys are providing that same kind of storage pool, and moving it to the Cloud is, What's the difference? and that pairing between the software and hardware the Rembrandt of motherboards. So everyone's pretty much in the software business. Which is the way Amazon does the branch. A 100% storage, API support, always encrypted, That's kind of the core premise. and made it so that you can actually replicate on the box which is on-premise, So you treat the Cloud as if it's a virtual device, and networking on the back end I just need to know where the storage is. Where as in the normal cloud offerings, And the agility is off the charts, it's amazing. You can go download the Cloud formation template and the words, the word flow. but all of the things we've built in with, is that the whole protection piece. And so it allows you to store an S3 de-duplicated So that's the store reduce product. And so you get the inherent abilities to even do, Okay so the data services are the; of the innovations within the Cloud, Okay so from the announcement standpoint, So for the skeptics out there who are- And so that's the most logical one The kind of the vitamin, if you will. that the enterprise deals with, You've seen from the beginning, now you're a public company And that's one of the things. a little bit of the Cloud. and give 'em the exact same experience in the Cloud. That's great and that's what you guys have done. and the success there, and bringing it to the Cloud.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Andy Bechtolsheim | PERSON | 0.99+ |
Chadd Kenney | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
PureStorage | ORGANIZATION | 0.99+ |
California | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
30 years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
mid 2019 | DATE | 0.99+ |
10 developers | QUANTITY | 0.99+ |
100% | QUANTITY | 0.99+ |
Chadd | PERSON | 0.99+ |
November 2018 | DATE | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Python | TITLE | 0.99+ |
Les Peckgesem | PERSON | 0.99+ |
both sites | QUANTITY | 0.99+ |
first one | QUANTITY | 0.99+ |
FlashArray | TITLE | 0.99+ |
Cloud Block Store | TITLE | 0.99+ |
VMworld | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
10 sandboxes | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Pure | ORGANIZATION | 0.98+ |
two environments | QUANTITY | 0.98+ |
two instances | QUANTITY | 0.98+ |
Arista | ORGANIZATION | 0.98+ |
Cloud | TITLE | 0.98+ |
first half of 2019 | DATE | 0.97+ |
Cloud Block Storage | TITLE | 0.97+ |
first | QUANTITY | 0.97+ |
mid part of next year | DATE | 0.97+ |
S3 | TITLE | 0.96+ |
Flash | TITLE | 0.96+ |
one | QUANTITY | 0.96+ |
GA | LOCATION | 0.95+ |
One area | QUANTITY | 0.95+ |
One | QUANTITY | 0.95+ |
CloudSnap | TITLE | 0.94+ |
two different availability zones | QUANTITY | 0.93+ |
second | QUANTITY | 0.93+ |
each one | QUANTITY | 0.93+ |
Chugg | ORGANIZATION | 0.92+ |
Cube Studios | ORGANIZATION | 0.91+ |