Ali Ghodsi, Databricks | Cube Conversation Partner Exclusive
(outro music) >> Hey, I'm John Furrier, here with an exclusive interview with Ali Ghodsi, who's the CEO of Databricks. Ali, great to see you. Preview for reinvent. We're going to launch this story, exclusive Databricks material on the notes, after the keynotes prior to the keynotes and after the keynotes that reinvent. So great to see you. You know, you've been a partner of AWS for a very, very long time. I think five years ago, I think I first interviewed you, you were one of the first to publicly declare that this was a place to build a company on and not just post an application, but refactor capabilities to create, essentially a platform in the cloud, on the cloud. Not just an ISV; Independent Software Vendor, kind of an old term, we're talking about real platform like capability to change the game. Can you talk about your experience as an AWS partner? >> Yeah, look, so we started in 2013. I swiped my personal credit card on AWS and some of my co-founders did the same. And we started building. And we were excited because we just thought this is a much better way to launch a company because you can just much faster get time to market and launch your thing and you can get the end users much quicker access to the thing you're building. So we didn't really talk to anyone at AWS, we just swiped a credit card. And eventually they told us, "Hey, do you want to buy extra support?" "You're asking a lot of advanced questions from us." "Maybe you want to buy our advanced support." And we said, no, no, no, no. We're very advanced ourselves, we know what we're doing. We're not going to buy any advanced support. So, you know, we just built this, you know, startup from nothing on AWS without even talking to anyone there. So at some point, I think around 2017, they suddenly saw this company with maybe a hundred million ARR pop up on their radar and it's driving massive amounts of compute, massive amounts of data. And it took a little bit in the beginning just us to get to know each other because as I said, it's like we were not on their radar and we weren't really looking, we were just doing our thing. And then over the years the partnership has deepened and deepened and deepened and then with, you know, Andy (indistinct) really leaning into the partnership, he mentioned us at Reinvent. And then we sort of figured out a way to really integrate the two service, the Databricks platform with AWS . And today it's an amazing partnership. You know, we directly connected with the general managers for the services. We're connected at the CEO level, you know, the sellers get compensated for pushing Databricks, we're, we have multiple offerings on their marketplace. We have a native offering on AWS. You know, we're prominently always sort of marketed and you know, we're aligned also vision wise in what we're trying to do. So yeah, we've come a very, very long way. >> Do you consider yourself a SaaS app or an ISV or do you see yourself more of a platform company because you have customers. How would you categorize your category as a company? >> Well, it's a data platform, right? And actually the, the strategy of the Databricks is take what's otherwise five, six services in the industry or five, six different startups, but do them as part of one data platform that's integrated. So in one word, the strategy of data bricks is "unification." We call it the data lake house. But really the idea behind the data lake house is that of unification, or in more words it's, "The whole is greater than the sum of its parts." So you could actually go and buy five, six services out there or actually use five, six services from the cloud vendors, stitch it together and it kind of resembles Databricks. Our power is in doing those integrated, together in a way in which it's really, really easy and simple to use for end users. So yeah, we're a data platform. I wouldn't, you know, ISV that's a old term, you know, Independent Software Vendor. You know, I think, you know, we have actually a whole slew of ISVs on top of Databricks, that integrate with our platform. And you know, in our marketplace as well as in our partner connect, we host those ISVs that then, you know, work on top of the data that we have in the Databricks, data lake house. >> You know, I think one of the things your journey has been great to document and watch from the beginning. I got to give you guys credit over there and props, congratulations. But I think you're the poster child as a company to what we see enterprises doing now. So go back in time when you guys swiped a credit card, you didn't need attending technical support because you guys had brains, you were refactoring, rethinking. It wasn't just banging out software, you had, you were doing some complex things. It wasn't like it was just write some software hosted on server. It was really a lot more. And as a result your business worth billions of dollars. I think 38 billion or something like that, big numbers, big numbers of great revenue growth as well, billions in revenue. You have customers, you have an ecosystem, you have data applications on top of Databricks. So in a way you're a cloud on top of the cloud. So is there a cloud on top of the cloud? So you have ISVs, Amazon has ISVs. Can you take us through what this means and at this point in history, because this seems to be an advanced version of benefits of platforming and refactoring, leveraging say AWS. >> Yeah, so look, when we started, there was really only one game in town. It was AWS. So it was one cloud. And the strategy of the company then was, well Amazon had this beautiful set of services that they're building bottom up, they have storage, compute, networking, and then they have databases and so on. But it's a lot of services. So let us not directly compete with AWS and try to take out one of their services. Let's not do that because frankly we can't. We were not of that size. They had the scale, they had the size and they were the only cloud vendor in town. So our strategy instead was, let's do something else. Let's not compete directly with say, a particular service they're building, let's take a different strategy. What if we had a unified holistic data platform, where it's just one integrated service end to end. So think of it as Microsoft office, which contains PowerPoint, and Word, and Excel and even Access, if you want to use it. What if we build that and AWS has this really amazing knack for releasing things, you know services, lots of them, every reinvent. And they're sort of a DevOps person's dream and you can stitch these together and you know you have to be technical. How do we elevate that and make it simpler and integrate it? That was our original strategy and it resonated with a segment of the market. And the reason it worked with AWS so that we wouldn't butt heads with AWS was because we weren't a direct replacement for this service or for that service, we were taking a different approach. And AWS, because credit goes to them, they're so customer obsessed, they would actually do what's right for the customer. So if the customer said we want this unified thing, their sellers would actually say, okay, so then you should use Databricks. So they truly are customer obsessed in that way. And I really mean it, John. Things have changed over the years. They're not the only cloud anymore. You know, Azure is real, GCP is real, there's also Alibaba. And now over 70% of our customers are on more than one cloud. So now what we hear from them is, not only want, do we want a simplified, unified thing, but we want it also to work across the clouds. Because those of them that are seriously considering multiple clouds, they don't want to use a service on cloud one and then use a similar service on cloud two. But it's a little bit different. And now they have to do twice the work to make it work. You know, John, it's hard enough as it is, like it's this data stuff and analytics. It's not a walk in the park, you know. You hire an administrator in the back office that clicks a button and its just, now you're a data driven digital transformed company. It's hard. If you now have to do it again on the second cloud with different set of services and then again on a third cloud with a different set of services. That's very, very costly. So the strategy then has changed that, how do we take that unified simple approach and make it also the same and standardize across the clouds, but then also integrate it as far down as we can on each of the clouds. So that you're not giving up any of the benefits that the particular cloud has. >> Yeah, I think one of the things that we see, and I want get your reaction to this, is this rise of the super cloud as we call it. I think you were involved in the Sky paper that I saw your position paper came out after we had introduced Super Cloud, which is great. Congratulations to the Berkeley team, wearing the hat here. But you guys are, I think a driver of this because you're creating the need for these things. You're saying, okay, we went on one cloud with AWS and you didn't hide that. And now you're publicly saying there's other clouds too, increased ham for your business. And customers have multiple clouds in their infrastructure for the best of breed that they have. Okay, get that. But there's still a challenge around the innovation, growth that's still around the corner. We still have a supply chain problem, we still have skill gaps. You know, you guys are unique at Databricks as other these big examples of super clouds that are developing. Enterprises don't have the Databricks kind of talent. They need, they need turnkey solutions. So Adam and the team at Amazon are promoting, you know, more solution oriented approaches higher up on the stack. You're starting to see kind of like, I won't say templates, but you know, almost like application specific headless like, low code, no code capability to accelerate clients who are wanting to write code for the modern error. Right, so this kind of, and then now you, as you guys pointed out with these common services, you're pushing the envelope. So you're saying, hey, I need to compete, I don't want to go to my customers and have them to have a staff or this cloud and this cloud and this cloud because they don't have the staff. Or if they do, they're very unique. So what's your reaction? Because this kind is the, it kind of shows your leadership as a partner of AWS and the clouds, but also highlights I think what's coming. But you share your reaction. >> Yeah, look, it's, first of all, you know, I wish I could take credit for this but I can't because it's really the customers that have decided to go on multiple clouds. You know, it's not Databricks that you know, push this or some other vendor, you know, that, Snowflake or someone who pushed this and now enterprises listened to us and they picked two clouds. That's not how it happened. The enterprises picked two clouds or three clouds themselves and we can get into why, but they did that. So this largely just happened in the market. We as data platforms responded to what they're then saying, which is they're saying, "I don't want to redo this again on the other cloud." So I think the writing is on the wall. I think it's super obvious what's going to happen next. They will say, "Any service I'm using, it better work exactly the same on all the clouds." You know, that's what's going to happen. So in the next five years, every enterprise will say, "I'm going to use the service, but you better make sure that this service works equally well on all of the clouds." And obviously the multicloud vendors like us, are there to do that. But I actually think that what you're going to see happening is that you're going to see the cloud vendors changing the existing services that they have to make them work on the other clouds. That's what's goin to happen, I think. >> Yeah, and I think I would add that, first of all, I agree with you. I think that's going to be a forcing function. Because I think you're driving it. You guys are in a way, one, are just an actor in the driving this because you're on the front end of this and there are others and there will be people following. But I think to me, I'm a cloud vendor, I got to differentiate. Adam, If I'm Adam Saleski, I got to say, "Hey, I got to differentiate." So I don't wan to get stuck in the middle, so to speak. Am I just going to innovate on the hardware AKA infrastructure or am I going to innovate at the higher level services? So what we're talking about here is the tail of two clouds within Amazon, for instance. So do I innovate on the silicon and get low level into the physics and squeeze performance out of the hardware and infrastructure? Or do I focus on ease of use at the top of the stack for the developers? So again, there's a channel of two clouds here. So I got to ask you, how do they differentiate? Number one and number two, I never heard a developer ever say, "I want to run my app or workload on the slower cloud." So I mean, you know, back when we had PCs you wanted to go, "I want the fastest processor." So again, you can have common level services, but where is that performance differentiation with the cloud? What do the clouds do in your opinion? >> Yeah, look, I think it's pretty clear. I think that it's, this is, you know, no surprise. Probably 70% or so of the revenue is in the lower infrastructure layers, compute, storage, networking. And they have to win that. They have to be competitive there. As you said, you can say, oh you know, I guess my CPUs are slower than the other cloud, but who cares? I have amazing other services which only work on my cloud by the way, right? That's not going to be a winning recipe. So I think all three are laser focused on, we going to have specialized hardware and the nuts and bolts of the infrastructure, we can do it better than the other clouds for sure. And you can see lots of innovation happening there, right? The Graviton chips, you know, we see huge price performance benefits in those chips. I mean it's real, right? It's basically a 20, 30% free lunch. You know, why wouldn't you, why wouldn't you go for it there? There's no downside. You know, there's no, "got you" or no catch. But we see Azure doing the same thing now, they're also building their own chips and we know that Google builds specialized machine learning chips, TPU, Tenor Processing Units. So their legs are focused on that. I don't think they can give up that or focused on higher levels if they had to pick bets. And I think actually in the next few years, most of us have to make more, we have to be more deliberate and calculated in the picks we do. I think in the last five years, most of us have said, "We'll do all of it." You know. >> Well you made a good bet with Spark, you know, the duke was pretty obvious trend that was, everyone was shut on that bandwagon and you guys picked a big bet with Spark. Look what happened with you guys? So again, I love this betting kind of concept because as the world matures, growth slows down and shifts and that next wave of value coming in, AKA customers, they're going to integrate with a new ecosystem. A new kind of partner network for AWS and the other clouds. But with aws they're going to need to nurture the next Databricks. They're going to need to still provide that SaaS, ISV like experience for, you know, a basic software hosting or some application. But I go to get your thoughts on this idea of multiple clouds because if I'm a developer, the old days was, old days, within our decade, full stack developer- >> It was two years ago, yeah (John laughing) >> This is a decade ago, full stack and then the cloud came in, you kind had the half stack and then you would do some things. It seems like the clouds are trying to say, we want to be the full stack or not. Or is it still going to be, you know, I'm an application like a PC and a Mac, I'm going to write the same application for both hardware. I mean what's your take on this? Are they trying to do full stack and you see them more like- >> Absolutely. I mean look, of course they're going, they have, I mean they have over 300, I think Amazon has over 300 services, right? That's not just compute, storage, networking, it's the whole stack, right? But my key point is, I think they have to nail the core infrastructure storage compute networking because the three clouds that are there competing, they're formidable companies with formidable balance sheets and it doesn't look like any of them is going to throw in the towel and say, we give up. So I think it's going to intensify. And given that they have a 70% revenue on that infrastructure layer, I think they, if they have to pick their bets, I think they'll focus it on that infrastructure layer. I think the layer above where they're also placing bets, they're doing that, the full stack, right? But there I think the demand will be, can you make that work on the other clouds? And therein lies an innovator's dilemma because if I make it work on the other clouds, then I'm foregoing that 70% revenue of the infrastructure. I'm not getting it. The other cloud vendor is going to get it. So should I do that or not? Second, is the other cloud vendor going to be welcoming of me making my service work on their cloud if I am a competing cloud, right? And what kind of terms of service are I giving me? And am I going to really invest in doing that? And I think right now we, you know, most, the vast, vast, vast majority of the services only work on the one cloud that you know, it's built on. It doesn't work on others, but this will shift. >> Yeah, I think the innovators dilemma is also very good point. And also add, it's an integrators dilemma too because now you talk about integration across services. So I believe that the super cloud movement's going to happen before Sky. And I think what explained by that, what you guys did and what other companies are doing by representing advanced, I call platform engineering, refactoring an existing market really fast, time to value and CAPEX is, I mean capital, market cap is going to be really fast. I think there's going to be an opportunity for those to emerge that's going to set the table for global multicloud ultimately in the future. So I think you're going to start to see the same pattern of what you guys did get in, leverage the hell out of it, use it, not in the way just to host, but to refactor and take down territory of markets. So number one, and then ultimately you get into, okay, I want to run some SLA across services, then there's a little bit more complication. I think that's where you guys put that beautiful paper out on Sky Computing. Okay, that makes sense. Now if you go to today's market, okay, I'm betting on Amazon because they're the best, this is the best cloud win scenario, not the most robust cloud. So if I'm a developer, I want the best. How do you look at their bet when it comes to data? Because now they've got machine learning, Swami's got a big keynote on Wednesday, I'm expecting to see a lot of AI and machine learning. I'm expecting to hear an end to end data story. This is what you do, so as a major partner, how do you view the moves Amazon's making and the bets they're making with data and machine learning and AI? >> First I want to lift off my hat to AWS for being customer obsessed. So I know that if a customer wants Databricks, I know that AWS and their sellers will actually help us get that customer deploy Databricks. Now which of the services is the customer going to pick? Are they going to pick ours or the end to end, what Swami is going to present on stage? Right? So that's the question we're getting. But I wanted to start with by just saying, their customer obsessed. So I think they're going to do the right thing for the customer and I see the evidence of it again and again and again. So kudos to them. They're amazing at this actually. Ultimately our bet is, customers want this to be simple, integrated, okay? So yes there are hundreds of services that together give you the end to end experience and they're very customizable that AWS gives you. But if you want just something simply integrated that also works across the clouds, then I think there's a special place for Databricks. And I think the lake house approach that we have, which is an integrated, completely integrated, we integrate data lakes with data warehouses, integrate workflows with machine learning, with real time processing, all these in one platform. I think there's going to be tailwinds because I think the most important thing that's going to happen in the next few years is that every customer is going to now be obsessed, given the recession and the environment we're in. How do I cut my costs? How do I cut my costs? And we learn this from the customers they're adopting the lake house because they're thinking, instead of using five vendors or three vendors, I can simplify it down to one with you and I can cut my cost. So I think that's going to be one of the main drivers of why people bet on the lake house because it helps them lower their TCO; Total Cost of Ownership. And it's as simple as that. Like I have three things right now. If I can get the same job done of those three with one, I'd rather do that. And by the way, if it's three or four across two clouds and I can just use one and it just works across two clouds, I'm going to do that. Because my boss is telling me I need to cut my budget. >> (indistinct) (John laughing) >> Yeah, and I'd rather not to do layoffs and they're asking me to do more. How can I get smaller budgets, not lay people off and do more? I have to cut, I have to optimize. What's happened in the last five, six years is there's been a huge sprawl of services and startups, you know, you know most of them, all these startups, all of them, all the activity, all the VC investments, well those companies sold their software, right? Even if a startup didn't make it big, you know, they still sold their software to some vendors. So the ecosystem is now full of lots and lots and lots and lots of different software. And right now people are looking, how do I consolidate, how do I simplify, how do I cut my costs? >> And you guys have a great solution. You're also an arms dealer and a innovator. So I have to ask this question, because you're a professor of the industry as well as at Berkeley, you've seen a lot of the historical innovations. If you look at the moment we're in right now with the recession, okay we had COVID, okay, it changed how people work, you know, people working at home, provisioning VLAN, all that (indistinct) infrastructure, okay, yeah, technology and cloud health. But we're in a recession. This is the first recession where the Amazon and the other cloud, mainly Amazon Web Services is a major economic puzzle in the piece. So they were never around before, even 2008, they were too small. They're now a major economic enabler, player, they're serving startups, enterprises, they have super clouds like you guys. They're a force and the people, their customers are cutting back but also they can also get faster. So agility is now an equation in the economic recovery. And I want to get your thoughts because you just brought that up. Customers can actually use the cloud and Databricks to actually get out of the recovery because no one's going to say, stop making profit or make more profit. So yeah, cut costs, be more efficient, but agility's also like, let's drive more revenue. So in this digital transformation, if you take this to conclusion, every company transforms, their company is the app. So their revenue is tied directly to their technology deployment. What's your reaction and comment to that because this is a new historical moment where cloud and scale and data, actually could be configured in a way to actually change the nature of a business in such a short time. And with the recession looming, no one's got time to wait. >> Yeah, absolutely. Look, the secular tailwind in the market is that of, you know, 10 years ago it was software is eating the world, now it's AI's going to eat all of software software. So more and more we're going to have, wherever you have software, which is everywhere now because it's eaten the world, it's going to be eaten up by AI and data. You know, AI doesn't exist without data so they're synonymous. You can't do machine learning if you don't have data. So yeah, you're going to see that everywhere and that automation will help people simplify things and cut down the costs and automate more things. And in the cloud you can also do that by changing your CAPEX to OPEX. So instead of I invest, you know, 10 million into a data center that I buy, I'm going to have headcount to manage the software. Why don't we change this to OPEX? And then they are going to optimize it. They want to lower the TCO because okay, it's in the cloud. but I do want the costs to be much lower that what they were in the previous years. Last five years, nobody cared. Who cares? You know what it costs. You know, there's a new brave world out there. Now there's like, no, it has to be efficient. So I think they're going to optimize it. And I think this lake house approach, which is an integration of the lakes and the warehouse, allows you to rationalize the two and simplify them. It allows you to basically rationalize away the data warehouse. So I think much faster we're going to see the, why do I need the data warehouse? If I can get the same thing done with the lake house for fraction of the cost, that's what's going to happen. I think there's going to be focus on that simplification. But I agree with you. Ultimately everyone knows, everybody's a software company. Every company out there is a software company and in the next 10 years, all of them are also going to be AI companies. So that is going to continue. >> (indistinct), dev's going to stop. And right sizing right now is a key economic forcing function. Final question for you and I really appreciate you taking the time. This year Reinvent, what's the bumper sticker in your mind around what's the most important industry dynamic, power dynamic, ecosystem dynamic that people should pay attention to as we move from the brave new world of okay, I see cloud, cloud operations. I need to really make it structurally change my business. How do I, what's the most important story? What's the bumper sticker in your mind for Reinvent? >> Bumper sticker? lake house 24. (John laughing) >> That's data (indistinct) bumper sticker. What's the- >> (indistinct) in the market. No, no, no, no. You know, it's, AWS talks about, you know, all of their services becoming a lake house because they want the center of the gravity to be S3, their lake. And they want all the services to directly work on that, so that's a lake house. We're Bumper see Microsoft with Synapse, modern, you know the modern intelligent data platform. Same thing there. We're going to see the same thing, we already seeing it on GCP with Big Lake and so on. So I actually think it's the how do I reduce my costs and the lake house integrates those two. So that's one of the main ways you can rationalize and simplify. You get in the lake house, which is the name itself is a (indistinct) of two things, right? Lake house, "lake" gives you the AI, "house" give you the database data warehouse. So you get your AI and you get your data warehousing in one place at the lower cost. So for me, the bumper sticker is lake house, you know, 24. >> All right. Awesome Ali, well thanks for the exclusive interview. Appreciate it and get to see you. Congratulations on your success and I know you guys are going to be fine. >> Awesome. Thank you John. It's always a pleasure. >> Always great to chat with you again. >> Likewise. >> You guys are a great team. We're big fans of what you guys have done. We think you're an example of what we call "super cloud." Which is getting the hype up and again your paper speaks to some of the innovation, which I agree with by the way. I think that that approach of not forcing standards is really smart. And I think that's absolutely correct, that having the market still innovate is going to be key. standards with- >> Yeah, I love it. We're big fans too, you know, you're doing awesome work. We'd love to continue the partnership. >> So, great, great Ali, thanks. >> Take care (outro music)
SUMMARY :
after the keynotes prior to the keynotes and you know, we're because you have customers. I wouldn't, you know, I got to give you guys credit over there So if the customer said we So Adam and the team at So in the next five years, But I think to me, I'm a cloud vendor, and calculated in the picks we do. But I go to get your thoughts on this idea Or is it still going to be, you know, And I think right now we, you know, So I believe that the super cloud I can simplify it down to one with you and startups, you know, and the other cloud, And in the cloud you can also do that I need to really make it lake house 24. That's data (indistinct) of the gravity to be S3, and I know you guys are going to be fine. It's always a pleasure. We're big fans of what you guys have done. We're big fans too, you know,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amazon | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
Ali Ghodsi | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
2013 | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
Alibaba | ORGANIZATION | 0.99+ |
2008 | DATE | 0.99+ |
five vendors | QUANTITY | 0.99+ |
Adam Saleski | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
Ali | PERSON | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
three vendors | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
Wednesday | DATE | 0.99+ |
Excel | TITLE | 0.99+ |
38 billion | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
Word | TITLE | 0.99+ |
three | QUANTITY | 0.99+ |
two clouds | QUANTITY | 0.99+ |
Andy | PERSON | 0.99+ |
three clouds | QUANTITY | 0.99+ |
10 million | QUANTITY | 0.99+ |
PowerPoint | TITLE | 0.99+ |
one | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
twice | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
over 300 services | QUANTITY | 0.99+ |
one game | QUANTITY | 0.99+ |
second cloud | QUANTITY | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
Sky | ORGANIZATION | 0.99+ |
one word | QUANTITY | 0.99+ |
OPEX | ORGANIZATION | 0.99+ |
two things | QUANTITY | 0.98+ |
two years ago | DATE | 0.98+ |
Access | TITLE | 0.98+ |
over 300 | QUANTITY | 0.98+ |
six years | QUANTITY | 0.98+ |
over 70% | QUANTITY | 0.98+ |
five years ago | DATE | 0.98+ |
Ali Ghosdi, Databricks | AWS Partner Exclusive
(outro music) >> Hey, I'm John Furrier, here with an exclusive interview with Ali Ghodsi, who's the CEO of Databricks. Ali, great to see you. Preview for reinvent. We're going to launch this story, exclusive Databricks material on the notes, after the keynotes prior to the keynotes and after the keynotes that reinvent. So great to see you. You know, you've been a partner of AWS for a very, very long time. I think five years ago, I think I first interviewed you, you were one of the first to publicly declare that this was a place to build a company on and not just post an application, but refactor capabilities to create, essentially a platform in the cloud, on the cloud. Not just an ISV; Independent Software Vendor, kind of an old term, we're talking about real platform like capability to change the game. Can you talk about your experience as an AWS partner? >> Yeah, look, so we started in 2013. I swiped my personal credit card on AWS and some of my co-founders did the same. And we started building. And we were excited because we just thought this is a much better way to launch a company because you can just much faster get time to market and launch your thing and you can get the end users much quicker access to the thing you're building. So we didn't really talk to anyone at AWS, we just swiped a credit card. And eventually they told us, "Hey, do you want to buy extra support?" "You're asking a lot of advanced questions from us." "Maybe you want to buy our advanced support." And we said, no, no, no, no. We're very advanced ourselves, we know what we're doing. We're not going to buy any advanced support. So, you know, we just built this, you know, startup from nothing on AWS without even talking to anyone there. So at some point, I think around 2017, they suddenly saw this company with maybe a hundred million ARR pop up on their radar and it's driving massive amounts of compute, massive amounts of data. And it took a little bit in the beginning just us to get to know each other because as I said, it's like we were not on their radar and we weren't really looking, we were just doing our thing. And then over the years the partnership has deepened and deepened and deepened and then with, you know, Andy (indistinct) really leaning into the partnership, he mentioned us at Reinvent. And then we sort of figured out a way to really integrate the two service, the Databricks platform with AWS . And today it's an amazing partnership. You know, we directly connected with the general managers for the services. We're connected at the CEO level, you know, the sellers get compensated for pushing Databricks, we're, we have multiple offerings on their marketplace. We have a native offering on AWS. You know, we're prominently always sort of marketed and you know, we're aligned also vision wise in what we're trying to do. So yeah, we've come a very, very long way. >> Do you consider yourself a SaaS app or an ISV or do you see yourself more of a platform company because you have customers. How would you categorize your category as a company? >> Well, it's a data platform, right? And actually the, the strategy of the Databricks is take what's otherwise five, six services in the industry or five, six different startups, but do them as part of one data platform that's integrated. So in one word, the strategy of data bricks is "unification." We call it the data lake house. But really the idea behind the data lake house is that of unification, or in more words it's, "The whole is greater than the sum of its parts." So you could actually go and buy five, six services out there or actually use five, six services from the cloud vendors, stitch it together and it kind of resembles Databricks. Our power is in doing those integrated, together in a way in which it's really, really easy and simple to use for end users. So yeah, we're a data platform. I wouldn't, you know, ISV that's a old term, you know, Independent Software Vendor. You know, I think, you know, we have actually a whole slew of ISVs on top of Databricks, that integrate with our platform. And you know, in our marketplace as well as in our partner connect, we host those ISVs that then, you know, work on top of the data that we have in the Databricks, data lake house. >> You know, I think one of the things your journey has been great to document and watch from the beginning. I got to give you guys credit over there and props, congratulations. But I think you're the poster child as a company to what we see enterprises doing now. So go back in time when you guys swiped a credit card, you didn't need attending technical support because you guys had brains, you were refactoring, rethinking. It wasn't just banging out software, you had, you were doing some complex things. It wasn't like it was just write some software hosted on server. It was really a lot more. And as a result your business worth billions of dollars. I think 38 billion or something like that, big numbers, big numbers of great revenue growth as well, billions in revenue. You have customers, you have an ecosystem, you have data applications on top of Databricks. So in a way you're a cloud on top of the cloud. So is there a cloud on top of the cloud? So you have ISVs, Amazon has ISVs. Can you take us through what this means and at this point in history, because this seems to be an advanced version of benefits of platforming and refactoring, leveraging say AWS. >> Yeah, so look, when we started, there was really only one game in town. It was AWS. So it was one cloud. And the strategy of the company then was, well Amazon had this beautiful set of services that they're building bottom up, they have storage, compute, networking, and then they have databases and so on. But it's a lot of services. So let us not directly compete with AWS and try to take out one of their services. Let's not do that because frankly we can't. We were not of that size. They had the scale, they had the size and they were the only cloud vendor in town. So our strategy instead was, let's do something else. Let's not compete directly with say, a particular service they're building, let's take a different strategy. What if we had a unified holistic data platform, where it's just one integrated service end to end. So think of it as Microsoft office, which contains PowerPoint, and Word, and Excel and even Access, if you want to use it. What if we build that and AWS has this really amazing knack for releasing things, you know services, lots of them, every reinvent. And they're sort of a DevOps person's dream and you can stitch these together and you know you have to be technical. How do we elevate that and make it simpler and integrate it? That was our original strategy and it resonated with a segment of the market. And the reason it worked with AWS so that we wouldn't butt heads with AWS was because we weren't a direct replacement for this service or for that service, we were taking a different approach. And AWS, because credit goes to them, they're so customer obsessed, they would actually do what's right for the customer. So if the customer said we want this unified thing, their sellers would actually say, okay, so then you should use Databricks. So they truly are customer obsessed in that way. And I really mean it, John. Things have changed over the years. They're not the only cloud anymore. You know, Azure is real, GCP is real, there's also Alibaba. And now over 70% of our customers are on more than one cloud. So now what we hear from them is, not only want, do we want a simplified, unified thing, but we want it also to work across the clouds. Because those of them that are seriously considering multiple clouds, they don't want to use a service on cloud one and then use a similar service on cloud two. But it's a little bit different. And now they have to do twice the work to make it work. You know, John, it's hard enough as it is, like it's this data stuff and analytics. It's not a walk in the park, you know. You hire an administrator in the back office that clicks a button and its just, now you're a data driven digital transformed company. It's hard. If you now have to do it again on the second cloud with different set of services and then again on a third cloud with a different set of services. That's very, very costly. So the strategy then has changed that, how do we take that unified simple approach and make it also the same and standardize across the clouds, but then also integrate it as far down as we can on each of the clouds. So that you're not giving up any of the benefits that the particular cloud has. >> Yeah, I think one of the things that we see, and I want get your reaction to this, is this rise of the super cloud as we call it. I think you were involved in the Sky paper that I saw your position paper came out after we had introduced Super Cloud, which is great. Congratulations to the Berkeley team, wearing the hat here. But you guys are, I think a driver of this because you're creating the need for these things. You're saying, okay, we went on one cloud with AWS and you didn't hide that. And now you're publicly saying there's other clouds too, increased ham for your business. And customers have multiple clouds in their infrastructure for the best of breed that they have. Okay, get that. But there's still a challenge around the innovation, growth that's still around the corner. We still have a supply chain problem, we still have skill gaps. You know, you guys are unique at Databricks as other these big examples of super clouds that are developing. Enterprises don't have the Databricks kind of talent. They need, they need turnkey solutions. So Adam and the team at Amazon are promoting, you know, more solution oriented approaches higher up on the stack. You're starting to see kind of like, I won't say templates, but you know, almost like application specific headless like, low code, no code capability to accelerate clients who are wanting to write code for the modern error. Right, so this kind of, and then now you, as you guys pointed out with these common services, you're pushing the envelope. So you're saying, hey, I need to compete, I don't want to go to my customers and have them to have a staff or this cloud and this cloud and this cloud because they don't have the staff. Or if they do, they're very unique. So what's your reaction? Because this kind is the, it kind of shows your leadership as a partner of AWS and the clouds, but also highlights I think what's coming. But you share your reaction. >> Yeah, look, it's, first of all, you know, I wish I could take credit for this but I can't because it's really the customers that have decided to go on multiple clouds. You know, it's not Databricks that you know, push this or some other vendor, you know, that, Snowflake or someone who pushed this and now enterprises listened to us and they picked two clouds. That's not how it happened. The enterprises picked two clouds or three clouds themselves and we can get into why, but they did that. So this largely just happened in the market. We as data platforms responded to what they're then saying, which is they're saying, "I don't want to redo this again on the other cloud." So I think the writing is on the wall. I think it's super obvious what's going to happen next. They will say, "Any service I'm using, it better work exactly the same on all the clouds." You know, that's what's going to happen. So in the next five years, every enterprise will say, "I'm going to use the service, but you better make sure that this service works equally well on all of the clouds." And obviously the multicloud vendors like us, are there to do that. But I actually think that what you're going to see happening is that you're going to see the cloud vendors changing the existing services that they have to make them work on the other clouds. That's what's goin to happen, I think. >> Yeah, and I think I would add that, first of all, I agree with you. I think that's going to be a forcing function. Because I think you're driving it. You guys are in a way, one, are just an actor in the driving this because you're on the front end of this and there are others and there will be people following. But I think to me, I'm a cloud vendor, I got to differentiate. Adam, If I'm Adam Saleski, I got to say, "Hey, I got to differentiate." So I don't wan to get stuck in the middle, so to speak. Am I just going to innovate on the hardware AKA infrastructure or am I going to innovate at the higher level services? So what we're talking about here is the tail of two clouds within Amazon, for instance. So do I innovate on the silicon and get low level into the physics and squeeze performance out of the hardware and infrastructure? Or do I focus on ease of use at the top of the stack for the developers? So again, there's a channel of two clouds here. So I got to ask you, how do they differentiate? Number one and number two, I never heard a developer ever say, "I want to run my app or workload on the slower cloud." So I mean, you know, back when we had PCs you wanted to go, "I want the fastest processor." So again, you can have common level services, but where is that performance differentiation with the cloud? What do the clouds do in your opinion? >> Yeah, look, I think it's pretty clear. I think that it's, this is, you know, no surprise. Probably 70% or so of the revenue is in the lower infrastructure layers, compute, storage, networking. And they have to win that. They have to be competitive there. As you said, you can say, oh you know, I guess my CPUs are slower than the other cloud, but who cares? I have amazing other services which only work on my cloud by the way, right? That's not going to be a winning recipe. So I think all three are laser focused on, we going to have specialized hardware and the nuts and bolts of the infrastructure, we can do it better than the other clouds for sure. And you can see lots of innovation happening there, right? The Graviton chips, you know, we see huge price performance benefits in those chips. I mean it's real, right? It's basically a 20, 30% free lunch. You know, why wouldn't you, why wouldn't you go for it there? There's no downside. You know, there's no, "got you" or no catch. But we see Azure doing the same thing now, they're also building their own chips and we know that Google builds specialized machine learning chips, TPU, Tenor Processing Units. So their legs are focused on that. I don't think they can give up that or focused on higher levels if they had to pick bets. And I think actually in the next few years, most of us have to make more, we have to be more deliberate and calculated in the picks we do. I think in the last five years, most of us have said, "We'll do all of it." You know. >> Well you made a good bet with Spark, you know, the duke was pretty obvious trend that was, everyone was shut on that bandwagon and you guys picked a big bet with Spark. Look what happened with you guys? So again, I love this betting kind of concept because as the world matures, growth slows down and shifts and that next wave of value coming in, AKA customers, they're going to integrate with a new ecosystem. A new kind of partner network for AWS and the other clouds. But with aws they're going to need to nurture the next Databricks. They're going to need to still provide that SaaS, ISV like experience for, you know, a basic software hosting or some application. But I go to get your thoughts on this idea of multiple clouds because if I'm a developer, the old days was, old days, within our decade, full stack developer- >> It was two years ago, yeah (John laughing) >> This is a decade ago, full stack and then the cloud came in, you kind had the half stack and then you would do some things. It seems like the clouds are trying to say, we want to be the full stack or not. Or is it still going to be, you know, I'm an application like a PC and a Mac, I'm going to write the same application for both hardware. I mean what's your take on this? Are they trying to do full stack and you see them more like- >> Absolutely. I mean look, of course they're going, they have, I mean they have over 300, I think Amazon has over 300 services, right? That's not just compute, storage, networking, it's the whole stack, right? But my key point is, I think they have to nail the core infrastructure storage compute networking because the three clouds that are there competing, they're formidable companies with formidable balance sheets and it doesn't look like any of them is going to throw in the towel and say, we give up. So I think it's going to intensify. And given that they have a 70% revenue on that infrastructure layer, I think they, if they have to pick their bets, I think they'll focus it on that infrastructure layer. I think the layer above where they're also placing bets, they're doing that, the full stack, right? But there I think the demand will be, can you make that work on the other clouds? And therein lies an innovator's dilemma because if I make it work on the other clouds, then I'm foregoing that 70% revenue of the infrastructure. I'm not getting it. The other cloud vendor is going to get it. So should I do that or not? Second, is the other cloud vendor going to be welcoming of me making my service work on their cloud if I am a competing cloud, right? And what kind of terms of service are I giving me? And am I going to really invest in doing that? And I think right now we, you know, most, the vast, vast, vast majority of the services only work on the one cloud that you know, it's built on. It doesn't work on others, but this will shift. >> Yeah, I think the innovators dilemma is also very good point. And also add, it's an integrators dilemma too because now you talk about integration across services. So I believe that the super cloud movement's going to happen before Sky. And I think what explained by that, what you guys did and what other companies are doing by representing advanced, I call platform engineering, refactoring an existing market really fast, time to value and CAPEX is, I mean capital, market cap is going to be really fast. I think there's going to be an opportunity for those to emerge that's going to set the table for global multicloud ultimately in the future. So I think you're going to start to see the same pattern of what you guys did get in, leverage the hell out of it, use it, not in the way just to host, but to refactor and take down territory of markets. So number one, and then ultimately you get into, okay, I want to run some SLA across services, then there's a little bit more complication. I think that's where you guys put that beautiful paper out on Sky Computing. Okay, that makes sense. Now if you go to today's market, okay, I'm betting on Amazon because they're the best, this is the best cloud win scenario, not the most robust cloud. So if I'm a developer, I want the best. How do you look at their bet when it comes to data? Because now they've got machine learning, Swami's got a big keynote on Wednesday, I'm expecting to see a lot of AI and machine learning. I'm expecting to hear an end to end data story. This is what you do, so as a major partner, how do you view the moves Amazon's making and the bets they're making with data and machine learning and AI? >> First I want to lift off my hat to AWS for being customer obsessed. So I know that if a customer wants Databricks, I know that AWS and their sellers will actually help us get that customer deploy Databricks. Now which of the services is the customer going to pick? Are they going to pick ours or the end to end, what Swami is going to present on stage? Right? So that's the question we're getting. But I wanted to start with by just saying, their customer obsessed. So I think they're going to do the right thing for the customer and I see the evidence of it again and again and again. So kudos to them. They're amazing at this actually. Ultimately our bet is, customers want this to be simple, integrated, okay? So yes there are hundreds of services that together give you the end to end experience and they're very customizable that AWS gives you. But if you want just something simply integrated that also works across the clouds, then I think there's a special place for Databricks. And I think the lake house approach that we have, which is an integrated, completely integrated, we integrate data lakes with data warehouses, integrate workflows with machine learning, with real time processing, all these in one platform. I think there's going to be tailwinds because I think the most important thing that's going to happen in the next few years is that every customer is going to now be obsessed, given the recession and the environment we're in. How do I cut my costs? How do I cut my costs? And we learn this from the customers they're adopting the lake house because they're thinking, instead of using five vendors or three vendors, I can simplify it down to one with you and I can cut my cost. So I think that's going to be one of the main drivers of why people bet on the lake house because it helps them lower their TCO; Total Cost of Ownership. And it's as simple as that. Like I have three things right now. If I can get the same job done of those three with one, I'd rather do that. And by the way, if it's three or four across two clouds and I can just use one and it just works across two clouds, I'm going to do that. Because my boss is telling me I need to cut my budget. >> (indistinct) (John laughing) >> Yeah, and I'd rather not to do layoffs and they're asking me to do more. How can I get smaller budgets, not lay people off and do more? I have to cut, I have to optimize. What's happened in the last five, six years is there's been a huge sprawl of services and startups, you know, you know most of them, all these startups, all of them, all the activity, all the VC investments, well those companies sold their software, right? Even if a startup didn't make it big, you know, they still sold their software to some vendors. So the ecosystem is now full of lots and lots and lots and lots of different software. And right now people are looking, how do I consolidate, how do I simplify, how do I cut my costs? >> And you guys have a great solution. You're also an arms dealer and a innovator. So I have to ask this question, because you're a professor of the industry as well as at Berkeley, you've seen a lot of the historical innovations. If you look at the moment we're in right now with the recession, okay we had COVID, okay, it changed how people work, you know, people working at home, provisioning VLAN, all that (indistinct) infrastructure, okay, yeah, technology and cloud health. But we're in a recession. This is the first recession where the Amazon and the other cloud, mainly Amazon Web Services is a major economic puzzle in the piece. So they were never around before, even 2008, they were too small. They're now a major economic enabler, player, they're serving startups, enterprises, they have super clouds like you guys. They're a force and the people, their customers are cutting back but also they can also get faster. So agility is now an equation in the economic recovery. And I want to get your thoughts because you just brought that up. Customers can actually use the cloud and Databricks to actually get out of the recovery because no one's going to say, stop making profit or make more profit. So yeah, cut costs, be more efficient, but agility's also like, let's drive more revenue. So in this digital transformation, if you take this to conclusion, every company transforms, their company is the app. So their revenue is tied directly to their technology deployment. What's your reaction and comment to that because this is a new historical moment where cloud and scale and data, actually could be configured in a way to actually change the nature of a business in such a short time. And with the recession looming, no one's got time to wait. >> Yeah, absolutely. Look, the secular tailwind in the market is that of, you know, 10 years ago it was software is eating the world, now it's AI's going to eat all of software software. So more and more we're going to have, wherever you have software, which is everywhere now because it's eaten the world, it's going to be eaten up by AI and data. You know, AI doesn't exist without data so they're synonymous. You can't do machine learning if you don't have data. So yeah, you're going to see that everywhere and that automation will help people simplify things and cut down the costs and automate more things. And in the cloud you can also do that by changing your CAPEX to OPEX. So instead of I invest, you know, 10 million into a data center that I buy, I'm going to have headcount to manage the software. Why don't we change this to OPEX? And then they are going to optimize it. They want to lower the TCO because okay, it's in the cloud. but I do want the costs to be much lower that what they were in the previous years. Last five years, nobody cared. Who cares? You know what it costs. You know, there's a new brave world out there. Now there's like, no, it has to be efficient. So I think they're going to optimize it. And I think this lake house approach, which is an integration of the lakes and the warehouse, allows you to rationalize the two and simplify them. It allows you to basically rationalize away the data warehouse. So I think much faster we're going to see the, why do I need the data warehouse? If I can get the same thing done with the lake house for fraction of the cost, that's what's going to happen. I think there's going to be focus on that simplification. But I agree with you. Ultimately everyone knows, everybody's a software company. Every company out there is a software company and in the next 10 years, all of them are also going to be AI companies. So that is going to continue. >> (indistinct), dev's going to stop. And right sizing right now is a key economic forcing function. Final question for you and I really appreciate you taking the time. This year Reinvent, what's the bumper sticker in your mind around what's the most important industry dynamic, power dynamic, ecosystem dynamic that people should pay attention to as we move from the brave new world of okay, I see cloud, cloud operations. I need to really make it structurally change my business. How do I, what's the most important story? What's the bumper sticker in your mind for Reinvent? >> Bumper sticker? lake house 24. (John laughing) >> That's data (indistinct) bumper sticker. What's the- >> (indistinct) in the market. No, no, no, no. You know, it's, AWS talks about, you know, all of their services becoming a lake house because they want the center of the gravity to be S3, their lake. And they want all the services to directly work on that, so that's a lake house. We're Bumper see Microsoft with Synapse, modern, you know the modern intelligent data platform. Same thing there. We're going to see the same thing, we already seeing it on GCP with Big Lake and so on. So I actually think it's the how do I reduce my costs and the lake house integrates those two. So that's one of the main ways you can rationalize and simplify. You get in the lake house, which is the name itself is a (indistinct) of two things, right? Lake house, "lake" gives you the AI, "house" give you the database data warehouse. So you get your AI and you get your data warehousing in one place at the lower cost. So for me, the bumper sticker is lake house, you know, 24. >> All right. Awesome Ali, well thanks for the exclusive interview. Appreciate it and get to see you. Congratulations on your success and I know you guys are going to be fine. >> Awesome. Thank you John. It's always a pleasure. >> Always great to chat with you again. >> Likewise. >> You guys are a great team. We're big fans of what you guys have done. We think you're an example of what we call "super cloud." Which is getting the hype up and again your paper speaks to some of the innovation, which I agree with by the way. I think that that approach of not forcing standards is really smart. And I think that's absolutely correct, that having the market still innovate is going to be key. standards with- >> Yeah, I love it. We're big fans too, you know, you're doing awesome work. We'd love to continue the partnership. >> So, great, great Ali, thanks. >> Take care (outro music)
SUMMARY :
after the keynotes prior to the keynotes and you know, we're because you have customers. I wouldn't, you know, I got to give you guys credit over there So if the customer said we So Adam and the team at So in the next five years, But I think to me, I'm a cloud vendor, and calculated in the picks we do. But I go to get your thoughts on this idea Or is it still going to be, you know, And I think right now we, you know, So I believe that the super cloud I can simplify it down to one with you and startups, you know, and the other cloud, And in the cloud you can also do that I need to really make it lake house 24. That's data (indistinct) of the gravity to be S3, and I know you guys are going to be fine. It's always a pleasure. We're big fans of what you guys have done. We're big fans too, you know,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Ali Ghodsi | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
2013 | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
Alibaba | ORGANIZATION | 0.99+ |
2008 | DATE | 0.99+ |
Ali Ghosdi | PERSON | 0.99+ |
five vendors | QUANTITY | 0.99+ |
Adam Saleski | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
Ali | PERSON | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
three vendors | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
Wednesday | DATE | 0.99+ |
Excel | TITLE | 0.99+ |
38 billion | QUANTITY | 0.99+ |
four | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
Word | TITLE | 0.99+ |
three | QUANTITY | 0.99+ |
two clouds | QUANTITY | 0.99+ |
Andy | PERSON | 0.99+ |
three clouds | QUANTITY | 0.99+ |
10 million | QUANTITY | 0.99+ |
PowerPoint | TITLE | 0.99+ |
one | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
twice | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
over 300 services | QUANTITY | 0.99+ |
one game | QUANTITY | 0.99+ |
second cloud | QUANTITY | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
Sky | ORGANIZATION | 0.99+ |
one word | QUANTITY | 0.99+ |
OPEX | ORGANIZATION | 0.99+ |
two things | QUANTITY | 0.98+ |
two years ago | DATE | 0.98+ |
Access | TITLE | 0.98+ |
over 300 | QUANTITY | 0.98+ |
six years | QUANTITY | 0.98+ |
over 70% | QUANTITY | 0.98+ |
five years ago | DATE | 0.98+ |
Accelerating Business Transformation with VMware Cloud on AWS 10 31
>>Hi everyone. Welcome to the Cube special presentation here in Palo Alto, California. I'm John Foer, host of the Cube. We've got two great guests, one for calling in from Germany, our videoing in from Germany, one from Maryland. We've got VMware and aws. This is the customer successes with VMware cloud on AWS showcase, accelerating business transformation here in the showcase with Samir Candu Worldwide. VMware strategic alliance solution, architect leader with AWS Samir. Great to have you and Daniel Re Myer, principal architect global AWS synergy at VMware. Guys, you guys are, are working together. You're the key players in the re relationship as it rolls out and continues to grow. So welcome to the cube. >>Thank you. Greatly appreciate it. >>Great to have you guys both on, As you know, we've been covering this since 2016 when Pat Geling, then CEO and then then CEO AWS at Andy Chasy did this. It kind of got people by surprise, but it really kind of cleaned out the positioning in the enterprise for the success. OFM workloads in the cloud. VMware's had great success with it since, and you guys have the great partnerships. So this has been like a really strategic, successful partnership. Where are we right now? You know, years later we got this whole inflection point coming. You're starting to see, you know, this idea of higher level services, more performance are coming in at the infrastructure side. More automation, more serverless, I mean, and a, I mean it's just getting better and better every year in the cloud. Kinda a whole nother level. Where are we, Samir? Let's start with you on, on the relationship. >>Yeah, totally. So I mean, there's several things to keep in mind, right? So in 2016, right, that's when the partnership between AWS and VMware was announced, and then less than a year later, that's when we officially launched VMware cloud on aws. Years later, we've been driving innovation, working with our customers, jointly engineering this between AWS and VMware day in, day out. As far as advancing VMware cloud on aws. You know, even if you look at the innovation that takes place with a solution, things have modernized, things have changed, there's been advancements, you know, whether it's security focus, whether it's platform focus, whether it's networking focus, there's been modifications along the way, even storage, right? More recently, one of the things to keep in mind is we're looking to deliver value to our customers together. These are our joint customers. So there's hundreds of VMware and AWS engineers working together on this solution. >>And then factor in even our sales teams, right? We have VMware and AWS sales teams interacting with each other on a constant daily basis. We're working together with our customers at the end of the day too. Then we're looking to even offer and develop jointly engineered solutions specific to VMware cloud on aws, and even with VMware's, other platforms as well. Then the other thing comes down to is where we have dedicated teams around this at both AWS and VMware. So even from solutions architects, even to our sales specialists, even to our account teams, even to specific engineering teams within the organizations, they all come together to drive this innovation forward with VMware cloud on AWS and the jointly engineered solution partnership as well. And then I think one of the key things to keep in mind comes down to we have nearly 600 channel partners that have achieved VMware cloud on AWS service competency. So think about it from the standpoint there's 300 certified or validated technology solutions, they're now available to our customers. So that's even innovation right off the top as well. >>Great stuff. Daniel, I wanna get to you in a second. Upon this principal architect position you have in your title, you're the global a synergy person. Synergy means bringing things together, making it work. Take us through the architecture, because we heard a lot of folks at VMware explore this year, formerly world, talking about how the, the workloads on it has been completely transforming into cloud and hybrid, right? This is where the action is. Where are you? Is your customers taking advantage of that new shift? You got AI ops, you got it. Ops changing a lot, you got a lot more automation edges right around the corner. This is like a complete transformation from where we were just five years ago. What's your thoughts on the >>Relationship? So at at, at first, I would like to emphasize that our collaboration is not just that we have dedicated teams to help our customers get the most and the best benefits out of VMware cloud on aws. We are also enabling US mutually. So AWS learns from us about the VMware technology, where VMware people learn about the AWS technology. We are also enabling our channel partners and we are working together on customer projects. So we have regular assembled globally and also virtually on Slack and the usual suspect tools working together and listening to customers, that's, that's very important. Asking our customers where are their needs? And we are driving the solution into the direction that our customers get the, the best benefits out of VMware cloud on aws. And over the time we, we really have involved the solution. As Samia mentioned, we just added additional storage solutions to VMware cloud on aws. We now have three different instance types that cover a broad range of, of workload. So for example, we just added the I four I host, which is ideally for workloads that require a lot of CPU power, such as you mentioned it, AI workloads. >>Yeah. So I wanna guess just specifically on the customer journey and their transformation. You know, we've been reporting on Silicon angle in the queue in the past couple weeks in a big way that the OPS teams are now the new devs, right? I mean that sounds OP a little bit weird, but operation IT operations is now part of the, a lot more data ops, security writing code composing, you know, with open source, a lot of great things are changing. Can you share specifically what customers are looking for when you say, as you guys come in and assess their needs, what are they doing? What are some of the things that they're doing with VMware on AWS specifically that's a little bit different? Can you share some of and highlights there? >>That, that's a great point because originally VMware and AWS came from very different directions when it comes to speaking people at customers. So for example, aws very developer focused, whereas VMware has a very great footprint in the IT ops area. And usually these are very different, very different teams, groups, different cultures, but it's, it's getting together. However, we always try to address the customers, right? There are customers that want to build up a new application from the scratch and build resiliency, availability, recoverability, scalability into the application. But there are still a lot of customers that say, well we don't have all of the skills to redevelop everything to refactor an application to make it highly available. So we want to have all of that as a service, recoverability as a service, scalability as a service. We want to have this from the infrastructure. That was one of the unique selling points for VMware on premise and now we are bringing this into the cloud. >>Samir, talk about your perspective. I wanna get your thoughts, and not to take a tangent, but we had covered the AWS remar of, actually it was Amazon res machine learning automation, robotics and space. It was really kinda the confluence of industrial IOT software physical. And so when you look at like the IT operations piece becoming more software, you're seeing things about automation, but the skill gap is huge. So you're seeing low code, no code automation, you know, Hey Alexa, deploy a Kubernetes cluster. Yeah, I mean, I mean that's coming, right? So we're seeing this kind of operating automation meets higher level services meets workloads. Can you unpack that and share your opinion on, on what you see there from an Amazon perspective and how it relates to this? >>Yeah, totally. Right. And you know, look at it from the point of view where we said this is a jointly engineered solution, but it's not migrating to one option or the other option, right? It's more or less together. So even with VMware cloud on aws, yes it is utilizing AWS infrastructure, but your environment is connected to that AWS VPC in your AWS account. So if you wanna leverage any of the native AWS services, so any of the 200 plus AWS services, you have that option to do so. So that's gonna give you that power to do certain things, such as, for example, like how you mentioned with iot, even with utilizing Alexa or if there's any other service that you wanna utilize, that's the joining point between both of the offerings. Right off the top though, with digital transformation, right? You, you have to think about where it's not just about the technology, right? There's also where you want to drive growth in the underlying technology. Even in your business leaders are looking to reinvent their business. They're looking to take different steps as far as pursuing a new strategy. Maybe it's a process, maybe it's with the people, the culture, like how you said before, where people are coming in from a different background, right? They may not be used to the cloud, they may not be used to AWS services, but now you have that capability to mesh them together. Okay. Then also, Oh, >>Go ahead, finish >>Your thought. No, no, I was gonna say, what it also comes down to is you need to think about the operating model too, where it is a shift, right? Especially for that VS four admin that's used to their on-premises at environment. Now with VMware cloud on aws, you have that ability to leverage a cloud, but the investment that you made and certain things as far as automation, even with monitoring, even with logging, yeah. You still have that methodology where you can utilize that in VMware cloud on AWS two. >>Danielle, I wanna get your thoughts on this because at at explore and, and, and after the event, now as we prep for Cuban and reinvent coming up the big AWS show, I had a couple conversations with a lot of the VMware customers and operators and it's like hundreds of thousands of, of, of, of users and millions of people talking about and and peaked on VM we're interested in v VMware. The common thread was one's one, one person said, I'm trying to figure out where I'm gonna put my career in the next 10 to 15 years. And they've been very comfortable with VMware in the past, very loyal, and they're kind of talking about, I'm gonna be the next cloud, but there's no like role yet architects, is it Solution architect sre. So you're starting to see the psychology of the operators who now are gonna try to make these career decisions, like how, what am I gonna work on? And it's, and that was kind of fuzzy, but I wanna get your thoughts. How would you talk to that persona about the future of VMware on, say, cloud for instance? What should they be thinking about? What's the opportunity and what's gonna happen? >>So digital transformation definitely is a huge change for many organizations and leaders are perfectly aware of what that means. And that also means in, in to to some extent, concerns with your existing employees. Concerns about do I have to relearn everything? Do I have to acquire new skills? And, and trainings is everything worthless I learned over the last 15 years of my career? And the, the answer is to make digital transformation a success. We need not just to talk about technology, but also about process people and culture. And this is where VMware really can help because if you are applying VMware cloud on a, on AWS to your infrastructure, to your existing on-premise infrastructure, you do not need to change many things. You can use the same tools and skills, you can manage your virtual machines as you did in your on-premise environment. You can use the same managing and monitoring tools. If you have written, and many customers did this, if you have developed hundreds of, of scripts that automate tasks and if you know how to troubleshoot things, then you can use all of that in VMware cloud on aws. And that gives not just leaders, but but also the architects at customers, the operators at customers, the confidence in, in such a complex project, >>The consistency, very key point, gives them the confidence to go and, and then now that once they're confident they can start committing themselves to new things. Samir, you're reacting to this because you know, on your side you've got higher level services, you got more performance at the hardware level. I mean, lot improvement. So, okay, nothing's changed. I can still run my job now I got goodness on the other side. What's the upside? What's in it for the, for the, for the customer there? >>Yeah, so I think what it comes down to is they've already been so used to or entrenched with that VMware admin mentality, right? But now extending that to the cloud, that's where now you have that bridge between VMware cloud on AWS to bridge that VMware knowledge with that AWS knowledge. So I will look at it from the point of view where now one has that capability and that ability to just learn about the cloud, but if they're comfortable with certain aspects, no one's saying you have to change anything. You can still leverage that, right? But now if you wanna utilize any other AWS service in conjunction with that VM that resides maybe on premises or even in VMware cloud on aws, you have that option to do so. So think about it where you have that ability to be someone who's curious and wants to learn. And then if you wanna expand on the skills, you certainly have that capability to do so. >>Great stuff. I love, love that. Now that we're peeking behind the curtain here, I'd love to have you guys explain, cuz people wanna know what's goes on in behind the scenes. How does innovation get happen? How does it happen with the relationship? Can you take us through a day in the life of kind of what goes on to make innovation happen with the joint partnership? You guys just have a zoom meeting, Do you guys fly out, you write go do you ship thing? I mean I'm making it up, but you get the idea, what's the, what's, how does it work? What's going on behind the scenes? >>So we hope to get more frequently together in person, but of course we had some difficulties over the last two to three years. So we are very used to zoom conferences and and Slack meetings. You always have to have the time difference in mind if we are working globally together. But what we try, for example, we have reg regular assembled now also in person geo based. So for emia, for the Americas, for aj. And we are bringing up interesting customer situations, architectural bits and pieces together. We are discussing it always to share and to contribute to our community. >>What's interesting, you know, as, as events are coming back to here, before you get, you weigh in, I'll comment, as the cube's been going back out to events, we are hearing comments like what, what pandemic we were more productive in the pandemic. I mean, developers know how to work remotely and they've been on all the tools there, but then they get in person, they're happy to see people, but there's no one's, no one's really missed the beat. I mean it seems to be very productive, you know, workflow, not a lot of disruption. More if anything, productivity gains. >>Agreed, right? I think one of the key things to keep in mind is, you know, even if you look at AWS's and even Amazon's leadership principles, right? Customer obsession, that's key. VMware is carrying that forward as well. Where we are working with our customers, like how Daniel said met earlier, right? We might have meetings at different time zones, maybe it's in person, maybe it's virtual, but together we're working to listen to our customers. You know, we're taking and capturing that feedback to drive innovation and VMware cloud on AWS as well. But one of the key things to keep in mind is yes, there have been, there has been the pandemic, we might have been disconnected to a certain extent, but together through technology we've been able to still communicate work with our customers. Even with VMware in between, with AWS and whatnot. We had that flexibility to innovate and continue that innovation. So even if you look at it from the point of view, right? VMware cloud on AWS outposts, that was something that customers have been asking for. We've been been able to leverage the feedback and then continue to drive innovation even around VMware cloud on AWS outposts. So even with the on premises environment, if you're looking to handle maybe data sovereignty or compliance needs, maybe you have low latency requirements, that's where certain advancements come into play, right? So the key thing is always to maintain that communication track. >>And our last segment we did here on the, on this showcase, we listed the accomplishments and they were pretty significant. I mean go, you got the global rollouts of the relationship. It's just really been interesting and, and people can reference that. We won't get into it here, but I will ask you guys to comment on, as you guys continue to evolve the relationship, what's in it for the customer? What can they expect next? Cuz again, I think right now we're in at a, an inflection point more than ever. What can people expect from the relationship and what's coming up with reinvent? Can you share a little bit of kind of what's coming down the pike? >>So one of the most important things we have announced this year, and we will continue to evolve into that direction, is independent scale of storage. That absolutely was one of the most important items customer asked us for over the last years. Whenever, whenever you are requiring additional storage to host your virtual machines, you usually in VMware cloud on aws, you have to add additional notes. Now we have three different note types with different ratios of compute, storage and memory. But if you only require additional storage, you always have to get also additional compute and memory and you have to pay. And now with two solutions which offer choice for the customers, like FS six one, NetApp onap, and VMware cloud Flex Storage, you now have two cost effective opportunities to add storage to your virtual machines. And that offers opportunities for other instance types maybe that don't have local storage. We are also very, very keen looking forward to announcements, exciting announcements at the upcoming events. >>Samir, what's your, what's your reaction take on the, on what's coming down on your side? >>Yeah, I think one of the key things to keep in mind is, you know, we're looking to help our customers be agile and even scale with their needs, right? So with VMware cloud on aws, that's one of the key things that comes to mind, right? There are gonna be announcements, innovations and whatnot with outcoming events. But together we're able to leverage that to advance VMware cloud on AWS to Daniel's point storage, for example, even with host offerings. And then even with decoupling storage from compute and memory, right now you have the flexibility where you can do all of that. So to look at it from the standpoint where now with 21 regions where we have VMware cloud on AWS available as well, where customers can utilize that as needed when needed, right? So it comes down to, you know, transformation will be there. Yes, there's gonna be maybe where workloads have to be adapted where they're utilizing certain AWS services, but you have that flexibility and option to do so. And I think with the continuing events that's gonna give us the options to even advance our own services together. >>Well you guys are in the middle of it, you're in the trenches, you're making things happen, you've got a team of people working together. My final question is really more of a kind of a current situation, kind of future evolutionary thing that you haven't seen this before. I wanna get both of your reaction to it. And we've been bringing this up in, in the open conversations on the cube is in the old days it was going back this generation, you had ecosystems, you had VMware had an ecosystem they did best, had an ecosystem. You know, we have a product, you have a product, biz dev deals happen, people sign relationships and they do business together and they, they sell to each other's products or do some stuff. Now it's more about architecture cuz we're now in a distributed large scale environment where the role of ecosystems are intertwining. >>And this, you guys are in the middle of two big ecosystems. You mentioned channel partners, you both have a lot of partners on both sides. They come together. So you have this now almost a three dimensional or multidimensional ecosystem, you know, interplay. What's your thoughts on this? And, and, and because it's about the architecture, integration is a value, not so much. Innovation is only, you gotta do innovation, but when you do innovation, you gotta integrate it, you gotta connect it. So what is, how do you guys see this as a, as an architectural thing, start to see more technical business deals? >>So we are, we are removing dependencies from individual ecosystems and from individual vendors. So a customer no longer has to decide for one vendor and then it is a very expensive and high effort project to move away from that vendor, which ties customers even, even closer to specific vendors. We are removing these obstacles. So with VMware cloud on aws moving to the cloud, firstly it's, it's not a dead end. If you decide at one point in time because of latency requirements or maybe it's some compliance requirements, you need to move back into on-premise. You can do this if you decide you want to stay with some of your services on premise and just run a couple of dedicated services in the cloud, you can do this and you can mana manage it through a single pane of glass. That's quite important. So cloud is no longer a dead and it's no longer a binary decision, whether it's on premise or the cloud. It it is the cloud. And the second thing is you can choose the best of both works, right? If you are migrating virtual machines that have been running in your on-premise environment to VMware cloud on aws, by the way, in a very, very fast cost effective and safe way, then you can enrich later on enrich these virtual machines with services that are offered by aws. More than 200 different services ranging from object based storage, load balancing and so on. So it's an endless, endless possibility. >>We, we call that super cloud in, in a, in a way that we be generically defining it where everyone's innovating, but yet there's some common services. But the differentiation comes from innovation where the lock in is the value, not some spec, right? Samir, this is gonna where cloud is right now, you guys are, are not commodity. Amazon's completely differentiating, but there's some commodity things. Having got storage, you got compute, but then you got now advances in all areas. But partners innovate with you on their terms. Absolutely. And everybody wins. >>Yeah. And a hundred percent agree with you. I think one of the key things, you know, as Daniel mentioned before, is where it it, it's a cross education where there might be someone who's more proficient on the cloud side with aws, maybe more proficient with the viewers technology, but then for partners, right? They bridge that gap as well where they come in and they might have a specific niche or expertise where their background, where they can help our customers go through that transformation. So then that comes down to, hey, maybe I don't know how to connect to the cloud. Maybe I don't know what the networking constructs are. Maybe I can leverage that partner. That's one aspect to go about it. Now maybe you migrated that workload to VMware cloud on aws. Maybe you wanna leverage any of the native AWS services or even just off the top 200 plus AWS services, right? But it comes down to that skill, right? So again, solutions architecture at the back of, back of the day, end of the day, what it comes down to is being able to utilize the best of both worlds. That's what we're giving our customers at the end of the >>Day. I mean, I just think it's, it's a, it's a refactoring and innovation opportunity at all levels. I think now more than ever, you can take advantage of each other's ecosystems and partners and technologies and change how things get done with keeping the consistency. I mean, Daniel, you nailed that, right? I mean, you don't have to do anything. You still run the fear, the way you working on it and now do new things. This is kind of a cultural shift. >>Yeah, absolutely. And if, if you look, not every, not every customer, not every organization has the resources to refactor and re-platform everything. And we gave, we give them a very simple and easy way to move workloads to the cloud. Simply run them and at the same time they can free up resources to develop new innovations and, and grow their business. >>Awesome. Samir, thank you for coming on. Danielle, thank you for coming to Germany, Octoberfest, I know it's evening over there, your weekend's here. And thank you for spending the time. Samir final give you the final word, AWS reinvents coming up. Preparing. We're gonna have an exclusive with Adam, but Fry, we do a curtain raise, a dual preview. What's coming down on your side with the relationship and what can we expect to hear about what you got going on at reinvent this year? The big show? >>Yeah, so I think, you know, Daniel hit upon some of the key points, but what I will say is we do have, for example, specific sessions, both that VMware's driving and then also that AWS is driving. We do have even where we have what I call a chalk talks. So I would say, and then even with workshops, right? So even with the customers, the attendees who are there, whatnot, if they're looking for to sit and listen to a session, yes that's there. But if they wanna be hands on, that is also there too. So personally for me as an IT background, you know, been in CIS admin world and whatnot, being hands on, that's one of the key things that I personally am looking forward. But I think that's one of the key ways just to learn and get familiar with the technology. Yeah, >>Reinvents an amazing show for the in person. You guys nail it every year. We'll have three sets this year at the cube. It's becoming popular. We more and more content. You guys got live streams going on, a lot of content, a lot of media, so thanks, thanks for sharing that. Samir Daniel, thank you for coming on on this part of the showcase episode of really the customer successes with VMware Cloud Ons, really accelerating business transformation withs and VMware. I'm John Fur with the cube, thanks for watching. Hello everyone. Welcome to this cube showcase, accelerating business transformation with VMware cloud on it's a solution innovation conversation with two great guests, Fred and VP of commercial services at aws and NA Ryan Bard, who's the VP and general manager of cloud solutions at VMware. Gentlemen, thanks for joining me on this showcase. >>Great to be here. >>Hey, thanks for having us on. It's a great topic. You know, we, we've been covering this VMware cloud on abus since, since the launch going back and it's been amazing to watch the evolution from people saying, Oh, it's the worst thing I've ever seen. It's what's this mean? And depress work were, we're kind of not really on board with kind of the vision, but as it played out as you guys had announced together, it did work out great for VMware. It did work out great for a D and it continues two years later and I want just get an update from you guys on where you guys see this has been going. I'll see multiple years. Where is the evolution of the solution as we are right now coming off VMware explorer just recently and going in to reinvent, which is only a couple weeks away, feels like tomorrow. But you know, as we prepare a lot going on, where are we with the evolution of the solution? >>I mean, first thing I wanna say is, you know, PBO 2016 was a someon moment and the history of it, right? When Pat Gelsinger and Andy Jessey came together to announce this and I think John, you were there at the time I was there, it was a great, great moment. We launched the solution in 2017, the year after that at VM Word back when we called it Word, I think we have gone from strength to strength. One of the things that has really mattered to us is we have learned froms also in the processes, this notion of working backwards. So we really, really focused on customer feedback as we build a service offering now five years old, pretty remarkable journey. You know, in the first years we tried to get across all the regions, you know, that was a big focus because there was so much demand for it. >>In the second year we started going really on enterprise grade features. We invented this pretty awesome feature called Stretch clusters, where you could stretch a vSphere cluster using VSA and NSX across two AZs in the same region. Pretty phenomenal four nine s availability that applications start started to get with that particular feature. And we kept moving forward all kinds of integration with AWS direct connect transit gateways with our own advanced networking capabilities. You know, along the way, disaster recovery, we punched out two, two new services just focused on that. And then more recently we launched our outposts partnership. We were up on stage at Reinvent, again with Pat Andy announcing AWS outposts and the VMware flavor of that VMware cloud and AWS outposts. I think it's been significant growth in our federal sector as well with our federal and high certification more recently. So all in all, we are super excited. We're five years old. The customer momentum is really, really strong and we are scaling the service massively across all geos and industries. >>That's great, great update. And I think one of the things that you mentioned was how the advantages you guys got from that relationship. And, and this has kind of been the theme for AWS since I can remember from day one. Fred, you guys do the heavy lifting as as, as you always say for the customers here, VMware comes on board, takes advantage of the AWS and kind of just doesn't miss a beat, continues to move their workloads that everyone's using, you know, vSphere and these are, these are big workloads on aws. What's the AWS perspective on this? How do you see it? >>Yeah, it's pretty fascinating to watch how fast customers can actually transform and move when you take the, the skill set that they're familiar with and the advanced capabilities that they've been using on Preem and then overlay it on top of the AWS infrastructure that's, that's evolving quickly and, and building out new hardware and new instances we'll talk about. But that combined experience between both of us on a jointly engineered solution to bring the best security and the best features that really matter for those workloads drive a lot of efficiency and speed for the, for the customer. So it's been well received and the partnership is stronger than ever from an engineering standpoint, from a business standpoint. And obviously it's been very interesting to look at just how we stay day one in terms of looking at new features and work and, and responding to what customers want. So pretty, pretty excited about just seeing the transformation and the speed that which customers can move to bmc. Yeah, >>That's what great value publish. We've been talking about that in context too. Anyone building on top of the cloud, they can have their own supercloud as we call it. If you take advantage of all the CapEx and and investment Amazon's made and AWS has made and, and and continues to make in performance IAS and pass all great stuff. I have to ask you guys both as you guys see this going to the next level, what are some of the differentiations you see around the service compared to other options on the market? What makes it different? What's the combination? You mentioned jointly engineered, what are some of the key differentiators of the service compared to others? >>Yeah, I think one of the key things Fred talked about is this jointly engineered notion right from day one. We were the earlier doctors of AWS Nitro platform, right? The reinvention of E two back five years ago. And so we have been, you know, having a very, very strong engineering partnership at that level. I think from a VMware customer standpoint, you get the full software defined data center or compute storage networking on EC two, bare metal across all regions. You can scale that elastically up and down. It's pretty phenomenal just having that consistency globally, right on aws EC two global regions. Now the other thing that's a real differentiator for us that customers tell us about is this whole notion of a managed service, right? And this was somewhat new to VMware, but we took away the pain of this undifferentiated heavy lifting where customers had to provision rack, stack hardware, configure the software on top, and then upgrade the software and the security batches on top. >>So we took, took away all of that pain as customers transitioned to VMware cloud and aws. In fact, my favorite story from last year when we were all going through the lock for j debacle industry was just going through that, right? Favorite proof point from customers was before they put even race this issue to us, we sent them a notification saying we already patched all of your systems, no action from you. The customers were super thrilled. I mean these are large banks, many other customers around the world, super thrilled they had to take no action, but a pretty incredible industry challenge that we were all facing. >>Nora, that's a great, so that's a great point. You know, the whole managed service piece brings up the security, you kind of teasing at it, but you know, there's always vulnerabilities that emerge when you are doing complex logic. And as you grow your solutions, there's more bits. You know, Fred, we were commenting before we came on camera, there's more bits than ever before and, and at at the physics layer too, as well as the software. So you never know when there's gonna be a zero day vulnerability out there. Just, it happens. We saw one with fornet this week, this came outta the woodwork. But moving fast on those patches, it's huge. This brings up the whole support angle. I wanted to ask you about how you guys are doing that as well, because to me we see the value when we, when we talk to customers on the cube about this, you know, it was a real, real easy understanding of how, what the cloud means to them with VMware now with the aws. But the question that comes up that we wanna get more clarity on is how do you guys handle support together? >>Well, what's interesting about this is that it's, it's done mutually. We have dedicated support teams on both sides that work together pretty seamlessly to make sure that whether there's a issue at any layer, including all the way up into the app layer, as you think about some of the other workloads like sap, we'll go end to end and make sure that we support the customer regardless of where the particular issue might be for them. And on top of that, we look at where, where we're improving reliability in, in as a first order of, of principle between both companies. So from an availability and reliability standpoint, it's, it's top of mind and no matter where the particular item might land, we're gonna go help the customer resolve. That works really well >>On the VMware side. What's been the feedback there? What's the, what are some of the updates? >>Yeah, I think, look, I mean, VMware owns and operates the service, but we have a phenomenal backend relationship with aws. Customers call VMware for the service for any issues and, and then we have a awesome relationship with AWS on the backend for support issues or any hardware issues. The BASKE management that we jointly do, right? All of the hard problems that customers don't have to worry about. I think on the front end, we also have a really good group of solution architects across the companies that help to really explain the solution. Do complex things like cloud migration, which is much, much easier with VMware cloud aws, you know, we are presenting that easy button to the public cloud in many ways. And so we have a whole technical audience across the two companies that are working with customers every single day. >>You know, you had mentioned, I've got a list here, some of the innovations the, you mentioned the stretch clustering, you know, getting the GOs working, Advanced network, disaster recovery, you know, fed, Fed ramp, public sector certifications, outposts, all good. You guys are checking the boxes every year. You got a good, good accomplishments list there on the VMware AWS side here in this relationship. The question that I'm interested in is what's next? What recent innovations are you doing? Are you making investments in what's on the lists this year? What items will be next year? How do you see the, the new things, the list of accomplishments, people wanna know what's next. They don't wanna see stagnant growth here, they wanna see more action, you know, as as cloud kind of continues to scale and modern applications cloud native, you're seeing more and more containers, more and more, you know, more CF C I C D pipe pipelining with with modern apps, put more pressure on the system. What's new, what's the new innovations? >>Absolutely. And I think as a five yearold service offering innovation is top of mind for us every single day. So just to call out a few recent innovations that we announced in San Francisco at VMware Explorer. First of all, our new platform i four I dot metal, it's isolate based, it's pretty awesome. It's the latest and greatest, all the speeds and feeds that we would expect from VMware and aws. At this point in our relationship. We announced two different storage options. This notion of working from customer feedback, allowing customers even more price reductions, really take off that storage and park it externally, right? And you know, separate that from compute. So two different storage offerings there. One is with AWS Fsx, with NetApp on tap, which brings in our NetApp partnership as well into the equation and really get that NetApp based, really excited about this offering as well. >>And the second storage offering for VMware cloud Flex Storage, VMware's own managed storage offering. Beyond that, we have done a lot of other innovations as well. I really wanted to talk about VMware cloud Flex Compute, where previously customers could only scale by hosts and a host is 36 to 48 cores, give or take. But with VMware cloud Flex Compute, we are now allowing this notion of a resource defined compute model where customers can just get exactly the V C P memory and storage that maps to the applications, however small they might be. So this notion of granularity is really a big innovation that that we are launching in the market this year. And then last but not least, talk about ransomware. Of course it's a hot topic in industry. We are seeing many, many customers ask for this. We are happy to announce a new ransomware recovery with our VMware cloud DR solution. >>A lot of innovation there and the way we are able to do machine learning and make sure the workloads that are covered from snapshots and backups are actually safe to use. So there's a lot of differentiation on that front as well. A lot of networking innovations with Project Knot star for ability to have layer flow through layer seven, you know, new SaaS services in that area as well. Keep in mind that the service already supports managed Kubernetes for containers. It's built in to the same clusters that have virtual machines. And so this notion of a single service with a great TCO for VMs and containers and sort of at the heart of our office, >>The networking side certainly is a hot area to keep innovating on. Every year it's the same, same conversation, get better, faster networking, more, more options there. The flex computes. Interesting. If you don't mind me getting a quick clarification, could you explain the Drew screen resource defined versus hardware defined? Because this is kind of what we had saw at Explore coming out, that notion of resource defined versus hardware defined. What's the, what does that mean? >>Yeah, I mean I think we have been super successful in this hardware defined notion. We we're scaling by the hardware unit that we present as software defined data centers, right? And so that's been super successful. But we, you know, customers wanted more, especially customers in different parts of the world wanted to start even smaller and grow even more incrementally, right? Lower their costs even more. And so this is the part where resource defined starts to be very, very interesting as a way to think about, you know, here's my bag of resources exactly based on what the customers request for fiber machines, five containers, its size exactly for that. And then as utilization grows, we elastically behind the scenes, we're able to grow it through policies. So that's a whole different dimension. It's a whole different service offering that adds value and customers are comfortable. They can go from one to the other, they can go back to that post based model if they so choose to. And there's a jump off point across these two different economic models. >>It's kind of cloud of flexibility right there. I like the name Fred. Let's get into some of the examples of customers, if you don't mind. Let's get into some of the ex, we have some time. I wanna unpack a little bit of what's going on with the customer deployments. One of the things we've heard again on the cube is from customers is they like the clarity of the relationship, they love the cloud positioning of it. And then what happens is they lift and shift the workloads and it's like, feels great. It's just like we're running VMware on AWS and then they would start consuming higher level services, kind of that adoption next level happens and because it it's in the cloud, so, So can you guys take us through some recent examples of customer wins or deployments where they're using VMware cloud on AWS on getting started, and then how do they progress once they're there? How does it evolve? Can you just walk us through a couple of use cases? >>Sure. There's a, well there's a couple. One, it's pretty interesting that, you know, like you said, as there's more and more bits you need better and better hardware and networking. And we're super excited about the I four and the capabilities there in terms of doubling and or tripling what we're doing around a lower variability on latency and just improving all the speeds. But what customers are doing with it, like the college in New Jersey, they're accelerating their deployment on a, on onboarding over like 7,400 students over a six to eight month period. And they've really realized a ton of savings. But what's interesting is where and how they can actually grow onto additional native services too. So connectivity to any other services is available as they start to move and migrate into this. The, the options there obviously are tied to all the innovation that we have across any services, whether it's containerized and with what they're doing with Tanu or with any other container and or services within aws. >>So there's, there's some pretty interesting scenarios where that data and or the processing, which is moved quickly with full compliance, whether it's in like healthcare or regulatory business is, is allowed to then consume and use things, for example, with tech extract or any other really cool service that has, you know, monthly and quarterly innovations. So there's things that you just can't, could not do before that are coming out and saving customers money and building innovative applications on top of their, their current app base in, in a rapid fashion. So pretty excited about it. There's a lot of examples. I think I probably don't have time to go into too, too many here. Yeah. But that's actually the best part is listening to customers and seeing how many net new services and new applications are they actually building on top of this platform. >>Nora, what's your perspective from the VMware sy? So, you know, you guys have now a lot of headroom to offer customers with Amazon's, you know, higher level services and or whatever's homegrown where's being rolled out? Cuz you now have a lot of hybrid too, so, so what's your, what's your take on what, what's happening in with customers? >>I mean, it's been phenomenal, the, the customer adoption of this and you know, banks and many other highly sensitive verticals are running production grade applications, tier one applications on the service over the last five years. And so, you know, I have a couple of really good examples. S and p Global is one of my favorite examples. Large bank, they merge with IHS market, big sort of conglomeration. Now both customers were using VMware cloud and AWS in different ways. And with the, with the use case, one of their use cases was how do I just respond to these global opportunities without having to invest in physical data centers? And then how do I migrate and consolidate all my data centers across the global, which there were many. And so one specific example for this company was how they migrated thousand 1000 workloads to VMware cloud AWS in just six weeks. Pretty phenomenal. If you think about everything that goes into a cloud migration process, people process technology and the beauty of the technology going from VMware point A to VMware point B, the the lowest cost, lowest risk approach to adopting VMware, VMware cloud, and aws. So that's, you know, one of my favorite examples. There are many other examples across other verticals that we continue to see. The good thing is we are seeing rapid expansion across the globe that constantly entering new markets with the limited number of regions and progressing our roadmap there. >>Yeah, it's great to see, I mean the data center migrations go from months, many, many months to weeks. It's interesting to see some of those success stories. So congratulations. One >>Of other, one of the other interesting fascinating benefits is the sustainability improvement in terms of being green. So the efficiency gains that we have both in current generation and new generation processors and everything that we're doing to make sure that when a customer can be elastic, they're also saving power, which is really critical in a lot of regions worldwide at this point in time. They're, they're seeing those benefits. If you're running really inefficiently in your own data center, that is just a, not a great use of power. So the actual calculators and the benefits to these workloads is, are pretty phenomenal just in being more green, which I like. We just all need to do our part there. And, and this is a big part of it here. >>It's a huge, it's a huge point about the sustainability. Fred, I'm glad you called that out. The other one I would say is supply chain issues. Another one you see that constrains, I can't buy hardware. And the third one is really obvious, but no one really talks about it. It's security, right? I mean, I remember interviewing Stephen Schmidt with that AWS and many years ago, this is like 2013, and you know, at that time people were saying the cloud's not secure. And he's like, listen, it's more secure in the cloud on premise. And if you look at the security breaches, it's all about the on-premise data center vulnerabilities, not so much hardware. So there's a lot you gotta to stay current on, on the isolation there is is hard. So I think, I think the security and supply chain, Fred is, is another one. Do you agree? >>I I absolutely agree. It's, it's hard to manage supply chain nowadays. We put a lot of effort into that and I think we have a great ability to forecast and make sure that we can lean in and, and have the resources that are available and run them, run them more efficiently. Yeah, and then like you said on the security point, security is job one. It is, it is the only P one. And if you think of how we build our infrastructure from Nitro all the way up and how we respond and work with our partners and our customers, there's nothing more important. >>And naron your point earlier about the managed service patching and being on top of things, it's really gonna get better. All right, final question. I really wanna thank you for your time on this showcase. It's really been a great conversation. Fred, you had made a comment earlier. I wanna kind of end with kind of a curve ball and put you eyes on the spot. We're talking about a modern, a new modern shift. It's another, we're seeing another inflection point, we've been documenting it, it's almost like cloud hitting another inflection point with application and open source growth significantly at the app layer. Continue to put a lot of pressure and, and innovation in the infrastructure side. So the question is for you guys each to answer is what's the same and what's different in today's market? So it's kind of like we want more of the same here, but also things have changed radically and better here. What are the, what's, what's changed for the better and where, what's still the same kind of thing hanging around that people are focused on? Can you share your perspective? >>I'll, I'll, I'll, I'll tackle it. You know, businesses are complex and they're often unique that that's the same. What's changed is how fast you can innovate. The ability to combine manage services and new innovative services and build new applications is so much faster today. Leveraging world class hardware that you don't have to worry about that's elastic. You, you could not do that even five, 10 years ago to the degree you can today, especially with innovation. So innovation is accelerating at a, at a rate that most people can't even comprehend and understand the, the set of services that are available to them. It's really fascinating to see what a one pizza team of of engineers can go actually develop in a week. It is phenomenal. So super excited about this space and it's only gonna continue to accelerate that. That's my take. All right. >>You got a lot of platform to compete on with, got a lot to build on then you're Ryan, your side, What's your, what's your answer to that question? >>I think we are seeing a lot of innovation with new applications that customers are constant. I think what we see is this whole notion of how do you go from desktop to production to the secure supply chain and how can we truly, you know, build on the agility that developers desire and build all the security and the pipelines to energize that motor production quickly and efficiently. I think we, we are seeing, you know, we are at the very start of that sort of of journey. Of course we have invested in Kubernetes the means to an end, but there's so much more beyond that's happening in industry. And I think we're at the very, very beginning of this transformations, enterprise transformation that many of our customers are going through and we are inherently part of it. >>Yeah. Well gentlemen, I really appreciate that we're seeing the same thing. It's more the same here on, you know, solving these complexities with distractions. Whether it's, you know, higher level services with large scale infrastructure at, at your fingertips. Infrastructures, code, infrastructure to be provisioned, serverless, all the good stuff happen in Fred with AWS on your side. And we're seeing customers resonate with this idea of being an operator, again, being a cloud operator and developer. So the developer ops is kind of, DevOps is kind of changing too. So all for the better. Thank you for spending the time and we're seeing again, that traction with the VMware customer base and of us getting, getting along great together. So thanks for sharing your perspectives, >>I appreciate it. Thank you so >>Much. Okay, thank you John. Okay, this is the Cube and AWS VMware showcase, accelerating business transformation. VMware cloud on aws, jointly engineered solution, bringing innovation to the VMware customer base, going to the cloud and beyond. I'm John Fur, your host. Thanks for watching. Hello everyone. Welcome to the special cube presentation of accelerating business transformation on vmc on aws. I'm John Furrier, host of the Cube. We have dawan director of global sales and go to market for VMware cloud on adb. This is a great showcase and should be a lot of fun. Ashish, thanks for coming on. >>Hi John. Thank you so much. >>So VMware cloud on AWS has been well documented as this big success for VMware and aws. As customers move their workloads into the cloud, IT operations of VMware customers has signaling a lot of change. This is changing the landscape globally is on cloud migration and beyond. What's your take on this? Can you open this up with the most important story around VMC on aws? >>Yes, John. The most important thing for our customers today is the how they can safely and swiftly move their ID infrastructure and applications through cloud. Now, VMware cloud AWS is a service that allows all vSphere based workloads to move to cloud safely, swiftly and reliably. Banks can move their core, core banking platforms, insurance companies move their core insurance platforms, telcos move their goss, bss, PLA platforms, government organizations are moving their citizen engagement platforms using VMC on aws because this is one platform that allows you to move it, move their VMware based platforms very fast. Migrations can happen in a matter of days instead of months. Extremely securely. It's a VMware manage service. It's very secure and highly reliably. It gets the, the reliability of the underlyings infrastructure along with it. So win-win from our customers perspective. >>You know, we reported on this big news in 2016 with Andy Chas, the, and Pat Geling at the time, a lot of people said it was a bad deal. It turned out to be a great deal because not only could VMware customers actually have a cloud migrate to the cloud, do it safely, which was their number one concern. They didn't want to have disruption to their operations, but also position themselves for what's beyond just shifting to the cloud. So I have to ask you, since you got the finger on the pulse here, what are we seeing in the market when it comes to migrating and modern modernizing in the cloud? Because that's the next step. They go to the cloud, you guys have done that, doing it, then they go, I gotta modernize, which means kind of upgrading or refactoring. What's your take on that? >>Yeah, absolutely. Look, the first step is to help our customers assess their infrastructure and licensing and entire ID operations. Once we've done the assessment, we then create their migration plans. A lot of our customers are at that inflection point. They're, they're looking at their real estate, ex data center, real estate. They're looking at their contracts with colocation vendors. They really want to exit their data centers, right? And VMware cloud and AWS is a perfect solution for customers who wanna exit their data centers, migrate these applications onto the AWS platform using VMC on aws, get rid of additional real estate overheads, power overheads, be socially and environmentally conscious by doing that as well, right? So that's the migration story, but to your point, it doesn't end there, right? Modernization is a critical aspect of the entire customer journey as as well customers, once they've migrated their ID applications and infrastructure on cloud get access to all the modernization services that AWS has. They can correct easily to our data lake services, to our AIML services, to custom databases, right? They can decide which applications they want to keep and which applications they want to refactor. They want to take decisions on containerization, make decisions on service computing once they've come to the cloud. But the most important thing is to take that first step. You know, exit data centers, come to AWS using vmc or aws, and then a whole host of modernization options available to them. >>Yeah, I gotta say, we had this right on this, on this story, because you just pointed out a big thing, which was first order of business is to make sure to leverage the on-prem investments that those customers made and then migrate to the cloud where they can maintain their applications, their data, their infrastructure operations that they're used to, and then be in position to start getting modern. So I have to ask you, how are you guys specifically, or how is VMware cloud on s addressing these needs of the customers? Because what happens next is something that needs to happen faster. And sometimes the skills might not be there because if they're running old school, IT ops now they gotta come in and jump in. They're gonna use a data cloud, they're gonna want to use all kinds of machine learning, and there's a lot of great goodness going on above the stack there. So as you move with the higher level services, you know, it's a no brainer, obviously, but they're not, it's not yesterday's higher level services in the cloud. So how are, how is this being addressed? >>Absolutely. I think you hit up on a very important point, and that is skills, right? When our customers are operating, some of the most critical applications I just mentioned, core banking, core insurance, et cetera, they're most of the core applications that our customers have across industries, like even, even large scale ERP systems, they're actually sitting on VMware's vSphere platform right now. When the customer wants to migrate these to cloud, one of the key bottlenecks they face is skill sets. They have the trained manpower for these core applications, but for these high level services, they may not, right? So the first order of business is to help them ease this migration pain as much as possible by not wanting them to, to upscale immediately. And we VMware cloud and AWS exactly does that. I mean, you don't have to do anything. You don't have to create new skill set for doing this, right? Their existing skill sets suffice, but at the same time, it gives them that, that leeway to build that skills roadmap for their team. DNS is invested in that, right? Yes. We want to help them build those skills in the high level services, be it aml, be it, be it i t be it data lake and analytics. We want to invest in them, and we help our customers through that. So that ultimately the ultimate goal of making them drop data is, is, is a front and center. >>I wanna get into some of the use cases and success stories, but I want to just reiterate, hit back your point on the skill thing. Because if you look at what you guys have done at aws, you've essentially, and Andy Chassey used to talk about this all the time when I would interview him, and now last year Adam was saying the same thing. You guys do all the heavy lifting, but if you're a VMware customer user or operator, you are used to things. You don't have to be relearn to be a cloud architect. Now you're already in the game. So this is like almost like a instant path to cloud skills for the VMware. There's hundreds of thousands of, of VMware architects and operators that now instantly become cloud architects, literally overnight. Can you respond to that? Do you agree with that? And then give an example. >>Yes, absolutely. You know, if you have skills on the VMware platform, you know, know, migrating to AWS using via by cloud and AWS is absolutely possible. You don't have to really change the skills. The operations are exactly the same. The management systems are exactly the same. So you don't really have to change anything but the advantages that you get access to all the other AWS services. So you are instantly able to integrate with other AWS services and you become a cloud architect immediately, right? You are able to solve some of the critical problems that your underlying IT infrastructure has immediately using this. And I think that's a great value proposition for our customers to use this service. >>And just one more point, I want just get into something that's really kind of inside baseball or nuanced VMC or VMware cloud on AWS means something. Could you take a minute to explain what on AWS means? Just because you're like hosting and using Amazon as a, as a work workload? Being on AWS means something specific in your world, being VMC on AWS mean? >>Yes. This is a great question, by the way, You know, on AWS means that, you know, VMware's vse platform is, is a, is an iconic enterprise virtualization software, you know, a disproportionately high market share across industries. So when we wanted to create a cloud product along with them, obviously our aim was for them, for the, for this platform to have the goodness of the AWS underlying infrastructure, right? And, and therefore, when we created this VMware cloud solution, it it literally use the AWS platform under the eighth, right? And that's why it's called a VMs VMware cloud on AWS using, using the, the, the wide portfolio of our regions across the world and the strength of the underlying infrastructure, the reliability and, and, and sustainability that it offers. And therefore this product is called VMC on aws. >>It's a distinction I think is worth noting, and it does reflect engineering and some levels of integration that go well beyond just having a SaaS app and, and basically platform as a service or past services. So I just wanna make sure that now super cloud, we'll talk about that a little bit in another interview, but I gotta get one more question in before we get into the use cases and customer success stories is in, in most of the VM world, VMware world, in that IT world, it used to, when you heard migration, people would go, Oh my God, that's gonna take months. And when I hear about moving stuff around and doing cloud native, the first reaction people might have is complexity. So two questions for you before we move on to the next talk. Track complexity. How are you addressing the complexity issue and how long these migrations take? Is it easy? Is it it hard? I mean, you know, the knee jerk reaction is month, You're very used to that. If they're dealing with Oracle or other old school vendors, like, they're, like the old guard would be like, takes a year to move stuff around. So can you comment on complexity and speed? >>Yeah. So the first, first thing is complexity. And you know, what makes what makes anything complex is if you're, if you're required to acquire new skill sets or you've gotta, if you're required to manage something differently, and as far as VMware cloud and AWS on both these aspects, you don't have to do anything, right? You don't have to acquire new skill sets. Your existing idea operation skill sets on, on VMware's platforms are absolutely fine and you don't have to manage it any differently like, than what you're managing your, your ID infrastructure today. So in both these aspects, it's exactly the same and therefore it is absolutely not complex as far as, as far as, as far as we cloud and AWS is concerned. And the other thing is speed. This is where the huge differentiation is. You have seen that, you know, large banks and large telcos have now moved their workloads, you know, literally in days instead of months. >>Because because of VMware cloud and aws, a lot of time customers come to us with specific deadlines because they want to exit their data centers on a particular date. And what happens, VMware cloud and AWS is called upon to do that migration, right? So speed is absolutely critical. The reason is also exactly the same because you are using the exactly the same platform, the same management systems, people are available to you, you're able to migrate quickly, right? I would just reference recently we got an award from President Zelensky of Ukraine for, you know, migrating their entire ID digital infrastructure and, and that that happened because they were using VMware cloud database and happened very swiftly. >>That's been a great example. I mean, that's one political, but the economic advantage of getting outta the data center could be national security. You mentioned Ukraine, I mean Oscar see bombing and death over there. So clearly that's a critical crown jewel for their running their operations, which is, you know, you know, world mission critical. So great stuff. I love the speed thing. I think that's a huge one. Let's get into some of the use cases. One of them is, the first one I wanted to talk about was we just hit on data, data center migration. It could be financial reasons on a downturn or our, or market growth. People can make money by shifting to the cloud, either saving money or making money. You win on both sides. It's a, it's a, it's almost a recession proof, if you will. Cloud is so use case for number one data center migration. Take us through what that looks like. Give an example of a success. Take us through a day, day in the life of a data center migration in, in a couple minutes. >>Yeah. You know, I can give you an example of a, of a, of a large bank who decided to migrate, you know, their, all their data centers outside their existing infrastructure. And they had, they had a set timeline, right? They had a set timeline to migrate the, the, they were coming up on a renewal and they wanted to make sure that this set timeline is met. We did a, a complete assessment of their infrastructure. We did a complete assessment of their IT applications, more than 80% of their IT applications, underlying v vSphere platform. And we, we thought that the right solution for them in the timeline that they wanted, right, is VMware cloud ands. And obviously it was a large bank, it wanted to do it safely and securely. It wanted to have it completely managed, and therefore VMware cloud and aws, you know, ticked all the boxes as far as that is concerned. >>I'll be happy to report that the large bank has moved to most of their applications on AWS exiting three of their data centers, and they'll be exiting 12 more very soon. So that's a great example of, of, of the large bank exiting data centers. There's another Corolla to that. Not only did they manage to manage to exit their data centers and of course use and be more agile, but they also met their sustainability goals. Their board of directors had given them goals to be carbon neutral by 2025. They found out that 35% of all their carbon foot footprint was in their data centers. And if they moved their, their ID infrastructure to cloud, they would severely reduce the, the carbon footprint, which is 35% down to 17 to 18%. Right? And that meant their, their, their, their sustainability targets and their commitment to the go to being carbon neutral as well. >>And that they, and they shift that to you guys. Would you guys take that burden? A heavy lifting there and you guys have a sustainability story, which is a whole nother showcase in and of itself. We >>Can Exactly. And, and cause of the scale of our, of our operations, we are able to, we are able to work on that really well as >>Well. All right. So love the data migration. I think that's got real proof points. You got, I can save money, I can, I can then move and position my applications into the cloud for that reason and other reasons as a lot of other reasons to do that. But now it gets into what you mentioned earlier was, okay, data migration, clearly a use case and you laid out some successes. I'm sure there's a zillion others. But then the next step comes, now you got cloud architects becoming minted every, and you got managed services and higher level services. What happens next? Can you give us an example of the use case of the modernization around the NextGen workloads, NextGen applications? We're starting to see, you know, things like data clouds, not data warehouses. We're not gonna data clouds, it's gonna be all kinds of clouds. These NextGen apps are pure digital transformation in action. Take us through a use case of how you guys make that happen with a success story. >>Yes, absolutely. And this is, this is an amazing success story and the customer here is s and p global ratings. As you know, s and p global ratings is, is the world leader as far as global ratings, global credit ratings is concerned. And for them, you know, the last couple of years have been tough as far as hardware procurement is concerned, right? The pandemic has really upended the, the supply chain. And it was taking a lot of time to procure hardware, you know, configure it in time, make sure that that's reliable and then, you know, distribute it in the wide variety of, of, of offices and locations that they have. And they came to us. We, we did, again, a, a, a alar, a fairly large comprehensive assessment of their ID infrastructure and their licensing contracts. And we also found out that VMware cloud and AWS is the right solution for them. >>So we worked there, migrated all their applications, and as soon as we migrated all their applications, they got, they got access to, you know, our high level services be our analytics services, our machine learning services, our, our, our, our artificial intelligence services that have been critical for them, for their growth. And, and that really is helping them, you know, get towards their next level of modern applications. Right Now, obviously going forward, they will have, they will have the choice to, you know, really think about which applications they want to, you know, refactor or which applications they want to go ahead with. That is really a choice in front of them. And, but you know, the, we VMware cloud and AWS really gave them the opportunity to first migrate and then, you know, move towards modernization with speed. >>You know, the speed of a startup is always the kind of the Silicon Valley story where you're, you know, people can make massive changes in 18 months, whether that's a pivot or a new product. You see that in startup world. Now, in the enterprise, you can see the same thing. I noticed behind you on your whiteboard, you got a slogan that says, are you thinking big? I know Amazon likes to think big, but also you work back from the customers and, and I think this modern application thing's a big deal because I think the mindset has always been constrained because back before they moved to the cloud, most IT, and, and, and on-premise data center shops, it's slow. You gotta get the hardware, you gotta configure it, you gotta, you gotta stand it up, make sure all the software is validated on it, and loading a database and loading oss, I mean, mean, yeah, it got easier and with scripting and whatnot, but when you move to the cloud, you have more scale, which means more speed, which means it opens up their capability to think differently and build product. What are you seeing there? Can you share your opinion on that epiphany of, wow, things are going fast, I got more time to actually think about maybe doing a cloud native app or transforming this or that. What's your, what's your reaction to that? Can you share your opinion? >>Well, ultimately we, we want our customers to utilize, you know, most of our modern services, you know, applications should be microservices based. When desired, they should use serverless applic. So list technology, they should not have monolithic, you know, relational database contracts. They should use custom databases, they should use containers when needed, right? So ultimately, we want our customers to use these modern technologies to make sure that their IT infrastructure, their licensing, their, their entire IT spend is completely native to cloud technologies. They work with the speed of a startup, but it's important for them to, to, to get to the first step, right? So that's why we create this journey for our customers, where you help them migrate, give them time to build the skills, they'll help them mo modernize, take our partners along with their, along with us to, to make sure that they can address the need for our customers. That's, that's what our customers need today, and that's what we are working backwards from. >>Yeah, and I think that opens up some big ideas. I'll just say that the, you know, we're joking, I was joking the other night with someone here in, in Palo Alto around serverless, and I said, you know, soon you're gonna hear words like architectural list. And that's a criticism on one hand, but you might say, Hey, you know, if you don't really need an architecture, you know, storage lists, I mean, at the end of the day, infrastructure is code means developers can do all the it in the coding cycles and then make the operations cloud based. And I think this is kind of where I see the dots connecting. Final thought here, take us through what you're thinking around how this new world is evolving. I mean, architecturals kind of a joke, but the point is, you know, you have to some sort of architecture, but you don't have to overthink it. >>Totally. No, that's a great thought, by the way. I know it's a joke, but it's a great thought because at the end of the day, you know, what do the customers really want? They want outcomes, right? Why did service technology come? It was because there was an outcome that they needed. They didn't want to get stuck with, you know, the, the, the real estate of, of a, of a server. They wanted to use compute when they needed to, right? Similarly, what you're talking about is, you know, outcome based, you know, desire of our customers and, and, and that's exactly where the word is going to, Right? Cloud really enforces that, right? We are actually, you know, working backwards from a customer's outcome and using, using our area the breadth and depth of our services to, to deliver those outcomes, right? And, and most of our services are in that path, right? When we use VMware cloud and aws, the outcome is a, to migrate then to modernize, but doesn't stop there, use our native services, you know, get the business outcomes using this. So I think that's, that's exactly what we are going through >>Actually, should actually, you're the director of global sales and go to market for VMware cloud on Aus. I wanna thank you for coming on, but I'll give you the final minute. Give a plug, explain what is the VMware cloud on Aus, Why is it great? Why should people engage with you and, and the team, and what ultimately is this path look like for them going forward? >>Yeah. At the end of the day, we want our customers to have the best paths to the cloud, right? The, the best path to the cloud is making sure that they migrate safely, reliably, and securely as well as with speed, right? And then, you know, use that cloud platform to, to utilize AWS's native services to make sure that they modernize their IT infrastructure and applications, right? We want, ultimately that our customers, customers, customer get the best out of, you know, utilizing the, that whole application experience is enhanced tremendously by using our services. And I think that's, that's exactly what we are working towards VMware cloud AWS is, is helping our customers in that journey towards migrating, modernizing, whether they wanna exit a data center or whether they wanna modernize their applications. It's a essential first step that we wanna help our customers with >>One director of global sales and go to market with VMware cloud on neighbors. He's with aws sharing his thoughts on accelerating business transformation on aws. This is a showcase. We're talking about the future path. We're talking about use cases with success stories from customers as she's thank you for spending time today on this showcase. >>Thank you, John. I appreciate it. >>Okay. This is the cube, special coverage, special presentation of the AWS Showcase. I'm John Furrier, thanks for watching.
SUMMARY :
Great to have you and Daniel Re Myer, principal architect global AWS synergy Greatly appreciate it. You're starting to see, you know, this idea of higher level services, More recently, one of the things to keep in mind is we're looking to deliver value Then the other thing comes down to is where we Daniel, I wanna get to you in a second. lot of CPU power, such as you mentioned it, AI workloads. composing, you know, with open source, a lot of great things are changing. So we want to have all of that as a service, on what you see there from an Amazon perspective and how it relates to this? And you know, look at it from the point of view where we said this to leverage a cloud, but the investment that you made and certain things as far How would you talk to that persona about the future And that also means in, in to to some extent, concerns with your I can still run my job now I got goodness on the other side. on the skills, you certainly have that capability to do so. Now that we're peeking behind the curtain here, I'd love to have you guys explain, You always have to have the time difference in mind if we are working globally together. I mean it seems to be very productive, you know, I think one of the key things to keep in mind is, you know, even if you look at AWS's guys to comment on, as you guys continue to evolve the relationship, what's in it for So one of the most important things we have announced this year, Yeah, I think one of the key things to keep in mind is, you know, we're looking to help our customers You know, we have a product, you have a product, biz dev deals happen, people sign relationships and they do business And this, you guys are in the middle of two big ecosystems. You can do this if you decide you want to stay with some of your services But partners innovate with you on their terms. I think one of the key things, you know, as Daniel mentioned before, You still run the fear, the way you working on it and And if, if you look, not every, And thank you for spending the time. So personally for me as an IT background, you know, been in CIS admin world and whatnot, thank you for coming on on this part of the showcase episode of really the customer successes with VMware we're kind of not really on board with kind of the vision, but as it played out as you guys had announced together, across all the regions, you know, that was a big focus because there was so much demand for We invented this pretty awesome feature called Stretch clusters, where you could stretch a And I think one of the things that you mentioned was how the advantages you guys got from that and move when you take the, the skill set that they're familiar with and the advanced capabilities that I have to ask you guys both as you guys see this going to the next level, you know, having a very, very strong engineering partnership at that level. put even race this issue to us, we sent them a notification saying we And as you grow your solutions, there's more bits. the app layer, as you think about some of the other workloads like sap, we'll go end to What's been the feedback there? which is much, much easier with VMware cloud aws, you know, they wanna see more action, you know, as as cloud kind of continues to And you know, separate that from compute. And the second storage offering for VMware cloud Flex Storage, VMware's own managed storage you know, new SaaS services in that area as well. If you don't mind me getting a quick clarification, could you explain the Drew screen resource defined versus But we, you know, because it it's in the cloud, so, So can you guys take us through some recent examples of customer The, the options there obviously are tied to all the innovation that we So there's things that you just can't, could not do before I mean, it's been phenomenal, the, the customer adoption of this and you know, Yeah, it's great to see, I mean the data center migrations go from months, many, So the actual calculators and the benefits So there's a lot you gotta to stay current on, Yeah, and then like you said on the security point, security is job one. So the question is for you guys each to Leveraging world class hardware that you don't have to worry production to the secure supply chain and how can we truly, you know, Whether it's, you know, higher level services with large scale Thank you so I'm John Furrier, host of the Cube. Can you open this up with the most important story around VMC on aws? platform that allows you to move it, move their VMware based platforms very fast. They go to the cloud, you guys have done that, So that's the migration story, but to your point, it doesn't end there, So as you move with the higher level services, So the first order of business is to help them ease Because if you look at what you guys have done at aws, the advantages that you get access to all the other AWS services. Could you take a minute to explain what on AWS on AWS means that, you know, VMware's vse platform is, I mean, you know, the knee jerk reaction is month, And you know, what makes what the same because you are using the exactly the same platform, the same management systems, which is, you know, you know, world mission critical. decided to migrate, you know, their, So that's a great example of, of, of the large bank exiting data And that they, and they shift that to you guys. And, and cause of the scale of our, of our operations, we are able to, We're starting to see, you know, things like data clouds, And for them, you know, the last couple of years have been tough as far as hardware procurement is concerned, And, and that really is helping them, you know, get towards their next level You gotta get the hardware, you gotta configure it, you gotta, you gotta stand it up, most of our modern services, you know, applications should be microservices based. I mean, architecturals kind of a joke, but the point is, you know, the end of the day, you know, what do the customers really want? I wanna thank you for coming on, but I'll give you the final minute. customers, customer get the best out of, you know, utilizing the, One director of global sales and go to market with VMware cloud on neighbors. I'm John Furrier, thanks for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Samir | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Maryland | LOCATION | 0.99+ |
Pat Geling | PERSON | 0.99+ |
John Foer | PERSON | 0.99+ |
Andy Chassey | PERSON | 0.99+ |
Adam | PERSON | 0.99+ |
Daniel | PERSON | 0.99+ |
Andy Jessey | PERSON | 0.99+ |
2017 | DATE | 0.99+ |
Daniel Re Myer | PERSON | 0.99+ |
Germany | LOCATION | 0.99+ |
Fred | PERSON | 0.99+ |
Samir Daniel | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Stephen Schmidt | PERSON | 0.99+ |
Danielle | PERSON | 0.99+ |
2016 | DATE | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Samia | PERSON | 0.99+ |
two companies | QUANTITY | 0.99+ |
2025 | DATE | 0.99+ |
Andy Chas | PERSON | 0.99+ |
John Fur | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
2013 | DATE | 0.99+ |
36 | QUANTITY | 0.99+ |
Pat Gelsinger | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
two questions | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Nora | PERSON | 0.99+ |
Madhura Maskasky & Sirish Raghuram | KubeCon + CloudNativeCon NA 2022
(upbeat synth intro music) >> Hey everyone and welcome to Detroit, Michigan. theCUBE is live at KubeCon CloudNativeCon, North America 2022. Lisa Martin here with John Furrier. John, this event, the keynote that we got out of a little while ago was, standing room only. The Solutions hall is packed. There's so much buzz. The community is continuing to mature. They're continuing to contribute. One of the big topics is Cloud Native at Scale. >> Yeah, I mean, this is a revolution happening. The developers are coming on board. They will be running companies. Developers, structurally, will be transforming companies with just, they got to get powered somewhere. And, I think, the Cloud Native at Scale speaks to getting everything under the covers, scaling up to support developers. In this next segment, we have two Kube alumnis. We're going to talk about Cloud Native at Scale. Some of the things that need to be there in a unified architecture, should be great. >> All right, it's going to be fantastic. Let's go under the covers here, as John mentioned, two alumni with us, Madhura Maskasky joins us, co-founder of Platform9. Sirish Raghuram, also co-founder of Platform9 joins us. Welcome back to theCUBE. Great to have you guys here at KubeCon on the floor in Detroit. >> Thank you for having us. >> Thank you for having us. >> Excited to be here >> So, talk to us. You guys have some news, Madhura, give us the sneak peak. What's going on? >> Definitely, we are very excited. So, we have John, not too long ago we spoke about our very new open source project called Arlon. And, we were talking about the launch of Arlon in terms of its first release and etcetera. And, just fresh hot of the press, we, Platform9 had its 5.6 release which is its most recent release of our product. And there's a number of key interesting announcements that we'd like to share as part of that. I think, the prominent one is, Platform9 added support for EKS Kubernetes cluster management. And, so, this is part of our vision of being able to add value, no matter where you run your Kubernetes clusters, because, Kubernetes or cluster management, is increasingly becoming commodity. And, so, I think the companies that succeed are going to add value on top, and are going to add value in a way that helps end users, developers, DevOps solve problems that they encounter as they start running these environments, with a lot of scale and a lot of diversity. So, towards that, key features in the 5.6 six release. First, is the very first package release of the product online, which is the open source project that we've kicked off to do cluster and application, entire cluster management at scale. And, then there's few other very interesting capabilities coming out of that. >> I want to just highlight something and then get your thoughts on this next, this release 5.6. First of all, 5.6, it's been around for a while, five reps, but, now, more than ever, you mentioned the application in Ops. You're seeing WebAssembly trends, you're seeing developers getting more and more advanced capability. It's going to accelerate their ability to write code and compose applications. So, you're seeing a application tsunami coming. So, the pressure is okay, they're going to need infrastructure to run all that stuff. And, so, you're seeing more clusters being spun up, more intelligence trying to automate. So you got the automation, so you got the dynamic, the power dynamic of developers and then under the covers. What does 5.6 do to push the mission forward for developers? How would you guys summarize that for people watching? what's in it for them right now? >> So it's, I think going back to what you just said, right, the breadth of applications that people are developing on top of something like Kubernetes and Cloud Native, is always growing. So, it's not just a number of clusters, but also the fact that different applications and different development groups need these clusters to be composed differently. So, a certain version of the application may require some set of build components, add-ons, and operators, and extensions. Whereas, a different application may require something entirely different. And, now, you take this in an enterprise context, right. Like, we had a major media company that worked with us. They have more than 10,000 pods being used by thousands of developers. And, you now think about the breadth of applications, the hundreds of different applications being built. how do you consistently build, and compose, and manage, a large number of communities clusters with a a large variety of extensions that these companies are trying to manage? That's really what I think 5.6 is bringing to the table. >> Scott Johnston just was on here early as the CEO of Docker. He said there's more applications being pushed now than in the history of application development combined. There's more and more apps coming, more and more pressure on the system. >> And, that's where, if you go, there's this famous landscape chart of the CNCF ecosystem technologies. And, the problem that people here have is, how do they put it all together? How do they make sense of it? And, what 5.6 and Arlon and what Platform9 is doing is, it's helping you declaratively capture blueprints of these clusters, using templates, and be able to manage a small number of blueprints that helps you make order out of the chaos of these hundreds of different projects, that are all very interesting and powerful. >> So Project Arlon really helping developers produce the configuration and the deployment complexities of Kubernetes at scale. >> That's exactly right. >> Talk about the, the impact on the business side. Ease of use, what's the benefits for 5.6? What's does it turn into for a benefit standpoint? >> Yeah, I think the biggest benefit, right, is being able to do Cloud Native at Scale faster, and while still keeping a very lean Ops team that is able to spend, let's say 70 plus percent of their time, caring for your actual business bread and butter applications, and not for the infrastructure that serves it, right. If you take the analogy of a restaurant, you don't want to spend 70% of your time in building the appliances or setting up your stoves etcetera. You want to spend 90 plus percent of your time cooking your own meal, because, that is your core key ingredient. But, what happens today in most enterprises is, because, of the level of automation, the level of hands-on available tooling, being there or not being there, majority of the ops time, I would say 50, 70% plus, gets spent in making that kitchen set up and ready, right. And, that is exactly what we are looking to solve, online. >> What would a customer look like, or prospect environment look like that would be really ready for platform9? What, is it more apps being pushed, big push on application development, or is it the toil of like really inefficient infrastructure, or gaps in skills of people? What does an environment look like? So, someone needs to look at their environment and say, okay, maybe I should call platform9. What's it look like? >> So, we generally see customers fall into two ends of the barbell, I would say. One, is the advanced communities users that are running, I would say, typically, 30 or more clusters already. These are the people that already know containers. They know, they've container wise... >> Savvy teams. >> They're savvy teams, a lot of them are out here. And for them, the problem is, how do I manage the complexity at scale? Because, now, the problem is how do I scale us? So, that's one end of the barbell. The other end of the barbell, is, how do we help make Kubernetes accessible to companies that, as what I would call the mainstream enterprise. We're in Detroit in Motown, right, And, we're outside of the echo chamber of the Silicon Valley. Here's the biggest truth, right. For all the progress that we made as a community, less than 20% of applications in the enterprise today are running on Kubernetes. So, what does it take? I would say it's probably less than 10%, okay. And, what does it take, to grow that in order of magnitude? That's the other kind of customer that we really serve, is, because, we have technologies like Kube Word, which helps them take their existing applications and start adopting Kubernetes as a directional roadmap, but, while using the existing applications that they have, without refactoring it. So, I would say those are the two ends of the barbell. The early adopters that are looking for an easier way to adopt Kubernetes as an architectural pattern. And, the advanced savvy users, for whom the problem is, how do they operationally solve the complexity of managing at scale. >> And, what is your differentiation message to both of those different user groups, as you talked about in terms of the number of users of Kubernetes so far? The community groundswell is tremendous, but, there's a lot of opportunity there. You talked about some of the barriers. What's your differentiation? What do you come in saying, this is why Platform9 is the right one for you, in the both of these groups. >> And it's actually a very simple message. We are the simplest and easiest way for a new user that is adopting Kubernetes as an architectural pattern, to get started with existing applications that they have, on the infrastructure that they have. Number one. And, for the savvy teams, our technology helps you operate with greater scale, with constrained operations teams. Especially, with the economy being the way it is, people are not going to get a lot more budget to go hire a lot more people, right. So, that all of them are being asked to do more with less. And, our team, our technology, and our teams, help you do more with less. >> I was talking with Phil Estes last night from AWS. He's here, he is one of their engineer open source advocates. He's always on the ground pumping up AWS. They've had great success, Amazon Web Services, with their EKS. A lot of people adopting clusters on the cloud and on-premises. But Amazon's doing well. You guys have, I think, a relationship with AWS. What's that, If I'm an Amazon customer, how do I get involved with Platform9? What's the hook? Where's the value? What's the product look like? >> Yeah, so, and it kind of goes back towards the point we spoke about, which is, Kubernetes is going to increasingly get commoditized. So, customers are going to find the right home whether it's hyperscalers, EKS, AKS, GKE, or their own infrastructure, to run Kubernetes. And, so, where we want to be at, is, with a project like Arlon, Sirish spoke about the barbell strategy, on one end there is these advanced Kubernetes users, majority of them are running Kubernetes on AKS, right? Because, that was the easiest platform that they found to get started with. So, now, they have a challenge of running these 50 to 100 clusters across various regions of Amazon, across their DevTest, their staging, their production. And, that results in a level of chaos that these DevOps or platform... >> So you come in and solve that. >> That is where we come in and we solve that. And it, you know, Amazon or EKS, doesn't give you tooling to solve that, right. It makes it very easy for you to create those number of clusters. >> Well, even in one hyperscale, let's say AWS, you got regions and locations... >> Exactly >> ...that's kind of a super cloud problem, we're seeing, opportunity problem, and opportunity is that, on Amazon, availability zones is one thing, but, now, also, you got regions. >> That is absolutely right. You're on point John. And the way we solve it, is by using infrastructure as a code, by using GitOps principles, right? Where you define it once, you define it in a yaml file, you define exactly how for your DevTest environment you want your entire infrastructure to look like, including EKS. And then you stamp it out. >> So let me, here's an analogy, I'll throw out this. You guys are like, someone learns how to drive a car, Kubernetes clusters, that's got a couple clusters. Then once they know how to drive a car, you give 'em the sports car. You allow them to stay on Amazon and all of a sudden go completely distributed, Edge, Global. >> I would say that a lot of people that we meet, we feel like they're figuring out how to build a car with the kit tools that they have. And we give them a car that's ready to go and doesn't require them to be trying to... ... they can focus on driving the car, rather than trying to build the car. >> You don't want people to stop, once they get the progressions, they hit that level up on Kubernetes, you guys give them the ability to go much bigger and stronger. >> That's right. >> To accelerate that applications. >> Building a car gets old for people at a certain point in time, and they really want to focus on is driving it and enjoying it. >> And we got four right behind us, so, we'll get them involved. So that's... >> But, you're not reinventing the wheel. >> We're not at all, because, what we are building is two very, very differentiated solutions, right. One, is, we're the simplest and easiest way to build and run Cloud Native private clouds. And, this is where the operational complexity of trying to do it yourself. You really have to be a car builder, to be able to do this with our Platform9. This is what we do uniquely that nobody else does well. And, the other end is, we help you operate at scale, in the hyperscalers, right. Those are the two problems that I feel, whether you're on-prem, or in the cloud, these are the two problems people face. How do you run a private cloud more easily, more efficiently? And, how do you govern at scale, especially in the public clouds? >> I want to get to two more points before we run out of time. Arlon and Argo CD as a service. We previously mentioned up coming into KubeCon, but, here, you guys couldn't be more relevant, 'cause Intuit was on stage on the keynote, getting an award for their work. You know, Argo, it comes from Intuit. That ArgoCon was in Mountain View. You guys were involved in that. You guys were at the center of all this super cloud action, if you will, or open source. How does Arlon fit into the Argo extension? What is Argo CD as a service? Who's going to take that one? I want to get that out there, because, Arlon has been talked about a lot. What's the update? >> I can talk about it. So, one of the things that Arlon uses behind the scenes, is it uses Argo CD, open source Argo CD as a service, as its key component to do the continuous deployment portion of its entire, the infrastructure management story, right. So, we have been very strongly partnering with Argo CD. We, really know and respect the Intuit team a lot. We, as part of this effort, in 5.6 release, we've also put out Argo CD as a service, in its GA version, right. Because, the power of running Arlon along with Argo CD as a service, in our mind, is enabling you to run on one end, your infrastructure as a scale, through GitOps, and infrastructure as a code practices. And on the other end, your entire application fleet, at scale, right. And, just marrying the two, really gives you the ability to perform that automation that we spoke about. >> But, and avoid the problem of sprawl when you have distributed teams, you have now things being bolted on, more apps coming out. So, this is really solves that problem, mainly. >> That is exactly right. And if you think of it, the way those problems are solved today, is, kind of in disconnected fashion, which is on one end you have your CI/CD tools, like Argo CD is an excellent one. There's some other choices, which are managed by a separate team to automate your application delivery. But, that team, is disconnected from the team that does the infrastructure management. And the infrastructure management is typically done through a bunch of Terraform scripts, or a bunch of ad hoc homegrown scripts, which are very difficult to manage. >> So, Arlon changes sure, as they change the complexity and also the sprawl. But, that's also how companies can die. They're growing fast, they're adding more capability. That's what trouble starts, right? >> I think in two ways, right. Like one is, as Madhura said, I think one of the common long-standing problems we've had, is, how do infrastructure and application teams communicate and work together, right. And, you've seen Argo's really get adopted by the application teams, but, it's now something that we are making accessible for the infrastructure teams to also bring the best practices of how application teams are managing applications. You can now use that to manage infrastructure, right. And, what that's going to do is, help you ultimately reduce waste, reduce inefficiency, and improve the developer experience. Because, that's what it's all about, ultimately. >> And, I know that you just released 5.6 today, congratulations on that. Any customer feedback yet? Any, any customers that you've been able to talk to, or have early access? >> Yeah, one of our large customers is a large SaaS retail company that is B2C SaaS. And, their feedback has been that this, basically, helps them bring exactly what I said in terms of bring some of the best practices that they wanted to adopt in the application space, down to the infrastructure management teams, right. And, we are also hearing a lot of customers, that I would say, large scale public cloud users, saying, they're really struggling with the complexity of how to tame the complexity of navigating that landscape and making it consumable for organizations that have thousands of developers or more. And that's been the feedback, is that this is the first open source standard mechanism that allows them to kind of reuse something, as opposed to everybody feels like they've had to build ad hoc solutions to solve this problem so far. >> Having a unified infrastructure is great. My final question, for me, before I end up, for Lisa to ask her last question is, if you had to explain Platform9, why you're relevant and cool today, what would you say? >> If I take that? I would say that the reason why Platform9, the reason why we exist, is, putting together a cloud, a hybrid cloud strategy for an enterprise today, historically, has required a lot of DIY, a lot of building your own car. Before you can drive a car, or you can enjoy the car, you really learn to build and operate the car. And that's great for maybe a 100 tech companies of the world, but, for the next 10,000 or 50,000 enterprises, they want to be able to consume a car. And that's why Platform9 exists, is, we are the only company that makes this delightfully simple and easy for companies that have a hybrid cloud strategy. >> Why you cool and relevant? How would you say it? >> Yeah, I think as Kubernetes becomes mainstream, as containers have become mainstream, I think automation at scale with ease, is going to be the key. And that's exactly what we help solve. Automation at scale and with ease. >> With ease and that differentiation. Guys, thank you so much for joining me. Last question, I guess, Madhura, for you, is, where can Devs go to learn more about 5.6 and get their hands on it? >> Absolutely. Go to platform9.com. There is info about 5.6 release, there's a press release, there's a link to it right on the website. And, if they want to learn about Arlon, it's an open source GitHub project. Go to GitHub and find out more about it. >> Excellent guys, thanks again for sharing what you're doing to really deliver Cloud Native at Scale in a differentiated way that adds ostensible value to your customers. John, and I, appreciate your insights and your time. >> Thank you for having us. >> Thanks so much >> Our pleasure. For our guests and John Furrier, I'm Lisa Martin. You're watching theCUBE Live from Detroit, Michigan at KubeCon CloudNativeCon 2022. Stick around, John and I will be back with our next guest. Just a minute. (light synth outro music)
SUMMARY :
One of the big topics is Some of the things that need to be there Great to have you guys here at KubeCon So, talk to us. And, just fresh hot of the press, So, the pressure is okay, they're to what you just said, right, as the CEO of Docker. of the CNCF ecosystem technologies. produce the configuration and impact on the business side. because, of the level of automation, or is it the toil of One, is the advanced communities users of the Silicon Valley. in the both of these groups. And, for the savvy teams, He's always on the ground pumping up AWS. that they found to get started with. And it, you know, Amazon or you got regions and locations... but, now, also, you got regions. And the way we solve it, Then once they know how to drive a car, of people that we meet, to go much bigger and stronger. and they really want to focus on And we got four right behind us, And, the other end is, What's the update? And on the other end, your But, and avoid the problem of sprawl that does the infrastructure management. and also the sprawl. for the infrastructure teams to also bring And, I know that you of bring some of the best practices today, what would you say? of the world, ease, is going to be the key. to learn more about 5.6 there's a link to it right on the website. to your customers. be back with our next guest.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Madhura Maskasky | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
John | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Sirish Raghuram | PERSON | 0.99+ |
Madhura | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Detroit | LOCATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Scott Johnston | PERSON | 0.99+ |
30 | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
Sirish | PERSON | 0.99+ |
50 | QUANTITY | 0.99+ |
Amazon Web Services | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
Platform9 | ORGANIZATION | 0.99+ |
two problems | QUANTITY | 0.99+ |
Phil Estes | PERSON | 0.99+ |
100 tech companies | QUANTITY | 0.99+ |
less than 20% | QUANTITY | 0.99+ |
less than 10% | QUANTITY | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Detroit, Michigan | LOCATION | 0.99+ |
First | QUANTITY | 0.99+ |
KubeCon | EVENT | 0.99+ |
both | QUANTITY | 0.99+ |
Motown | LOCATION | 0.99+ |
first release | QUANTITY | 0.99+ |
more than 10,000 pods | QUANTITY | 0.99+ |
Docker | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
two alumni | QUANTITY | 0.99+ |
two ways | QUANTITY | 0.99+ |
Arlon | ORGANIZATION | 0.99+ |
5.6 | QUANTITY | 0.98+ |
Mountain View | LOCATION | 0.98+ |
One | QUANTITY | 0.98+ |
two more points | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
EKS | ORGANIZATION | 0.98+ |
last night | DATE | 0.98+ |
Cloud Native | TITLE | 0.98+ |
70 plus percent | QUANTITY | 0.97+ |
one end | QUANTITY | 0.97+ |
four | QUANTITY | 0.97+ |
90 plus percent | QUANTITY | 0.97+ |
DevTest | TITLE | 0.97+ |
Argo | ORGANIZATION | 0.97+ |
50,000 enterprises | QUANTITY | 0.96+ |
Kube | ORGANIZATION | 0.96+ |
two ends | QUANTITY | 0.96+ |
Intuit | ORGANIZATION | 0.96+ |
five reps | QUANTITY | 0.96+ |
today | DATE | 0.96+ |
Kubernetes | TITLE | 0.95+ |
GitOps | TITLE | 0.95+ |
Cloud Native | TITLE | 0.95+ |
platform9.com | OTHER | 0.95+ |
hundreds of different applications | QUANTITY | 0.95+ |
David Flynn Supercloud Audio
>> From every ISV to solve the problems. You want there to be tools in place that you can use, either open source tools or whatever it is that help you build it. And slowly over time, that building will become easier and easier. So my question to you was, where do you see you playing? Do you see yourself playing to ISVs as a set of tools, which will make their life a lot easier and provide that work? >> Absolutely. >> If they don't have, so they don't have to do it. Or you're providing this for the end users? Or both? >> So it's a progression. If you go to the ISVs first, you're doomed to starved before you have time for that other option. >> Yeah. >> Right? So it's a question of phase, the phasing of it. And also if you go directly to end users, you can demonstrate the power of it and get the attention of the ISVs. I believe that the ISVs, especially those with the biggest footprints and the most, you know, coveted estates, they have already made massive investments at trying to solve decentralization of their software stack. And I believe that they have used it as a hook to try to move to a software as a service model and rope people into leasing their infrastructure. So if you look at the clouds that have been propped up by Autodesk or by Adobe, or you name the company, they are building proprietary makeshift solutions for decentralizing or hybrid clouding. Or maybe they're not even doing that at all and all they're is saying hey, if you want to get location agnosticness, then what you should just, is just move into our cloud. >> Right. >> And then they try to solve on the background how to decentralize it between different regions so they can have decent offerings in each region. But those who are more advanced have already made larger investments and will be more averse to, you know, throwing that stuff away, all of their makeshift machinery away, and using a platform that gives them high performance parallel, low level file system access, while at the same time having metadata-driven, you know, policy-based, intent-based orchestration to manage the diffusion of data across a decentralized infrastructure. They are not going to be as open because they've made such an investment and they're going to look at how do they monetize it. So what we have found with like the movie studios who are using us already, many of the app they're using, many of those software offerings, the ISVs have their own cloud that offers that software for the cloud. But what we got when I asked about this, 'cause I was dealt specifically into this question because I'm very interested to know how we're going to make that leap from end user upstream into the ISVs where I believe we need to, and they said, look, we cannot use these software ISV-specific SAS clouds for two reasons. Number one is we lose control of the data. We're giving it to them. That's security and other issues. And here you're talking about we're doing work for Disney, we're doing work for Netflix, and they're not going to let us put our data on those software clouds, on those SAS clouds. Secondly, in any reasonable pipeline, the data is shared by many different applications. We need to be agnostic as to the application. 'Cause the inputs to one application, you know, the output for one application provides the input to the next, and it's not necessarily from the same vendor. So they need to have a data platform that lets them, you know, go from one software stack, and you know, to run it on another. Because they might do the rendering with this and yet, they do the editing with that, and you know, et cetera, et cetera. So I think the further you go up the stack in the structured data and dedicated applications for specific functions in specific verticals, the further up the stack you go, the harder it is to justify a SAS offering where you're basically telling the end users you need to park all your data with us and then you can run your application in our cloud and get this. That ultimately is a dead end path versus having the data be open and available to many applications across this supercloud layer. >> Okay, so-- >> Is that making any sense? >> Yes, so if I could just ask a clarifying question. So, if I had to take Snowflake as an example, I think they're doing exactly what you're saying is a dead end, put everything into our proprietary system and then we'll figure out how to distribute it. >> Yeah. >> And and I think if you're familiar with Zhamak Dehghaniis' data mesh concept. Are you? >> A little bit, yeah. >> But in her model, Snowflake, a Snowflake warehouse is just a node on the mesh and that mesh is-- >> That's right. >> Ultimately the supercloud and you're an enabler of that is what I'm hearing. >> That's right. What they're doing up at the structured level and what they're talking about at the structured level we're doing at the underlying, unstructured level, which by the way has implications for how you implement those distributed database things. In other words, implementing a Snowflake on top of Hammerspace would have made building stuff like in the first place easier. It would allow you to easily shift and run the database engine anywhere. You still have to solve how to shard and distribute at the transaction layer above, so I'm not saying we're a substitute for what you need to do at the app layer. By the way, there is another example of that and that's Microsoft Office, right? It's one thing to share that, to have a file share where you can share all the docs. It's something else to have Word and PowerPoint, Excel know how to allow people to be simultaneously editing the same doc. That's always going to happen in the app layer. But not all applications need that level of, you know, in-app decentralization. You know, many of them, many workflows are pipelined, especially the ones that are very data intensive where you're doing drug discovery or you're doing rendering, or you're doing machine learning training. These things are human in the loop with large stages of processing across tens of thousands of cores. And I think that kind of data processing pipeline is what we're focusing on first. Not so much the Microsoft Office or the Snowflake, you know, parking a relational database because that takes a lot of application layer stuff and that's what they're good at. >> Right. >> But I think... >> Go ahead, sorry. >> Later entrance in these markets will find Hammerspace as a way to accelerate their work so they can focus more narrowly on just the stuff that's app-specific, higher level sharing in the app. >> Yes, Snowflake founders-- >> I think it might be worth mentioning also, just keep this confidential guys, but one of our customers is Blue Origin. And one of the things that we have found is kind of the point of what you're talking about with our customers. They're needing to build this and since it's not commercially available or they don't know where to look for it to be commercially available, they're all building themselves. So this layer is needed. And Blue is just one of the examples of quite a few we're now talking to. And like manufacturing, HPC, research where they're out trying to solve this problem with their own scripting tools and things like that. And I just, I don't know if there's anything you want to add, David, but you know, but there's definitely a demand here and customers are trying to figure out how to solve it beyond what Hammerspace is doing. Like the need is so great that they're just putting developers on trying to do it themselves. >> Well, and you know, Snowflake founders, they didn't have a Hammerspace to lean on. But, one of the things that's interesting about supercloud is we feel as though industry clouds will emerge, that as part of company's digital transformations, they will, you know, every company's a software company, they'll begin to build their own clouds and they will be able to use a Hammerspace to do that. >> A super pass layer. >> Yes. It's really, I don't know if David's speaking, I don't want to speak over him, but we can't hear you. May be going through a bad... >> Well, a regional, regional talks that make that possible. And so they're doing these render farms and editing farms, and it's a cloud-specific to the types of workflows in the median entertainment world. Or clouds specifically to workflows in the chip design world or in the drug and bio and life sciences exploration world. There are large organizations that are kind of a blend of end users, like the Broad, which has their own kind of cloud where they're asking collaborators to come in and work with them. So it starts to even blur who's an end user versus an ISV. >> Yes. >> Right? When you start talking about the massive data is the main gravity is to having lots of people participate. >> Yep, and that's where the value is. And that's where the value is. And this is a megatrend that we see. And so it's really important for us to get to the point of what is and what is not a supercloud and, you know, that's where we're trying to evolve. >> Let's talk about this for a second 'cause I want to, I want to challenge you on something and it's something that I got challenged on and it has led me to thinking differently than I did at first, which Molly can attest to. Okay? So, we have been looking for a way to talk about the concept of cloud of utility computing, run anything anywhere that isn't addressed in today's realization of cloud. 'Cause today's cloud is not run anything anywhere, it's quite the opposite. You park your data in AWS and that's where you run stuff. And you pretty much have to. Same with with Azure. They're using data gravity to keep you captive there, just like the old infrastructure guys did. But now it's even worse because it's coupled back with the software to some degree, as well. And you have to use their storage, networking, and compute. It's not, I mean it fell back to the mainframe era. Anyhow, so I love the concept of supercloud. By the way, I was going to suggest that a better term might be hyper cloud since hyper speaks to the multidimensionality of it and the ability to be in a, you know, be in a different dimension, a different plane of existence kind of thing like hyperspace. But super and hyper are somewhat synonyms. I mean, you have hyper cars and you have super cars and blah, blah, blah. I happen to like hyper maybe also because it ties into the whole Hammerspace notion of a hyper-dimensional, you know, reality, having your data centers connected by a wormhole that is Hammerspace. But regardless, what I got challenged on is calling it something different at all versus simply saying, this is what cloud has always meant to be. This is the true cloud, this is real cloud, this is cloud. And I think back to what happened, you'll remember, at Fusion IO we talked about IO memory and we did that because people had a conceptualization of what an SSD was. And an SSD back then was low capacity, low endurance, made to go military, aerospace where things needed to be rugged but was completely useless in the data center. And we needed people to imagine this thing as being able to displace entire SAND, with the kind of capacity density, performance density, endurance. And so we talked IO memory, we could have said enterprise SSD, and that's what the industry now refers to for that concept. What will people be saying five and 10 years from now? Will they simply say, well this is cloud as it was always meant to be where you are truly able to run anything anywhere and have not only the same APIs, but you're same data available with high performance access, all forms of access, block file and object everywhere. So yeah. And I wonder, and this is just me throwing it out there, I wonder if, well, there's trade offs, right? Giving it a new moniker, supercloud, versus simply talking about how cloud is always intended to be and what it was meant to be, you know, the real cloud or true cloud, there are trade-offs. By putting a name on it and branding it, that lets people talk about it and understand they're talking about something different. But it also is that an affront to people who thought that that's what they already had. >> What's different, what's new? Yes, and so we've given a lot of thought to this. >> Right, it's like you. >> And it's because we've been asked that why does the industry need a new term, and we've tried to address some of that. But some of the inside baseball that we haven't shared is, you remember the Web 2.0, back then? >> Yep. >> Web 2.0 was the same thing. And I remember Tim Burners Lee saying, "Why do we need Web 2.0? "This is what the Web was always supposed to be." But the truth is-- >> I know, that was another perfect-- >> But the truth is it wasn't, number one. Number two, everybody hated the Web 2.0 term. John Furrier was actually in the middle of it all. And then it created this groundswell. So one of the things we wrote about is that supercloud is an evocative term that catalyzes debate and conversation, which is what we like, of course. And maybe that's self-serving. But yeah, HyperCloud, Metacloud, super, meaning, it's funny because super came from Latin supra, above, it was never the superlative. But the superlative was a convenient byproduct that caused a lot of friction and flack, which again, in the media business is like a perfect storm brewing. >> The bad thing to have to, and I think you do need to shake people out of their, the complacency of the limitations that they're used to. And I'll tell you what, the fact that you even have the terms hybrid cloud, multi-cloud, private cloud, edge computing, those are all just referring to the different boundaries that isolate the silo that is the current limited cloud. >> Right. >> So if I heard correctly, what just, in terms of us defining what is and what isn't in supercloud, you would say traditional applications which have to run in a certain place, in a certain cloud can't run anywhere else, would be the stuff that you would not put in as being addressed by supercloud. And over time, you would want to be able to run the data where you want to and in any of those concepts. >> Or even modern apps, right? Or even modern apps that are siloed in SAS within an individual cloud, right? >> So yeah, I guess it's twofold. Number one, if you're going at the high application layers, there's lots of ways that you can give the appearance of anything running anywhere. The ISV, the SAS vendor can engineer stuff to have the ability to serve with low enough latency to different geographies, right? So if you go too high up the stack, it kind of loses its meaning because there's lots of different ways to make due and give the appearance of omni-presence of the service. Okay? As you come down more towards the platform layer, it gets harder and harder to mask the fact that supercloud is something entirely different than just a good regionally-distributed SAS service. So I don't think you, I don't think you can distinguish supercloud if you go too high up the stack because it's just SAS, it's just a good SAS service where the SAS vendor has done the hard work to give you low latency access from different geographic regions. >> Yeah, so this is one of the hardest things, David. >> Common among them. >> Yeah, this is really an important point. This is one of the things I've had the most trouble with is why is this not just SAS? >> So you dilute your message when you go up to the SAS layer. If you were to focus most of this around the super pass layer, the how can you host applications and run them anywhere and not host this, not run a service, not have a service available everywhere. So how can you take any application, even applications that are written, you know, in a traditional legacy data center fashion and be able to run them anywhere and have them have their binaries and their datasets and the runtime environment and the infrastructure to start them and stop them? You know, the jobs, the, what the Kubernetes, the job scheduler? What we're really talking about here, what I think we're really talking about here is building the operating system for a decentralized cloud. What is the operating system, the operating environment for a decentralized cloud? Where you can, and that the main two functions of an operating system or an operating environment are the process scheduler, the thing that's scheduling what is running where and when and so forth, and the file system, right? The thing that's supplying a common view and access to data. So when we talk about this, I think that the strongest argument for supercloud is made when you go down to the platform layer and talk of it, talk about it as an operating environment on which you can run all forms of applications. >> Would you exclude--? >> Not a specific application that's been engineered as a SAS. (audio distortion) >> He'll come back. >> Are you there? >> Yeah, yeah, you just cut out for a minute. >> I lost your last statement when you broke up. >> We heard you, you said that not the specific application. So would you exclude Snowflake from supercloud? >> Frankly, I would. I would. Because, well, and this is kind of hard to do because Snowflake doesn't like to, Frank doesn't like to talk about Snowflake as a SAS service. It has a negative connotation. >> But it is. >> I know, we all know it is. We all know it is and because it is, yes, I would exclude them. >> I think I actually have him on camera. >> There's nothing in common. >> I think I have him on camera or maybe Benoit as saying, "Well, we are a SAS." I think it's Slootman. I think I said to Slootman, "I know you don't like to say you're a SAS." And I think he said, "Well, we are a SAS." >> Because again, if you go to the top of the application stack, there's any number of ways you can give it location agnostic function or you know, regional, local stuff. It's like let's solve the location problem by having me be your one location. How can it be decentralized if you're centralizing on (audio distortion)? >> Well, it's more decentralized than if it's all in one cloud. So let me actually, so the spectrum. So again, in the spirit of what is and what isn't, I think it's safe to say Hammerspace is supercloud. I think there's no debate there, right? Certainly among this crowd. And I think we can all agree that Dell, Dell Storage is not supercloud. Where it gets fuzzy is this Snowflake example or even, how about a, how about a Cohesity that instantiates its stack in different cloud regions in different clouds, and synchronizes, however magic sauce it does that. Is that a supercloud? I mean, so I'm cautious about having too strict of a definition 'cause then only-- >> Fair enough, fair enough. >> But I could use your help and thoughts on that. >> So I think we're talking about two different spectrums here. One is the spectrum of platform to application-specific. As you go up the application stack and it becomes this specific thing. Or you go up to the more and more structured where it's serving a specific application function where it's more of a SAS thing. I think it's harder to call a SAS service a supercloud. And I would argue that the reason there, and what you're lacking in the definition is to talk about it as general purpose. Okay? Now, that said, a data warehouse is general purpose at the structured data level. So you could make the argument for why Snowflake is a supercloud by saying that it is a general purpose platform for doing lots of different things. It's just one at a higher level up at the structured data level. So one spectrum is the high level going from platform to, you know, unstructured data to structured data to very application-specific, right? Like a specific, you know, CAD/CAM mechanical design cloud, like an Autodesk would want to give you their cloud for running, you know, and sharing CAD/CAM designs, doing your CAD/CAM anywhere stuff. Well, the other spectrum is how well does the purported supercloud technology actually live up to allowing you to run anything anywhere with not just the same APIs but with the local presence of data with the exact same runtime environment everywhere, and to be able to correctly manage how to get that runtime environment anywhere. So a Cohesity has some means of running things in different places and some means of coordinating what's where and of serving diff, you know, things in different places. I would argue that it is a very poor approximation of what Hammerspace does in providing the exact same file system with local high performance access everywhere with metadata ability to control where the data is actually instantiated so that you don't have to wait for it to get orchestrated. But even then when you do have to wait for it, it happens automatically and so it's still only a matter of, well, how quick is it? And on the other end of the spectrum is you could look at NetApp with Flexcache and say, "Is that supercloud?" And I would argue, well kind of because it allows you to run things in different places because it's a cache. But you know, it really isn't because it presumes some central silo from which you're cacheing stuff. So, you know, is it or isn't it? Well, it's on a spectrum of exactly how fully is it decoupling a runtime environment from specific locality? And I think a cache doesn't, it stretches a specific silo and makes it have some semblance of similar access in other places. But there's still a very big difference to the central silo, right? You can't turn off that central silo, for example. >> So it comes down to how specific you make the definition. And this is where it gets kind of really interesting. It's like cloud. Does IBM have a cloud? >> Exactly. >> I would say yes. Does it have the kind of quality that you would expect from a hyper-scale cloud? No. Or see if you could say the same thing about-- >> But that's a problem with choosing a name. That's the problem with choosing a name supercloud versus talking about the concept of cloud and how true up you are to that concept. >> For sure. >> Right? Because without getting a name, you don't have to draw, yeah. >> I'd like to explore one particular or bring them together. You made a very interesting observation that from a enterprise point of view, they want to safeguard their store, their data, and they want to make sure that they can have that data running in their own workflows, as well as, as other service providers providing services to them for that data. So, and in in particular, if you go back to, you go back to Snowflake. If Snowflake could provide the ability for you to have your data where you wanted, you were in charge of that, would that make Snowflake a supercloud? >> I'll tell you, in my mind, they would be closer to my conceptualization of supercloud if you can instantiate Snowflake as software on your own infrastructure, and pump your own data to Snowflake that's instantiated on your own infrastructure. The fact that it has to be on their infrastructure or that it's on their, that it's on their account in the cloud, that you're giving them the data and they're, that fundamentally goes against it to me. If they, you know, they would be a pure, a pure plate if they were a software defined thing where you could instantiate Snowflake machinery on the infrastructure of your choice and then put your data into that machinery and get all the benefits of Snowflake. >> So did you see--? >> In other words, if they were not a SAS service, but offered all of the similar benefits of being, you know, if it were a service that you could run on your own infrastructure. >> So did you see what they announced, that--? >> I hope that's making sense. >> It does, did you see what they announced at Dell? They basically announced the ability to take non-native Snowflake data, read it in from an object store on-prem, like a Dell object store. They do the same thing with Pure, read it in, running it in the cloud, and then push it back out. And I was saying to Dell, look, that's fine. Okay, that's interesting. You're taking a materialized view or an extended table, whatever you're doing, wouldn't it be more interesting if you could actually run the query locally with your compute? That would be an extension that would actually get my attention and extend that. >> That is what I'm talking about. That's what I'm talking about. And that's why I'm saying I think Hammerspace is more progressive on that front because with our technology, anybody who can instantiate a service, can make a service. And so I, so MSPs can use Hammerspace as a way to build a super pass layer and host their clients on their infrastructure in a cloud-like fashion. And their clients can have their own private data centers and the MSP or the public clouds, and Hammerspace can be instantiated, get this, by different parties in these different pieces of infrastructure and yet linked together to make a common file system across all of it. >> But this is data mesh. If I were HPE and Dell it's exactly what I'd be doing. I'd be working with Hammerspace to create my own data. I'd work with Databricks, Snowflake, and any other-- >> Data mesh is a good way to put it. Data mesh is a good way to put it. And this is at the lowest level of, you know, the underlying file system that's mountable by the operating system, consumed as a real file system. You can't get lower level than that. That's why this is the foundation for all of the other apps and structured data systems because you need to have a data mesh that can at least mesh the binary blob. >> Okay. >> That hold the binaries and that hold the datasets that those applications are running. >> So David, in the third week of January, we're doing supercloud 2 and I'm trying to convince John Furrier to make it a data slash data mesh edition. I'm slowly getting him to the knothole. I would very much, I mean you're in the Bay Area, I'd very much like you to be one of the headlines. As Zhamak Dehghaniis going to speak, she's the creator of Data Mesh, >> Sure. >> I'd love to have you come into our studio as well, for the live session. If you can't make it, we can pre-record. But you're right there, so I'll get you the dates. >> We'd love to, yeah. No, you can count on it. No, definitely. And you know, we don't typically talk about what we do as Data Mesh. We've been, you know, using global data environment. But, you know, under the covers, that's what the thing is. And so yeah, I think we can frame the discussion like that to line up with other, you know, with the other discussions. >> Yeah, and Data Mesh, of course, is one of those evocative names, but she has come up with some very well defined principles around decentralized data, data as products, self-serve infrastructure, automated governance, and and so forth, which I think your vision plugs right into. And she's brilliant. You'll love meeting her. >> Well, you know, and I think.. Oh, go ahead. Go ahead, Peter. >> Just like to work one other interface which I think is important. How do you see yourself and the open source? You talked about having an operating system. Obviously, Linux is the operating system at one level. How are you imagining that you would interface with cost community as part of this development? >> Well, it's funny you ask 'cause my CTO is the kernel maintainer of the storage networking stack. So how the Linux operating system perceives and consumes networked data at the file system level, the network file system stack is his purview. He owns that, he wrote most of it over the last decade that he's been the maintainer, but he's the gatekeeper of what goes in. And we have leveraged his abilities to enhance Linux to be able to use this decentralized data, in particular with decoupling the control plane driven by metadata from the data access path and the many storage systems on which the data gets accessed. So this factoring, this splitting of control plane from data path, metadata from data, was absolutely necessary to create a data mesh like we're talking about. And to be able to build this supercloud concept. And the highways on which the data runs and the client which knows how to talk to it is all open source. And we have, we've driven the NFS 4.2 spec. The newest NFS spec came from my team. And it was specifically the enhancements needed to be able to build a spanning file system, a data mesh at a file system level. Now that said, our file system itself and our server, our file server, our data orchestration, our data management stuff, that's all closed source, proprietary Hammerspace tech. But the highways on which the mesh connects are actually all open source and the client that knows how to consume it. So we would, honestly, I would welcome competitors using those same highways. They would be at a major disadvantage because we kind of built them, but it would still be very validating and I think only increase the potential adoption rate by more than whatever they might take of the market. So it'd actually be good to split the market with somebody else to come in and share those now super highways for how to mesh data at the file system level, you know, in here. So yeah, hopefully that answered your question. Does that answer the question about how we embrace the open source? >> Right, and there was one other, just that my last one is how do you enable something to run in every environment? And if we take the edge, for example, as being, as an environment which is much very, very compute heavy, but having a lot less capability, how do you do a hold? >> Perfect question. Perfect question. What we do today is a software appliance. We are using a Linux RHEL 8, RHEL 8 equivalent or a CentOS 8, or it's, you know, they're all roughly equivalent. But we have bundled and a software appliance which can be instantiated on bare metal hardware on any type of VM system from VMware to all of the different hypervisors in the Linux world, to even Nutanix and such. So it can run in any virtualized environment and it can run on any cloud instance, server instance in the cloud. And we have it packaged and deployable from the marketplaces within the different clouds. So you can literally spin it up at the click of an API in the cloud on instances in the cloud. So with all of these together, you can basically instantiate a Hammerspace set of machinery that can offer up this file system mesh. like we've been using the terminology we've been using now, anywhere. So it's like being able to take and spin up Snowflake and then just be able to install and run some VMs anywhere you want and boom, now you have a Snowflake service. And by the way, it is so complete that some of our customers, I would argue many aren't even using public clouds at all, they're using this just to run their own data centers in a cloud-like fashion, you know, where they have a data service that can span it all. >> Yeah and to Molly's first point, we would consider that, you know, cloud. Let me put you on the spot. If you had to describe conceptually without a chalkboard what an architectural diagram would look like for supercloud, what would you say? >> I would say it's to have the same runtime environment within every data center and defining that runtime environment as what it takes to schedule the execution of applications, so job scheduling, runtime stuff, and here we're talking Kubernetes, Slurm, other things that do job scheduling. We're talking about having a common way to, you know, instantiate compute resources. So a global compute environment, having a common compute environment where you can instantiate things that need computing. Okay? So that's the first part. And then the second is the data platform where you can have file block and object volumes, and have them available with the same APIs in each of these distributed data centers and have the exact same data omnipresent with the ability to control where the data is from one moment to the next, local, where all the data is instantiate. So my definition would be a common runtime environment that's bifurcate-- >> Oh. (attendees chuckling) We just lost them at the money slide. >> That's part of the magic makes people listen. We keep someone on pin and needles waiting. (attendees chuckling) >> That's good. >> Are you back, David? >> I'm on the edge of my seat. Common runtime environment. It was like... >> And just wait, there's more. >> But see, I'm maybe hyper-focused on the lower level of what it takes to host and run applications. And that's the stuff to schedule what resources they need to run and to get them going and to get them connected through to their persistence, you know, and their data. And to have that data available in all forms and have it be the same data everywhere. On top of that, you could then instantiate applications of different types, including relational databases, and data warehouses and such. And then you could say, now I've got, you know, now I've got these more application-level or structured data-level things. I tend to focus less on that structured data level and the application level and am more focused on what it takes to host any of them generically on that super pass layer. And I'll admit, I'm maybe hyper-focused on the pass layer and I think it's valid to include, you know, higher levels up the stack like the structured data level. But as soon as you go all the way up to like, you know, a very specific SAS service, I don't know that you would call that supercloud. >> Well, and that's the question, is there value? And Marianna Tessel from Intuit said, you know, we looked at it, we did it, and it just, it was actually negative value for us because connecting to all these separate clouds was a real pain in the neck. Didn't bring us any additional-- >> Well that's 'cause they don't have this pass layer underneath it so they can't even shop around, which actually makes it hard to stand up your own SAS service. And ultimately they end up having to build their own infrastructure. Like, you know, I think there's been examples like Netflix moving away from the cloud to their own infrastructure. Basically, if you're going to rent it for more than a few months, it makes sense to build it yourself, if it's at any kind of scale. >> Yeah, for certain components of that cloud. But if the Goldman Sachs came to you, David, and said, "Hey, we want to collaborate and we want to build "out a cloud and essentially build our SAS system "and we want to do that with Hammerspace, "and we want to tap the physical infrastructure "of not only our data centers but all the clouds," then that essentially would be a SAS, would it not? And wouldn't that be a Super SAS or a supercloud? >> Well, you know, what they may be using to build their service is a supercloud, but their service at the end of the day is just a SAS service with global reach. Right? >> Yeah. >> You know, look at, oh shoot. What's the name of the company that does? It has a cloud for doing bookkeeping and accounting. I forget their name, net something. NetSuite. >> NetSuite. NetSuite, yeah, Oracle. >> Yeah. >> Yep. >> Oracle acquired them, right? Is NetSuite a supercloud or is it just a SAS service? You know? I think under the covers you might ask are they using supercloud under the covers so that they can run their SAS service anywhere and be able to shop the venue, get elasticity, get all the benefits of cloud in the, to the benefit of their service that they're offering? But you know, folks who consume the service, they don't care because to them they're just connecting to some endpoint somewhere and they don't have to care. So the further up the stack you go, the more location-agnostic it is inherently anyway. >> And I think it's, paths is really the critical layer. We thought about IAS Plus and we thought about SAS Minus, you know, Heroku and hence, that's why we kind of got caught up and included it. But SAS, I admit, is the hardest one to crack. And so maybe we exclude that as a deployment model. >> That's right, and maybe coming down a level to saying but you can have a structured data supercloud, so you could still include, say, Snowflake. Because what Snowflake is doing is more general purpose. So it's about how general purpose it is. Is it hosting lots of other applications or is it the end application? Right? >> Yeah. >> So I would argue general purpose nature forces you to go further towards platform down-stack. And you really need that general purpose or else there is no real distinguishing. So if you want defensible turf to say supercloud is something different, I think it's important to not try to wrap your arms around SAS in the general sense. >> Yeah, and we've kind of not really gone, leaned hard into SAS, we've just included it as a deployment model, which, given the constraints that you just described for structured data would apply if it's general purpose. So David, super helpful. >> Had it sign. Define the SAS as including the hybrid model hold SAS. >> Yep. >> Okay, so with your permission, I'm going to add you to the list of contributors to the definition. I'm going to add-- >> Absolutely. >> I'm going to add this in. I'll share with Molly. >> Absolutely. >> We'll get on the calendar for the date. >> If Molly can share some specific language that we've been putting in that kind of goes to stuff we've been talking about, so. >> Oh, great. >> I think we can, we can share some written kind of concrete recommendations around this stuff, around the general purpose, nature, the common data thing and yeah. >> Okay. >> Really look forward to it and would be glad to be part of this thing. You said it's in February? >> It's in January, I'll let Molly know. >> Oh, January. >> What the date is. >> Excellent. >> Yeah, third week of January. Third week of January on a Tuesday, whatever that is. So yeah, we would welcome you in. But like I said, if it doesn't work for your schedule, we can prerecord something. But it would be awesome to have you in studio. >> I'm sure with this much notice we'll be able to get something. Let's make sure we have the dates communicated to Molly and she'll get my admin to set it up outside so that we have it. >> I'll get those today to you, Molly. Thank you. >> By the way, I am so, so pleased with being able to work with you guys on this. I think the industry needs it very bad. They need something to break them out of the box of their own mental constraints of what the cloud is versus what it's supposed to be. And obviously, the more we get people to question their reality and what is real, what are we really capable of today that then the more business that we're going to get. So we're excited to lend the hand behind this notion of supercloud and a super pass layer in whatever way we can. >> Awesome. >> Can I ask you whether your platforms include ARM as well as X86? >> So we have not done an ARM port yet. It has been entertained and won't be much of a stretch. >> Yeah, it's just a matter of time. >> Actually, entertained doing it on behalf of NVIDIA, but it will absolutely happen because ARM in the data center I think is a foregone conclusion. Well, it's already there in some cases, but not quite at volume. So definitely will be the case. And I'll tell you where this gets really interesting, discussion for another time, is back to my old friend, the SSD, and having SSDs that have enough brains on them to be part of that fabric. Directly. >> Interesting. Interesting. >> Very interesting. >> Directly attached to ethernet and able to create a data mesh global file system, that's going to be really fascinating. Got to run now. >> All right, hey, thanks you guys. Thanks David, thanks Molly. Great to catch up. Bye-bye. >> Bye >> Talk to you soon.
SUMMARY :
So my question to you was, they don't have to do it. to starved before you have I believe that the ISVs, especially those the end users you need to So, if I had to take And and I think Ultimately the supercloud or the Snowflake, you know, more narrowly on just the stuff of the point of what you're talking Well, and you know, Snowflake founders, I don't want to speak over So it starts to even blur who's the main gravity is to having and, you know, that's where to be in a, you know, a lot of thought to this. But some of the inside baseball But the truth is-- So one of the things we wrote the fact that you even have that you would not put in as to give you low latency access the hardest things, David. This is one of the things I've the how can you host applications Not a specific application Yeah, yeah, you just statement when you broke up. So would you exclude is kind of hard to do I know, we all know it is. I think I said to Slootman, of ways you can give it So again, in the spirit But I could use your to allowing you to run anything anywhere So it comes down to how quality that you would expect and how true up you are to that concept. you don't have to draw, yeah. the ability for you and get all the benefits of Snowflake. of being, you know, if it were a service They do the same thing and the MSP or the public clouds, to create my own data. for all of the other apps and that hold the datasets So David, in the third week of January, I'd love to have you come like that to line up with other, you know, Yeah, and Data Mesh, of course, is one Well, you know, and I think.. and the open source? and the client which knows how to talk and then just be able to we would consider that, you know, cloud. and have the exact same data We just lost them at the money slide. That's part of the I'm on the edge of my seat. And that's the stuff to schedule Well, and that's the Like, you know, I think But if the Goldman Sachs Well, you know, what they may be using What's the name of the company that does? NetSuite, yeah, Oracle. So the further up the stack you go, But SAS, I admit, is the to saying but you can have a So if you want defensible that you just described Define the SAS as including permission, I'm going to add you I'm going to add this in. We'll get on the calendar to stuff we've been talking about, so. nature, the common data thing and yeah. to it and would be glad to have you in studio. and she'll get my admin to set it up I'll get those today to you, Molly. And obviously, the more we get people So we have not done an ARM port yet. because ARM in the data center I think is Interesting. that's going to be really fascinating. All right, hey, thanks you guys.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Slootman | PERSON | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
Molly | PERSON | 0.99+ |
Marianna Tessel | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Frank | PERSON | 0.99+ |
Disney | ORGANIZATION | 0.99+ |
Goldman Sachs | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
January | DATE | 0.99+ |
John Furrier | PERSON | 0.99+ |
February | DATE | 0.99+ |
Peter | PERSON | 0.99+ |
Zhamak Dehghaniis | PERSON | 0.99+ |
Hammerspace | ORGANIZATION | 0.99+ |
Word | TITLE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
RHEL 8 | TITLE | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Benoit | PERSON | 0.99+ |
Excel | TITLE | 0.99+ |
second | QUANTITY | 0.99+ |
Autodesk | ORGANIZATION | 0.99+ |
CentOS 8 | TITLE | 0.99+ |
David Flynn | PERSON | 0.99+ |
one | QUANTITY | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
PowerPoint | TITLE | 0.99+ |
first point | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Tuesday | DATE | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
first part | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
each region | QUANTITY | 0.98+ |
Linux | TITLE | 0.98+ |
One | QUANTITY | 0.98+ |
Intuit | ORGANIZATION | 0.98+ |
Tim Burners Lee | PERSON | 0.98+ |
Zhamak Dehghaniis' | PERSON | 0.98+ |
Blue Origin | ORGANIZATION | 0.98+ |
Bay Area | LOCATION | 0.98+ |
two reasons | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
one application | QUANTITY | 0.98+ |
Snowflake | TITLE | 0.98+ |
first | QUANTITY | 0.98+ |
more than a few months | QUANTITY | 0.97+ |
SAS | ORGANIZATION | 0.97+ |
ARM | ORGANIZATION | 0.97+ |
Microsoft | ORGANIZATION | 0.97+ |
Michael Rogers, CrowdStrike | CrowdStrike Fal.Con 2022
foreign okay we're back at Falcon 2022 crowdstrike's big user conference first time in a couple of years obviously because of kova this is thecube's coverage Dave vellante and Dave Nicholson wall-to-wall coverage two days in a row Michael Rogers the series the newly minted vice president of global alliances at crowdstrike Michael first of all congratulations on the new appointment and welcome to the cube thank you very much it's an honor to be here so dial back just a bit like think about your first hundred days in this new role what was it like who'd you talk to what'd you learn wow well the first hundred days were filled with uh excitement uh I would say 18 plus hours a day getting to know the team across the globe a wonderful team across all of the partner types that we cover and um just digging in and spending time with people and understanding uh what the partner needs were and and and and it was just a it was a blur but a blast I agree with any common patterns that you heard that you could sort of coalesce around yeah I mean I think that uh really what a common thing that we hear at crowdstrike whether it's internal is extra external is getting to the market as fast as possible there's so much opportunity and every time we open a door the resource investment we need we continue to invest in resources and that was an area that we identified and quickly pivoted and started making some of those new investments in a structure of the organization how we cover Partners uh how we optimize uh the different routes to Market with our partners and yeah just a just a it's been a wonderful experience and in my 25 years of cyber security uh actually 24 and a half as of Saturday uh I can tell you that I have never felt and had a better experience in terms of culture people and a greater mission for our customers and our partners you'll Max funny a lot of times Dave we talk about this is we you know we learned a lot from Amazon AWS with the cloud you know taking something you did internally pointing it externally to Pizza teams there's shared responsibility model we talk about that and and one of the things is blockers you know Amazon uses that term blocker so were there any blockers that you identified that you're you're sort of working with the partner ecosystem to knock down to accelerate that go to market well I mean if I think about what we had put in place prior and I had the benefit of being vice president of America's prior to the appointment um and had the pleasure of succeeding my dear friend and Mentor Matthew Pauley um a lot of that groundwork was put in place and we work collectively as a leadership team to knock down a lot of those blockers and I think it really as I came into the opportunity and we made new Investments going into the fiscal year it's really getting to Market as fast as possible it's a massive Target addressable market and identifying the right routes and how to how to harness that power of we to drive the most value to the marketplace yeah what is it what does that look like in terms of alliances alliances can take a lot of shape we've we've talked to uh service providers today as an example um our Global Systems integrators in that group also what what is what does the range look like yeah I mean alliances at crowdstrike and it's a great question because a lot of times people think alliances and they only think of Technology alliances and for us it spans really any and all routes to Market it could be your traditional solution providers which might be regionally focused it could be nationally focused larger solution providers or Lars as you noted service providers and telcos global system integrators mssps iot Partners OEM Partners um and store crouchstrike store Partners so you look across that broad spectrum and we cover it all so the mssps we heard a lot about that on the recent earnings call we've heard this is a consistent theme we've interviewed a couple here today what's driving that I mean is it the fact that csos are just you know drowning for talent um and why crowdstrike why is there such an affinity between mssps and crowdstrike yeah a great question we um and you noted that uh succinctly that csos today are faced with the number one challenge is lack of resources and cyber security the last that I heard was you know in the hundreds of thousands like 350 000 and that's an old stat so I would venture to Guess that the open positions in cyber security are north of a half a million uh as we sit here today and um service providers and mssps are focused on providing service to those customers that are understaffed and have that Personnel need and they are harnessing the crowdstrike platform to bring a cloud native best of breed solution to their customers to augment and enhance the services that they bring to those customers so partner survey what tell us about the I love surveys I love data you know this what was the Genesis of the survey who took it give us the breakdown yeah that's a great question no uh nothing is more important than the feedback that we get from our partners so every single year we do a partner survey it reaches all partner types in the uh in the ecosystem and we use the net promoter score model and so we look at ourselves in terms of how we how we uh rate against other SAS solution providers and then we look at how we did last year and in the next year and so I'm happy to say that we increased our net promoter score by 16 percent year over year but my philosophy is there's always room for improvement so the feedback from our partners on the positive side they love the Falcon platform they love the crowdstrike technology they love the people that they work with at crowdstrike and they like our enablement programs the areas that they like us to see more investment in is the partner program uh better and enhanced enablement making it easier to work with crowdstrike and more opportunities to offer services enhance services to their customers dramatic differences between the types of Partners and and if so you know why do you think those were I mean like you mentioned you know iot Partners that's kind of a new area you know so maybe maybe there was less awareness there were there any sort of differences that you noticed by type of partner I would say that you know the areas or the part the partners that identified areas for improvement were the partners that that uh either were new to crowdstrike or they're areas that we're just investing in uh as as we expand as a company and a demand from the market is you know pull this thing into these new routes to Market um not not one in particular I mean iot is something that we're looking to really blow up in the next uh 12 to 18 months um but no no Common Thread uh consistent feedback across the partner base speaking of iot he brought it up before it's is it in a you see it as an adjacency to i-team it seems like it and OT used to never talk to each other and now they're increasingly doing so but they're still it still seems like different worlds what have you found and learned in that iot partner space yeah I mean I think the key and we the way we look at the journey is it starts with um Discovery discovering the assets that are in the OT environment um it then uh transitions to uh detection and response and really prevention and once you can solve that and you build that trust through certifications in the industry um you know it really is a game changer anytime you have Global in your job title first word that comes to mind for me anyway is sovereignty issues is that something that you deal with in this space uh in terms of partners that you're working with uh focusing on Partners in certain regions so that they can comply with any governance or sovereignty yeah that's that's a great question Dave I mean we have a fantastic and deep bench on our compliance team and there are certain uh you know parameters and processes that have been put in place to make sure that we have a solid understanding in all markets in terms of sovereignty and and uh where we're able to play and how that were you North America before or Americas uh Americas America so you're familiar with the sovereignty issue yeah a little already Latin America is certainly uh exposed me plenty of plenty of that yes 100 so you mentioned uh uh Tam before I think it was total available Market you had a different word for the t uh total addressable Mark still addressable Market okay fine so I'm hearing Global that's a tam expansion opportunity iot is definitely you know the OT piece and then just working better um you know better Groove swing with the partners for higher velocity when you think about the total available total addressable market and and accelerating penetration and growing your Tam I've seen the the charts in your investor presentation and you know starts out small and then grows to you know I think it could be 100 billion I do a lot of Tam analysis but just my back a napkin had you guys approaching 100 billion anyway how do you think about the Tam and what role do Partners play in terms of uh increasing your team yeah that's a great question I mean if you think about it today uh George announced on the day after our 11th anniversary as a company uh 20 000 customers and and if you look at that addressable Market just in the SMB space it's north of 50 million companies that are running on Legacy on-prem Solutions and it really provides us an opportunity to provide those customers with uh Next Generation uh threat protection and and detection and and response partners are the route to get there there is no doubt that we cannot cover 50 50 million companies requires a span of of uh of of of a number of service providers and mssps to get to that market and that's where we're making our bets what what's an SMB that is a candidate for crowdstrike like employee size or how do you look at that like what's the sort of minimum range yeah the way we segment out the SMB space it's 250 seats or endpoints and below 250 endpoints yes right and so it's going to be fairly significant so math changes with xdr with the X and xdr being extended the greater number of endpoints means that a customer today when you talk about total addressable Market that market can expand even without expanding the number of net new customers is that a fair yeah Fair assessment yep yeah you got that way in that way but but map that to like company size can you roughly what's the what's the smallest s that would do business with crowdstrike yeah I mean we have uh companies as small as five employees that will leverage crowd strike yeah 100 and they've got hundreds of endpoints oh no I'm sorry five uh five endpoints is oh okay so it's kind of 250 endpoints as well like the app that's the sweets that's it's that's kind of the Top Line we look at and then we focus oh okay when we Define SMB it's below so five to 250 endpoints right yes and so roughly so you're talking to companies with less than 100 employees right yeah yeah so I mean this is what I was talking about before I say I look around the the ecosystem myself it kind of reminds me of service now in 2013 but servicenow never had a SMB play right and and you know very kind of proprietary closed platform not that you don't have a lot of propriety in your platform you do but you they were never going to get down Market there and their Tam is not as big in my view but I mean your team is when you start bringing an iot it's it's mind-boggling it's endless how large it could be yeah all right so what's your vision for the Elevate program partner program well I I look at uh a couple things that we've we've have in place today one is um one is we've we've established for the first time ever at crowdstrike the Alliance program management office apmo and that team is focused on building out our next Generation partner program and that's you know processes it's you know uh it's it's ring fencing but it's most important importantly identifying capabilities for partners to expand to reduce friction and uh grow their business together with crowdstrike we also look at uh what we call program Harmony and that's taking all of the partner types or the majority of the partner types and starting to look at it with the customer in the middle and so multiple partners can play a role on the journey to bringing a customer on board initially to supporting that customer going forward and they can all participate and be rewarded for their contribution to that opportunity so it's really a key area for us going forward Hub and spoke model with the center of the that model is the customer you're saying that's good okay so you're not like necessarily fighting each other for for a sort of ownership of that model but uh cool Michael Rogers thanks so much for coming on thecube it was great to have you my pleasure thank you for having me you're welcome all right keep it right there Dave Nicholson and Dave vellante we'll be right back to Falcon 22 from the Aria in Las Vegas you're watching thecube foreign [Music]
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Michael Rogers | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Dave vellante | PERSON | 0.99+ |
2013 | DATE | 0.99+ |
250 seats | QUANTITY | 0.99+ |
25 years | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
100 billion | QUANTITY | 0.99+ |
16 percent | QUANTITY | 0.99+ |
Michael | PERSON | 0.99+ |
two days | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
next year | DATE | 0.99+ |
less than 100 employees | QUANTITY | 0.99+ |
hundreds of thousands | QUANTITY | 0.99+ |
Dave vellante | PERSON | 0.99+ |
today | DATE | 0.99+ |
100 | QUANTITY | 0.99+ |
250 | QUANTITY | 0.99+ |
first hundred days | QUANTITY | 0.98+ |
Americas | LOCATION | 0.98+ |
five | QUANTITY | 0.98+ |
five employees | QUANTITY | 0.98+ |
North America | LOCATION | 0.97+ |
first time | QUANTITY | 0.97+ |
250 endpoints | QUANTITY | 0.97+ |
18 plus hours a day | QUANTITY | 0.97+ |
first time | QUANTITY | 0.97+ |
24 and a half | QUANTITY | 0.97+ |
Mentor Matthew Pauley | PERSON | 0.97+ |
Saturday | DATE | 0.96+ |
hundreds of endpoints | QUANTITY | 0.96+ |
Las Vegas | LOCATION | 0.96+ |
Latin America | LOCATION | 0.95+ |
first hundred days | QUANTITY | 0.95+ |
50 50 million companies | QUANTITY | 0.95+ |
first | QUANTITY | 0.93+ |
north of a half a million | QUANTITY | 0.93+ |
first word | QUANTITY | 0.92+ |
12 | QUANTITY | 0.92+ |
11th anniversary | QUANTITY | 0.91+ |
18 months | QUANTITY | 0.91+ |
telcos | ORGANIZATION | 0.91+ |
iot | ORGANIZATION | 0.89+ |
five endpoints | QUANTITY | 0.88+ |
Global | ORGANIZATION | 0.88+ |
20 000 customers | QUANTITY | 0.88+ |
one | QUANTITY | 0.85+ |
north of 50 million companies | QUANTITY | 0.85+ |
CrowdStrike | EVENT | 0.85+ |
couple | QUANTITY | 0.85+ |
crowdstrike | ORGANIZATION | 0.84+ |
America | LOCATION | 0.83+ |
Falcon 22 | ORGANIZATION | 0.81+ |
number one | QUANTITY | 0.76+ |
Falcon 2022 crowdstrike | EVENT | 0.75+ |
couple of years | QUANTITY | 0.74+ |
350 000 | QUANTITY | 0.74+ |
SAS | ORGANIZATION | 0.72+ |
vice president | PERSON | 0.7+ |
every single year | QUANTITY | 0.7+ |
uh Next Generation | ORGANIZATION | 0.68+ |
Fal.Con 2022 | EVENT | 0.63+ |
Falcon | ORGANIZATION | 0.62+ |
big user conference | EVENT | 0.61+ |
CrowdStrike | ORGANIZATION | 0.6+ |
Elevate | TITLE | 0.57+ |
below | QUANTITY | 0.56+ |
Lars | PERSON | 0.54+ |
Aria | LOCATION | 0.53+ |
Alliance | ORGANIZATION | 0.53+ |
president | PERSON | 0.51+ |
George | PERSON | 0.51+ |
America | ORGANIZATION | 0.5+ |
Legacy | ORGANIZATION | 0.49+ |
csos | TITLE | 0.47+ |
AWS | ORGANIZATION | 0.45+ |
Generation | OTHER | 0.44+ |
Paula Hansen and Jacqui van der Leij Greyling | Democratizing Analytics Across the Enterprise
(light upbeat music) (mouse clicks) >> Hey, everyone. Welcome back to the program. Lisa Martin here. I've got two guests joining me. Please welcome back to The Cube, Paula Hansen, the chief revenue officer and president at Alteryx. And Jacqui Van der Leij - Greyling joins us as well, the global head of tax technology at eBay. They're going to share with you how Alteryx is helping eBay innovate with analytics. Ladies, welcome. It's great to have you both on the program. >> Thank you, Lisa. >> Thank you, Lisa. >> It's great to be here. >> Yeah, Paula. We're going to start with you. In this program, we've heard from Jason Klein, we've heard from Alan Jacobson, they talked about the need to democratize analytics across any organization to really drive innovation. With analytics as they talked about at the forefront of software investments, how's Alteryx helping its customers to develop roadmaps for success with analytics? >> Well, thank you, Lisa. It absolutely is about our customer's success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts, of course, with our innovative technology and platform but ultimately, we help our customers to create a culture of data literacy and analytics from the top of the organization, starting with the C-suite. And we partner with our customers to build their roadmaps for scaling that culture of analytics through things like enablement programs, skills assessments, hackathons, setting up centers of excellence to help their organizations scale and drive governance of this analytics capability across the enterprise. So at the end of the day, it's really about helping our customers to move up their analytics maturity curve with proven technologies and best practices so they can make better business decisions and compete in their respective industries. >> Excellent. Sounds like a very strategic program. We're going to unpack that. Jacqui let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the IDC report that showed that 93% of organizations are not utilizing the analytics skills of their employees, but then there's eBay. How, Jacqui, did eBay become one of the 7% of organizations who's really maturing and how are you using analytics across the organization at eBay? >> So I think the main thing for us is just when we started out was, is that, you know, our, especially in finance they became spreadsheet professionals, instead of the things that we really want our employees to add value to. And we realized we had to address that. And we also knew we couldn't wait for all our data to be centralized until we actually start using the data or start automating and be more effective. So ultimately, we really started very, very actively embedding analytics in our people and our data and our processes. >> Starting with people is really critical. Jacqui, continuing with you, what were some of the roadblocks to analytics adoption that you faced and how did you overcome them? >> So I think, you know, eBay is a very data driven company. We have a lot of data. I think we are 27 years around this year so we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and just finding those data sources and finding ways to connect to them to move forward. The other thing is, is that you know, people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals, right? And there was no, we're not independent. You couldn't move forward. You would've been dependent on somebody else's roadmap to get to data and to get the information you wanted. So really finding something that everybody could access analytics or access data. And finally, we have to realize is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy and that is not so daunting on somebody who's brand new to the field? And I would call those out as your major roadblocks because you always have, not always, but most of the times you have support from the top in our case, we have, but in the end of the day, it's our people that need to actually really embrace it and making that accessible for them, I would say is definitely not per se, a roadblock but basically some, a block you want to be able to move. >> It's really all about putting people first. Question for both of you, and Paula will start with you, and then Jacqui will go to you. I think the message in this program that the audience is watching with us is very clear. Analytics is for everyone, should be for everyone. Let's talk now about how both of your organizations are empowering people those in the organization that may not have technical expertise to be able to leverage data so that they can actually be data driven? Paula? >> Yes. Well, we leverage our platform across all of our business functions here at Alteryx. And just like Jacqui explained at eBay finance is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jacqui mentioned, we have this huge amount of data flowing through our enterprise and the opportunity to leverage that into insights and analytics is really endless. So our CFO, Kevin Rubin has been a key sponsor for using our own technology. We use Alteryx for forecasting, all of our key performance metrics for business planning across our audit function to help with compliance and regulatory requirements, tax and even to close our books at the end of each quarter so it's really remained across our business. And at the end of the day, it comes to how do you train users? How do you engage users to lean into this analytic opportunity to discover use cases. And so one of the other things that we've seen many companies do is to gamify that process to build a game that brings users into the experience for training and to work with each other, to problem solve, and along the way, maybe earn badges depending on the capabilities and trainings that they take. And just have a little healthy competition as an employee base around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jacqui mentioned, it's really about ensuring that people feel comfortable, that they feel supported that they have access to the training that they need. And ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >> That confidence is key. Jacqui, talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >> Yeah, I think it means to what Paula has said in terms of you know, getting people excited about it but it's also understanding that this is a journey. And everybody is the different place in their journey. You have folks that's already really advanced who has done this every day, and then you have really some folks that this is brand new and, or maybe somewhere in between. And it's about how you could get everybody in their different phases to get to the initial destination. I say initially, because I believe the journey is never really complete. What we have done is that we decided to invest in a... We build a proof of concepts and we got our CFO to sponsor a hackathon. We opened it up to everybody in finance in the middle of the pandemic. So everybody was on Zoom. And we told people, "Listen, we're going to teach you this tool, super easy. And let's just see what you can do." We ended up having 70 entries. We had only three weeks. So, and these are people that has... They do not have a background. They are not engineers, they're not data scientists. And we ended up with a 25,000 hour savings at the end of that hackathon. From the 70 entries with people that have never, ever done anything like this before and there you had the result. And then it just went from there. It was people had a proof of concept, they knew that it worked, and they overcame that initial barrier of change. And that's where we are seeing things really, really picking up now. >> That's fantastic. And the business outcome that you mentioned there, the business impact is massive helping folks get that confidence to be able to overcome sometimes the cultural barriers is key here. I think another thing that this program has really highlighted is there is a clear demand for data literacy in the job market, regardless of organization. Can each of you share more about how you're empowering the next generation of data workers? Paula will start with you. >> Absolutely. And Jacqui says it so well, which is that it really is a journey that organizations are on. And we, as people in society are on in terms of upskilling our capabilities. So one of the things that we're doing here at Alteryx to help address this skillset gap on a global level is through a program that we call SparkED, which is essentially a no-cost analytics education program that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay, and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this program's really developed to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with SparkED, we started last May, but we currently have over 850 educational institutions globally engaged across 47 countries. And we're going to continue to invest here because there's so much opportunity for people, for society and for enterprises, when we close gap and empower more people with the necessary analytics skills to solve all the problems that data can help solve. >> So SparkED just made a really big impact in such a short time period. It's going to be fun to watch the progress of that. Jacqui let's go over to you now. Talk about some of the things that eBay is doing to empower the next generation of data workers. >> So we basically wanted to make sure that we kicked that momentum from the hackathon. Like we don't lose that excitement, right? So we just launched a program called eBay Masterminds. And what it basically is, it's an inclusive innovation initiative, where we firmly believe that innovation is for upscaling for all analytics role. So it doesn't matter your background, doesn't matter which function you are in, come and participate in this, where we really focus on innovation, introducing new technologies and upscaling our people. We are... Apart from that, we also said... Well, we should just keep it to inside eBay. We have to share this innovation with the community. So we are actually working on developing an analytics high school program, which we hope to pilot by the end of this year, where we will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, but also how to use alter Alteryx. And we're working with actually, we're working with SparkED and they're helping us develop that program. And we really hope that, let us say, by the end of the year have a pilot and then also next, was hoping to roll it out in multiple locations, in multiple countries, and really, really focus on that whole concept of analytics role. >> Analytics role, sounds like Alteryx and eBay have a great synergistic relationship there, that is jointly aimed at, especially, kind of, going down the stuff and getting people when they're younger interested and understanding how they can be empowered with data across any industry. Paula let's go back to you. You were recently on The Cube's Supercloud event just a couple of weeks ago. And you talked about the challenges the companies are facing as they're navigating what is by default a multi-cloud world? How does the Alteryx Analytics Cloud platform enable CIOs to democratize analytics across their organization? >> Yes, business leaders and CIOs across all industries are realizing that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organizations. Last, I check there was 2 million data scientists in the world. So that's woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. (Paula clears throat) So what we're seeing now with CIOs, with business leaders is that they're integrating data analysis and the skillset of data analysis into virtually every job function. And that is what we think of when we think of analytics for all. And so our mission with Alteryx Analytics Cloud, is to empower all of those people in every job function regardless of their skillset. As Jacqui pointed out from people that would, you know are just getting started all the way to the most sophisticated of technical users. Every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organizations. So that's our goal with Alteryx Analytics Cloud and it operates in a multi-cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyze and report out so that we can break down data silos across the enterprise and help drive real business outcomes as a result of unlocking the potential of data. >> As well as really lessening that skills gap as you were saying, there's only 2 million data scientists. You don't need to be a data scientist. That's the beauty of what Alteryx is enabling and eBay is a great example of that. Jacqui let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where Alteryx fits in as that analytics maturity journey continues. And what are some of the things that you are most excited about as analytics truly gets democratized across eBay? >> When we started about getting excited about things when it comes to analytics, I can go on all day but I'll keep it short and sweet for you. I do think we are on the topic full of data scientists. And I really feel that that is your next step, for us anyways, it's just that, how do we get folks to not see data scientists as this big thing, like a rocket scientist, it's something completely different. And it's something that is in everybody in a certain extent. So again, partnering with Alteryx would just release the AI/ML solution, allowing, you know, folks to not have a data scientist program but actually build models and be able to solve problems that way. So we have engaged with Alteryx and we purchased the licenses quite a few. And right now, through our mastermind program we're actually running a four-months program for all skill levels. Teaching them AI/ML and machine learning and how they can build their own models. We are really excited about that. We have over 50 participants without the background from all over the organization. We have members from our customer services, we have even some of our engineers, are actually participating in the program. We just kicked it off. And I really believe that that is our next step. I want to give you a quick example of the beauty of this is where we actually just allow people to go out and think about ideas and come up with things. And one of the people in our team who doesn't have a data scientist background at all was able to develop a solution where, you know, there is a checkout feedback, checkout functionality on the eBay site, where sellers or buyers can verbatim add information. And she build a model to be able to determine what relates to tax specific, what is the type of problem, and even predict how that problem can be solved before we, as a human even step in. And now instead of us or somebody going to the bay to try to figure out what's going on there, we can focus on fixing the error versus actually just reading through things and not adding any value. And it's a beautiful tool, and I'm very impressed when you saw the demo and they've been developing that further. >> That sounds fantastic. And I think just the one word that keeps coming to mind and we've said this a number of times in the program today is, empowerment. What you're actually really doing to truly empower people across the organization with varying degrees of skill level going down to the high school level, really exciting. We'll have to stay tuned to see what some of the great things are that come from this continued partnership. Ladies, I want to thank you so much for joining me on the program today and talking about how Alteryx and eBay are really partnering together to democratize analytics and to facilitate its maturity. It's been great talking to you >> Thank you, Lisa. >> Thank you so much. (light upbeat music) >> As you heard over the course of our program, organizations where more people are using analytics who have deeper capabilities in each of the four E's that's, everyone, everything, everywhere and easy analytics. Those organizations achieve more ROI from their respective investments in analytics and automation than those who don't. We also heard a great story from eBay, great example of an enterprise that is truly democratizing analytics across its organization. It's enabling an empowering line of business users to use analytics. Not only focused on key aspects of their job, but develop new skills rather than doing the same repetitive tasks. We want to thank you so much for watching the program today. Remember you can find all of the content on thecube.net. You can find all of the news from today on siliconangle.com, and of course alteryx.com. We also want to thank Alteryx for making this program possible and for sponsoring The Cube. For all of my guests, I'm Lisa Martin. We want to thank you for watching and bye for now. (light upbeat music)
SUMMARY :
the global head of tax technology at eBay. going to start with you. So at the end of the day, one of the things that we talked about instead of the things that that you faced and how but most of the times you that the audience is watching and the confidence to be able to be a part Jacqui, talk about some of the ways And everybody is the different get that confidence to be able to overcome that it's difficult to find Jacqui let's go over to you now. that momentum from the hackathon. And you talked about the in the opportunity to unlock and eBay is a great example of that. example of the beauty of this is It's been great talking to you Thank you so much. in each of the four E's
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jacqui | PERSON | 0.99+ |
Paula | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Jason Klein | PERSON | 0.99+ |
eBay | ORGANIZATION | 0.99+ |
Kevin Rubin | PERSON | 0.99+ |
Alteryx | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Paula Hansen | PERSON | 0.99+ |
Alan Jacobson | PERSON | 0.99+ |
70 entries | QUANTITY | 0.99+ |
Jacqui Van der Leij - Greyling | PERSON | 0.99+ |
25,000 hour | QUANTITY | 0.99+ |
four-months | QUANTITY | 0.99+ |
last May | DATE | 0.99+ |
today | DATE | 0.99+ |
27 years | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
93% | QUANTITY | 0.99+ |
two guests | QUANTITY | 0.99+ |
IDC | ORGANIZATION | 0.99+ |
over 50 participants | QUANTITY | 0.99+ |
Jacqui van der Leij Greyling | PERSON | 0.99+ |
one | QUANTITY | 0.98+ |
siliconangle.com | OTHER | 0.98+ |
each | QUANTITY | 0.98+ |
2 million data scientists | QUANTITY | 0.98+ |
over 850 educational institutions | QUANTITY | 0.97+ |
47 countries | QUANTITY | 0.97+ |
7% | QUANTITY | 0.97+ |
thecube.net | OTHER | 0.97+ |
Supercloud | EVENT | 0.95+ |
three weeks | QUANTITY | 0.95+ |
end of this year | DATE | 0.93+ |
Paula Hansen Jacqui van der Leij Greyling Alteryx
>>Hey everyone. Welcome back to the program. Lisa Martin here, I've got two guests joining me, please. Welcome back to the cube. Paula Hansen, the chief revenue officer and president at Al alters and Jackie Vander lake grayling joins us as well. The global head of tax technology at eBay. They're gonna share with you how an alter Ricks is helping eBay innovate with analytics. Ladies. Welcome. It's great to have you both on the program. >>Thank you, Lisa. It's great to be here. >>Yeah, Paula, we're gonna start with you in this program. We've heard from Jason Klein, we've heard from Alan Jacobson, they talked about the need to democratize analytics across any organization to really drive innovation with analytics. As they talked about at the forefront of software investments, how's alters helping its customers to develop roadmaps for success with analytics. >>Well, thank you, Lisa. It absolutely is about our customer's success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts of course, with our innovative technology and platform, but ultimately we help our customers to create a culture of data literacy and analytics from the top of the organization, starting with the C-suite. And we partner with our customers to build their roadmaps for scaling that culture of analytics through things like enablement programs, skills, assessments, hackathons, setting up centers of excellence to help their organizations scale and drive governance of this analytics capability across the enterprise. So at the end of the day, it's really about helping our customers to move up their analytics, maturity curve with proven technologies and best practices so they can make better business decisions and compete in their respective industries. >>Excellent. Sounds like a very strategic program. We're gonna unpack that Jackie, let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the IDC report that showed that 93% of organizations are not utilizing the analytics skills of their employees, but then there's eBay. How Jackie did eBay become one of the 7% of organizations who's really maturing and how are you using analytics across the organization at eBay? >>So I think the main thing for us is just when we started out was is that, you know, our, especially in finance, they became spreadsheet professionals instead of the things that we really want our employees to add value to. And we realized we had to address that. And we also knew we couldn't wait for all our data to be centralized until we actually start using the data or start automating and be more effective. So ultimately we really started very, very actively embedding analytics in our people and our data and our processes, >>Starting with people is really critical. Jackie, continuing with you, what were some of the roadblocks to analytics adoption that you faced and how did you overcome them? >>So I think, you know, eBay is a very data driven company. We have a lot of data. I think we are 27 years around this year, so we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and, and just finding those data sources and finding ways to connect to them to move forward. The other thing is, is that, you know, people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals, right? And we, there was no, we're not independent. You couldn't move forward. You would've opinion on somebody else's roadmap to get to data and to get the information you wanted. So really finding something that everybody could access analytics or access data. >>And finally we have to realize is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy? And that is not so daunting on somebody who's brand new to the field. And I would, I would call those out as your, as your major roadblocks, because you always have not always, but most of the times you have support from the top in our case, we have, but in the end of the day, it's, it's our people that need to actually really embrace it and, and making that accessible for them, I would say is definitely not per se, a roadblock, but basically some, a block you wanna be able to move. >>It's really all about putting people. First question for both of you and Paula will start with you. And then Jackie will go to you. I think the message in this program that the audience is watching with us is very clear. Analytics is for everyone should be for everyone. Let's talk now about how both of your organizations are empowering people, those in the organization that may not have technical expertise to be able to leverage data so that they can actually be data driven Paula. >>Yes. Well, we leverage our platform across all of our business functions here at Altrix and just like Jackie explained it, eBay finances is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jackie mentioned, we have this huge amount of data flowing through our enterprise and the opportunity to leverage that into insights and analytics is really endless. So our CFO, Kevin Rubin has been a, a key sponsor for using our own technology. We use Altrix for forecasting, all of our key performance metrics for business planning across our audit function, to help with compliance and regulatory requirements tax, and even to close our books at the end of each quarter. So it's really remain across our business. And at the end of the day, it comes to how do you train users? How do you engage users to lean into this analytic opportunity to discover use cases? >>And so one of the other things that we've seen many companies do is to gamify that process, to build a game that brings users into the experience for training and to work with each other, to problem solve and along the way, maybe earn badges depending on the capabilities and trainings that they take. And just have a little healthy competition as an employee base around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jackie mentioned, it's really about ensuring that people feel comfortable, that they feel supported, that they have access to the training that they need. And ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >>That confidence is key. Jackie, talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >>Yeah, I think it means to what Paula has said in terms of, you know, you know, getting people excited about it, but it's also understanding that this is a journey and everybody's the different place in their journey. You have folks that's already really advanced who has done this every day. And then you have really some folks that this is brand new and, or maybe somewhere in between. And it's about how you put, get everybody in their different phases to get to the, the initial destination. I say initially, because I believe the journey is never really complete. What we have done is, is that we decided to invest in an Ebola group of concept. And we got our CFO to sponsor a hackathon. We opened it up to everybody in finance, in the middle of the pandemic. So everybody was on zoom and we had, and we told people, listen, we're gonna teach you this tool super easy. >>And let's just see what you can do. We ended up having 70 entries. We had only three weeks. So, and these are people that has N that do not have a background. They are not engineers, they're not data scientists. And we ended up with a 25,000 hour savings at the end of that hackathon from the 70 inches with people that have never, ever done anything like this before and there you had the result. And then it just went from there. It was, people had a proof of concept. They, they knew that it worked and they overcame the initial barrier of change. And that's where we are seeing things really, really picking up. Now >>That's fantastic. And the, the business outcome that you mentioned there, the business impact is massive helping folks get that confidence to be able to overcome. Sometimes the, the cultural barriers is key. I think another thing that this program has really highlighted is there is a clear demand for data literacy in the job market, regardless of organization. Can each of you share more about how you are empowering the next generation of data workers, Paula will start with you? >>Absolutely. And, and Jackie says it so well, which is that it really is a journey that organizations are on. And, and we, as people in society are on in terms of upskilling our capabilities. So one of the things that we're doing here at Altrix to help address this skillset gap on a global level is through a program that we call sparked, which is essentially a, no-cost a no cost analytics education program that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this program's really developed just to, to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with sparked. We started last may, but we currently have over 850 educational institutions globally engaged across 47 countries. And we're gonna continue to invest here because there's so much opportunity for people, for society and for enterprises, when we close gap and empower more people within necessary analytics skills to solve all the problems that data can help solve. >>So spark has made a really big impact in such a short time period. And it's gonna be fun to watch the progress of that. Jackie, let's go over to you now talk about some of the things that eBay is doing to empower the next generation of data workers. >>So we basically wanted to make sure that we keep that momentum from the hackathon that we don't lose that excitement, right? So we just launched a program called Ebo masterminds. And what it basically is, it's an inclusive innovation initiative where we firmly believe that innovation is all up scaling for all analytics for. So it doesn't matter. Your background doesn't matter which function you are in, come and participate in, in this where we really focus on innovation, introducing new technologies and upskilling our people. We are apart from that, we also say, well, we should just keep it to inside eBay. We, we have to share this innovation with the community. So we are actually working on developing an analytics high school program, which we hope to pilot by the end of this year, where we will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, but also how to use alter alter. And we're working with actually, we're working with spark and they're helping us develop that program. And we really hope that as a say, by the end of the year, have a pilot and then also make you, so we roll it out in multiple locations in multiple countries and really, really focus on, on that whole concept of analytics, role >>Analytics for all sounds like ultra and eBay have a great synergistic relationship there that is jointly aimed at, especially kind of going down the staff and getting people when they're younger, interested, and understanding how they can be empowered with data across any industry. Paula, let's go back to you. You were recently on the Cube's super cloud event just a couple of weeks ago. And you talked about the challenges the companies are facing as they're navigating. What is by default a multi-cloud world? How does the alters analytics cloud platform enable CIOs to democratize analytics across their organization? >>Yes, business leaders and CIOs across all industries are realizing that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organizations. Last I check there was 2 million data scientists in the world. So that's woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. So what we're seeing now with CIOs with business leaders is that they're integrating data analysis and the skill of data analysis into virtually every job function. And that is what we think of when we think of analytics for all. And so our mission with Altrics analytics cloud is to empower all of those people in every job function, regardless of their skillset. As Jackie pointed out from people that would, you know, are just getting started all the way to the most sophisticated of technical users. Every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organizations. So that's our goal with Altrics analytics cloud, and it operates in a multi cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyze, and report out so that we can break down data silos across the enterprise and drive real business outcomes. As a result of unlocking the potential of data, >>As well as really re lessening that skill gap. As you were saying, there's only 2 million data scientists. You don't need to be a data scientist. That's the, the beauty of what Altrics is enabling. And, and eBay is a great example of that. Jackie, let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where alters fits in on as that analytics maturity journey continues and what are some of the things that you are most excited about as analytics truly gets democratized across eBay? >>When we start about getting excited about things, when it comes to analytics, I can go on all day, but I I'll keep it short and sweet for you. I do think we are on the topic full of, of, of data scientists. And I really feel that that is your next step for us anyways, is that, how do we get folks to not see data scientists as this big thing, like a rocket scientist, it's, it's something completely different. And it's something that, that is in everybody to a certain extent. So again, partner with three X would just released the AI ML solution, allowing, you know, folks to not have a data scientist program, but actually build models and be able to solve problems that way. So we have engaged with alters and we, we purchased a license, this quite a few. And right now through our mastermind program, we're actually running a four months program for all skill levels, teaching, teaching them AI ML and machine learning and how they can build their own models. >>We are really excited about that. We have over 50 participants without the background from all over the organization. We have members from our customer services. We have even some of our engineers are actually participating in the program. We just kicked it off. And I really believe that that is our next step. I wanna give you a quick example of, of the beauty of this is where we actually just allow people to go out and think about ideas and come up with things. And one of the people in our team who doesn't have a data scientist background at all, was able to develop a solution where, you know, there is a checkout feedback checkout functionality on the eBay site where sellers or buyers can verbatim add information. And she build a model to be able to determine what relates to tax specific, what is the type of problem, and even predict how that problem can be solved before we, as a human even step in, and now instead of us or somebody going to verbatim and try to figure out what's going on there, we can focus on fixing the error versus actually just reading through things and not adding any value. >>And it's a beautiful tool and very impressed. You saw the demo and they developing that further. >>That sounds fantastic. And I think just the one word that keeps coming to mind, and we've said this a number of times in the program today is empowerment. What you're actually really doing to truly empower people across the organization with, with varying degrees of skill level, going down to the high school level, really exciting, we'll have to stay tuned to see what some of the great things are that come from this continued partnership. Ladies, I wanna thank you so much for joining me on the program today and talking about how alters and eBay are really partnering together to democratize analytics and to facilitate its maturity. It's been great talking to you. >>Thank you. >>As you heard over the course of our program organizations, where more people are using analytics who have the deeper capabilities in each of the four E's, that's, everyone, everything everywhere and easy analytics, those organizations achieve more ROI from their respective investments in analytics and automation than those who don't. We also heard a great story from eBay, great example of an enterprise that is truly democratizing analytics across its organization. It's enabling an empowering line of business users to use analytics, not only focused on key aspects of their job, but develop new skills rather than doing the same repetitive tasks. We wanna thank you so much for watching the program today. Remember you can find all of the content on the cue.net. You can find all of the news from today on Silicon angle.com and of course, alter.com. We also wanna thank alt alters for making this program possible and for sponsored in the queue for all of my guests. I'm Lisa Martin. We wanna thank you for watching and bye for now.
SUMMARY :
It's great to have you both on the program. Yeah, Paula, we're gonna start with you in this program. end of the day, it's really about helping our customers to move up their analytics, Speaking of analytics maturity, one of the things that we talked about in this event is the IDC instead of the things that we really want our employees to add value to. adoption that you faced and how did you overcome them? data and to get the information you wanted. And finally we have to realize is that this is uncharted territory. those in the organization that may not have technical expertise to be able to leverage data it comes to how do you train users? that people feel comfortable, that they feel supported, that they have access to the training that they need. expertise to really be data driven. And then you have really some folks that this is brand new and, And we ended up with a 25,000 folks get that confidence to be able to overcome. and colleges globally to help build the next generation of data workers. Jackie, let's go over to you now talk about some of the things that eBay is doing to empower And we really hope that as a say, by the end of the year, And you talked about the challenges the companies are facing as in terms of the opportunity for people to be a part of the analytics solution. It obviously has the right culture to adapt to that. And it's something that, that is in everybody to a certain extent. And she build a model to be able to determine what relates to tax specific, You saw the demo and they developing that skill level, going down to the high school level, really exciting, we'll have to stay tuned to see what some of We wanna thank you so much for watching the program today.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Paula | PERSON | 0.99+ |
Jason Klein | PERSON | 0.99+ |
Kevin Rubin | PERSON | 0.99+ |
eBay | ORGANIZATION | 0.99+ |
Paula Hansen | PERSON | 0.99+ |
Jackie | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Altrix | ORGANIZATION | 0.99+ |
Alan Jacobson | PERSON | 0.99+ |
25,000 hour | QUANTITY | 0.99+ |
70 entries | QUANTITY | 0.99+ |
Al alters | ORGANIZATION | 0.99+ |
27 years | QUANTITY | 0.99+ |
93% | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
70 inches | QUANTITY | 0.99+ |
First question | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
four months | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
over 50 participants | QUANTITY | 0.99+ |
two guests | QUANTITY | 0.99+ |
IDC | ORGANIZATION | 0.98+ |
2 million data scientists | QUANTITY | 0.97+ |
47 countries | QUANTITY | 0.97+ |
7% | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
over 850 educational institutions | QUANTITY | 0.97+ |
each | QUANTITY | 0.96+ |
three weeks | QUANTITY | 0.95+ |
each quarter | QUANTITY | 0.95+ |
Altrics analytics | ORGANIZATION | 0.95+ |
three X | ORGANIZATION | 0.94+ |
Jackie Vander | PERSON | 0.93+ |
couple of weeks ago | DATE | 0.93+ |
alter Ricks | ORGANIZATION | 0.93+ |
Altrics analytics | ORGANIZATION | 0.93+ |
pandemic | EVENT | 0.92+ |
last may | DATE | 0.91+ |
Altrics | ORGANIZATION | 0.89+ |
one word | QUANTITY | 0.89+ |
Jacqui van der Leij Greyling Alteryx | PERSON | 0.86+ |
end of this year | DATE | 0.81+ |
ultra | ORGANIZATION | 0.76+ |
four E's | QUANTITY | 0.73+ |
cue.net | ORGANIZATION | 0.73+ |
end | DATE | 0.71+ |
Silicon angle.com | OTHER | 0.64+ |
Alteryx Democratizing Analytics Across the Enterprise Full Episode V1b
>> It's no surprise that 73% of organizations indicate analytics spend will outpace other software investments in the next 12 to 18 months. After all as we know, data is changing the world and the world is changing with it. But is everyone's spending resulting in the same ROI? This is Lisa Martin. Welcome to "theCUBE"'s presentation of democratizing analytics across the enterprise, made possible by Alteryx. An Alteryx commissioned IDC info brief entitled, "Four Ways to Unlock Transformative Business Outcomes from Analytics Investments" found that 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. On this special "CUBE" presentation, Jason Klein, product marketing director of Alteryx, will join me to share key findings from the new Alteryx commissioned IDC brief and uncover how enterprises can derive more value from their data. In our second segment, we'll hear from Alan Jacobson, chief data and analytics officer at Alteryx. He's going to discuss how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. And then in our final segment, Paula Hansen, who is the president and chief revenue officer of Alteryx, and Jacqui Van der Leij Greyling, who is the global head of tax technology at eBay, they'll join me. They're going to share how Alteryx is helping the global eCommerce company innovate with analytics. Let's get the show started. (upbeat music) Jason Klein joins me next, product marketing director at Alteryx. Jason, welcome to the program. >> Hello, nice to be here. >> Excited to talk with you. What can you tell me about the new Alteryx IDC research, which spoke with about 1500 leaders, what nuggets were in there? >> Well, as the business landscape changes over the next 12 to 18 months, we're going to see that analytics is going to be a key component to navigating this change. 73% of the orgs indicated that analytics spend will outpace other software investments. But just putting more money towards technology, it isn't going to solve everything. And this is why everyone's spending is resulting in different ROIs. And one of the reasons for this gap is because 93% of organizations, they're still not fully using the analytics skills of their employees, and this widening analytics gap, it's threatening operational progress by wasting workers' time, harming business productivity and introducing costly errors. So in this research, we developed a framework of enterprise analytics proficiency that helps organizations reap greater benefits from their investments. And we based this framework on the behaviors of organizations that saw big improvements across financial, customer, and employee metrics, and we're able to focus on the behaviors driving higher ROI. >> So the info brief also revealed that nearly all organizations are planning to increase their analytics spend. And it looks like from the info brief that nearly three quarters plan on spending more on analytics than any other software. And can you unpack, what's driving this demand, this need for analytics across organizations? >> Sure, well first there's more data than ever before, the data's changing the world, and the world is changing data. Enterprises across the world, they're accelerating digital transformation to capitalize on new opportunities, to grow revenue, to increase margins and to improve customer experiences. And analytics along with automation and AI is what's making digital transformation possible. They're providing the fuel to new digitally enabled lines of business. >> One of the things that the study also showed was that not all analytics spending is resulting in the same ROI. What are some of the discrepancies that the info brief uncovered with respect to the changes in ROI that organizations are achieving? >> Our research with IDC revealed significant roadblocks across people, processes, and technologies. They're preventing companies from reaping greater benefits from their investments. So for example, on the people side, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy as compared to the technology itself. And next, while data is everywhere, most organizations, 63% from our survey, are still not using the full breadth of data types available. Yet data's never been this prolific, it's going to continue to grow, and orgs should be using it to their advantage. And lastly organizations, they need to provide the right analytics tools to help everyone unlock the power of data. >> So they- >> They instead rely on outdated spreadsheet technology. In our survey, nine out of 10 respondents said less than half of their knowledge workers are active users of analytics software beyond spreadsheets. But true analytic transformation can't happen for an organization in a few select pockets or silos. We believe everyone regardless of skill level should be able to participate in the data and analytics process and be driving value. >> Should we retake that, since I started talking over Jason accidentally? >> Yep, absolutely we can do so. We'll just go, yep, we'll go back to Lisa's question. Let's just, let's do the, retake the question and the answer, that'll be able to. >> It'll be not all analytics spending results in the same ROI, what are some of the discrepancies? >> Yes, Lisa, so we'll go from your ISO, just so we get that clean question and answer. >> Okay. >> Thank you for that. On your ISO, we're still speeding, Lisa, so give it a beat in your head and then on you. >> Yet not all analytics spending is resulting in the same ROI. So what are some of the discrepancies that the info brief uncovered with respect to ROI? >> Well, our research with IDC revealed significant roadblocks across people, processes, and technologies, all preventing companies from reaping greater benefits from their investments. So on the people side, for example, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy as compared to the technology itself. And next, while data is everywhere, most organizations, 63% in our survey, are still not using the full breadth of data types available. Data has never been this prolific. It's going to continue to grow and orgs should be using it to their advantage. And lastly, organizations, they need to provide the right analytic tools to help everyone unlock the power of data, yet instead they're relying on outdated spreadsheet technology. Nine of 10 survey respondents said that less than half of their knowledge workers are active users of analytics software. True analytics transformation can't happen for an organization in a few select pockets or silos. We believe everyone regardless of skill level should be able to participate in the data and analytics process and drive value. >> So if I look at this holistically, then what would you say organizations need to do to make sure that they're really deriving value from their investments in analytics? >> Yeah, sure. So overall, the enterprises that derive more value from their data and analytics and achieve more ROI, they invested more aggressively in the four dimensions of enterprise analytics proficiency. So they've invested in the comprehensiveness of analytics across all data sources and data types, meaning they're applying analytics to everything. They've invested in the flexibility of analytics across deployment scenarios and departments, meaning they're putting analytics everywhere. They've invested in the ubiquity of analytics and insights for every skill level, meaning they're making analytics for everyone. And they've invested in the usability of analytics software, meaning they're prioritizing easy technology to accelerate analytics democratization. >> So very strategic investments. Did the survey uncover any specific areas where most companies are falling short, like any black holes that organizations need to be aware of at the outset? >> It did, it did. So organizations, they need to build a data-centric culture. And this begins with people. But what the survey told us is that the people aspect of analytics is the most heavily skewed towards low proficiency. In order to maximize ROI, organizations need to make sure everyone in the organization has access to the data and analytics technology they need. And then the organizations also have to align their investments with upskilling in data literacy to enjoy that higher ROI. Companies who did so experience higher ROI than companies who underinvested in analytics literacy. So among the high ROI achievers, 78% have a good or great alignment between analytics investment and workforce upskilling compared to only 64% among those without positive ROI. And as more orgs adopt cloud data warehouses or cloud data lakes, in order to manage the massively increasing workloads- Can I start that one over. >> Sure. >> Can I redo this one? >> Yeah. >> Of course, stand by. >> Tongue tied. >> Yep, no worries. >> One second. >> If we could do the same, Lisa, just have a clean break, we'll go your question. >> Yep, yeah. >> On you Lisa. Just give that a count and whenever you're ready. Here, I'm going to give us a little break. On you Lisa. >> So are there any specific areas that the survey uncovered where most companies are falling short? Like any black holes organizations need to be aware of from the outset? >> It did. You need to build a data-centric culture and this begins with people, but we found that the people aspect of analytics is most heavily skewed towards low proficiency. In order to maximize ROI organizations need to make sure everyone has access to the data and analytics technology they need. Organizations that align their analytics investments with upskilling enjoy higher ROI than orgs that are less aligned. For example, among the high ROI achievers in our survey, 78% had good or great alignment between analytics investments and workforce upskilling, compared to only 64% among those without positive ROI. And as more enterprises adopt cloud data warehouses or cloud data lakes to manage increasingly massive data sets, analytics needs to exist everywhere, especially for those cloud environments. And what we found is organizations that use more data types and more data sources generate higher ROI from their analytics investments. Among those with improved customer metrics, 90% were good or great at utilizing all data sources, compared to only 67% among the ROI laggards. >> So interesting that you mentioned people, I'm glad that you mentioned people. Data scientists, everybody talks about data scientists. They're in high demand, we know that, but there aren't enough to meet the needs of all enterprises. So given that discrepancy, how can organizations fill the gap and really maximize the investments that they're making in analytics? >> Right, so analytics democratization, it's no longer optional, but it doesn't have to be complex. So we at Alteryx, we're democratizing analytics by empowering every organization to upskill every worker into a data worker. And the data from this survey shows this is the optimal approach. Organizations with a higher percentage of knowledge workers who are actively using analytics software enjoy higher returns from their analytics investment than orgs still stuck on spreadsheets. Among those with improved financial metrics, AKA the high ROI achievers, nearly 70% say that at least a quarter of their knowledge workers are using analytics software other than spreadsheets compared to only 56% in the low ROI group. Also among the high ROI performers, 63% said data and analytic workers collaborate well or extremely well compared to only 51% in the low ROI group. The data from the survey shows that supporting more business domains with analytics and providing cross-functional analytics correlates with higher ROI. So to maximize ROI, orgs should be transitioning workers from spreadsheets to analytics software. They should be letting them collaborate effectively and letting them do so cross-functionally. >> Yeah, that cross-functional collaboration is essential for anyone in any organization and in any discipline. Another key thing that jumped out from the survey was around shadow IT. The business side is using more data science tools than the IT side. And it's expected to spend more on analytics than other IT. What risks does this present to the overall organization, if IT and the lines of business guys and gals aren't really aligned? >> Well, there needs to be better collaboration and alignment between IT and the line of business. The data from the survey, however, shows that business managers, they're expected to spend more on analytics and use more analytics tools than IT is aware of. And this isn't because the lines of business have recognized the value of analytics and plan to invest accordingly, but a lack of alignment between IT and business. This will negatively impact governance, which ultimately impedes democratization and hence ROI. >> So Jason, where can organizations that are maybe at the outset of their analytics journey, or maybe they're in environments where there's multiple analytics tools across shadow IT, where can they go to Alteryx to learn more about how they can really simplify, streamline, and dial up the value on their investment? >> Well, they can learn more on our website. I also encourage them to explore the Alteryx community, which has lots of best practices, not just in terms of how you do the analytics, but how you stand up in Alteryx environment, but also to take a look at your analytics stack and prioritize technologies that can snap to and enhance your organization's governance posture. It doesn't have to change it, but it should be able to align to and enhance it. >> And of course, as you mentioned, it's about people, process, and technologies. Jason, thank you so much for joining me today, unpacking the IDC info brief and the great nuggets in there. Lots that organizations can learn and really become empowered to maximize their analytics investments. We appreciate your time. >> Thank you, it's been a pleasure. >> In a moment, Alan Jacobson, who's the chief data and analytics officer at Alteryx is going to join me. He's going to be here to talk about how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. You're watching "theCUBE", the leader in tech enterprise coverage. >> Somehow many have come to believe that data analytics is for the few, for the scientists, the PhDs, the MBAs. Well, it is for them, but that's not all. You don't have to have an advanced degree to do amazing things with data. You don't even have to be a numbers person. You can be just about anything. A titan of industry or a future titan of industry. You could be working to change the world, your neighborhood, or the course of your business. You can be saving lives or just looking to save a little time. The power of data analytics shouldn't be limited to certain job titles or industries or organizations because when more people are doing more things with data, more incredible things happen. Analytics makes us smarter and faster and better at what we do. It's practically a superpower. That's why we believe analytics is for everyone, and everything, and should be everywhere. That's why we believe in analytics for all. (upbeat music) >> Hey, everyone. Welcome back to "Accelerating Analytics Maturity". I'm your host, Lisa Martin. Alan Jacobson joins me next. The chief of data and analytics officer at Alteryx. Alan, it's great to have you on the program. >> Thanks, Lisa. >> So Alan, as we know, everyone knows that being data driven is very important. It's a household term these days, but 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. What's your advice, your recommendations for organizations who are just starting out with analytics? >> You're spot on, many organizations really aren't leveraging the full capability of their knowledge workers. And really the first step is probably assessing where you are on the journey, whether that's you personally, or your organization as a whole. We just launched an assessment tool on our website that we built with the International Institute of Analytics, that in a very short period of time, in about 15 minutes, you can go on and answer some questions and understand where you sit versus your peer set versus competitors and kind of where you are on the journey. >> So when people talk about data analytics, they often think, ah, this is for data science experts like people like you. So why should people in the lines of business like the finance folks, the marketing folks, why should they learn analytics? >> So domain experts are really in the best position. They know where the gold is buried in their companies. They know where the inefficiencies are. And it is so much easier and faster to teach a domain expert a bit about how to automate a process or how to use analytics than it is to take a data scientist and try to teach them to have the knowledge of a 20 year accounting professional or a logistics expert of your company. Much harder to do that. And really, if you think about it, the world has changed dramatically in a very short period of time. If you were a marketing professional 30 years ago, you likely didn't need to know anything about the internet, but today, do you know what you would call that marketing professional if they didn't know anything about the internet, probably unemployed or retired. And so knowledge workers are having to learn more and more skills to really keep up with their professions. And analytics is really no exception. Pretty much in every profession, people are needing to learn analytics to stay current and be capable for their companies. And companies need people who can do that. >> Absolutely, it seems like it's table stakes these days. Let's look at different industries now. Are there differences in how you see analytics in automation being employed in different industries? I know Alteryx is being used across a lot of different types of organizations from government to retail. I also see you're now with some of the leading sports teams. Any differences in industries? >> Yeah, there's an incredible actually commonality between the domains industry to industry. So if you look at what an HR professional is doing, maybe attrition analysis, it's probably quite similar, whether they're in oil and gas or in a high tech software company. And so really the similarities are much larger than you might think. And even on the sports front, we see many of the analytics that sports teams perform are very similar. So McLaren is one of the great partners that we work with and they use Alteryx across many areas of their business from finance to production, extreme sports, logistics, wind tunnel engineering, the marketing team analyzes social media data, all using Alteryx, and if I take as an example, the finance team, the finance team is trying to optimize the budget to make sure that they can hit the very stringent targets that F1 Sports has, and I don't see a ton of difference between the optimization that they're doing to hit their budget numbers and what I see Fortune 500 finance departments doing to optimize their budget, and so really the commonality is very high, even across industries. >> I bet every Fortune 500 or even every company would love to be compared to the same department within McLaren F1. Just to know that wow, what they're doing is so incredibly important as is what we're doing. >> So talk- >> Absolutely. >> About lessons learned, what lessons can business leaders take from those organizations like McLaren, who are the most analytically mature? >> Probably first and foremost, is that the ROI with analytics and automation is incredibly high. Companies are having a ton of success. It's becoming an existential threat to some degree, if your company isn't going on this journey and your competition is, it can be a huge problem. IDC just did a recent study about how companies are unlocking the ROI using analytics. And the data was really clear, organizations that have a higher percentage of their workforce using analytics are enjoying a much higher return from their analytic investment, and so it's not about hiring two double PhD statisticians from Oxford. It really is how widely you can bring your workforce on this journey, can they all get 10% more capable? And that's having incredible results at businesses all over the world. An another key finding that they had is that the majority of them said that when they had many folks using analytics, they were going on the journey faster than companies that didn't. And so picking technologies that'll help everyone do this and do this fast and do it easily. Having an approachable piece of software that everyone can use is really a key. >> So faster, able to move faster, higher ROI. I also imagine analytics across the organization is a big competitive advantage for organizations in any industry. >> Absolutely the IDC, or not the IDC, the International Institute of Analytics showed huge correlation between companies that were more analytically mature versus ones that were not. They showed correlation to growth of the company, they showed correlation to revenue and they showed correlation to shareholder values. So across really all of the key measures of business, the more analytically mature companies simply outperformed their competition. >> And that's key these days, is to be able to outperform your competition. You know, one of the things that we hear so often, Alan, is people talking about democratizing data and analytics. You talked about the line of business workers, but I got to ask you, is it really that easy for the line of business workers who aren't trained in data science to be able to jump in, look at data, uncover and extract business insights to make decisions? >> So in many ways, it really is that easy. I have a 14 and 16 year old kid. Both of them have learned Alteryx, they're Alteryx certified and it was quite easy. It took 'em about 20 hours and they were off to the races, but there can be some hard parts. The hard parts have more to do with change management. I mean, if you're an accountant that's been doing the best accounting work in your company for the last 20 years, and all you happen to know is a spreadsheet for those 20 years, are you ready to learn some new skills? And I would suggest you probably need to, if you want to keep up with your profession. The big four accounting firms have trained over a hundred thousand people in Alteryx. Just one firm has trained over a hundred thousand. You can't be an accountant or an auditor at some of these places without knowing Alteryx. And so the hard part, really in the end, isn't the technology and learning analytics and data science, the harder part is this change management, change is hard. I should probably eat better and exercise more, but it's hard to always do that. And so companies are finding that that's the hard part. They need to help people go on the journey, help people with the change management to help them become the digitally enabled accountant of the future, the logistics professional that is E enabled, that's the challenge. >> That's a huge challenge. Cultural shift is a challenge, as you said, change management. How do you advise customers if you might be talking with someone who might be early in their analytics journey, but really need to get up to speed and mature to be competitive, how do you guide them or give them recommendations on being able to facilitate that change management? >> Yeah, that's a great question. So people entering into the workforce today, many of them are starting to have these skills. Alteryx is used in over 800 universities around the globe to teach finance and to teach marketing and to teach logistics. And so some of this is happening naturally as new workers are entering the workforce, but for all of those who are already in the workforce, have already started their careers, learning in place becomes really important. And so we work with companies to put on programmatic approaches to help their workers do this. And so it's, again, not simply putting a box of tools in the corner and saying free, take one. We put on hackathons and analytic days, and it can be great fun. We have a great time with many of the customers that we work with, helping them do this, helping them go on the journey, and the ROI, as I said, is fantastic. And not only does it sometimes affect the bottom line, it can really make societal changes. We've seen companies have breakthroughs that have really made great impact to society as a whole. >> Isn't that so fantastic, to see the difference that that can make. It sounds like you guys are doing a great job of democratizing access to Alteryx to everybody. We talked about the line of business folks and the incredible importance of enabling them and the ROI, the speed, the competitive advantage. Can you share some specific examples that you think of Alteryx customers that really show data breakthroughs by the lines of business using the technology? >> Yeah, absolutely, so many to choose from. I'll give you two examples quickly. One is Armor Express. They manufacture life saving equipment, defensive equipments, like armor plated vests, and they were needing to optimize their supply chain, like many companies through the pandemic. We see how important the supply chain is. And so adjusting supply to match demand is really vital. And so they've used Alteryx to model some of their supply and demand signals and built a predictive model to optimize the supply chain. And it certainly helped out from a dollar standpoint. They cut over a half a million dollars of inventory in the first year, but more importantly, by matching that demand and supply signal, you're able to better meet customer demand. And so when people have orders and are looking to pick up a vest, they don't want to wait. And it becomes really important to get that right. Another great example is British Telecom. They're a company that services the public sector. They have very strict reporting regulations that they have to meet and they had, and this is crazy to think about, over 140 legacy spreadsheet models that they had to run to comply with these regulatory processes and report, and obviously running 140 legacy models that had to be done in a certain order and length, incredibly challenging. It took them over four weeks each time that they had to go through that process. And so to save time and have more efficiency in doing that, they trained 50 employees over just a two week period to start using Alteryx and learn Alteryx. And they implemented an all new reporting process that saw a 75% reduction in the number of man hours it took to run in a 60% run time performance. And so, again, a huge improvement. I can imagine it probably had better quality as well, because now that it was automated, you don't have people copying and pasting data into a spreadsheet. And that was just one project that this group of folks were able to accomplish that had huge ROI, but now those people are moving on and automating other processes and performing analytics in other areas. So you can imagine the impact by the end of the year that they will have on their business, potentially millions upon millions of dollars. And this is what we see again and again, company after company, government agency after government agency, is how analytics are really transforming the way work is being done. >> That was the word that came to mind when you were describing the all three customer examples, transformation, this is transformative. The ability to leverage Alteryx, to truly democratize data and analytics, give access to the lines of business is transformative for every organization. And also the business outcome you mentioned, those are substantial metrics based business outcomes. So the ROI in leveraging a technology like Alteryx seems to be right there, sitting in front of you. >> That's right, and to be honest, it's not only important for these businesses. It's important for the knowledge workers themselves. I mean, we hear it from people that they discover Alteryx, they automate a process, they finally get to get home for dinner with their families, which is fantastic, but it leads to new career paths. And so knowledge workers that have these added skills have so much larger opportunity. And I think it's great when the needs of businesses to become more analytic and automate processes actually matches the needs of the employees, and they too want to learn these skills and become more advanced in their capabilities. >> Huge value there for the business, for the employees themselves to expand their skillset, to really open up so many opportunities for not only the business to meet the demands of the demanding customer, but the employees to be able to really have that breadth and depth in their field of service. Great opportunities there, Alan. Is there anywhere that you want to point the audience to go to learn more about how they can get started? >> Yeah, so one of the things that we're really excited about is how fast and easy it is to learn these tools. So any of the listeners who want to experience Alteryx, they can go to the website, there's a free download on the website. You can take our analytic maturity assessment, as we talked about at the beginning, and see where you are on the journey and just reach out. We'd love to work with you and your organization to see how we can help you accelerate your journey on analytics and automation. >> Alan, it was a pleasure talking to you about democratizing data and analytics, the power in it for organizations across every industry. We appreciate your insights and your time. >> Thank you so much. >> In a moment, Paula Hansen, who is the president and chief revenue officer of Alteryx, and Jacqui Van der Leij Greyling, who's the global head of tax technology at eBay, will join me. You're watching "theCUBE", the leader in high tech enterprise coverage. >> 1200 hours of wind tunnel testing, 30 million race simulations, 2.4 second pit stops. >> Make that 2.3. >> Sector times out the wazoo. >> Way too much of this. >> Velocities, pressures, temperatures, 80,000 components generating 11.8 billion data points and one analytics platform to make sense of it all. When McLaren needs to turn complex data into winning insights, they turn to Alteryx. Alteryx, analytics automation. (upbeat music) >> Hey, everyone, welcome back to the program. Lisa Martin here, I've got two guests joining me. Please welcome back to "theCUBE" Paula Hansen, the chief revenue officer and president at Alteryx, and Jacqui Van der Leij Greyling joins us as well, the global head of tax technology at eBay. They're going to share with you how Alteryx is helping eBay innovate with analytics. Ladies, welcome, it's great to have you both on the program. >> Thank you, Lisa, it's great to be here. >> Yeah, Paula, we're going to start with you. In this program, we've heard from Jason Klein, we've heard from Alan Jacobson. They talked about the need to democratize analytics across any organization to really drive innovation. With analytics, as they talked about, at the forefront of software investments, how's Alteryx helping its customers to develop roadmaps for success with analytics? >> Well, thank you, Lisa. It absolutely is about our customers' success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts of course with our innovative technology and platform, but ultimately we help our customers to create a culture of data literacy and analytics from the top of the organization, starting with the C-suite. And we partner with our customers to build their roadmaps for scaling that culture of analytics, through things like enablement programs, skills assessments, hackathons, setting up centers of excellence to help their organization scale and drive governance of this analytics capability across the enterprise. So at the end of the day, it's really about helping our customers to move up their analytics maturity curve with proven technologies and best practices, so they can make better business decisions and compete in their respective industries. >> Excellent, sounds like a very strategic program, we're going to unpack that. Jacqui, let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the IDC report that showed that 93% of organizations are not utilizing the analytics skills of their employees, but then there's eBay. How Jacqui did eBay become one of the 7% of organizations who's really maturing and how are you using analytics across the organization at eBay? >> So I think the main thing for us is when we started out was is that, our, especially in finance, they became spreadsheet professionals instead of the things that we really want our employees to add value to. And we realized we had to address that. And we also knew we couldn't wait for all our data to be centralized until we actually start using the data or start automating and being more effective. So ultimately we really started very, very actively embedding analytics in our people and our data and our processes. >> Starting with people is really critical. Jacqui, continuing with you, what were some of the roadblocks to analytics adoption that you faced and how did you overcome them? >> So I think eBay is a very data driven company. We have a lot of data. I think we are 27 years around this year, so we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and just finding those data sources and finding ways to connect to them to move forward. The other thing is that people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals. And there was no, we were not independent. You couldn't move forward, you would've put it on somebody else's roadmap to get the data and to get the information if you want it. So really finding something that everybody could access analytics or access data. And finally we have to realize is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy, and that is not so daunting on somebody who's brand new to the field? And I would call those out as your major roadblocks, because you always have, not always, but most of the times you have support from the top, and in our case we have, but at the end of the day, it's our people that need to actually really embrace it, and making that accessible for them, I would say is definitely not per se, a roadblock, but basically a block you want to be able to move. >> It's really all about putting people first. Question for both of you, and Paula we'll start with you, and then Jacqui we'll go to you. I think the message in this program that the audience is watching with us is very clear. Analytics is for everyone, should be for everyone. Let's talk now about how both of your organizations are empowering people, those in the organization that may not have technical expertise to be able to leverage data, so that they can actually be data driven. Paula. >> Yes, well, we leverage our platform across all of our business functions here at Alteryx. And just like Jacqui explained, at eBay finance is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jacqui mentioned, we have this huge amount of data flowing through our enterprise and the opportunity to leverage that into insights and analytics is really endless. So our CFO Kevin Rubin has been a key sponsor for using our own technology. We use Alteryx for forecasting all of our key performance metrics, for business planning, across our audit function, to help with compliance and regulatory requirements, tax, and even to close our books at the end of each quarter. So it's really going to remain across our business. And at the end of the day, it comes to how do you train users? How do you engage users to lean into this analytic opportunity to discover use cases? And so one of the other things that we've seen many companies do is to gamify that process, to build a game that brings users into the experience for training and to work with each other, to problem solve and along the way, maybe earn badges depending on the capabilities and trainings that they take. And just have a little healthy competition as an employee base around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jacqui mentioned, it's really about ensuring that people feel comfortable, that they feel supported, that they have access to the training that they need, and ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >> That confidence is key. Jacqui, talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >> Yeah, I think it means to what Paula has said in terms of getting people excited about it, but it's also understanding that this is a journey and everybody is at a different place in their journey. You have folks that's already really advanced who has done this every day. And then you have really some folks that this is brand new or maybe somewhere in between. And it's about how you get everybody in their different phases to get to the initial destination. I say initial, because I believe a journey is never really complete. What we have done is that we decided to invest, and built a proof of concept, and we got our CFO to sponsor a hackathon. We opened it up to everybody in finance in the middle of the pandemic. So everybody was on Zoom and we told people, listen, we're going to teach you this tool, it's super easy, and let's just see what you can do. We ended up having 70 entries. We had only three weeks. So and these are people that do not have a background. They are not engineers, they're not data scientists. And we ended up with a 25,000 hour savings at the end of that hackathon from the 70 entries with people that have never, ever done anything like this before. And there you have the result. And then it just went from there. People had a proof of concept. They knew that it worked and they overcame the initial barrier of change. And that's where we are seeing things really, really picking up now. >> That's fantastic. And the business outcome that you mentioned there, the business impact is massive, helping folks get that confidence to be able to overcome sometimes the cultural barriers is key here. I think another thing that this program has really highlighted is there is a clear demand for data literacy in the job market, regardless of organization. Can each of you share more about how you're empowering the next generation of data workers? Paula, we'll start with you. >> Absolutely, and Jacqui says it so well, which is that it really is a journey that organizations are on and we as people in society are on in terms of upskilling our capabilities. So one of the things that we're doing here at Alteryx to help address this skillset gap on a global level is through a program that we call SparkED, which is essentially a no-cost analytics education program that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this program's really developed just to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with SparkED. We started last May, but we currently have over 850 educational institutions globally engaged across 47 countries, and we're going to continue to invest here because there's so much opportunity for people, for society and for enterprises, when we close the gap and empower more people with the necessary analytics skills to solve all the problems that data can help solve. >> So SparkED has made a really big impact in such a short time period. It's going to be fun to watch the progress of that. Jacqui, let's go over to you now. Talk about some of the things that eBay is doing to empower the next generation of data workers. >> So we basically wanted to make sure that we kept that momentum from the hackathon, that we don't lose that excitement. So we just launched the program called eBay Masterminds. And what it basically is, is it's an inclusive innovation in each other, where we firmly believe that innovation is for upskilling for all analytics roles. So it doesn't matter your background, doesn't matter which function you are in, come and participate in in this where we really focus on innovation, introducing new technologies and upskilling our people. We are, apart from that, we also said, well, we shouldn't just keep it to inside eBay. We have to share this innovation with the community. So we are actually working on developing an analytics high school program, which we hope to pilot by the end of this year, where we will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, but also how to use Alteryx. And we're working with, actually, we're working with SparkED and they're helping us develop that program. And we really hope that at, say, by the end of the year, we have a pilot and then also next year, we want to roll it out in multiple locations in multiple countries and really, really focus on that whole concept of analytics for all. >> Analytics for all, sounds like Alteryx and eBay have a great synergistic relationship there that is jointly aimed at especially going down the stuff and getting people when they're younger interested, and understanding how they can be empowered with data across any industry. Paula, let's go back to you, you were recently on "theCUBE"'s Supercloud event just a couple of weeks ago. And you talked about the challenges the companies are facing as they're navigating what is by default a multi-cloud world. How does the Alteryx Analytics Cloud platform enable CIOs to democratize analytics across their organization? >> Yes, business leaders and CIOs across all industries are realizing that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organizations. Last I checked, there was 2 million data scientists in the world, so that's woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. So what we're seeing now with CIOs, with business leaders is that they're integrating data analysis and the skillset of data analysis into virtually every job function, and that is what we think of when we think of analytics for all. And so our mission with Alteryx Analytics Cloud is to empower all of those people in every job function, regardless of their skillset, as Jacqui pointed out from people that are just getting started all the way to the most sophisticated of technical users. Every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organizations. So that's our goal with Alteryx Analytics Cloud, and it operates in a multi cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyze, and report out so that we can break down data silos across the enterprise and help drive real business outcomes as a result of unlocking the potential of data. >> As well as really lessening that skill gap. As you were saying, there's only 2 million data scientists. You don't need to be a data scientist, that's the beauty of what Alteryx is enabling and eBay is a great example of that. Jacqui, let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where Alteryx fits in as that analytics maturity journey continues and what are some of the things that you are most excited about as analytics truly gets democratized across eBay? >> When we're starting up and getting excited about things when it comes to analytics, I can go on all day, but I'll keep it short and sweet for you. I do think we are on the top of the pool of data scientists. And I really feel that that is your next step, for us anyways, is that how do we get folks to not see data scientists as this big thing, like a rocket scientist, it's something completely different. And it's something that is in everybody in a certain extent. So again, partnering with Alteryx who just released the AI ML solution, allowing folks to not have a data scientist program, but actually build models and be able to solve problems that way. So we have engaged with Alteryx and we purchased the licenses, quite a few. And right now through our Masterminds program, we're actually running a four month program for all skill levels, teaching them AI ML and machine learning and how they can build their own models. We are really excited about that. We have over 50 participants without a background from all over the organization. We have members from our customer services. We have even some of our engineers are actually participating in the program. We just kicked it off. And I really believe that that is our next step. I want to give you a quick example of the beauty of this is where we actually just allow people to go out and think about ideas and come up with things. And one of the people in our team who doesn't have a data scientist background at all, was able to develop a solution where there is a checkout feedback functionality on the eBay side where sellers or buyers can verbatim add information. And she built a model to be able to determine what relates to tax specific, what is the type of problem, and even predict how that problem can be solved before we as a human even step in, and now instead of us or somebody going to verbatim and try to figure out what's going on there, we can focus on fixing the error versus actually just reading through things and not adding any value, and it's a beautiful tool and I was very impressed when I saw the demo and definitely developing that sort of thing. >> That sounds fantastic. And I think just the one word that keeps coming to mind, and we've said this a number of times in the program today is empowerment. What you're actually really doing to truly empower people across the organization with varying degrees of skill level, going down to the high school level, really exciting. We'll have to stay tuned to see what some of the great things are that come from this continued partnership. Ladies, I want to thank you so much for joining me on the program today and talking about how Alteryx and eBay are really partnering together to democratize analytics and to facilitate its maturity. It's been great talking to you. >> Thank you, Lisa. >> Thank you so much. (cheerful electronic music) >> As you heard over the course of our program, organizations where more people are using analytics who have deeper capabilities in each of the four Es, that's everyone, everything, everywhere, and easy analytics, those organizations achieve more ROI from their respective investments in analytics and automation than those who don't. We also heard a great story from eBay, great example of an enterprise that is truly democratizing analytics across its organization. It's enabling and empowering line of business users to use analytics, not only focused on key aspects of their job, but develop new skills rather than doing the same repetitive tasks. We want to thank you so much for watching the program today. Remember you can find all of the content on thecube.net. You can find all of the news from today on siliconangle.com and of course alteryx.com. We also want to thank Alteryx for making this program possible and for sponsoring "theCUBE". For all of my guests, I'm Lisa Martin. We want to thank you for watching and bye for now. (upbeat music)
SUMMARY :
in the next 12 to 18 months. Excited to talk with you. over the next 12 to 18 months, And it looks like from the info brief and the world is changing data. that the info brief uncovered with respect So for example, on the people side, in the data and analytics and the answer, that'll be able to. just so we get that clean Thank you for that. that the info brief uncovered as compared to the technology itself. So overall, the enterprises to be aware of at the outset? is that the people aspect of analytics If we could do the same, Lisa, Here, I'm going to give us a little break. to the data and analytics and really maximize the investments And the data from this survey shows this And it's expected to spend more and plan to invest accordingly, that can snap to and the great nuggets in there. Alteryx is going to join me. that data analytics is for the few, Alan, it's great to that being data driven is very important. And really the first step the lines of business and more skills to really keep of the leading sports teams. between the domains industry to industry. to be compared to the same is that the majority of them said So faster, able to So across really all of the is to be able to outperform that is E enabled, that's the challenge. and mature to be competitive, around the globe to teach finance and the ROI, the speed, that they had to run to comply And also the business of the employees, and they of the demanding customer, to see how we can help you the power in it for organizations and Jacqui Van der Leij 1200 hours of wind tunnel testing, to make sense of it all. back to the program. going to start with you. So at the end of the day, one of the 7% of organizations to be centralized until we of the roadblocks to analytics adoption and to get the information if you want it. that the audience is watching and the confidence to be able to be a part to really be data driven. in their different phases to And the business outcome and to work hand in hand Jacqui, let's go over to you now. We have to share this Paula, let's go back to in the opportunity to unlock and eBay is a great example of that. and be able to solve problems that way. that keeps coming to mind, Thank you so much. in each of the four Es,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jacqui | PERSON | 0.99+ |
Paula | PERSON | 0.99+ |
Jason Klein | PERSON | 0.99+ |
Paula Hansen | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Paula Hansen | PERSON | 0.99+ |
Alan Jacobson | PERSON | 0.99+ |
Alteryx | ORGANIZATION | 0.99+ |
eBay | ORGANIZATION | 0.99+ |
Jason | PERSON | 0.99+ |
International Institute of Analytics | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Alan | PERSON | 0.99+ |
Alan Jacobson | PERSON | 0.99+ |
60% | QUANTITY | 0.99+ |
Kevin Rubin | PERSON | 0.99+ |
Jacqui Van der Leij Greyling | PERSON | 0.99+ |
14 | QUANTITY | 0.99+ |
International Institute of Analytics | ORGANIZATION | 0.99+ |
10% | QUANTITY | 0.99+ |
50 employees | QUANTITY | 0.99+ |
63% | QUANTITY | 0.99+ |
93% | QUANTITY | 0.99+ |
90% | QUANTITY | 0.99+ |
nine | QUANTITY | 0.99+ |
75% | QUANTITY | 0.99+ |
70 entries | QUANTITY | 0.99+ |
16 year | QUANTITY | 0.99+ |
1200 hours | QUANTITY | 0.99+ |
Paula Hansen & Jacqui van der Leij Greyling
>>Hey, everyone, welcome back to the programme. Lisa Martin here. I've got two guests joining me. Please welcome back to the Q. Paula Hanson, the chief Revenue officer and president at all tricks. And Jackie Vanderlei Grayling joins us as well. The global head of tax technology at eBay. They're gonna share with you how an all tricks is helping eBay innovate with analytics. Ladies, welcome. It's great to have you both on the programme. >>Thank you, Lisa. Not great to be >>here. >>Yeah, Paula, we're gonna start with you in this programme. We've heard from Jason Klein. We've heard from Allan Jacobsen. They talked about the need to democratise analytics across any organisation to really drive innovation with analytics as they talked about at the forefront of software investments. House all tricks, helping its customers to develop roadmaps for success with analytics. >>Well, thank you, Lisa. Absolutely is about our customers success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts, of course, with our innovative technology and platform. But ultimately we help our customers to create a culture of data literacy and analytics from the top of the organisation starting with the C suite and we partner with our customers to build their road maps for scaling that culture of analytics through things like enablement programmes, skills assessments, hackathons, uh, setting up centres of excellence to help their organisation scale and drive governance of this, uh, analytics capability across the Enterprise. So at the end of the day, it's really about helping our customers to move up their analytics maturity curve with proven technologies and best practises so they can make better business decisions and compete in their respective industries. >>Excellent. Sounds like a very strategic programme. We're gonna unpack that, Jackie, let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the I. D. C report that showed that 93% of organisations are not utilising the analytic skills of their employees. But then there's eBay. How Jackie did eBay become one of the 7% of organisations who's really maturing and how are you using analytics across the organisation at bay? >>So I think the main thing for us is when we started out was is that you know our especially in finance. They became spreadsheet professionals instead of the things that we really want our influence to add value to. And we realised we have to address that. And we also knew we couldn't wait for all our data to be centralised until we actually start using the data or start automating and be more effective. Um, so ultimately, we really started very, very actively embedding analytics in our people and our data and our processes. >>Starting with people is really critical jacket continuing with you. What was in the roadblocks to analytics adoption that you faced and how did you overcome them? >>So I think you know, Eva is a very data driven company. We have a lot of data. I think we are 27 years around this year. So we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and just finding those data sources and finding ways to connect to them, um, to move forward. The other thing is that you know, people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals, right? And there was no we're not independent. You couldn't move forward. You're dependent on somebody else's roadmap to get to data to get the information you want it. So really finding something that everybody could access analytics or access data. And finally we have to realise, is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy and that is not so daunting on somebody who's brand new to the field? And I would I would call those out as your as your major roadblocks, because you always have always. But most of the times you have support from the top. In our case we have. But in the end of the day, it's it's our people that need to actually really embrace it and making that accessible for them. I would say it's not to say a road block a block you want to be able to do. >>It's really all about putting people first question for both of you and Paula will start with you and then Jackie will go to you. I think the message in this programme that the audience is watching with us is very clear. Analytics is for everyone should be for everyone. Let's talk now about how both of your organisations are empowering people, those in the organisation that may not have technical expertise to be able to leverage data so that they can actually be data driven colour. >>Yes, well, we leverage our platform across all of our business functions here at all tricks. And just like Jackie explained that eBay finance is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jackie mentioned, we have this huge amount of data, uh, flowing through our enterprise, and the opportunity to leverage that into insights and analytics is really endless. So our CFO, Kevin Ruben has been a key sponsor for using our own technology. We use all tricks for forecasting all of our key performance metrics for business planning across our audit function, uh, to help with compliance and regulatory requirements, tax and even to close our books at the end of each quarter. So it's really remain across our business. And at the end of the day, it comes to How do you train users? How do you engage users to lean into this analytic opportunity to discover use cases? And so one of the other things that we've seen many companies do is to gamify that process, to build a game that brings users into the experience for training and to work with each other to problem solve and, along the way, maybe earn badges, depending on the capabilities and trainings that they take and just have a little healthy competition, Uh, as an employee based around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jackie mentioned, it's really about ensuring that people feel comfortable that they feel supportive, that they have access to the training that they need, and ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >>That confidence is key. Jackie talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >>I think it means to what Paula has said in terms of, you know, getting people excited about it. But it's also understanding that this is a journey and everybody is the different place in their journey. You have folks that's already really advanced. Who's done this every day. And then you have really some folks that this is brand new and, um, or maybe somewhere in between. And it's about how you could get everybody in their different phases to get to the the initial destination. And I say initial because I believe the journey is never really complete. Um, what we have done is that we decided to invest in a group of concept when we got our CFO to sponsor a hackathon. Um, we open it up to everybody in finance, um, in the middle of the pandemic. So everybody was on Zoom, um, and we had and we told people, Listen, we're gonna teach you this tool. It's super easy, and let's just see what you can do. We ended up having 70 injuries. We had only three weeks. So these are people that that do not have a background. They are not engineers and not data scientists and we ended up with 25,000 our savings at the end of the hackathon. Um, from the 70 countries with people that I've never, ever done anything like this before. And there you have the results. And they just went from there because people had a proof of concept. They knew that it worked and they overcame the initial barrier of change. Um, and that's what we are seeing things really, really picking up now >>that's fantastic. And the business outcome that you mentioned that the business impact is massive, helping folks get that confidence to be able to overcome. Sometimes the cultural barriers is key there. I think another thing that this programme has really highlighted is there is a clear demand for data literacy in the job market, regardless of organisation. Can each of you share more about how your empowering the next generation of data workers Paula will start with you? >>Absolutely. And Jackie says it so well, which is that it really is a journey that organisations are on and we, as people in society, are on in terms of up skilling our capabilities. Uh, so one of the things that we're doing here at all tricks to help address the skill set gap on a global level is through a programme that we call Sparked, which is essentially a no cost analyst education programme that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this programme is really developed just to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with sparked we started last May, but we currently have over 850 educational institutions globally engaged across 47 countries, and we're going to continue to invest here because there's so much opportunity for people, for society and for enterprises when we close gap and empower more people with the necessary analytic skills to solve all the problems that data can help solve. >>So >>I just made a really big impact in such a short time period is gonna be fun to watch the progress of that. Jackie, let's go over to you now Talk about some of the things that eBay is doing to empower the next generation of data workers. >>So we definitely wanted to make sure that we kept implemented from the hackathon that we don't lose that excitement life. So we just launched a programme for evil masterminds and what it basically is. It's an inclusive innovation initiative where we firmly believe that innovation is all upscaling for all analytics role. So it doesn't matter. Your background doesn't matter which function you are in. Come and participate in this where we really focus on innovation, introducing these technologies and upscaling of people. Um, we are apart from that. We also said, Well, we should just keep it to inside the way we have to share this innovation with the community. So we are actually working on developing an analytics high school programme which we hope to pilot by the end of this year. We will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, But also, um, how to use all tricks and we're working with Actually, we're working with spark and they're helping us develop that programme. And we really hope that it is said by the end of the year, have a pilot and then also makes you must have been rolled out in multiple locations in multiple countries and really, really, uh, focused on that whole concept of analytic school >>analytics. Girl sounds like ultra and everybody have a great synergistic relationship there that is jointly aimed at especially kind of going down the stock and getting people when they're younger, interested and understanding how they can be empowered with data across any industry. Paula, let's go back to you. You were recently on the cubes Super Cloud event just a couple of weeks ago and you talked about the challenges the companies are facing as they are navigating what is by default, a multi cloud world. How does the all tricks analytics cloud platform enable CEO s to democratise analytics across their organisation? >>Yes, business leaders and CEO s across all industries are realising that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organisations. Last I checked, there was two million data scientists in the world. So that's, uh, woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. So what we're seeing now with CEO s with business leaders is that they are integrating data analysis and the skill set of data analysis into virtually every job function. Uh, and that is what we think of when we think of analytics for all. And so our mission with all tricks analytics cloud is to empower all of those people in every job function, regardless of their skill set, as Jackie pointed out, from people that would are just getting started all the way to the most sophisticated of technical users. Um, every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organisations. So that's our goal with all tricks, analytics cloud and it operates in a multi cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyse and report out so that we can break down data silos across the Enterprise and Dr Real Business Outcomes. As a result, of unlocking the potential of data >>as well as really listening that skills gap. As you were saying, There's only two million data scientists. You don't need to be a data scientist. That's the beauty of what all tricks is enabling. And eBay is a great example of that. Jackie, let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where all tricks fits in as that analytics maturity journey continues. And what are some of the things that you're most excited about as analytics truly gets democratised across eBay >>when we start about getting excited about things when it comes to analytics, I can go on all day, but I'll keep it short and sweet for you. Um, I do think we're on the topic full of data scientists, and I really feel that that is your next step for us, anyway. Is that how do we get folks to not see data scientist as this big thing like a rocket scientist it's something completely different and it's something that is in everybody in a certain extent. So, um, game partnering with all tricks to just release uh, ai ml um, solution allowing. You know, folks do not have a data scientist programme but actually build models and be able to solve problems that way. So we have engaged with all turrets and we purchase the licence is quite a few. And right now, through our masterminds programme, we're actually running a four months programme. Um, for all skill levels, um, teaching them ai ml and machine learning and how they can build their own models. Um, we are really excited about that. We have over 50 participants without the background from all over the organisation. We have members from our customer services. We have even some of our engineers are actually participating in the programme will just kick it off. And I really believe that that is our next step. Um, I want to give you a quick example of the beauty of this is where we actually, um, just allow people to go out and think about ideas and come up with things and one of the people in our team who doesn't have a data scientist background at all, was able to develop a solution. Where, um, you know there is a checkout feedback checkout functionality on the eBay side, There's sellers or buyers can pervade them at information. And she built a model to be able to determine what relates to tax specific what is the type of problem and even predict how that problem can be solved before we as human, even stepped in. And now, instead of us or somebody going to debate and try to figure out what's going on there, we can focus on fixing their versus, um, actually just reading through things and not adding any value and its a beautiful tool. And I'm very impressed when we saw the demo and they've been developing that further. >>That sounds fantastic. And I think just the one word that keeps coming to mind. And we've said this a number of times in the programme. Today's empowerment, what you're actually really doing to truly empower people across the organisation with with varying degrees of skill level, going down to the high school level really exciting. We'll have so stay tuned to see what some of the great things are that come from this continued partnership? Ladies, I wanna thank you so much for joining me on the programme today and talking about how all tricks and eBay are really partnering together to democratise analytics and to facilitate its maturity. It's been great talking to you. >>Thank you. >>Thank you so much.
SUMMARY :
It's great to have you both on the programme. They talked about the need to democratise analytics So at the end of the day, it's really about helping our customers to move Speaking of analytics maturity, one of the things that we talked about in this event is the I. instead of the things that we really want our influence to add value to. adoption that you faced and how did you overcome them? But most of the times you have support from the top. those in the organisation that may not have technical expertise to be able to leverage data And at the end of the day, it comes to How do you train users? Jackie talk about some of the ways that you're empowering folks without that technical and we had and we told people, Listen, we're gonna teach you this tool. And the business outcome that you mentioned that the business impact is massive, And so this programme is really developed just to Jackie, let's go over to you now Talk about some of the things that eBay is doing to empower the next And we really hope that it is said by the end of the year, have a pilot and then also that is jointly aimed at especially kind of going down the stock and getting people when they're younger, have a meaningful role in the opportunity to unlock the potential of the data for It obviously has the right culture to adapt to that. And she built a model to be able to determine of the great things are that come from this continued partnership?
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jason Klein | PERSON | 0.99+ |
Paula | PERSON | 0.99+ |
Jackie | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Kevin Ruben | PERSON | 0.99+ |
eBay | ORGANIZATION | 0.99+ |
Allan Jacobsen | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
25,000 | QUANTITY | 0.99+ |
Jackie Vanderlei Grayling | PERSON | 0.99+ |
Paula Hansen | PERSON | 0.99+ |
93% | QUANTITY | 0.99+ |
27 years | QUANTITY | 0.99+ |
70 injuries | QUANTITY | 0.99+ |
Jacqui van der Leij Greyling | PERSON | 0.99+ |
both | QUANTITY | 0.99+ |
70 countries | QUANTITY | 0.99+ |
two guests | QUANTITY | 0.99+ |
over 50 participants | QUANTITY | 0.99+ |
four months | QUANTITY | 0.99+ |
first question | QUANTITY | 0.99+ |
47 countries | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
over 850 educational institutions | QUANTITY | 0.98+ |
last May | DATE | 0.98+ |
Today | DATE | 0.97+ |
7% | QUANTITY | 0.97+ |
this year | DATE | 0.96+ |
Eva | ORGANIZATION | 0.96+ |
two million data scientists | QUANTITY | 0.95+ |
Super Cloud | EVENT | 0.95+ |
pandemic | EVENT | 0.95+ |
three weeks | QUANTITY | 0.94+ |
each | QUANTITY | 0.94+ |
one word | QUANTITY | 0.94+ |
couple of weeks ago | DATE | 0.91+ |
Paula Hanson | PERSON | 0.87+ |
Sparked | TITLE | 0.86+ |
end of this year | DATE | 0.84+ |
each quarter | QUANTITY | 0.83+ |
Q. | PERSON | 0.78+ |
D. C | TITLE | 0.68+ |
Zoom | ORGANIZATION | 0.64+ |
sparked | TITLE | 0.58+ |
Brent Meadows, Expedient & Bryan Smith, Expedient | VMware Explore 2022
(upbeat music) >> Hey everyone. Welcome to theCUBE's coverage of VMware Explore 2022. We are at Moscone West. Lisa Martin and Dave Nicholson here. Excited, really excited, whereas they were saying in the VMware keynote, pumped and jacked and jazzed to be back in-person with a lot of folks here. Keynote with standing room only. We've just come from that. We've got a couple of guests here from Expedient, going to unpack their relationship with VMware. Please welcome Brian Smith, the Senior Vice President and Chief Strategy Officer at Expedient. And Brent Meadows, the Vice President of Advanced Solution Architecture at Expedient. Guys it's great to have you on the program. >> Appreciate it bringing us on. >> Yep, welcome. >> Isn't it great to be back in person? >> It is phenomenal to be back. >> So let's talk about obviously three years since the last, what was called VMworld, so many dynamics in the market. Talk to us about what's going on at Expedient, we want to dig into Cloud Different, but kind of give us a lay of the land of what's going on and then we're going to uncrack the VMware partnership as well. >> Sure, so Expedient we're a full stack cloud service provider. So we have physical data centers that we run and then have VMware-based cloud and we've seen a huge shift from the client perspective during the pandemic in how they've really responded from everything pre-pandemic was very focused with Cloud First and trying to go that route only with hyper scaler. And there's been a big evolution with how people have to change how they think about their transformation to get the end result they're looking for. >> Talk about Cloud Different and what it's helping customers to achieve as everyone's in this accelerated transformation. >> Yeah. So, Cloud Different is something that Expedient branded. It's really about how the transformation works. And traditionally, companies thought about doing their transformation, at first they kept everything in house that they were doing and they started building their new applications out into a hyper scale cloud. And what that really is like is, a good analogy would be, it's like living in a house while you're renovating it. And I know what that's like from my relationship versus if you build a new house, or move to a new property that's completed already. And that's really the difference in that experience from a Cloud Different approach from transformation is you think of all the things that you have internally, and there's a lot of technical debt there, and that's a lot of weight that you're carrying when you're trying to do that transformation. So if you kind of flip that around and instead look to make that transformation and move all that technical debt into a cloud that's already built to run those same types of applications, a VMware-based cloud, now you can remove all of that noise, move into a curated stack of technology and everything just works. It has the security in place, your teams know how to run it, and then you can take that time you really reclaim and apply that towards new applications and new things that are strategic to the business. >> That's really critical, Brent, to get folks in the IT organization across the business, really focused on strategic initiatives rather than a lot of the mundane tasks that they just don't have time for. Brent, what are you hearing in the last couple of years with the dynamics we talked about, what are you hearing from the customer? >> Right. So, one of the big things and the challenges in the current dynamic is kind of that staffing part. So as people have built their infrastructure over the years, there's a lot of tribal knowledge that's been created during that process and every day more and more of that knowledge is walking out the door. So taking some of that technical debt that Brian mentioned and kind of removing that so you don't have to have all that tribal knowledge, really standardizing on the foundational infrastructure pieces, allows them to make that transition and not have to carry that technical debt along with them as they make their digital transformations. >> We heard a lot this morning in the keynote guys about customers going, most of them still being in cloud chaos, but VMware wanting them to get to cloud smart. What does that mean, Brian, from Expedient's perspective? What does cloud smart look like to Expedient and its customers? >> Yeah, we completely agree with that message. And it's something we've been preaching for a couple years in part of that Cloud Different story. And it's really about having a consistent wrapper across all of your environments. It doesn't matter if it's things that you're running on-premises that's legacy to things that are in a VMware-based cloud, like an Expedient cloud or things that are in a hyper scale, but having one consistent security, one consistent automation, one consistent cost management, really gives you the governance so that you can get the value out of cloud that you are hoping for and remove a lot of the noise and think less about the technology and more about what the business is getting out of the technology. >> So what does that look like as a practical matter? I imagine you have customers whose on-premises VMware environments look different than what you've created within Expedient data centers. I'm thinking of things like the level of adoption of NSX, how well a customer may embrace VSAN on-prem as an example. Is part of this transmogrification into your data center, kind of nudging people to adopt frameworks that are really necessary for success in the future? >> It's less of a nudge because a lot of times as a service provider, we don't talk about the technology, we talk more about the outcome. So the nice thing with VMware is we can move that same virtual machine or that container into the platform and the client doesn't always know exactly what's underneath because we have that standardized VMware stack and it just works. And that's part of the beauty of the process. I dunno if you want to talk about a specific client or... >> Yeah, so one of the ones we worked with is Bob Evans Foods. So they were in that transformation stage of refreshing, not only their office space and their data center, but also their VMware environment. So we helped them go through and first thing is looking at their existing environment, figuring out what they currently have, because you can't really make a good decision of what you need to change until you know where you're starting from. So we worked with them through that process, completely evacuated their data center. And from a business perspective, what that allowed them to do as well is have more flexibility in the choice of their next corporate office, because they didn't have to have a data center attached to it. So just from that data center perspective, we gave them some flexibility there. But then from an operations perspective, really standardize that process, offloaded some of those menial tasks that you mentioned earlier, and allow them to really look more towards business-driving projects, instead of just trying to keep those lights on, keeping the backups running, et cetera. >> Brian, question for you, here we are, the theme of the event is "The Center of the Multi-cloud Universe" which seems like a Marvel movie, I haven't seen any new superheroes yet, but I suspect there might be some here. But as customers end up and land in multi-cloud by default not by strategy, how does Expedient and VMware help them actually take the environment that they have and make it strategic so that the business can achieve the outcomes, improving revenue, finding new revenue streams, new products, new routes to market to delight those customers. How do you turn that kind of cloud chaos into a strategy? >> Yeah. I'd say there's a couple different components. One is really time. How can you give them time back for things that are creating noise and aren't really strategic to the business? And so if you can give that time back, that's the first way that you can really impact the business. And the second is through that standardization, but also a lot of times when people think of that new standard, they're only thinking if you're building from scratch. And what VMware has really helped is by taking those existing workloads and giving a standard that works for those applications and what you're building new and brings those together under a common platform and so had a really significant impact to the speed that somebody can get to that cloud operating model, that used to be a multi-year process and most of our clients can go from really everything or almost everything on-prem and a little bit in a cloud to a complete cloud operating model, on average, in four to six months. >> Wow! >> So if I have an on-premises environment and some of my workloads are running in a VMware context, VMware would make the pitch in an agnostic way that, "Well, you can go and deploy that "on top of a stack of infrastructure "and anybody and anywhere now." Why do customers come to you instead of saying, "Oh, we'll go to "pick your flavor of hyper scale cloud provider." What's kind of your superpower? You've mentioned a couple of things, but really hone it in on, why would someone want to go to Expedient? >> Yeah. In a single word, service. I mean, we have a 99% client retention rate and have for well over a decade. So it's really that expertise that wraps around all the different technology so that you're not worried about what's happening and you're not worried about trying to keep the lights on and doing the firefighting. You're really focused on the business. And the other way to, I guess another analogy is, if you think about a lot of the technology and the way people go to cloud, it's like if you got a set of Legos without the box or the instructions. So you can build stuff, it could be cool, but you're not going to get to that end state-- >> Hold on. That's how Legos used to work. Just maybe you're too young to remember a time-- >> You see their sales go up because now you buy a different set for this-- >> I build those sets with my son, but I do it grudgingly. >> Do you ever step on one? >> Of course I do. >> Yeah, there's some pain involved. Same thing happens in the transformation. So when they're buying services from an Expedient, you're buying that box set where you have a picture of what your outcome's going to be, the instructions are there. So you also have confidence that you're going to get to the end outcome much faster than you would if you're trying to assemble everything yourself. (David laughing) >> In my mind, I'm imagining the things that I built with Lego, before there were instructions. >> No death star? >> No. Nothing close with the death star. Definitely something that you would not want your information technology to depend upon. >> Got it. >> Brent, we've seen obviously, it seems like every customer these days, regardless of industry has a cloud first initiative. They have competitors in the rear view mirror who are, if they're able to be more agile and faster to market, are potential huge competitive threat. As we see the rise of multi-cloud in the last 12 months, there's also been a lot of increased analyst coverage for alternate specialty hybrid cloud. Talk to us about, Expedient was in the recent Gartner market guide for specialty cloud. How are these related? What's driving this constant change out in the customer marketplace? >> Sure. So a lot of that agility that clients are getting and trying to do that digital transformation or refactor their applications requires a lot of effort from the developers and the internal IT practitioners. So by moving to a model with an enterprise kind of like Expedient, that allows them to get a consistent foundational level for those technical debt, the 'traditional workloads' where they can start focusing their efforts more on that refactoring of their applications, to get that agility, to get the flexibility, to get the market advantage of time to market with their new refactored applications. That takes them much faster to market, allows them to get ahead of those competitors, if they're not already ahead of them, get further ahead of them or catch up the ones that may have already made that transition. >> And I would add that the analyst coverage you've seen in the last 9 to 12 months, really accelerate for our type of cloud because before everything was hyper scale, everything's going to be hyper scale and they realized that companies have been trying to go to the cloud really for over a decade, really 15 years, that digital transformation, but most companies, when you look at the analysts say they're about 30% there, they've hit a plateau. So they need to look at a different way to approach that. And they're realizing that a VMware-based cloud or the specialty cloud providers give a different mode of cloud. Because you had of a pendulum that everything was on-premises, everything swung to cloud first and then it swung to multi-cloud, which meant multiple hyper scale providers and now it's really landing at that equilibrium where you have different modes of cloud. So it's similar like if you want to travel the world, you don't use one mode of transportation to get from one continent to the other. You have to use different modes. Same thing to get all the way to that cloud transformation, you need to use different modes of cloud, an enterprise cloud, a hyper scale cloud, working them together with that common management plan. >> And with that said Brian, where have customer conversations gone in the last couple of years? Obviously this has got to be an executive level, maybe even a board level conversation. Talk to us about how your customer conversations have changed. Have the stakeholders changed? Has things gone up to stack? >> Yeah. The business is much more involved than what it's been in the past and some of the drivers, even through the pandemic, as people reevaluate office space, a lot of times data centers were part of the same building. Or they were added into a review that nobody ever asked, "Well, why are you only using 20% of your data center?" So now that conversation is very active and they're reevaluating that and then the conversation shifts to "Where's the best place?" And that's a lot of, the conference also talks about the best place for your application for the workload in the right location. >> My role here is to dive down into the weeds constantly to stay away from business outcomes and things like that. But somewhere in the middle there's this question of how what you provide is consumed. So fair to assume that often people are moving from CapEx model to an OPEX model where they're consuming by the glass, by the drink. What does that mean organizationally for your customers? And do you help them work through that journey, reorganizing their internal organization to take advantage of cloud? Is that something that Expedient is a part of, or do you have partners that help them through that? How does that work? >> Yeah. There's some unique things that an enterprise doesn't understand when they think about what they've done on-prem versus a service provider is. There's whole models that they can purchase with us in consumption, not just the physical hardware, but licensing as well. Do you want to talk about how clients actually step in and start to do that evaluation? >> Sure. So it really kind of starts on the front end of evaluating what they have. So going through an assessment process, because traditionally, if you have a big data center full of hardware, you've already paid for it. So as you're deploying new workloads, it's "free to deploy." But when you go to that cloud operating model, you're paying for each drink that you're taking. So we want to make sure that as they're going into that cloud operating model, that they are right sized on the front end. They're not over-provisioned on anything that they're going to just waste money and resources on after they make that transition. So it's really about giving them great data on the front end, doing all that collection from a foundational level, from a infrastructure level, but also from a business and IT operations perspective and figuring out where they're spending, not just their money, but also their time and effort and helping them streamline and simplify those IT operations. >> Let's talk about one of the other elephants in the room and that is the remote hybrid workforce. Obviously it's been two and a half years, which is hard to believe. I think I'm one of the only people that hates working from home. Most people, do you too? Okay, good. Thank you, we're normal. >> Absolutely. (Lisa laughing) But VMware was talking about desktop as a service, there was so much change and quick temporary platform set up to accommodate offsite workers during the pandemic. What are some of the experiences that your clients are having and how is Expedient plus VMware helping businesses adapt and really create them the right hybrid model for them going forward? >> Sure. So as part of being that full sack cloud service provider, desktop in that remote user has to be part of that consideration. And one of the biggest things we saw with the pandemic was people stood up what we call pandemic VDI, very temporary solutions. And you saw the news articles that they said, "We did it in 10 days." And how many big transformational events do people plan and execute in 10 days that transform their workforce? So now they're having to come back and say, "Okay, what's the right way to deploy it?" And do you want to talk about some of the specifics of what we're seeing in the adjustments that they're doing? >> Sure. So it is, when you look at it from the end user perspective, it's how they're operating, how they're getting their tools through their day to day job, but it's also the IT administrators that are having to provide that service to the end users. So it's really kind of across the board, it's affecting everyone. So it's really kind of going through and helping them figure out how they're going to support their users going forward. So we've spun up things like VMware desktop as a service providing that multi-tenant ability to consume on a per desktop basis, but then we've also wrapped around with a lot of security features. So one of the big things is as people are going and distributing where they're working from, that data and access to data is also opened up to those locations. So putting those protections in place to be able to protect the environment and then be able, if something does get in, to be able to detect what's going on. And then of course, with a lot of the other components, being able to recover those environments. So building the desktops, the end user access into the disaster recovery plans. >> And talk more, a little bit Brent, about the security aspect. We've seen the threat landscape change dramatically in the last couple of years, ransomware is a household word. I'm pretty sure even my mom knows what that means, to some degree. Where is that in customer conversations? I can imagine in certain industries like financial services and healthcare with PII, it's absolutely critical to ensure that that data is, they know where it is. It's protected and it's recoverable, 'cause everyone's talking about cyber resilience these days. >> Right. And if it's not conversation 1, it's conversation 1A. So it's really kind of core to everything that we do when we're talking to clients. It's whether it's production DR or the desktops, is building that security in place to help them build their security practice up. So when you think about it, it's doing it at layers. So starting with things like more advanced antivirus to see what's actually going on the desktop and then kind of layering above there. So even up to micro-segmentation, where you can envelop each individual desktop in their own quasi network, so that they're only allowed kind of that zero trust model where, Hey, if you can get to a file share, that's the only place you should be going or do I need web apps to get my day to day job done, but really restricting that access and making sure that everything is more good traffic versus unknown traffic. >> Yeah. >> And also on the, you asked about the clouds smarter earlier. And you can really weave the desktop into that because when you're thinking of your production compute environment and your remote desktop environment, and now you can actually share storage together, you can share security together and you start to get economies of scale across those different environments as well. >> So as we are in August, I think still yeah, 2022, barely for a couple more days, lot of change going on at VMware. Expedient has been VMware America's partner of the year before. Talk to us about some of the things that you think from a strategic perspective are next for the partnership. >> That it's definitely the multi-cloud world is here. And it's how we can go deeper, how we're going to see that really mature. You know, one of the things that we've actually done together this year was we worked on a project and evaluated over 30 different companies of what they spend on IT. Everything from the physical data center to the entire stack, to people and actually build a cloud transformation calculator that allows you to compare strategies, so that if you look at Strategy A over a five year period, doing your current transformation, versus that Cloud Different approach, it can actually help quantify the number of hours difference that you can get, the total cost of ownership and the speed that you can get there. So it's things like that that help people make easier decisions and simplify information are going to be part of it. But without a doubt, it's going to be how you can have that wrapper across all of your different environments that really delivers that cloud-like environment that panacea people have been looking for. >> Yeah. That panacea, that seems like it's critical for every organization to achieve. Last question for you. When customers come to you, when they've hit that plateau. They come to Expedient saying, "Guys, with VMware, help us accelerate past this. "We don't have the time, we need to get this done quickly." How do you advise them to move forward? >> Sure. So it goes back to that, what's causing them to hit that plateau? Is it more on the development side of things? Is it the infrastructure teams, not being able to respond fast enough to the developers? And really putting a plan in place to really get rid of those plateaus. It could be getting rid of the technical debt. It could be changing the IT operations and kind of that, the way that they're looking at a cloud transformation model, to help them kind of get accelerated and get them back on the right path. >> Back on the right path. I think we all want to get back on the right path. Guys, thank you so much for joining David and me on theCUBE today, talking about Expedient Cloud Different, what you're seeing in the marketplace, and how Expedient and VMware are helping customers to succeed. We appreciate your time. >> Yep. >> Thanks for having us. >> For our guests and Dave Nicholson, I'm Lisa Martin. You're watching theCUBE live from VMware Explorer '22, stick around, Dave and I will be back shortly with our next guest. (gentle upbeat music)
SUMMARY :
And Brent Meadows, the Vice President the land of what's going on to get the end result they're looking for. and what it's helping customers to achieve and instead look to in the last couple of years and kind of removing that to get to cloud smart. so that you can get the value out of cloud kind of nudging people to adopt frameworks or that container into the platform and allow them to really look more towards so that the business can that you can really impact the business. Why do customers come to and the way people go to cloud, Just maybe you're too I build those sets with my son, So you also have confidence I'm imagining the things that you would not want agile and faster to market, that allows them to get a and then it swung to multi-cloud, in the last couple of years? and some of the drivers, So fair to assume that and start to do that evaluation? that they're going to just and that is the remote hybrid workforce. What are some of the experiences And one of the biggest things that service to the end users. in the last couple of years, that's the only place you should be going and now you can actually that you think from a and the speed that you can get there. "We don't have the time, we of the technical debt. Back on the right path. with our next guest.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Brian | PERSON | 0.99+ |
Dave Nicholson | PERSON | 0.99+ |
Brent Meadows | PERSON | 0.99+ |
Brian Smith | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
August | DATE | 0.99+ |
20% | QUANTITY | 0.99+ |
Brent | PERSON | 0.99+ |
Expedient | ORGANIZATION | 0.99+ |
15 years | QUANTITY | 0.99+ |
99% | QUANTITY | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Lego | ORGANIZATION | 0.99+ |
Bryan Smith | PERSON | 0.99+ |
two and a half years | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
four | QUANTITY | 0.99+ |
2022 | DATE | 0.99+ |
three years | QUANTITY | 0.99+ |
each drink | QUANTITY | 0.99+ |
VMware America | ORGANIZATION | 0.99+ |
10 days | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
first initiative | QUANTITY | 0.98+ |
six months | QUANTITY | 0.98+ |
Legos | ORGANIZATION | 0.98+ |
Gartner | ORGANIZATION | 0.98+ |
One | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
pandemic | EVENT | 0.98+ |
VMworld | ORGANIZATION | 0.98+ |
today | DATE | 0.97+ |
over a decade | QUANTITY | 0.97+ |
about 30% | QUANTITY | 0.97+ |
this year | DATE | 0.96+ |
first thing | QUANTITY | 0.96+ |
over 30 different companies | QUANTITY | 0.96+ |
one continent | QUANTITY | 0.95+ |
single word | QUANTITY | 0.95+ |
Moscone West | LOCATION | 0.95+ |
first way | QUANTITY | 0.94+ |
Marvel | ORGANIZATION | 0.94+ |
each individual desktop | QUANTITY | 0.94+ |
Bob Evans Foods | ORGANIZATION | 0.94+ |
VMware | TITLE | 0.93+ |
The Center of the Multi | TITLE | 0.93+ |
this morning | DATE | 0.92+ |
Lisa | PERSON | 0.91+ |
last couple of years | DATE | 0.9+ |
VMware Explore 2022 | TITLE | 0.89+ |
12 months | QUANTITY | 0.89+ |
theCUBE | ORGANIZATION | 0.88+ |
last 12 months | DATE | 0.88+ |
Shishir Shrivastava, TEKsystems & Devang Pandya, TEKsystems | Snowflake Summit 2022
>>Welcome back everyone to the Cube's live coverage of snowflake summit 22, we are live in Las Vegas. Caesar's forum, Lisa Martin, Dave Valante, Dave. This is day one of a lot of wall action on the, >>Yeah. A lot of content on day one. It, it feels like, you know, the, the reinvent fire hose yes. Of announcements feels like a little mini version of that. >>It does. That's a good, that's a good way of putting it. We've been unpacking a lot of the news. That's come out, stick around, lots more coming. We've got two guests joining us from tech systems global services. Please welcome Devon. Pania managing director and Shai Sheva of us senior and Shire. Shrivastava senior manager, guys. Great to have you on the cube. >>Thank you so much. Good to see you. And it's great to be in person. Finally, it's been a long UE, so excited to be here. >>Agree. The keynote this morning was not only standing room only, but there was an overflow area. >>Oh my goodness. We have a hard time getting in and it is unbelievable announcement that we have heard looking forward for an exciting time. Next two days here >>Absolutely exciting. The, the cannon shotgun of announcements this morning was amazing. The innovation that has been happening at snowflake and you know, this clearly as partner has been, it just seems like it's the innovation flywheel is getting faster and faster and faster. Talk to us a little bit, Devon about tech systems. Give us the audience a little bit of an overview of the company, and then talk to us about the partnership with snowflake. >>Sure. Thank you. Lisa tech system global services is a full stack global system integrator working with 8% of fortune 500 customers helping in accelerating their business as well as technology modernization journey. We have been a snowflake partner since 2019, and we are one of the highest accredited sales and technical certification with snowflake. And that's what we have earned as a elite partner or sorry, emerging partner with snowflake last year. And we are one of the top elite partner as well. >>Yeah. So since 2019, I mean, in the keynote this morning, Frank showed it. I think Christian showed it as well in terms of the amount of, of change innovation that's happened since 2019 Ellen, we were talking before we went live to share about the, the last two years, the acceleration of innovation cloud adoption digital transformation. The last two years is kind of knock your head back. You need a yeah. A whiplash collar to deal with that. Talk about what you've seen in the last three years, particularly with the partnership and how quickly they are moving and listening to their customers. >>Yeah. Yeah. I think last two years really has given pretty much every organization, including us and our customers a complete different perspective. And that's, that's the exact thing which Christian was talking about, you know, disruption, that's the that's that has been the core message, which we have seen and we've got it from the customers. And we have worked on that right from the get go. We have, you know, all our tools and technology. We are working hand in hand with snowflake in terms of our offerings, working with customers, we have tools. We talk about, you know, accelerators quote unquote that's that helps our customers, you know, to take it from on-prem systems to all the way to the snowflake data cloud and that too, you know, fraction of seconds. You talk about data, you talk about, you know, code conversion, you talk about data validation. So, you know, there are ample amount of things, you know, in terms of, you know, innovation, all workload, I've heard, you know, those are the buzzwords today, and those are like such an exciting time out here. >>So before the pandemic, you know, digital transformation, it was, it was sort of a thing, but it was, it was also a lot of complacency around it. And then of course, if you weren't in a digital business, you were out of the business and boom. So you talked to bang about the stack. You guys obviously do a lot in cloud migration. What's changed in cloud migration. And how is the stack evolving to accommodate that? >>That's a great question there when last two years, it's absolutely a game changer in terms of the digital transformation. Can we believe that 90% of world's data that we have produced and captured is in last two years? It's, isn't that amazing? Right. And what IDC is predicting by 20 25, 200 terabytes of data is going to be generated. And most of them is going to be unstructured. And what we are fascinated about is only 0.5% of unstructured data is currently analyzed by the organization to look at the immense opportunity in front of us and with Snowflake's data cloud, as well as some of the retail data cloud finance and healthcare data cloud launching, it's going to immensely help in processing that unstructured data and really bring life to the data in making organization and market leader. >>Quick, quick fall, if I could, why is, is such a small, why is so much data dark and not accessible to organizations? What's >>The, that's a, that's a great question. I think it's a legacy that we have been trained such a way that data has to be structured. It needs to be modeled, but last decade or so we have seen note it hasn't required that way. And all the social media data being generated, how we communicate in a world is all arm structure, right? We don't create structured data and put it into the CSV and things like that. It's just a natural human behavior. And I think that's where we see a lot of potential in mining that dataset and bringing, you know, AI ML capabilities from descriptive to diagnostic analysis, moving forward with prescriptive and predictive analytics. And that's what we heard from snowflake in Christian announce, Hey, machine learning workload is going to be the key lot of investment happening last 10 years. Now it's going to, you know, capitalize on those ROI in making quick decisions. >>Should you talk to me about those customer conversations? Obviously they have they've transformed and evolved considerably. Yeah. But for customers that have this tremendous amount of unstructured data, a lot of potential as you talked about dung, but there's gotta be, it's gotta be a daunting task. Oh yeah. But these days, every company has to be a data company to be successful, to be competitive and to deliver the experience that the demanding consumers expect. Yeah. How do you start with customers? Where do they start? What's that conversation like and how can tech systems help them get rid of that kind of that daunting iceberg, if you will and get around >>It. Yeah, yeah, yeah, exactly. And I think you got the right point there. Unstructured data is just the tip of the iceberg we are talking about and we have just scratched little surface of it, you know, it's it's and as the one was mentioning earlier, it's, it's gone out those days, you know, where we are talking about, you know, gigabytes of data or, you know, terabytes. Now we are talking about petabytes and Zab bytes of data, and there are so many, and that's, that's the data insight we are looking for and what else, you know, what best platform you can get better than, you know, snowflake data cloud. You have everything in there. You talk about programmability today. You know, Christian was talking about snow park, you know, that, that gives you all the cutting edge languages. You talk about Java, you talk about scale, you talk about Python, you know, all those languages. >>I mean, there were days when these languages, you need to bring that data to a separate platform, process it and then connect it. Now it is right there. You can connect it and just process it. So I think that's, that's the beginning. And to start the conversation, we always, you know, go ahead and talk to the customers and, you know, understand their perspective, know where they want to start, you know, what are their pain points and where they, they want to go, you know, what's their end goal, you know, how they want to pro proceed, you know, how they want to mature in terms of, you know, data agility and flexibility and you know, how do they want to offer their customers? So that's, that's the basically, you know, that's our, the path forward and that's how we see it. >>And just, >>Just to add on top of that, Dave, sorry about that. What we have seen with our customers, the legacy mindset of creating the data silos, primarily because it's not that they wanted it that way, but there were limitations in terms of either the infrastructure or the unlimited scalability and flexibility and accent extensibility, right? That's why those kind of, you know, work around has been built. But with snowflake unified data cloud platform, you have everything in unified platform and what we are telling our customers, we need to eliminate the Datalog. Yes, data is a new oil, but we need to make sure that you eliminate the Datalog within the enterprise, as well as outside the enterprise to really combine then and get a, you know, valuable insight to be the market leader. >>You know, when the cube started, it was 2010. And I remember we went to Hadoop world and it was a lot of excitement around big data and yes, and it turned out, it didn't quite live up to the expectations. That's an understatement, but we, we learned a lot and we made some strides and, and now we're sort of entering this, this new era, but you know, the, the, the last era was largely this big batch job right now, today. You're seeing real time, you know, we've, we've projected out real time in, is gonna become more and more of a thing. How do you guys see the, the sort of data patterns changing and again, where do you see snowflake fitting in? >>Yeah. Great question. And they, what I would have to say, just in a one word is removing the complexity and moving towards the simplicity. Why the legacy solutions such as big data didn't really work out well, it had all the capabilities, but it was a complex environment. You need to really be, you know, knowing a lot of technical aspect of it. And your data analyst were struggling with that kind of a tool set. So with snowflake simplicity, you can bring citizen data scientists, you can bring your data scientists, you can bring your data analysts, all of them under one platform, and they can all mine the data because it's all sitting in the one environment, are >>You seeing organizations change the way they architect their data teams? And specifically, are you seeing a decentralization of data teams or you see, you mentioned citizen data scientists, are you seeing lines of business take more ownership of the data or is it still cuz again, that big data era created this data science role, the data engineering role, the data pipeline, and it was sort of an extension of the sort of EDW. We had a, a few people, maybe one or two experts who knew how to use the system and you build cubes. And it was sort of a, you know, in order of magnitude more complex than that could maybe do more, but are you seeing it being pushed out to the lines of business? >>That's a great question. And I think what we are seeing in the organization today is this time is absolutely both it and business coming together, hand in hand. It's not that, Hey, it, you do this data pipeline work. And then I will analyze this data. And then we'll, you know, share the dashboards to the CEO. We are seeing more and more cohesiveness within the organization in making a path forward in making the decision intelligence very, very rapid. So I think that's a great change. We don't need to operate in silos. I think it's coming together. And I think it's going to create a win-win combination for our >>Customers. Just to add one more point, what the one has mentioned. I think it's the world of data democratization we are talking about, you know, data is available there, insights. We need to pull it out and you know, just give it to every consumer of the organization and they're ready to consume it. They are, they are hungry. They are ready to take it. You know, that's, that's, that's something, you know, we need to look forward for. >>Well, absolutely look forward to it. And as you talked about, there's so much potential it's we see the tip of the iceberg, right? There's so much underneath that guys. I wish we had more time to continue unpacking this, but thank you so much for joining Dave and me on the program, talking about tech systems and snowflake, what you guys are doing together and what you're enabling those end customers to achieve. We appreciate your insights. >>Yeah. Thank you so much. It's an exciting time for us. And we have been, you know, partnering with snowflake on retail data cloud launch, as well as some upcoming opportunity with manufacturing and also the financial competency that we have earned. So I think it's a great time for us ahead in future. So >>Excellent. Lots to come from Texas systems guys. Thank you. We appreciate your time. Thank you. >>Appreciate it. Thank you. Let it snow. I would say let >>It snow, snow. Let it snow. I like that. You're heard of your life from hot Las Vegas for our guests and Dave ante. I'm Lisa Martin. We are live in Las Vegas. It's not snowing. It's very hot here. We're at the snowflake summit, 22 covering that stick around Dave and I will be joined where next guests in just a moment.
SUMMARY :
Welcome back everyone to the Cube's live coverage of snowflake summit 22, It, it feels like, you know, the, the reinvent fire hose yes. Great to have you on the cube. Thank you so much. The keynote this morning was not only standing room only, but there was an overflow area. We have a hard time getting in and it is unbelievable announcement that we have The innovation that has been happening at snowflake and you know, this clearly as partner has been, And we are one of the top elite partner as well. I think Christian showed it as well in terms of the amount of, of change innovation that's happened since that's the exact thing which Christian was talking about, you know, disruption, that's the that's that has been the So before the pandemic, you know, digital transformation, it was, it was sort of a thing, And most of them is going to be unstructured. in mining that dataset and bringing, you know, AI ML capabilities from descriptive a lot of potential as you talked about dung, but there's gotta be, it's gotta be a daunting task. of the iceberg we are talking about and we have just scratched little surface of it, you know, it's it's and as the one was mentioning And to start the conversation, we always, you know, go ahead and talk to the customers and, That's why those kind of, you know, work around has been built. and now we're sort of entering this, this new era, but you know, the, the, the last era was largely this big you know, knowing a lot of technical aspect of it. And it was sort of a, you know, in order of magnitude more And then we'll, you know, share the dashboards to the CEO. We need to pull it out and you know, And as you talked about, there's so much potential it's we see the And we have been, you know, partnering with snowflake on Lots to come from Texas systems guys. Let it snow. We're at the snowflake summit, 22 covering that stick around Dave and I will be
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Shai Sheva | PERSON | 0.99+ |
Dave Valante | PERSON | 0.99+ |
Frank | PERSON | 0.99+ |
2010 | DATE | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
Shishir Shrivastava | PERSON | 0.99+ |
Shire | PERSON | 0.99+ |
Texas | LOCATION | 0.99+ |
two guests | QUANTITY | 0.99+ |
Java | TITLE | 0.99+ |
Python | TITLE | 0.99+ |
Shrivastava | PERSON | 0.99+ |
last year | DATE | 0.99+ |
2019 | DATE | 0.99+ |
Pania | PERSON | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
90% | QUANTITY | 0.99+ |
8% | QUANTITY | 0.99+ |
both | QUANTITY | 0.98+ |
two experts | QUANTITY | 0.98+ |
one word | QUANTITY | 0.98+ |
Devang Pandya | PERSON | 0.98+ |
Caesar | PERSON | 0.98+ |
Ellen | PERSON | 0.98+ |
one platform | QUANTITY | 0.98+ |
Devon | PERSON | 0.98+ |
0.5% | QUANTITY | 0.97+ |
Snowflake Summit 2022 | EVENT | 0.97+ |
today | DATE | 0.97+ |
Christian | ORGANIZATION | 0.96+ |
day one | QUANTITY | 0.96+ |
20 25, 200 terabytes | QUANTITY | 0.94+ |
TEKsystems | ORGANIZATION | 0.92+ |
this morning | DATE | 0.91+ |
pandemic | EVENT | 0.9+ |
one environment | QUANTITY | 0.89+ |
IDC | ORGANIZATION | 0.88+ |
one more point | QUANTITY | 0.87+ |
last two years | DATE | 0.87+ |
last three years | DATE | 0.86+ |
last 10 years | DATE | 0.85+ |
TEK | PERSON | 0.84+ |
last decade | DATE | 0.8+ |
Lisa | PERSON | 0.74+ |
snowflake summit 22 | EVENT | 0.74+ |
fortune 500 customers | QUANTITY | 0.71+ |
Cube | ORGANIZATION | 0.66+ |
two days | QUANTITY | 0.61+ |
highest accredited sales | QUANTITY | 0.59+ |
petabytes | QUANTITY | 0.55+ |
Devon | ORGANIZATION | 0.52+ |
terabytes | QUANTITY | 0.52+ |
22 | QUANTITY | 0.37+ |
Jason Buffington, Veeam | VeeamON 2022
(upbeat music) >> Welcome back to theCUBE's coverage of VEEMON 2022. We're here at the Aria in Las Vegas. Dave Vellante with David Nicholson, my co-host for the week, two days at wall to wall coverage. Jason Buffington is here, JBuff, who does some amazing work for VEEAM, former Analyst from the Enterprise Strategy Group. So he's got a real appreciation for independence data, and we're going to dig into some data. You guys, I got to say, Jason, first of all, welcome back to theCUBE. It's great to see you again. >> Yeah, two and a half years, thanks for having me back. >> Yeah, that's right. (Jason laughs) Seems like a blur. >> No doubt. >> But so here's the thing as analysts, you can appreciate this, the trend is your friend, right? and everybody just inundates you with now, ransomware. It's the trend. So you get everybody's talking about the ransomware, cyber resiliency, immutability, air gaps, et cetera. Okay, great. Technology's there, it's kind of like the NFL, everybody kind of does the same thing. >> There's a lot of wonderful buzzwords in that sentence. >> Absolutely, but what you guys have done that's different is you brought in some big time thought leadership, with data and survey work which of course as an analyst we love, but you drive strategies off of this. So you got to, I'll set it up. You got a new study out that's pivoted off of February study of 3,600 organizations, and then you follow that up with a thousand organizations that actually got hit with ransomware. So tell us more about the study and the work that you've done there. >> Yeah, I got to say I have the best job ever. So I spent seven years as an analyst. And when I decided I didn't want to be an analyst anymore, I called VEEAM and said, I'd like to get in the fight and they let me in. But they let me do independent research on their behalf. So it's kind of like being an in-house counsel. I'm an in-house analyst. And for the beginning of this year, in February, we published a report called the Data Protection Trends Report. And it was over 3000 responses, right? 28 countries around the world looking at digital transformation, the effects of COVID, where are they are on BAS and DRS. But one of the new areas we wanted to look at was how pervasive is ransomware? How does that align with BCDR overall? So some of those just big thought questions that everyone's trying to solve for. And out of that, we said, "Wow, this is really worth double clicking." And so today, actually about an hour ago we published the Ransomware Trends Report and it's a thousand organizations all of which have all been survived. They all had a ransomware attack. One of the things I think I'm most proud of for VEEAM in this particular project, we use an independent research firm. So no one knows it's VEEAM that's asking the questions. We don't have any access to the respondents along the way. I wish we did, right? >> Yeah, I bet >> Go sell 'em back up software. But of the thousands 200 were CISOs, 400 were security professionals which we don't normally interact with, 200 backup admins, 200 IT ops, and the idea was, "Okay, you've all been through a really bad day. Tell us from your four different views, how did that go? What did you solve for? What did you learn? What are you moving forward with?" And so, yeah, some great learnings all around helping us understand how do we deliver solutions that meet their needs? >> I mean, there's just not enough time here to cover all this data. And I think I like about it is, like you said, it's a blind survey. You used an independent third party whom I know they're really good. And you guys are really honest about it. It's like, it was funny that the analyst called today for the analyst meeting when Danny was saying if 54% and Dave Russell was like, it's 52%, actually ended up being 53%. (Jason laughs) So, whereas many companies would say 75%. So anyway, what were some of the more striking findings of that study? Let's get into it a little bit. >> So a couple of the ones that were really startling for me, on average about one in four organizations say they have not been hit. But since we know that ransomware has a gestation for around 200 days from first intrusions, so when you have that attack, 25% may be wrong. That's 25% in best case. Another 16% said they only got hit once in the last year. And that means 60%, right on the money got hit more than once per year. And so when you think about it's like that school bully Once they take your lunch money once and they want lunch money, again, they just come right back again. Did you fix this hole? Did you fix that hole? Cool, payday. And so that was really, really scary. Once they get in, on average organizations said 47% of their production data was encrypted. Think about that. So, and we tested for, hey, was it in the, maybe it's just in the ROBO. So on the edge where the tech isn't as good, or maybe it's in the cloud because it's in a broad attack surface. Whatever it is, turns out, doesn't matter. >> So this isn't just nibbling around the edges. >> No. >> This is going straight to the heart of the enterprise. >> 47% of production data, regardless of where it's stored, data center ROBO or cloud, on average was encrypted. But what I thought was really interesting was when you look at the four personas, the security professional and the backup admin. The person responsible for prevention or mediation, they saw a much higher rate of infection than the CSOs and the IT pros, which I think the meta point there is the closer you are to the problem. the worst this is. 47% is bad. it's worse than that. As you get closer to it. >> The other thing that struck me is that a large proportion of, I think it was a third of the companies that paid ransom. >> Oh yeah. >> Weren't able to recover it. Maybe got the keys and it didn't work or maybe they never got the keys. >> That's crazy too. And I think one thing that a lot of folks, you watch the movies and stuff and you think, "Oh, I'm going to pay the Bitcoin. I'm going to get this magic incantation key and all of a sudden it's like it never happened. That is not how this works. And so yeah. So the question actually was did you pay and did it work right? And so 52%, just at half of organization said, yes. I paid and I was able to recover it. A third of folks, 27%. So a third of those that paid, they paid they cut the check, they did the ransom, whatever, and they still couldn't get back. Almost even money by the way. So 24% paid, but could not get back. 19% did not pay, but recovered from backup. VEEAM's whole job for all of 2022 and 23 needs to be invert that number and help the other 81% say, "No, I didn't pay I just recovered." >> Well, in just a huge number of cases they attacked the backup Corpus. >> Yes. >> I mean, that's was... >> 94% >> 94%? >> 94% of the time, one of the first intrusions is to attempt to get rid of the backup repository. And in two thirds of all cases the back repository is impacted. And so when I describe this, I talk about it this way. The ransomware thief, they're selling a product. They're selling your survivability as a product. And how do you increase the likelihood that you will buy what they're selling? Get rid of the life preserver. Get rid of their only other option 'cause then they got nothing left. So yeah, two thirds, the backup password goes away. That's why VEEAM is so important around cloud and disk and tape, immutable at every level. How we do what we do. >> So what's the answer here. We hear things like immutability. We hear terms like air gap. We heard, which we don't hear often, is orchestrated recovery and automated recovery. I wonder if you could get, I want to come back to... So, okay. So you're differentiating with some thought leadership, that's nice. >> Yep. >> Okay, good. Thank you. The industry thanks you for that free service. But how about product and practices? How does VEEAM differentiate in that regard? >> Sure. Now full disclosure. So when you download that report, for every five or six pages of research, the marketing department is allowed to put in one paragraph. It says, this is our answer. They call the VEEAM perspective. That's their rebuttal. To five pages of research, they get one paragraph, 250 word count and you're done. And so there is actually a commercial... >> We're here to buy here in. (chuckles) >> To the back of that. It's how we pay for the research. >> Everybody sells an onset. (laughs) >> All right. So let's talk about the tech that actually matters though, because there actually are some good insights there. Certainly the first one is immutability. So if you don't have a survivable repository you have no options. And so we provide air gaping, whether you are cloud based. So your favorite hyper-scale or one of the tens of thousands of cloud service providers that offer VEEAM products. So you can have, immutability at the cloud layer. You can certainly have immutability at the object layer on-prem or disk. We're happy to use all your favorite DDoS and then tape. It is hard to get more air-gaped and take the tape out drive, stick it on a shelf or stick it in a white van and have it shipped down the street. So, and the fact that we aren't dependent on any architecture, means choose your favorite cloud, choose your favorite disc, choose your favorite tape and we'll make all of 'em usable and defendable. So that's super key Number one. Super key number two there's three. >> So Platform agnostic essentially. >> Yeah. >> Cloud platform agenda, >> Any cloud, any physical, we work happily with everybody. Just here for your data. So, now you know you have at least a repository, which is not affectable. The next thing is you need to know, do you actually have recoverable data? And that's two different questions. >> How do you know? Right, I mean... >> You don't. So one of my colleagues, Chris Hoff, talks about how you can have this Nalgene bottle that makes sure that no water spills. Do you know that that's water? Is it vodka? Is it poison? You don't know. You just know that nothing's spilling out of it. That's an immutable repository. Then you got to know, can you actually restore the data? And so automating test restores every night, not just did the backup log work. Only 16% actually test their backups. That breaks my heart. That means 84% got it wrong. >> And that's because it just don't have the resource or sometimes testing is dangerous. >> It can be dangerous. It can also just be hard. I mean, how do you spend something up without breaking what's already live. So several years ago, VEEAM created the sandbox is what we call a data lab. And so we create a whole framework for you with a proxy that goes in you can stand up whatever you want. You can, if file exists, you can ping it, you can ODBC SQL, you can map the exchange. I mean, you can, did it actually come up. >> You can actually run water through the recovery pipes. >> Yes. >> And tweak it so that it actually works. >> Exactly. So that's the second thing. And only 16% of organizations do. >> Wow. >> And then the third thing is orchestration. So there's a lot of complexity that happens when you recover one workload. There is a stupid amount of complexity happens when you try cover a whole site or old system, or I don't know, 47% of your infrastructure. And so what can you do to orchestrate that to remediate that time? Those are the three things we found. >> So, and that orchestration piece, a number of customers that were in the survey were trying to recover manually. Which is a formula for failure. A number of, I think the largest percentage were scripts which I want you to explain why scripts are problematic. And then there was a portion that was actually doing it right. Maybe it was bigger, maybe it was a quarter that was doing orchestrated recovery. But talk about why scripts are not the right approach. >> So there were two numbers in there. So there was 16% test the ability to recover, 25% use orchestration as part of the recovery process. And so the problem where it is, is that okay, if I'm doing it manually, think about, okay, I've stood back up these databases. Now I have to reconnect the apps. Now I have to re IP. I mean, there's lots of stuff to stand up any given application. Scripts says, "Hey, I'm going to write those steps down." But we all know that, that IT and infrastructure is a living breathing thing. And so those scripts are good for about the day after you put the application in, and after that they start to gather dust pretty quick. The thing about orchestration is, if you only have a script, it's as frequently as you run the script that's all you know. But if you do a workflow, have it run the workflow every night, every week, every month. Test it the same way. That's why that's such a key to success. And for us that's VEEAM disaster recovery orchestra tour. That's a product that orchestrates all the stuff that VEEAM users know and love about our backend recovery engine. >> So imagine you're, you are an Excel user, you're using macros. And I got to go in here, click on that, doing this, sort of watching you and it repeats that, but then something changes. New data or new compliance issue, whatever... >> That got renamed directly. >> So you're going to have to go in and manually change that. How do you, what's the technology behind automated orchestration? What's the magic there? >> The magic is a product that we call orchestrator. And so it actually takes all of those steps and you actually define each step along the way. You define the IP addresses. You define the paths. You define where it's going to go. And then it runs the job in test mode every night, every week, whatever. And so if there's a problem with any step along the way, it gives you the report. Fix those things before you need it. That's the power of orchestrator. >> So what are you guys doing with this study? What can we expect? >> So the report came out today. In a couple weeks, we'll release regional versions of the same data. The reason that we survey at scale is because we want to know what's different in a PJ versus the Americas versus Europe and all those different personas. So we'll be releasing regional versions of the data along the way. And then we'll enable road shows and events and all the other stuff that happens and our partners get it so they can use it for consulting, et cetera. >> So you saw differences in persona. In terms of their perception, the closer you were to the problem, the more obvious it was, did you have enough end to discern its pearly? I know that's why you're due the drill downs but did you sense any preliminary data you can share on regions as West getting hit harder or? >> So attack rate's actually pretty consistent. Especially because so many criminals now use ransomware as a service. I mean, you're standing it up and you're spreading wide and you're seeing what hits. Where we actually saw pretty distinct geographic problems is the cloud is not of as available in all segments. Expertise around preventative measures and remediation is not available in all segments, in all regions. And so really geographic split and segment split and the lack of expertise in some of the more advanced technologies you want to use, that's really where things break down. Common attack plane, uncommon disadvantage in recovery. >> Great stuff. I want to dig in more. I probably have a few more questions if you don't mind, I can email you or give you a call. It's Jason Buffington. Thanks so much for coming on theCUBE. >> Thanks for having me. >> All right, keep it right there. You're watching theCUBE's live coverage of VEEAMON 2022. We're here in person in Las Vegas, huge hybrid audience. Keep it right there, be right back. (upbeat music)
SUMMARY :
It's great to see you again. Yeah, two and a half years, Yeah, that's right. But so here's the thing as analysts, buzzwords in that sentence. and the work that you've done there. And for the beginning of But of the thousands 200 were CISOs, And you guys are really honest about it. So a couple of the ones that nibbling around the edges. straight to the heart of the enterprise. is the closer you are to the problem. is that a large proportion of, Maybe got the keys and it didn't work So the question actually was Well, in just a huge number of cases And how do you increase the likelihood I wonder if you could get, The industry thanks you So when you download that report, We're here to buy here in. To the back of that. So, and the fact that we aren't dependent The next thing is you need to know, How do you know? not just did the backup log work. just don't have the resource And so we create a whole framework for you You can actually run water So that's the second thing. And so what can you do to orchestrate that are not the right approach. And so the problem where it is, And I got to go in here, What's the magic there? and you actually define So the report came out today. the closer you were to the problem, and the lack of expertise I can email you or give you a call. live coverage of VEEAMON 2022.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jason | PERSON | 0.99+ |
Dave Russell | PERSON | 0.99+ |
Danny | PERSON | 0.99+ |
David Nicholson | PERSON | 0.99+ |
Chris Hoff | PERSON | 0.99+ |
Jason Buffington | PERSON | 0.99+ |
JBuff | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
25% | QUANTITY | 0.99+ |
February | DATE | 0.99+ |
16% | QUANTITY | 0.99+ |
seven years | QUANTITY | 0.99+ |
3,600 organizations | QUANTITY | 0.99+ |
five pages | QUANTITY | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
47% | QUANTITY | 0.99+ |
Excel | TITLE | 0.99+ |
84% | QUANTITY | 0.99+ |
54% | QUANTITY | 0.99+ |
75% | QUANTITY | 0.99+ |
53% | QUANTITY | 0.99+ |
52% | QUANTITY | 0.99+ |
two numbers | QUANTITY | 0.99+ |
24% | QUANTITY | 0.99+ |
one paragraph | QUANTITY | 0.99+ |
60% | QUANTITY | 0.99+ |
27% | QUANTITY | 0.99+ |
six pages | QUANTITY | 0.99+ |
19% | QUANTITY | 0.99+ |
VEEAM | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
Data Protection Trends Report | TITLE | 0.99+ |
two days | QUANTITY | 0.99+ |
Europe | LOCATION | 0.99+ |
81% | QUANTITY | 0.99+ |
four personas | QUANTITY | 0.99+ |
over 3000 responses | QUANTITY | 0.99+ |
200 backup admins | QUANTITY | 0.99+ |
250 word | QUANTITY | 0.99+ |
each step | QUANTITY | 0.99+ |
2022 | DATE | 0.99+ |
28 countries | QUANTITY | 0.98+ |
DRS. | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
two different questions | QUANTITY | 0.98+ |
third thing | QUANTITY | 0.98+ |
two thirds | QUANTITY | 0.98+ |
two and a half years | QUANTITY | 0.98+ |
second thing | QUANTITY | 0.98+ |
Americas | LOCATION | 0.98+ |
94% | QUANTITY | 0.98+ |
several years ago | DATE | 0.97+ |
Enterprise Strategy Group | ORGANIZATION | 0.97+ |
three | QUANTITY | 0.97+ |
first one | QUANTITY | 0.97+ |
Ransomware Trends Report | TITLE | 0.97+ |
thousands | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.97+ |
last year | DATE | 0.96+ |
One | QUANTITY | 0.96+ |
BAS | ORGANIZATION | 0.96+ |
around 200 days | QUANTITY | 0.96+ |
COVID | OTHER | 0.95+ |
200 IT ops | QUANTITY | 0.95+ |
third | QUANTITY | 0.94+ |
four organizations | QUANTITY | 0.94+ |
NFL | ORGANIZATION | 0.94+ |
400 | QUANTITY | 0.94+ |
about an hour ago | DATE | 0.94+ |
four different views | QUANTITY | 0.94+ |
first intrusions | QUANTITY | 0.93+ |
once | QUANTITY | 0.93+ |
ROBO | ORGANIZATION | 0.92+ |
Webb Brown, Kubecost | CUBE Conversation
>>Welcome to this cube conversation. I'm Dave Nicholson, and this is part of the AWS startup showcase season two. I'm very happy to have with me Webb brown CEO of Qube cost web. Welcome to the program. How are you? I'm doing >>Great. It's great to be here, Dave. Thank you so much for having me really excited for the discussion. >>Good to see you. I guess we saw each other last down in Los Angeles for, for coop con, >>Right? Exactly. Right. Still feeling the energy from that event. Hoping we can be back together in person. Not, not too long from now. >>Yeah. Well I'll second that, well, let, let's get straight to it. Tell us, tell us about Q cost. What do you guys do? And I think just central to that question is what gives you guys the right to exist? What problem are you solving? >>Yeah, I love the question. So first and foremost coupe costs, we provide cost monitoring and cost management solutions for teams running Kubernetes or cloud native workloads. Everything we do is, is built on open source. Our founding team was working on infrastructure monitoring solutions at Google before this. And, and what we saw was as we had several teammates join the Kubernetes effort very early days at Google, we saw teams really struggling even just to, to monitor and understand Kubernetes costs, right? There's lots of complexity with the Kubernetes scheduler and being able to answer the question of what is the cost of an application or what is the cost of, you know, a team department, et cetera. And the workloads that they're deploying was really hard for most teams. If you look at CNCF study from late last year, still today, about two thirds of teams, can't answer where they are spending money. And what we saw when digging in there is that when you can't answer that question, it's really hard to be efficient. And by be efficient, we, we mean get the right balance between cost and performance and reliability. So we help teams in, in these areas and more where, you know, now have thousands of teams using our product. You know, we feel where we're just getting started on our mission as well. >>So when people hear it, when people think of coop costs, they w they naturally associate that with Kubernetes. And they think, well, Kubernetes is open-source wait, isn't that free? So what, so what costs are you tracking? Exactly. >>Yeah. Great question. We would track costs in any environment where you can run Kubernetes. So if that's on-prem, you can bring a custom pricing sheet to monitor, say the cost of your underlying CPU course, you know, GPU's memory, et cetera. If you're running in a cloud environment, we have integrations with Azure, GCP and AWS, where we would be able to reflect all the complexity of, you know, whatever deployment you have, whether you're using a spot and multiple regions where you have complex enterprise discounts are eyes savings plans, you name it, we'd be reflecting it. So it's really about, you know, not just generic prices, it's about getting the right price for your organization. >>So the infrastructure that goes into this calculation can be on premises or off premises in the form of cloud. I heard that, right? >>Yeah, that's exactly right. So all of those environments, we'd give you a visibility into all the resources that your Kubernetes clusters are consuming. Again, that's, you know, nodes, load balancers, every resource that it's directly touching also have the ability for you to pull in external costs, right? So if you have Kubernetes tenants that are using S3 or cloud sequel, or, you know, another external cloud service, we would make that connection for you. And then lastly, if you have shared costs, sometimes even like the cost of a dev ops team, we'd give you the ability to kind of allocate that back to your core infrastructure, which may be used for showback or even charged back across your, your, >>So who are the folks in an organization that are tapping into this, are these, you know, our, our, our, our developers being encouraged to be cognizant of these costs throughout the process, or is this just sort of a CFO on down visibility tool? >>Yeah, it's a great, it's a great question. And what we see is a major transformation here where, you know, kind of shift left from a cost perspective where more and more engineering teams are interested in just being aware or having transparency. So they can build a culture of accountability with costs, right, with the amazing ability to rapidly push to production and iterate, you know, with microservices and Kubernetes, it's hard to have this kind of, you know, just wait for say the finance team to review this at the end of the month or the end of the quarter. We see this increasingly be being viewed in real time by infrastructure teams, by engineering teams. Now finance is still a very important stakeholder and, you know, absolutely has a very important like seat at the table in these conversations. But increasingly these are, again, real time or near real time engineering decisions that are really moving the needle on cost and cost efficiency, overtime and performance as well. >>Now, can you use this to model what costs might be, or is this, or is this, you know, you, you mentioned monitoring in real time, is this only for pulling information as it exists, or could you do, could you use some of the aspects of, of, of your toolset to make a decision, whether something makes more sense to run on your existing infrastructure on premises versus moving into, you know, working in a cloud? Is that something that is designed for or not? >>Great question. So we do have the ability to predict cost cost going forward, based on everything we've learned about your environment, whether you're in multi-cloud hybrid cloud, et cetera. So some really interesting functionality there and a lot more coming later this year, because we do see more and more teams wanting to model the state of the future, right? As you deploy really complex technologies, like say the cluster auto scale or, or HPA in different environments, it can really challenging to do an apples to apples comparison, and we help teams do exactly that. And again, gonna have a lot more interesting announcements here later this year. >>So later that later this year, meaning not in the next few minutes while we're together, >>Nothing new to announce on that front today, but I would say, you know, expect later this quarter for us to have more. >>Okay, that sounds good. Now, now you touched on this a little bit, but I want to hone in on why this is particularly relevant now and moving into the future. You know, we've always tracking costs has always been important, you know, even before the Dawn of cloud, but why is it increasingly important? And, and, you know, there are, there are alternatives for cost tracking legacy alternatives that are out there. So talk about why it's particularly relevant now and tell us what your super power is. You know, what's the, all right. All right. >>Secrets, >>Secret sauce is something you can't share super power. You can talk about >>Absolutely >>NDA. So yes, >>Your superpower. Yeah. Great questions. So for support, just to, to, to touch on, what's fundamentally changing to make a company like ours, you know, impactful or relevant. There's really three things here first and foremost is the new abstractions or complexities that come with Kubernetes, right. Super powerful, but from a cost standpoint, make it considerably harder to accurately track costs. And the big transformation here is, you know, with Kubernetes, you can, at any given moment have 50 applications running on a single node or a single VM, you can fast forward five minutes and there could be 50 entirely new applications, right? So just assigning that VM or, you know, tagging that VM back to an application or team or department really is not relevant in those places. So just the new complexity related to costs makes this problem harder for teams. Second is what we touch on. >>Just again, the power of Cooney. Kubernetes is the ability to allow distributed engineering teams to work on many microservices concurrently. So you're no longer in a lot of ways managing this problem where they centralized kind of single point of decision-making. Oftentimes these decisions are distributed across not only your infrastructure team, but your engineering team. So just the way these decisions and, you know, innovation is happening is changing how you manage these. And lastly, it's just scale, right? The, the cloud and, you know, Kubernetes continue to be incredibly successful. You know, where as goop costs now managing billions of dollars as these numbers get bigger and bigger just becomes more of a business focus and business critical issue. So those are the, you know, the three kind of underlying themes that are changing. When I talk about what we do, that makes us special. It's really this like foundational layer of visibility that we build. >>And what we can do is in real time with a very high degree of accuracy at the largest Kubernetes clusters in the world, give you visibility at any dimension. And so from there, you can do things like have real-time monitoring. You can have real-time insights, you can allow automation to make decisions on these, you know, inputs or data feeds. You can set alerts, you can set recurring reports. All of these things are made possible because of, you know, the, the, I would say really hard work that we've done to, again, give this real-time visibility with a high degree of accuracy at, at crazy scale. >>So if we were to play little make-believe for a moment, pretend like I'm a skeptical sitting on the fence. Not sure if I want to go down this path kind of person. And I say, you know what, web, I think I have a really good handle on all of my costs so far. What would you hit me with as, as, as an example of something that people really didn't expect until they, until they were running coup costs and they had actually had that visibility, what are some of the things that people are surprised by? >>Yeah. Great question. There'd be a number, number one. I'd have, you know, one data point I want to get from you, which is, you know, for your organization or for all of your clusters, what is your cost efficiency? Can you answer that with a high degree of accuracy and by cost efficiency? >>And the answer is now. So tell me, tell me, tell me how to sign up for coupons. >>Yeah. And so the answer, the answer there is you can go get our community version, you know, you can be up and running in minutes, you don't have to share any data, right? Like it is, you know, simply a helmet install, but cost efficiency is this notion of, of every dollar that you are spending on provision resources. What percentage of those dollars are you actually utilizing? And we have, you know, we, we now have, you know, thousands of teams using our product and we've worked with, you know, hundreds of them really closely, you know, this is, you know, that's not the entire market, but in our large sample sizes, we regularly see teams start in the low 20% cost efficiency, meaning that approximately 80% is quote waste time and time. Again, we see teams just be shocked by this number. And again, most of it is not because they were measuring it and accurately or anything like that. Most teams again today still just don't have that visibility until they start working with this. >>So is that, is that sort of the, I in my house household, certain members seem to only believe that there is one position for a light switch, and that would be the on position. Is there, is this a bit of a parallel where, where folks are, are spinning up resources and then just out of sight, out of mind, maybe not spinning them down when not needed. Yeah. >>Yeah. It's, it's, that's definitely one class of the challenges I would say, you know, so today, if you look at our product, we have 14 different insights across like different dimensions of your infrastructure one, or, or I would say several of those relate to exactly what you just described, which is you spin up a VM, you spend a bit load balancer, you spin up an external IP address. You're using it. You're not paying for it. Another class is this notion of, again, I don't have an understanding of what my resources cost. I also don't have a great sense for how much my microservice or application will need. So I'm just going to turn on all the lights, which is, or I'm going to drastically over provision again, I don't know the cost, so I'm just going to kind of set it and forget it. And if my application is performing, you know, then you know, we're doing well here. Again, with this visibility, you can get much more specific, much more accurate, much more actionable with making that trade off, you know, again, down to the individual pod workload, you know, deployment, et cetera. >>So we've, we've touched on this a bit peripherally, but give me an example. You know, you, you run into someone who happens to be a happy user of coop cost. What's the dream story that you love to hear from them about what life was before was before coop costs and what life was like after? >>Yeah, there's a lot, a lot of different dimensions there. You know, one, one is, you know, working with an infrastructure team that, that used to get asked these questions a lot about, you know, why does this cost so much, or why are we spending this and Kubernetes or, or wire expenses growing the rate that they are, you know, like when this, when this works, you know, engineering teams or infrastructure teams, aren't getting asked those questions, right? The tool could cost itself is getting asked that and answering that. So I think one is infrastructure teams, not fielding those types of questions as much. Secondly, is just, you know, more and more teams rolling this out throughout their organization. And ultimately just getting, building a culture of awareness, like ownership, accountability. And then, you know, we just increasingly are seeing teams, you know, find this right balance between cost and performance again. So, you know, in certain cases, improving performance, when are resource bottlenecks in places and other places, you know, reducing costs, you know, by 10 plus million dollars, ultimately at the end of the day, we like to see just teams being more comfortable running their workloads in Kubernetes, right? That is the ultimate sign of success is just an organization, feels comfortable with how they're deploying, how they're managing, how they're spending in Kubernetes. Again, whether that be, you know, on-prem or transitioning from on-prem to a cloud in multiple clouds, et cetera. >>So we're talking to you today as part of the second season of the AWS startup showcase. What's, what's the relationship there with, with AWS? >>So it is the, the largest platform for coop costs being run today. So I believe, you know, at this point, at least a thousand different organizations running our product on AWS hosted clusters, whether they're, you know, ETS or, or self-managed, but you know, a growing number of those on, on EKS. And, you know, we've just, you know, absolutely loved working with the team across, I think at this point, you know, six or seven different groups from marketplace to their containers team, you know, obviously, you know, ETS and others, and just very much see them continuing to push the boundaries on what's possible from a scale and, you know, ease of use and, you know, just breadth of, of offering to this market. >>Well, we really look forward to having you back and hearing about some of these announcements, things that are, that are coming down the line. So we'll definitely have to touch base in the future, but just one, one final, more general question for you, where do you see Kubernetes in general going in 2022? Is it sort of a linear growth? Is there some, is there an inflection point that we see, you know, a good percentage of software that's running enterprises right now is already in that open source category, but what are your thoughts on Kubernetes in 2022? >>Yeah, I think, you know, the one word is everywhere is where I see Kubernetes in 2022, like very deep in the like large and really complex enterprises. Right. So I think you'll see just, you know, major bets there. And I think you'll continue to see more engineers adopted. And I think you'll also continue to see, you know, more and more flavors of it, right? So, you know, some teams find that running Kubernetes anymore serverless fashion is, is right for them. Others find that, you know, having full control, you know, at every part of the stack, including running their own autoscaler for example is really powerful. So I think just, you know, you'll see more and more options. And again, I think teams increasingly adopting the right, you know, abstraction level on top of Kubernetes that works for their workloads and their organizations >>Sounds good. We'll we'll, we'll come back in 2023 and we'll check and see how that, how that all panned out. Well, it's been great talking to you today as part of the startup showcase. Really appreciate it. Great to see you again. It's right about the time where I can still tell you happy new year, because we're still, we're still in January here. Hope you have a great 2022 with that from me, Dave Nicholson, part of the cube part of AWS startup showcase season two, I'd like to thank everyone for joining and stay with us for the best in hybrid tech coverage.
SUMMARY :
I'm Dave Nicholson, and this is part of the AWS startup showcase Thank you so much for having me really excited for the discussion. Good to see you. Still feeling the energy from that event. And I think just central to that question is what gives you guys in, in these areas and more where, you know, now have thousands of teams using our so what costs are you tracking? all the complexity of, you know, whatever deployment you have, whether you're using a spot So the infrastructure that goes into this calculation can be on premises or cloud sequel, or, you know, another external cloud service, we would make that connection this kind of, you know, just wait for say the finance team to review this at the end of As you deploy really say, you know, expect later this quarter for us to have more. we've always tracking costs has always been important, you know, even before the Dawn of cloud, Secret sauce is something you can't share super power. So yes, So just assigning that VM or, you know, tagging that VM The, the cloud and, you know, Kubernetes continue to be incredibly decisions on these, you know, inputs or data feeds. And I say, you know what, web, I think I have a really good handle you know, one data point I want to get from you, which is, you know, for your organization So tell me, tell me, tell me how to sign up for coupons. you know, hundreds of them really closely, you know, this is, So is that, is that sort of the, I in my house And if my application is performing, you know, then you know, What's the dream story that you love to hear from them about what And then, you know, we just increasingly So we're talking to you today as part of the second season of the AWS startup So I believe, you know, at this point, at least a thousand we see, you know, a good percentage of software that's running enterprises right now is already in that open source So I think just, you know, you'll see more and more options. Well, it's been great talking to you today as part of the startup showcase.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Nicholson | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
six | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
50 | QUANTITY | 0.99+ |
50 applications | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
January | DATE | 0.99+ |
2022 | DATE | 0.99+ |
Webb Brown | PERSON | 0.99+ |
five minutes | QUANTITY | 0.99+ |
hundreds | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
2023 | DATE | 0.99+ |
one | QUANTITY | 0.99+ |
Los Angeles | LOCATION | 0.99+ |
today | DATE | 0.99+ |
late last year | DATE | 0.99+ |
10 plus million dollars | QUANTITY | 0.99+ |
ETS | ORGANIZATION | 0.99+ |
later this year | DATE | 0.99+ |
S3 | TITLE | 0.99+ |
billions of dollars | QUANTITY | 0.98+ |
Kubernetes | TITLE | 0.98+ |
approximately 80% | QUANTITY | 0.98+ |
EKS | ORGANIZATION | 0.98+ |
14 different insights | QUANTITY | 0.98+ |
three things | QUANTITY | 0.98+ |
20% | QUANTITY | 0.98+ |
CNCF | ORGANIZATION | 0.98+ |
second season | QUANTITY | 0.97+ |
apples | ORGANIZATION | 0.97+ |
first | QUANTITY | 0.97+ |
thousands of teams | QUANTITY | 0.96+ |
one word | QUANTITY | 0.96+ |
Secondly | QUANTITY | 0.95+ |
later this quarter | DATE | 0.95+ |
one position | QUANTITY | 0.94+ |
Kubecost | PERSON | 0.93+ |
Webb brown | PERSON | 0.92+ |
seven different groups | QUANTITY | 0.92+ |
single point | QUANTITY | 0.91+ |
Kubernetes | PERSON | 0.91+ |
season two | QUANTITY | 0.9+ |
about two thirds | QUANTITY | 0.9+ |
one class | QUANTITY | 0.9+ |
three kind | QUANTITY | 0.89+ |
Kubernetes | ORGANIZATION | 0.87+ |
new applications | QUANTITY | 0.84+ |
Qube cost web | ORGANIZATION | 0.83+ |
single node | QUANTITY | 0.83+ |
end | DATE | 0.82+ |
Q cost | ORGANIZATION | 0.82+ |
single VM | QUANTITY | 0.81+ |
Azure | TITLE | 0.77+ |
HPA | ORGANIZATION | 0.73+ |
GCP | TITLE | 0.7+ |
new year | EVENT | 0.65+ |
Analyst Predictions 2022: The Future of Data Management
[Music] in the 2010s organizations became keenly aware that data would become the key ingredient in driving competitive advantage differentiation and growth but to this day putting data to work remains a difficult challenge for many if not most organizations now as the cloud matures it has become a game changer for data practitioners by making cheap storage and massive processing power readily accessible we've also seen better tooling in the form of data workflows streaming machine intelligence ai developer tools security observability automation new databases and the like these innovations they accelerate data proficiency but at the same time they had complexity for practitioners data lakes data hubs data warehouses data marts data fabrics data meshes data catalogs data oceans are forming they're evolving and exploding onto the scene so in an effort to bring perspective to the sea of optionality we've brought together the brightest minds in the data analyst community to discuss how data management is morphing and what practitioners should expect in 2022 and beyond hello everyone my name is dave vellante with the cube and i'd like to welcome you to a special cube presentation analyst predictions 2022 the future of data management we've gathered six of the best analysts in data and data management who are going to present and discuss their top predictions and trends for 2022 in the first half of this decade let me introduce our six power panelists sanjeev mohan is former gartner analyst and principal at sanjamo tony bear is principal at db insight carl olufsen is well-known research vice president with idc dave meninger is senior vice president and research director at ventana research brad shimon chief analyst at ai platforms analytics and data management at omnia and doug henschen vice president and principal analyst at constellation research gentlemen welcome to the program and thanks for coming on thecube today great to be here thank you all right here's the format we're going to use i as moderator are going to call on each analyst separately who then will deliver their prediction or mega trend and then in the interest of time management and pace two analysts will have the opportunity to comment if we have more time we'll elongate it but let's get started right away sanjeev mohan please kick it off you want to talk about governance go ahead sir thank you dave i i believe that data governance which we've been talking about for many years is now not only going to be mainstream it's going to be table stakes and all the things that you mentioned you know with data oceans data lakes lake houses data fabric meshes the common glue is metadata if we don't understand what data we have and we are governing it there is no way we can manage it so we saw informatica when public last year after a hiatus of six years i've i'm predicting that this year we see some more companies go public uh my bet is on colibra most likely and maybe alation we'll see go public this year we we i'm also predicting that the scope of data governance is going to expand beyond just data it's not just data and reports we are going to see more transformations like spark jaws python even airflow we're going to see more of streaming data so from kafka schema registry for example we will see ai models become part of this whole governance suite so the governance suite is going to be very comprehensive very detailed lineage impact analysis and then even expand into data quality we already seen that happen with some of the tools where they are buying these smaller companies and bringing in data quality monitoring and integrating it with metadata management data catalogs also data access governance so these so what we are going to see is that once the data governance platforms become the key entry point into these modern architectures i'm predicting that the usage the number of users of a data catalog is going to exceed that of a bi tool that will take time and we already seen that that trajectory right now if you look at bi tools i would say there are 100 users to a bi tool to one data catalog and i i see that evening out over a period of time and at some point data catalogs will really become you know the main way for us to access data data catalog will help us visualize data but if we want to do more in-depth analysis it'll be the jumping-off point into the bi tool the data science tool and and that is that is the journey i see for the data governance products excellent thank you some comments maybe maybe doug a lot a lot of things to weigh in on there maybe you could comment yeah sanjeev i think you're spot on a lot of the trends uh the one disagreement i think it's it's really still far from mainstream as you say we've been talking about this for years it's like god motherhood apple pie everyone agrees it's important but too few organizations are really practicing good governance because it's hard and because the incentives have been lacking i think one thing that deserves uh mention in this context is uh esg mandates and guidelines these are environmental social and governance regs and guidelines we've seen the environmental rags and guidelines imposed in industries particularly the carbon intensive industries we've seen the social mandates particularly diversity imposed on suppliers by companies that are leading on this topic we've seen governance guidelines now being imposed by banks and investors so these esgs are presenting new carrots and sticks and it's going to demand more solid data it's going to demand more detailed reporting and solid reporting tighter governance but we're still far from mainstream adoption we have a lot of uh you know best of breed niche players in the space i think the signs that it's going to be more mainstream are starting with things like azure purview google dataplex the big cloud platform uh players seem to be uh upping the ante and and addressing starting to address governance excellent thank you doug brad i wonder if you could chime in as well yeah i would love to be a believer in data catalogs um but uh to doug's point i think that it's going to take some more pressure for for that to happen i recall metadata being something every enterprise thought they were going to get under control when we were working on service oriented architecture back in the 90s and that didn't happen quite the way we we anticipated and and uh to sanjeev's point it's because it is really complex and really difficult to do my hope is that you know we won't sort of uh how do we put this fade out into this nebulous nebula of uh domain catalogs that are specific to individual use cases like purview for getting data quality right or like data governance and cyber security and instead we have some tooling that can actually be adaptive to gather metadata to create something i know is important to you sanjeev and that is this idea of observability if you can get enough metadata without moving your data around but understanding the entirety of a system that's running on this data you can do a lot to help with with the governance that doug is talking about so so i just want to add that you know data governance like many other initiatives did not succeed even ai went into an ai window but that's a different topic but a lot of these things did not succeed because to your point the incentives were not there i i remember when starbucks oxley had come into the scene if if a bank did not do service obviously they were very happy to a million dollar fine that was like you know pocket change for them instead of doing the right thing but i think the stakes are much higher now with gdpr uh the floodgates open now you know california you know has ccpa but even ccpa is being outdated with cpra which is much more gdpr like so we are very rapidly entering a space where every pretty much every major country in the world is coming up with its own uh compliance regulatory requirements data residence is becoming really important and and i i think we are going to reach a stage where uh it won't be optional anymore so whether we like it or not and i think the reason data catalogs were not successful in the past is because we did not have the right focus on adoption we were focused on features and these features were disconnected very hard for business to stop these are built by it people for it departments to to take a look at technical metadata not business metadata today the tables have turned cdo's are driving this uh initiative uh regulatory compliances are beating down hard so i think the time might be right yeah so guys we have to move on here and uh but there's some some real meat on the bone here sanjeev i like the fact that you late you called out calibra and alation so we can look back a year from now and say okay he made the call he stuck it and then the ratio of bi tools the data catalogs that's another sort of measurement that we can we can take even though some skepticism there that's something that we can watch and i wonder if someday if we'll have more metadata than data but i want to move to tony baer you want to talk about data mesh and speaking you know coming off of governance i mean wow you know the whole concept of data mesh is decentralized data and then governance becomes you know a nightmare there but take it away tony we'll put it this way um data mesh you know the the idea at least is proposed by thoughtworks um you know basically was unleashed a couple years ago and the press has been almost uniformly almost uncritical um a good reason for that is for all the problems that basically that sanjeev and doug and brad were just you know we're just speaking about which is that we have all this data out there and we don't know what to do about it um now that's not a new problem that was a problem we had enterprise data warehouses it was a problem when we had our hadoop data clusters it's even more of a problem now the data's out in the cloud where the data is not only your data like is not only s3 it's all over the place and it's also including streaming which i know we'll be talking about later so the data mesh was a response to that the idea of that we need to debate you know who are the folks that really know best about governance is the domain experts so it was basically data mesh was an architectural pattern and a process my prediction for this year is that data mesh is going to hit cold hard reality because if you if you do a google search um basically the the published work the articles and databases have been largely you know pretty uncritical um so far you know that you know basically learning is basically being a very revolutionary new idea i don't think it's that revolutionary because we've talked about ideas like this brad and i you and i met years ago when we were talking about so and decentralizing all of us was at the application level now we're talking about at the data level and now we have microservices so there's this thought of oh if we manage if we're apps in cloud native through microservices why don't we think of data in the same way um my sense this year is that you know this and this has been a very active search if you look at google search trends is that now companies are going to you know enterprises are going to look at this seriously and as they look at seriously it's going to attract its first real hard scrutiny it's going to attract its first backlash that's not necessarily a bad thing it means that it's being taken seriously um the reason why i think that that uh that it will you'll start to see basically the cold hard light of day shine on data mesh is that it's still a work in progress you know this idea is basically a couple years old and there's still some pretty major gaps um the biggest gap is in is in the area of federated governance now federated governance itself is not a new issue uh federated governance position we're trying to figure out like how can we basically strike the balance between getting let's say you know between basically consistent enterprise policy consistent enterprise governance but yet the groups that understand the data know how to basically you know that you know how do we basically sort of balance the two there's a huge there's a huge gap there in practice and knowledge um also to a lesser extent there's a technology gap which is basically in the self-service technologies that will help teams essentially govern data you know basically through the full life cycle from developed from selecting the data from you know building the other pipelines from determining your access control determining looking at quality looking at basically whether data is fresh or whether or not it's trending of course so my predictions is that it will really receive the first harsh scrutiny this year you are going to see some organization enterprises declare premature victory when they've uh when they build some federated query implementations you're going to see vendors start to data mesh wash their products anybody in the data management space they're going to say that whether it's basically a pipelining tool whether it's basically elt whether it's a catalog um or confederated query tool they're all going to be like you know basically promoting the fact of how they support this hopefully nobody is going to call themselves a data mesh tool because data mesh is not a technology we're going to see one other thing come out of this and this harks back to the metadata that sanji was talking about and the catalogs that he was talking about which is that there's going to be a new focus on every renewed focus on metadata and i think that's going to spur interest in data fabrics now data fabrics are pretty vaguely defined but if we just take the most elemental definition which is a common metadata back plane i think that if anybody is going to get serious about data mesh they need to look at a data fabric because we all at the end of the day need to speak you know need to read from the same sheet of music so thank you tony dave dave meninger i mean one of the things that people like about data mesh is it pretty crisply articulates some of the flaws in today's organizational approaches to data what are your thoughts on this well i think we have to start by defining data mesh right the the term is already getting corrupted right tony said it's going to see the cold hard uh light of day and there's a problem right now that there are a number of overlapping terms that are similar but not identical so we've got data virtualization data fabric excuse me for a second sorry about that data virtualization data fabric uh uh data federation right uh so i i think that it's not really clear what each vendor means by these terms i see data mesh and data fabric becoming quite popular i've i've interpreted data mesh as referring primarily to the governance aspects as originally you know intended and specified but that's not the way i see vendors using i see vendors using it much more to mean data fabric and data virtualization so i'm going to comment on the group of those things i think the group of those things is going to happen they're going to happen they're going to become more robust our research suggests that a quarter of organizations are already using virtualized access to their data lakes and another half so a total of three quarters will eventually be accessing their data lakes using some sort of virtualized access again whether you define it as mesh or fabric or virtualization isn't really the point here but this notion that there are different elements of data metadata and governance within an organization that all need to be managed collectively the interesting thing is when you look at the satisfaction rates of those organizations using virtualization versus those that are not it's almost double 68 of organizations i'm i'm sorry um 79 of organizations that were using virtualized access express satisfaction with their access to the data lake only 39 expressed satisfaction if they weren't using virtualized access so thank you uh dave uh sanjeev we just got about a couple minutes on this topic but i know you're speaking or maybe you've spoken already on a panel with jamal dagani who sort of invented the concept governance obviously is a big sticking point but what are your thoughts on this you are mute so my message to your mark and uh and to the community is uh as opposed to what dave said let's not define it we spent the whole year defining it there are four principles domain product data infrastructure and governance let's take it to the next level i get a lot of questions on what is the difference between data fabric and data mesh and i'm like i can compare the two because data mesh is a business concept data fabric is a data integration pattern how do you define how do you compare the two you have to bring data mesh level down so to tony's point i'm on a warp path in 2022 to take it down to what does a data product look like how do we handle shared data across domains and govern it and i think we are going to see more of that in 2022 is operationalization of data mesh i think we could have a whole hour on this topic couldn't we uh maybe we should do that uh but let's go to let's move to carl said carl your database guy you've been around that that block for a while now you want to talk about graph databases bring it on oh yeah okay thanks so i regard graph database as basically the next truly revolutionary database management technology i'm looking forward to for the graph database market which of course we haven't defined yet so obviously i have a little wiggle room in what i'm about to say but that this market will grow by about 600 percent over the next 10 years now 10 years is a long time but over the next five years we expect to see gradual growth as people start to learn how to use it problem isn't that it's used the problem is not that it's not useful is that people don't know how to use it so let me explain before i go any further what a graph database is because some of the folks on the call may not may not know what it is a graph database organizes data according to a mathematical structure called a graph a graph has elements called nodes and edges so a data element drops into a node the nodes are connected by edges the edges connect one node to another node combinations of edges create structures that you can analyze to determine how things are related in some cases the nodes and edges can have properties attached to them which add additional informative material that makes it richer that's called a property graph okay there are two principal use cases for graph databases there's there's semantic proper graphs which are used to break down human language text uh into the semantic structures then you can search it organize it and and and answer complicated questions a lot of ai is aimed at semantic graphs another kind is the property graph that i just mentioned which has a dazzling number of use cases i want to just point out is as i talk about this people are probably wondering well we have relational databases isn't that good enough okay so a relational database defines it uses um it supports what i call definitional relationships that means you define the relationships in a fixed structure the database drops into that structure there's a value foreign key value that relates one table to another and that value is fixed you don't change it if you change it the database becomes unstable it's not clear what you're looking at in a graph database the system is designed to handle change so that it can reflect the true state of the things that it's being used to track so um let me just give you some examples of use cases for this um they include uh entity resolution data lineage uh um social media analysis customer 360 fraud prevention there's cyber security there's strong supply chain is a big one actually there's explainable ai and this is going to become important too because a lot of people are adopting ai but they want a system after the fact to say how did the ai system come to that conclusion how did it make that recommendation right now we don't have really good ways of tracking that okay machine machine learning in general um social network i already mentioned that and then we've got oh gosh we've got data governance data compliance risk management we've got recommendation we've got personalization anti-money money laundering that's another big one identity and access management network and i.t operations is already becoming a key one where you actually have mapped out your operation your your you know whatever it is your data center and you you can track what's going on as things happen there root cause analysis fraud detection is a huge one a number of major credit card companies use graph databases for fraud detection risk analysis tracking and tracing churn analysis next best action what-if analysis impact analysis entity resolution and i would add one other thing or just a few other things to this list metadata management so sanjay here you go this is your engine okay because i was in metadata management for quite a while in my past life and one of the things i found was that none of the data management technologies that were available to us could efficiently handle metadata because of the kinds of structures that result from it but grass can okay grafts can do things like say this term in this context means this but in that context it means that okay things like that and in fact uh logistics management supply chain it also because it handles recursive relationships by recursive relationships i mean objects that own other objects that are of the same type you can do things like bill materials you know so like parts explosion you can do an hr analysis who reports to whom how many levels up the chain and that kind of thing you can do that with relational databases but yes it takes a lot of programming in fact you can do almost any of these things with relational databases but the problem is you have to program it it's not it's not supported in the database and whenever you have to program something that means you can't trace it you can't define it you can't publish it in terms of its functionality and it's really really hard to maintain over time so carl thank you i wonder if we could bring brad in i mean brad i'm sitting there wondering okay is this incremental to the market is it disruptive and replaceable what are your thoughts on this space it's already disrupted the market i mean like carl said go to any bank and ask them are you using graph databases to do to get fraud detection under control and they'll say absolutely that's the only way to solve this problem and it is frankly um and it's the only way to solve a lot of the problems that carl mentioned and that is i think it's it's achilles heel in some ways because you know it's like finding the best way to cross the seven bridges of konigsberg you know it's always going to kind of be tied to those use cases because it's really special and it's really unique and because it's special and it's unique uh it it still unfortunately kind of stands apart from the rest of the community that's building let's say ai outcomes as the great great example here the graph databases and ai as carl mentioned are like chocolate and peanut butter but technologically they don't know how to talk to one another they're completely different um and you know it's you can't just stand up sql and query them you've got to to learn um yeah what is that carlos specter or uh special uh uh yeah thank you uh to actually get to the data in there and if you're gonna scale that data that graph database especially a property graph if you're gonna do something really complex like try to understand uh you know all of the metadata in your organization you might just end up with you know a graph database winter like we had the ai winter simply because you run out of performance to make the thing happen so i i think it's already disrupted but we we need to like treat it like a first-class citizen in in the data analytics and ai community we need to bring it into the fold we need to equip it with the tools it needs to do that the magic it does and to do it not just for specialized use cases but for everything because i i'm with carl i i think it's absolutely revolutionary so i had also identified the principal achilles heel of the technology which is scaling now when these when these things get large and complex enough that they spill over what a single server can handle you start to have difficulties because the relationships span things that have to be resolved over a network and then you get network latency and that slows the system down so that's still a problem to be solved sanjeev any quick thoughts on this i mean i think metadata on the on the on the word cloud is going to be the the largest font uh but what are your thoughts here i want to like step away so people don't you know associate me with only meta data so i want to talk about something a little bit slightly different uh dbengines.com has done an amazing job i think almost everyone knows that they chronicle all the major databases that are in use today in january of 2022 there are 381 databases on its list of ranked list of databases the largest category is rdbms the second largest category is actually divided into two property graphs and rdf graphs these two together make up the second largest number of data databases so talking about accolades here this is a problem the problem is that there's so many graph databases to choose from they come in different shapes and forms uh to bright's point there's so many query languages in rdbms is sql end of the story here we've got sci-fi we've got gremlin we've got gql and then your proprietary languages so i think there's a lot of disparity in this space but excellent all excellent points sanji i must say and that is a problem the languages need to be sorted and standardized and it needs people need to have a road map as to what they can do with it because as you say you can do so many things and so many of those things are unrelated that you sort of say well what do we use this for i'm reminded of the saying i learned a bunch of years ago when somebody said that the digital computer is the only tool man has ever devised that has no particular purpose all right guys we gotta we gotta move on to dave uh meninger uh we've heard about streaming uh your prediction is in that realm so please take it away sure so i like to say that historical databases are to become a thing of the past but i don't mean that they're going to go away that's not my point i mean we need historical databases but streaming data is going to become the default way in which we operate with data so in the next say three to five years i would expect the data platforms and and we're using the term data platforms to represent the evolution of databases and data lakes that the data platforms will incorporate these streaming capabilities we're going to process data as it streams into an organization and then it's going to roll off into historical databases so historical databases don't go away but they become a thing of the past they store the data that occurred previously and as data is occurring we're going to be processing it we're going to be analyzing we're going to be acting on it i mean we we only ever ended up with historical databases because we were limited by the technology that was available to us data doesn't occur in batches but we processed it in batches because that was the best we could do and it wasn't bad and we've continued to improve and we've improved and we've improved but streaming data today is still the exception it's not the rule right there's there are projects within organizations that deal with streaming data but it's not the default way in which we deal with data yet and so that that's my prediction is that this is going to change we're going to have um streaming data be the default way in which we deal with data and and how you label it what you call it you know maybe these databases and data platforms just evolve to be able to handle it but we're going to deal with data in a different way and our research shows that already about half of the participants in our analytics and data benchmark research are using streaming data you know another third are planning to use streaming technologies so that gets us to about eight out of ten organizations need to use this technology that doesn't mean they have to use it throughout the whole organization but but it's pretty widespread in its use today and has continued to grow if you think about the consumerization of i.t we've all been conditioned to expect immediate access to information immediate responsiveness you know we want to know if an uh item is on the shelf at our local retail store and we can go in and pick it up right now you know that's the world we live in and that's spilling over into the enterprise i.t world where we have to provide those same types of capabilities um so that's my prediction historical database has become a thing of the past streaming data becomes the default way in which we we operate with data all right thank you david well so what what say you uh carl a guy who's followed historical databases for a long time well one thing actually every database is historical because as soon as you put data in it it's now history it's no longer it no longer reflects the present state of things but even if that history is only a millisecond old it's still history but um i would say i mean i know you're trying to be a little bit provocative in saying this dave because you know as well as i do that people still need to do their taxes they still need to do accounting they still need to run general ledger programs and things like that that all involves historical data that's not going to go away unless you want to go to jail so you're going to have to deal with that but as far as the leading edge functionality i'm totally with you on that and i'm just you know i'm just kind of wondering um if this chain if this requires a change in the way that we perceive applications in order to truly be manifested and rethinking the way m applications work um saying that uh an application should respond instantly as soon as the state of things changes what do you say about that i i think that's true i think we do have to think about things differently that's you know it's not the way we design systems in the past uh we're seeing more and more systems designed that way but again it's not the default and and agree 100 with you that we do need historical databases you know that that's clear and even some of those historical databases will be used in conjunction with the streaming data right so absolutely i mean you know let's take the data warehouse example where you're using the data warehouse as context and the streaming data as the present you're saying here's a sequence of things that's happening right now have we seen that sequence before and where what what does that pattern look like in past situations and can we learn from that so tony bear i wonder if you could comment i mean if you when you think about you know real-time inferencing at the edge for instance which is something that a lot of people talk about um a lot of what we're discussing here in this segment looks like it's got great potential what are your thoughts yeah well i mean i think you nailed it right you know you hit it right on the head there which is that i think a key what i'm seeing is that essentially and basically i'm going to split this one down the middle is i don't see that basically streaming is the default what i see is streaming and basically and transaction databases um and analytics data you know data warehouses data lakes whatever are converging and what allows us technically to converge is cloud native architecture where you can basically distribute things so you could have you can have a note here that's doing the real-time processing that's also doing it and this is what your leads in we're maybe doing some of that real-time predictive analytics to take a look at well look we're looking at this customer journey what's happening with you know you know with with what the customer is doing right now and this is correlated with what other customers are doing so what i so the thing is that in the cloud you can basically partition this and because of basically you know the speed of the infrastructure um that you can basically bring these together and or and so and kind of orchestrate them sort of loosely coupled manner the other part is that the use cases are demanding and this is part that goes back to what dave is saying is that you know when you look at customer 360 when you look at let's say smart you know smart utility grids when you look at any type of operational problem it has a real-time component and it has a historical component and having predictives and so like you know you know my sense here is that there that technically we can bring this together through the cloud and i think the use case is that is that we we can apply some some real-time sort of you know predictive analytics on these streams and feed this into the transactions so that when we make a decision in terms of what to do as a result of a transaction we have this real time you know input sanjeev did you have a comment yeah i was just going to say that to this point you know we have to think of streaming very different because in the historical databases we used to bring the data and store the data and then we used to run rules on top uh aggregations and all but in case of streaming the mindset changes because the rules normally the inference all of that is fixed but the data is constantly changing so it's a completely reverse way of thinking of uh and building applications on top of that so dave menninger there seemed to be some disagreement about the default or now what kind of time frame are you are you thinking about is this end of decade it becomes the default what would you pin i i think around you know between between five to ten years i think this becomes the reality um i think you know it'll be more and more common between now and then but it becomes the default and i also want sanjeev at some point maybe in one of our subsequent conversations we need to talk about governing streaming data because that's a whole other set of challenges we've also talked about it rather in a two dimensions historical and streaming and there's lots of low latency micro batch sub second that's not quite streaming but in many cases it's fast enough and we're seeing a lot of adoption of near real time not quite real time as uh good enough for most for many applications because nobody's really taking the hardware dimension of this information like how do we that'll just happen carl so near real time maybe before you lose the customer however you define that right okay um let's move on to brad brad you want to talk about automation ai uh the the the pipeline people feel like hey we can just automate everything what's your prediction yeah uh i'm i'm an ai fiction auto so apologies in advance for that but uh you know um i i think that um we've been seeing automation at play within ai for some time now and it's helped us do do a lot of things for especially for practitioners that are building ai outcomes in the enterprise uh it's it's helped them to fill skills gaps it's helped them to speed development and it's helped them to to actually make ai better uh because it you know in some ways provides some swim lanes and and for example with technologies like ottawa milk and can auto document and create that sort of transparency that that we talked about a little bit earlier um but i i think it's there's an interesting kind of conversion happening with this idea of automation um and and that is that uh we've had the automation that started happening for practitioners it's it's trying to move outside of the traditional bounds of things like i'm just trying to get my features i'm just trying to pick the right algorithm i'm just trying to build the right model uh and it's expanding across that full life cycle of building an ai outcome to start at the very beginning of data and to then continue on to the end which is this continuous delivery and continuous uh automation of of that outcome to make sure it's right and it hasn't drifted and stuff like that and because of that because it's become kind of powerful we're starting to to actually see this weird thing happen where the practitioners are starting to converge with the users and that is to say that okay if i'm in tableau right now i can stand up salesforce einstein discovery and it will automatically create a nice predictive algorithm for me um given the data that i that i pull in um but what's starting to happen and we're seeing this from the the the companies that create business software so salesforce oracle sap and others is that they're starting to actually use these same ideals and a lot of deep learning to to basically stand up these out of the box flip a switch and you've got an ai outcome at the ready for business users and um i i'm very much you know i think that that's that's the way that it's going to go and what it means is that ai is is slowly disappearing uh and i don't think that's a bad thing i think if anything what we're going to see in 2022 and maybe into 2023 is this sort of rush to to put this idea of disappearing ai into practice and have as many of these solutions in the enterprise as possible you can see like for example sap is going to roll out this quarter this thing called adaptive recommendation services which which basically is a cold start ai outcome that can work across a whole bunch of different vertical markets and use cases it's just a recommendation engine for whatever you need it to do in the line of business so basically you're you're an sap user you look up to turn on your software one day and you're a sales professional let's say and suddenly you have a recommendation for customer churn it's going that's great well i i don't know i i think that's terrifying in some ways i think it is the future that ai is going to disappear like that but i am absolutely terrified of it because um i i think that what it what it really does is it calls attention to a lot of the issues that we already see around ai um specific to this idea of what what we like to call it omdia responsible ai which is you know how do you build an ai outcome that is free of bias that is inclusive that is fair that is safe that is secure that it's audible etc etc etc etc that takes some a lot of work to do and so if you imagine a customer that that's just a sales force customer let's say and they're turning on einstein discovery within their sales software you need some guidance to make sure that when you flip that switch that the outcome you're going to get is correct and that's that's going to take some work and so i think we're going to see this let's roll this out and suddenly there's going to be a lot of a lot of problems a lot of pushback uh that we're going to see and some of that's going to come from gdpr and others that sam jeeve was mentioning earlier a lot of it's going to come from internal csr requirements within companies that are saying hey hey whoa hold up we can't do this all at once let's take the slow route let's make ai automated in a smart way and that's going to take time yeah so a couple predictions there that i heard i mean ai essentially you disappear it becomes invisible maybe if i can restate that and then if if i understand it correctly brad you're saying there's a backlash in the near term people can say oh slow down let's automate what we can those attributes that you talked about are non trivial to achieve is that why you're a bit of a skeptic yeah i think that we don't have any sort of standards that companies can look to and understand and we certainly within these companies especially those that haven't already stood up in internal data science team they don't have the knowledge to understand what that when they flip that switch for an automated ai outcome that it's it's gonna do what they think it's gonna do and so we need some sort of standard standard methodology and practice best practices that every company that's going to consume this invisible ai can make use of and one of the things that you know is sort of started that google kicked off a few years back that's picking up some momentum and the companies i just mentioned are starting to use it is this idea of model cards where at least you have some transparency about what these things are doing you know so like for the sap example we know for example that it's convolutional neural network with a long short-term memory model that it's using we know that it only works on roman english uh and therefore me as a consumer can say oh well i know that i need to do this internationally so i should not just turn this on today great thank you carl can you add anything any context here yeah we've talked about some of the things brad mentioned here at idc in the our future of intelligence group regarding in particular the moral and legal implications of having a fully automated you know ai uh driven system uh because we already know and we've seen that ai systems are biased by the data that they get right so if if they get data that pushes them in a certain direction i think there was a story last week about an hr system that was uh that was recommending promotions for white people over black people because in the past um you know white people were promoted and and more productive than black people but not it had no context as to why which is you know because they were being historically discriminated black people being historically discriminated against but the system doesn't know that so you know you have to be aware of that and i think that at the very least there should be controls when a decision has either a moral or a legal implication when when you want when you really need a human judgment it could lay out the options for you but a person actually needs to authorize that that action and i also think that we always will have to be vigilant regarding the kind of data we use to train our systems to make sure that it doesn't introduce unintended biases and to some extent they always will so we'll always be chasing after them that's that's absolutely carl yeah i think that what you have to bear in mind as a as a consumer of ai is that it is a reflection of us and we are a very flawed species uh and so if you look at all the really fantastic magical looking supermodels we see like gpt three and four that's coming out z they're xenophobic and hateful uh because the people the data that's built upon them and the algorithms and the people that build them are us so ai is a reflection of us we need to keep that in mind yeah we're the ai's by us because humans are biased all right great okay let's move on doug henson you know a lot of people that said that data lake that term's not not going to not going to live on but it appears to be have some legs here uh you want to talk about lake house bring it on yes i do my prediction is that lake house and this idea of a combined data warehouse and data lake platform is going to emerge as the dominant data management offering i say offering that doesn't mean it's going to be the dominant thing that organizations have out there but it's going to be the predominant vendor offering in 2022. now heading into 2021 we already had cloudera data bricks microsoft snowflake as proponents in 2021 sap oracle and several of these fabric virtualization mesh vendors join the bandwagon the promise is that you have one platform that manages your structured unstructured and semi-structured information and it addresses both the beyond analytics needs and the data science needs the real promise there is simplicity and lower cost but i think end users have to answer a few questions the first is does your organization really have a center of data gravity or is it is the data highly distributed multiple data warehouses multiple data lakes on-premises cloud if it if it's very distributed and you you know you have difficulty consolidating and that's not really a goal for you then maybe that single platform is unrealistic and not likely to add value to you um you know also the fabric and virtualization vendors the the mesh idea that's where if you have this highly distributed situation that might be a better path forward the second question if you are looking at one of these lake house offerings you are looking at consolidating simplifying bringing together to a single platform you have to make sure that it meets both the warehouse need and the data lake need so you have vendors like data bricks microsoft with azure synapse new really to the data warehouse space and they're having to prove that these data warehouse capabilities on their platforms can meet the scaling requirements can meet the user and query concurrency requirements meet those tight slas and then on the other hand you have the or the oracle sap snowflake the data warehouse uh folks coming into the data science world and they have to prove that they can manage the unstructured information and meet the needs of the data scientists i'm seeing a lot of the lake house offerings from the warehouse crowd managing that unstructured information in columns and rows and some of these vendors snowflake in particular is really relying on partners for the data science needs so you really got to look at a lake house offering and make sure that it meets both the warehouse and the data lake requirement well thank you doug well tony if those two worlds are going to come together as doug was saying the analytics and the data science world does it need to be some kind of semantic layer in between i don't know weigh in on this topic if you would oh didn't we talk about data fabrics before common metadata layer um actually i'm almost tempted to say let's declare victory and go home in that this is actually been going on for a while i actually agree with uh you know much what doug is saying there which is that i mean we i remembered as far back as i think it was like 2014 i was doing a a study you know it was still at ovum predecessor omnia um looking at all these specialized databases that were coming up and seeing that you know there's overlap with the edges but yet there was still going to be a reason at the time that you would have let's say a document database for json you'd have a relational database for tran you know for transactions and for data warehouse and you had you know and you had basically something at that time that that resembles to do for what we're considering a day of life fast fo and the thing is what i was saying at the time is that you're seeing basically blur you know sort of blending at the edges that i was saying like about five or six years ago um that's all and the the lake house is essentially you know the amount of the the current manifestation of that idea there is a dichotomy in terms of you know it's the old argument do we centralize this all you know you know in in in in in a single place or do we or do we virtualize and i think it's always going to be a yin and yang there's never going to be a single single silver silver bullet i do see um that they're also going to be questions and these are things that points that doug raised they're you know what your what do you need of of of your of you know for your performance there or for your you know pre-performance characteristics do you need for instance hiking currency you need the ability to do some very sophisticated joins or is your requirement more to be able to distribute and you know distribute our processing is you know as far as possible to get you know to essentially do a kind of brute force approach all these approaches are valid based on you know based on the used case um i just see that essentially that the lake house is the culmination of it's nothing it's just it's a relatively new term introduced by databricks a couple years ago this is the culmination of basically what's been a long time trend and what we see in the cloud is that as we start seeing data warehouses as a checkbox item say hey we can basically source data in cloud and cloud storage and s3 azure blob store you know whatever um as long as it's in certain formats like you know like you know parquet or csv or something like that you know i see that as becoming kind of you know a check box item so to that extent i think that the lake house depending on how you define it is already reality um and in some in some cases maybe new terminology but not a whole heck of a lot new under the sun yeah and dave menger i mean a lot of this thank you tony but a lot of this is going to come down to you know vendor marketing right some people try to co-opt the term we talked about data mesh washing what are your thoughts on this yeah so um i used the term data platform earlier and and part of the reason i use that term is that it's more vendor neutral uh we've we've tried to uh sort of stay out of the the vendor uh terminology patenting world right whether whether the term lake house is what sticks or not the concept is certainly going to stick and we have some data to back it up about a quarter of organizations that are using data lakes today already incorporate data warehouse functionality into it so they consider their data lake house and data warehouse one in the same about a quarter of organizations a little less but about a quarter of organizations feed the data lake from the data warehouse and about a quarter of organizations feed the data warehouse from the data lake so it's pretty obvious that three quarters of organizations need to bring this stuff together right the need is there the need is apparent the technology is going to continue to verge converge i i like to talk about you know you've got data lakes over here at one end and i'm not going to talk about why people thought data lakes were a bad idea because they thought you just throw stuff in a in a server and you ignore it right that's not what a data lake is so you've got data lake people over here and you've got database people over here data warehouse people over here database vendors are adding data lake capabilities and data lake vendors are adding data warehouse capabilities so it's obvious that they're going to meet in the middle i mean i think it's like tony says i think we should there declare victory and go home and so so i it's just a follow-up on that so are you saying these the specialized lake and the specialized warehouse do they go away i mean johnny tony data mesh practitioners would say or or advocates would say well they could all live as just a node on the on the mesh but based on what dave just said are we going to see those all morph together well number one as i was saying before there's always going to be this sort of you know kind of you know centrifugal force or this tug of war between do we centralize the data do we do it virtualize and the fact is i don't think that work there's ever going to be any single answer i think in terms of data mesh data mesh has nothing to do with how you physically implement the data you could have a data mesh on a basically uh on a data warehouse it's just that you know the difference being is that if we use the same you know physical data store but everybody's logically manual basically governing it differently you know um a data mission is basically it's not a technology it's a process it's a governance process um so essentially um you know you know i basically see that you know as as i was saying before that this is basically the culmination of a long time trend we're essentially seeing a lot of blurring but there are going to be cases where for instance if i need let's say like observe i need like high concurrency or something like that there are certain things that i'm not going to be able to get efficiently get out of a data lake um and you know we're basically i'm doing a system where i'm just doing really brute forcing very fast file scanning and that type of thing so i think there always will be some delineations but i would agree with dave and with doug that we are seeing basically a a confluence of requirements that we need to essentially have basically the element you know the ability of a data lake and a data laid out their warehouse we these need to come together so i think what we're likely to see is organizations look for a converged platform that can handle both sides for their center of data gravity the mesh and the fabric vendors the the fabric virtualization vendors they're all on board with the idea of this converged platform and they're saying hey we'll handle all the edge cases of the stuff that isn't in that center of data gradient that is off distributed in a cloud or at a remote location so you can have that single platform for the center of of your your data and then bring in virtualization mesh what have you for reaching out to the distributed data bingo as they basically said people are happy when they virtualize data i i think yes at this point but to this uh dave meningas point you know they have convert they are converging snowflake has introduced support for unstructured data so now we are literally splitting here now what uh databricks is saying is that aha but it's easy to go from data lake to data warehouse than it is from data warehouse to data lake so i think we're getting into semantics but we've already seen these two converge so is that so it takes something like aws who's got what 15 data stores are they're going to have 15 converged data stores that's going to be interesting to watch all right guys i'm going to go down the list and do like a one i'm going to one word each and you guys each of the analysts if you wouldn't just add a very brief sort of course correction for me so sanjeev i mean governance is going to be the maybe it's the dog that wags the tail now i mean it's coming to the fore all this ransomware stuff which really didn't talk much about security but but but what's the one word in your prediction that you would leave us with on governance it's uh it's going to be mainstream mainstream okay tony bear mesh washing is what i wrote down that's that's what we're going to see in uh in in 2022 a little reality check you you want to add to that reality check is i hope that no vendor you know jumps the shark and calls their offering a data mesh project yeah yeah let's hope that doesn't happen if they do we're going to call them out uh carl i mean graph databases thank you for sharing some some you know high growth metrics i know it's early days but magic is what i took away from that it's the magic database yeah i would actually i've said this to people too i i kind of look at it as a swiss army knife of data because you can pretty much do anything you want with it it doesn't mean you should i mean that's definitely the case that if you're you know managing things that are in a fixed schematic relationship probably a relational database is a better choice there are you know times when the document database is a better choice it can handle those things but maybe not it may not be the best choice for that use case but for a great many especially the new emerging use cases i listed it's the best choice thank you and dave meninger thank you by the way for bringing the data in i like how you supported all your comments with with some some data points but streaming data becomes the sort of default uh paradigm if you will what would you add yeah um i would say think fast right that's the world we live in you got to think fast fast love it uh and brad shimon uh i love it i mean on the one hand i was saying okay great i'm afraid i might get disrupted by one of these internet giants who are ai experts so i'm gonna be able to buy instead of build ai but then again you know i've got some real issues there's a potential backlash there so give us the there's your bumper sticker yeah i i would say um going with dave think fast and also think slow uh to to talk about the book that everyone talks about i would say really that this is all about trust trust in the idea of automation and of a transparent invisible ai across the enterprise but verify verify before you do anything and then doug henson i mean i i look i think the the trend is your friend here on this prediction with lake house is uh really becoming dominant i liked the way you set up that notion of you know the the the data warehouse folks coming at it from the analytics perspective but then you got the data science worlds coming together i still feel as though there's this piece in the middle that we're missing but your your final thoughts we'll give you the last well i think the idea of consolidation and simplification uh always prevails that's why the appeal of a single platform is going to be there um we've already seen that with uh you know hadoop platforms moving toward cloud moving toward object storage and object storage becoming really the common storage point for whether it's a lake or a warehouse uh and that second point uh i think esg mandates are uh are gonna come in alongside uh gdpr and things like that to uh up the ante for uh good governance yeah thank you for calling that out okay folks hey that's all the time that that we have here your your experience and depth of understanding on these key issues and in data and data management really on point and they were on display today i want to thank you for your your contributions really appreciate your time enjoyed it thank you now in addition to this video we're going to be making available transcripts of the discussion we're going to do clips of this as well we're going to put them out on social media i'll write this up and publish the discussion on wikibon.com and siliconangle.com no doubt several of the analysts on the panel will take the opportunity to publish written content social commentary or both i want to thank the power panelist and thanks for watching this special cube presentation this is dave vellante be well and we'll see you next time [Music] you
SUMMARY :
the end of the day need to speak you
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
381 databases | QUANTITY | 0.99+ |
2014 | DATE | 0.99+ |
2022 | DATE | 0.99+ |
2021 | DATE | 0.99+ |
january of 2022 | DATE | 0.99+ |
100 users | QUANTITY | 0.99+ |
jamal dagani | PERSON | 0.99+ |
last week | DATE | 0.99+ |
dave meninger | PERSON | 0.99+ |
sanji | PERSON | 0.99+ |
second question | QUANTITY | 0.99+ |
15 converged data stores | QUANTITY | 0.99+ |
dave vellante | PERSON | 0.99+ |
microsoft | ORGANIZATION | 0.99+ |
three | QUANTITY | 0.99+ |
sanjeev | PERSON | 0.99+ |
2023 | DATE | 0.99+ |
15 data stores | QUANTITY | 0.99+ |
siliconangle.com | OTHER | 0.99+ |
last year | DATE | 0.99+ |
sanjeev mohan | PERSON | 0.99+ |
six | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
carl | PERSON | 0.99+ |
tony | PERSON | 0.99+ |
carl olufsen | PERSON | 0.99+ |
six years | QUANTITY | 0.99+ |
david | PERSON | 0.99+ |
carlos specter | PERSON | 0.98+ |
both sides | QUANTITY | 0.98+ |
2010s | DATE | 0.98+ |
first backlash | QUANTITY | 0.98+ |
five years | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
dave | PERSON | 0.98+ |
each | QUANTITY | 0.98+ |
three quarters | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
single platform | QUANTITY | 0.98+ |
lake house | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
doug | PERSON | 0.97+ |
one word | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
wikibon.com | OTHER | 0.97+ |
one platform | QUANTITY | 0.97+ |
39 | QUANTITY | 0.97+ |
about 600 percent | QUANTITY | 0.97+ |
two analysts | QUANTITY | 0.97+ |
ten years | QUANTITY | 0.97+ |
single platform | QUANTITY | 0.96+ |
five | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
three quarters | QUANTITY | 0.96+ |
california | LOCATION | 0.96+ |
ORGANIZATION | 0.96+ | |
single | QUANTITY | 0.95+ |
Predictions 2022: Top Analysts See the Future of Data
(bright music) >> In the 2010s, organizations became keenly aware that data would become the key ingredient to driving competitive advantage, differentiation, and growth. But to this day, putting data to work remains a difficult challenge for many, if not most organizations. Now, as the cloud matures, it has become a game changer for data practitioners by making cheap storage and massive processing power readily accessible. We've also seen better tooling in the form of data workflows, streaming, machine intelligence, AI, developer tools, security, observability, automation, new databases and the like. These innovations they accelerate data proficiency, but at the same time, they add complexity for practitioners. Data lakes, data hubs, data warehouses, data marts, data fabrics, data meshes, data catalogs, data oceans are forming, they're evolving and exploding onto the scene. So in an effort to bring perspective to the sea of optionality, we've brought together the brightest minds in the data analyst community to discuss how data management is morphing and what practitioners should expect in 2022 and beyond. Hello everyone, my name is Dave Velannte with theCUBE, and I'd like to welcome you to a special Cube presentation, analysts predictions 2022: the future of data management. We've gathered six of the best analysts in data and data management who are going to present and discuss their top predictions and trends for 2022 in the first half of this decade. Let me introduce our six power panelists. Sanjeev Mohan is former Gartner Analyst and Principal at SanjMo. Tony Baer, principal at dbInsight, Carl Olofson is well-known Research Vice President with IDC, Dave Menninger is Senior Vice President and Research Director at Ventana Research, Brad Shimmin, Chief Analyst, AI Platforms, Analytics and Data Management at Omdia and Doug Henschen, Vice President and Principal Analyst at Constellation Research. Gentlemen, welcome to the program and thanks for coming on theCUBE today. >> Great to be here. >> Thank you. >> All right, here's the format we're going to use. I as moderator, I'm going to call on each analyst separately who then will deliver their prediction or mega trend, and then in the interest of time management and pace, two analysts will have the opportunity to comment. If we have more time, we'll elongate it, but let's get started right away. Sanjeev Mohan, please kick it off. You want to talk about governance, go ahead sir. >> Thank you Dave. I believe that data governance which we've been talking about for many years is now not only going to be mainstream, it's going to be table stakes. And all the things that you mentioned, you know, the data, ocean data lake, lake houses, data fabric, meshes, the common glue is metadata. If we don't understand what data we have and we are governing it, there is no way we can manage it. So we saw Informatica went public last year after a hiatus of six. I'm predicting that this year we see some more companies go public. My bet is on Culebra, most likely and maybe Alation we'll see go public this year. I'm also predicting that the scope of data governance is going to expand beyond just data. It's not just data and reports. We are going to see more transformations like spark jawsxxxxx, Python even Air Flow. We're going to see more of a streaming data. So from Kafka Schema Registry, for example. We will see AI models become part of this whole governance suite. So the governance suite is going to be very comprehensive, very detailed lineage, impact analysis, and then even expand into data quality. We already seen that happen with some of the tools where they are buying these smaller companies and bringing in data quality monitoring and integrating it with metadata management, data catalogs, also data access governance. So what we are going to see is that once the data governance platforms become the key entry point into these modern architectures, I'm predicting that the usage, the number of users of a data catalog is going to exceed that of a BI tool. That will take time and we already seen that trajectory. Right now if you look at BI tools, I would say there a hundred users to BI tool to one data catalog. And I see that evening out over a period of time and at some point data catalogs will really become the main way for us to access data. Data catalog will help us visualize data, but if we want to do more in-depth analysis, it'll be the jumping off point into the BI tool, the data science tool and that is the journey I see for the data governance products. >> Excellent, thank you. Some comments. Maybe Doug, a lot of things to weigh in on there, maybe you can comment. >> Yeah, Sanjeev I think you're spot on, a lot of the trends the one disagreement, I think it's really still far from mainstream. As you say, we've been talking about this for years, it's like God, motherhood, apple pie, everyone agrees it's important, but too few organizations are really practicing good governance because it's hard and because the incentives have been lacking. I think one thing that deserves mention in this context is ESG mandates and guidelines, these are environmental, social and governance, regs and guidelines. We've seen the environmental regs and guidelines and posts in industries, particularly the carbon-intensive industries. We've seen the social mandates, particularly diversity imposed on suppliers by companies that are leading on this topic. We've seen governance guidelines now being imposed by banks on investors. So these ESGs are presenting new carrots and sticks, and it's going to demand more solid data. It's going to demand more detailed reporting and solid reporting, tighter governance. But we're still far from mainstream adoption. We have a lot of, you know, best of breed niche players in the space. I think the signs that it's going to be more mainstream are starting with things like Azure Purview, Google Dataplex, the big cloud platform players seem to be upping the ante and starting to address governance. >> Excellent, thank you Doug. Brad, I wonder if you could chime in as well. >> Yeah, I would love to be a believer in data catalogs. But to Doug's point, I think that it's going to take some more pressure for that to happen. I recall metadata being something every enterprise thought they were going to get under control when we were working on service oriented architecture back in the nineties and that didn't happen quite the way we anticipated. And so to Sanjeev's point it's because it is really complex and really difficult to do. My hope is that, you know, we won't sort of, how do I put this? Fade out into this nebula of domain catalogs that are specific to individual use cases like Purview for getting data quality right or like data governance and cybersecurity. And instead we have some tooling that can actually be adaptive to gather metadata to create something. And I know its important to you, Sanjeev and that is this idea of observability. If you can get enough metadata without moving your data around, but understanding the entirety of a system that's running on this data, you can do a lot. So to help with the governance that Doug is talking about. >> So I just want to add that, data governance, like any other initiatives did not succeed even AI went into an AI window, but that's a different topic. But a lot of these things did not succeed because to your point, the incentives were not there. I remember when Sarbanes Oxley had come into the scene, if a bank did not do Sarbanes Oxley, they were very happy to a million dollar fine. That was like, you know, pocket change for them instead of doing the right thing. But I think the stakes are much higher now. With GDPR, the flood gates opened. Now, you know, California, you know, has CCPA but even CCPA is being outdated with CPRA, which is much more GDPR like. So we are very rapidly entering a space where pretty much every major country in the world is coming up with its own compliance regulatory requirements, data residents is becoming really important. And I think we are going to reach a stage where it won't be optional anymore. So whether we like it or not, and I think the reason data catalogs were not successful in the past is because we did not have the right focus on adoption. We were focused on features and these features were disconnected, very hard for business to adopt. These are built by IT people for IT departments to take a look at technical metadata, not business metadata. Today the tables have turned. CDOs are driving this initiative, regulatory compliances are beating down hard, so I think the time might be right. >> Yeah so guys, we have to move on here. But there's some real meat on the bone here, Sanjeev. I like the fact that you called out Culebra and Alation, so we can look back a year from now and say, okay, he made the call, he stuck it. And then the ratio of BI tools to data catalogs that's another sort of measurement that we can take even though with some skepticism there, that's something that we can watch. And I wonder if someday, if we'll have more metadata than data. But I want to move to Tony Baer, you want to talk about data mesh and speaking, you know, coming off of governance. I mean, wow, you know the whole concept of data mesh is, decentralized data, and then governance becomes, you know, a nightmare there, but take it away, Tony. >> We'll put this way, data mesh, you know, the idea at least as proposed by ThoughtWorks. You know, basically it was at least a couple of years ago and the press has been almost uniformly almost uncritical. A good reason for that is for all the problems that basically Sanjeev and Doug and Brad we're just speaking about, which is that we have all this data out there and we don't know what to do about it. Now, that's not a new problem. That was a problem we had in enterprise data warehouses, it was a problem when we had over DoOP data clusters, it's even more of a problem now that data is out in the cloud where the data is not only your data lake, is not only us three, it's all over the place. And it's also including streaming, which I know we'll be talking about later. So the data mesh was a response to that, the idea of that we need to bait, you know, who are the folks that really know best about governance? It's the domain experts. So it was basically data mesh was an architectural pattern and a process. My prediction for this year is that data mesh is going to hit cold heart reality. Because if you do a Google search, basically the published work, the articles on data mesh have been largely, you know, pretty uncritical so far. Basically loading and is basically being a very revolutionary new idea. I don't think it's that revolutionary because we've talked about ideas like this. Brad now you and I met years ago when we were talking about so and decentralizing all of us, but it was at the application level. Now we're talking about it at the data level. And now we have microservices. So there's this thought of have we managed if we're deconstructing apps in cloud native to microservices, why don't we think of data in the same way? My sense this year is that, you know, this has been a very active search if you look at Google search trends, is that now companies, like enterprise are going to look at this seriously. And as they look at it seriously, it's going to attract its first real hard scrutiny, it's going to attract its first backlash. That's not necessarily a bad thing. It means that it's being taken seriously. The reason why I think that you'll start to see basically the cold hearted light of day shine on data mesh is that it's still a work in progress. You know, this idea is basically a couple of years old and there's still some pretty major gaps. The biggest gap is in the area of federated governance. Now federated governance itself is not a new issue. Federated governance decision, we started figuring out like, how can we basically strike the balance between getting let's say between basically consistent enterprise policy, consistent enterprise governance, but yet the groups that understand the data and know how to basically, you know, that, you know, how do we basically sort of balance the two? There's a huge gap there in practice and knowledge. Also to a lesser extent, there's a technology gap which is basically in the self-service technologies that will help teams essentially govern data. You know, basically through the full life cycle, from develop, from selecting the data from, you know, building the pipelines from, you know, determining your access control, looking at quality, looking at basically whether the data is fresh or whether it's trending off course. So my prediction is that it will receive the first harsh scrutiny this year. You are going to see some organization and enterprises declare premature victory when they build some federated query implementations. You going to see vendors start with data mesh wash their products anybody in the data management space that they are going to say that where this basically a pipelining tool, whether it's basically ELT, whether it's a catalog or federated query tool, they will all going to get like, you know, basically promoting the fact of how they support this. Hopefully nobody's going to call themselves a data mesh tool because data mesh is not a technology. We're going to see one other thing come out of this. And this harks back to the metadata that Sanjeev was talking about and of the catalog just as he was talking about. Which is that there's going to be a new focus, every renewed focus on metadata. And I think that's going to spur interest in data fabrics. Now data fabrics are pretty vaguely defined, but if we just take the most elemental definition, which is a common metadata back plane, I think that if anybody is going to get serious about data mesh, they need to look at the data fabric because we all at the end of the day, need to speak, you know, need to read from the same sheet of music. >> So thank you Tony. Dave Menninger, I mean, one of the things that people like about data mesh is it pretty crisply articulate some of the flaws in today's organizational approaches to data. What are your thoughts on this? >> Well, I think we have to start by defining data mesh, right? The term is already getting corrupted, right? Tony said it's going to see the cold hard light of day. And there's a problem right now that there are a number of overlapping terms that are similar but not identical. So we've got data virtualization, data fabric, excuse me for a second. (clears throat) Sorry about that. Data virtualization, data fabric, data federation, right? So I think that it's not really clear what each vendor means by these terms. I see data mesh and data fabric becoming quite popular. I've interpreted data mesh as referring primarily to the governance aspects as originally intended and specified. But that's not the way I see vendors using it. I see vendors using it much more to mean data fabric and data virtualization. So I'm going to comment on the group of those things. I think the group of those things is going to happen. They're going to happen, they're going to become more robust. Our research suggests that a quarter of organizations are already using virtualized access to their data lakes and another half, so a total of three quarters will eventually be accessing their data lakes using some sort of virtualized access. Again, whether you define it as mesh or fabric or virtualization isn't really the point here. But this notion that there are different elements of data, metadata and governance within an organization that all need to be managed collectively. The interesting thing is when you look at the satisfaction rates of those organizations using virtualization versus those that are not, it's almost double, 68% of organizations, I'm sorry, 79% of organizations that were using virtualized access express satisfaction with their access to the data lake. Only 39% express satisfaction if they weren't using virtualized access. >> Oh thank you Dave. Sanjeev we just got about a couple of minutes on this topic, but I know you're speaking or maybe you've always spoken already on a panel with (indistinct) who sort of invented the concept. Governance obviously is a big sticking point, but what are your thoughts on this? You're on mute. (panelist chuckling) >> So my message to (indistinct) and to the community is as opposed to what they said, let's not define it. We spent a whole year defining it, there are four principles, domain, product, data infrastructure, and governance. Let's take it to the next level. I get a lot of questions on what is the difference between data fabric and data mesh? And I'm like I can't compare the two because data mesh is a business concept, data fabric is a data integration pattern. How do you compare the two? You have to bring data mesh a level down. So to Tony's point, I'm on a warpath in 2022 to take it down to what does a data product look like? How do we handle shared data across domains and governance? And I think we are going to see more of that in 2022, or is "operationalization" of data mesh. >> I think we could have a whole hour on this topic, couldn't we? Maybe we should do that. But let's corner. Let's move to Carl. So Carl, you're a database guy, you've been around that block for a while now, you want to talk about graph databases, bring it on. >> Oh yeah. Okay thanks. So I regard graph database as basically the next truly revolutionary database management technology. I'm looking forward for the graph database market, which of course we haven't defined yet. So obviously I have a little wiggle room in what I'm about to say. But this market will grow by about 600% over the next 10 years. Now, 10 years is a long time. But over the next five years, we expect to see gradual growth as people start to learn how to use it. The problem is not that it's not useful, its that people don't know how to use it. So let me explain before I go any further what a graph database is because some of the folks on the call may not know what it is. A graph database organizes data according to a mathematical structure called a graph. The graph has elements called nodes and edges. So a data element drops into a node, the nodes are connected by edges, the edges connect one node to another node. Combinations of edges create structures that you can analyze to determine how things are related. In some cases, the nodes and edges can have properties attached to them which add additional informative material that makes it richer, that's called a property graph. There are two principle use cases for graph databases. There's semantic property graphs, which are use to break down human language texts into the semantic structures. Then you can search it, organize it and answer complicated questions. A lot of AI is aimed at semantic graphs. Another kind is the property graph that I just mentioned, which has a dazzling number of use cases. I want to just point out as I talk about this, people are probably wondering, well, we have relation databases, isn't that good enough? So a relational database defines... It supports what I call definitional relationships. That means you define the relationships in a fixed structure. The database drops into that structure, there's a value, foreign key value, that relates one table to another and that value is fixed. You don't change it. If you change it, the database becomes unstable, it's not clear what you're looking at. In a graph database, the system is designed to handle change so that it can reflect the true state of the things that it's being used to track. So let me just give you some examples of use cases for this. They include entity resolution, data lineage, social media analysis, Customer 360, fraud prevention. There's cybersecurity, there's strong supply chain is a big one actually. There is explainable AI and this is going to become important too because a lot of people are adopting AI. But they want a system after the fact to say, how do the AI system come to that conclusion? How did it make that recommendation? Right now we don't have really good ways of tracking that. Machine learning in general, social network, I already mentioned that. And then we've got, oh gosh, we've got data governance, data compliance, risk management. We've got recommendation, we've got personalization, anti money laundering, that's another big one, identity and access management, network and IT operations is already becoming a key one where you actually have mapped out your operation, you know, whatever it is, your data center and you can track what's going on as things happen there, root cause analysis, fraud detection is a huge one. A number of major credit card companies use graph databases for fraud detection, risk analysis, tracking and tracing turn analysis, next best action, what if analysis, impact analysis, entity resolution and I would add one other thing or just a few other things to this list, metadata management. So Sanjeev, here you go, this is your engine. Because I was in metadata management for quite a while in my past life. And one of the things I found was that none of the data management technologies that were available to us could efficiently handle metadata because of the kinds of structures that result from it, but graphs can, okay? Graphs can do things like say, this term in this context means this, but in that context, it means that, okay? Things like that. And in fact, logistics management, supply chain. And also because it handles recursive relationships, by recursive relationships I mean objects that own other objects that are of the same type. You can do things like build materials, you know, so like parts explosion. Or you can do an HR analysis, who reports to whom, how many levels up the chain and that kind of thing. You can do that with relational databases, but yet it takes a lot of programming. In fact, you can do almost any of these things with relational databases, but the problem is, you have to program it. It's not supported in the database. And whenever you have to program something, that means you can't trace it, you can't define it. You can't publish it in terms of its functionality and it's really, really hard to maintain over time. >> Carl, thank you. I wonder if we could bring Brad in, I mean. Brad, I'm sitting here wondering, okay, is this incremental to the market? Is it disruptive and replacement? What are your thoughts on this phase? >> It's already disrupted the market. I mean, like Carl said, go to any bank and ask them are you using graph databases to get fraud detection under control? And they'll say, absolutely, that's the only way to solve this problem. And it is frankly. And it's the only way to solve a lot of the problems that Carl mentioned. And that is, I think it's Achilles heel in some ways. Because, you know, it's like finding the best way to cross the seven bridges of Koenigsberg. You know, it's always going to kind of be tied to those use cases because it's really special and it's really unique and because it's special and it's unique, it's still unfortunately kind of stands apart from the rest of the community that's building, let's say AI outcomes, as a great example here. Graph databases and AI, as Carl mentioned, are like chocolate and peanut butter. But technologically, you think don't know how to talk to one another, they're completely different. And you know, you can't just stand up SQL and query them. You've got to learn, know what is the Carl? Specter special. Yeah, thank you to, to actually get to the data in there. And if you're going to scale that data, that graph database, especially a property graph, if you're going to do something really complex, like try to understand you know, all of the metadata in your organization, you might just end up with, you know, a graph database winter like we had the AI winter simply because you run out of performance to make the thing happen. So, I think it's already disrupted, but we need to like treat it like a first-class citizen in the data analytics and AI community. We need to bring it into the fold. We need to equip it with the tools it needs to do the magic it does and to do it not just for specialized use cases, but for everything. 'Cause I'm with Carl. I think it's absolutely revolutionary. >> Brad identified the principal, Achilles' heel of the technology which is scaling. When these things get large and complex enough that they spill over what a single server can handle, you start to have difficulties because the relationships span things that have to be resolved over a network and then you get network latency and that slows the system down. So that's still a problem to be solved. >> Sanjeev, any quick thoughts on this? I mean, I think metadata on the word cloud is going to be the largest font, but what are your thoughts here? >> I want to (indistinct) So people don't associate me with only metadata, so I want to talk about something slightly different. dbengines.com has done an amazing job. I think almost everyone knows that they chronicle all the major databases that are in use today. In January of 2022, there are 381 databases on a ranked list of databases. The largest category is RDBMS. The second largest category is actually divided into two property graphs and IDF graphs. These two together make up the second largest number databases. So talking about Achilles heel, this is a problem. The problem is that there's so many graph databases to choose from. They come in different shapes and forms. To Brad's point, there's so many query languages in RDBMS, in SQL. I know the story, but here We've got cipher, we've got gremlin, we've got GQL and then we're proprietary languages. So I think there's a lot of disparity in this space. >> Well, excellent. All excellent points, Sanjeev, if I must say. And that is a problem that the languages need to be sorted and standardized. People need to have a roadmap as to what they can do with it. Because as you say, you can do so many things. And so many of those things are unrelated that you sort of say, well, what do we use this for? And I'm reminded of the saying I learned a bunch of years ago. And somebody said that the digital computer is the only tool man has ever device that has no particular purpose. (panelists chuckle) >> All right guys, we got to move on to Dave Menninger. We've heard about streaming. Your prediction is in that realm, so please take it away. >> Sure. So I like to say that historical databases are going to become a thing of the past. By that I don't mean that they're going to go away, that's not my point. I mean, we need historical databases, but streaming data is going to become the default way in which we operate with data. So in the next say three to five years, I would expect that data platforms and we're using the term data platforms to represent the evolution of databases and data lakes, that the data platforms will incorporate these streaming capabilities. We're going to process data as it streams into an organization and then it's going to roll off into historical database. So historical databases don't go away, but they become a thing of the past. They store the data that occurred previously. And as data is occurring, we're going to be processing it, we're going to be analyzing it, we're going to be acting on it. I mean we only ever ended up with historical databases because we were limited by the technology that was available to us. Data doesn't occur in patches. But we processed it in patches because that was the best we could do. And it wasn't bad and we've continued to improve and we've improved and we've improved. But streaming data today is still the exception. It's not the rule, right? There are projects within organizations that deal with streaming data. But it's not the default way in which we deal with data yet. And so that's my prediction is that this is going to change, we're going to have streaming data be the default way in which we deal with data and how you label it and what you call it. You know, maybe these databases and data platforms just evolved to be able to handle it. But we're going to deal with data in a different way. And our research shows that already, about half of the participants in our analytics and data benchmark research, are using streaming data. You know, another third are planning to use streaming technologies. So that gets us to about eight out of 10 organizations need to use this technology. And that doesn't mean they have to use it throughout the whole organization, but it's pretty widespread in its use today and has continued to grow. If you think about the consumerization of IT, we've all been conditioned to expect immediate access to information, immediate responsiveness. You know, we want to know if an item is on the shelf at our local retail store and we can go in and pick it up right now. You know, that's the world we live in and that's spilling over into the enterprise IT world We have to provide those same types of capabilities. So that's my prediction, historical databases become a thing of the past, streaming data becomes the default way in which we operate with data. >> All right thank you David. Well, so what say you, Carl, the guy who has followed historical databases for a long time? >> Well, one thing actually, every database is historical because as soon as you put data in it, it's now history. They'll no longer reflect the present state of things. But even if that history is only a millisecond old, it's still history. But I would say, I mean, I know you're trying to be a little bit provocative in saying this Dave 'cause you know, as well as I do that people still need to do their taxes, they still need to do accounting, they still need to run general ledger programs and things like that. That all involves historical data. That's not going to go away unless you want to go to jail. So you're going to have to deal with that. But as far as the leading edge functionality, I'm totally with you on that. And I'm just, you know, I'm just kind of wondering if this requires a change in the way that we perceive applications in order to truly be manifested and rethinking the way applications work. Saying that an application should respond instantly, as soon as the state of things changes. What do you say about that? >> I think that's true. I think we do have to think about things differently. It's not the way we designed systems in the past. We're seeing more and more systems designed that way. But again, it's not the default. And I agree 100% with you that we do need historical databases you know, that's clear. And even some of those historical databases will be used in conjunction with the streaming data, right? >> Absolutely. I mean, you know, let's take the data warehouse example where you're using the data warehouse as its context and the streaming data as the present and you're saying, here's the sequence of things that's happening right now. Have we seen that sequence before? And where? What does that pattern look like in past situations? And can we learn from that? >> So Tony Baer, I wonder if you could comment? I mean, when you think about, you know, real time inferencing at the edge, for instance, which is something that a lot of people talk about, a lot of what we're discussing here in this segment, it looks like it's got a great potential. What are your thoughts? >> Yeah, I mean, I think you nailed it right. You know, you hit it right on the head there. Which is that, what I'm seeing is that essentially. Then based on I'm going to split this one down the middle is that I don't see that basically streaming is the default. What I see is streaming and basically and transaction databases and analytics data, you know, data warehouses, data lakes whatever are converging. And what allows us technically to converge is cloud native architecture, where you can basically distribute things. So you can have a node here that's doing the real-time processing, that's also doing... And this is where it leads in or maybe doing some of that real time predictive analytics to take a look at, well look, we're looking at this customer journey what's happening with what the customer is doing right now and this is correlated with what other customers are doing. So the thing is that in the cloud, you can basically partition this and because of basically the speed of the infrastructure then you can basically bring these together and kind of orchestrate them sort of a loosely coupled manner. The other parts that the use cases are demanding, and this is part of it goes back to what Dave is saying. Is that, you know, when you look at Customer 360, when you look at let's say Smart Utility products, when you look at any type of operational problem, it has a real time component and it has an historical component. And having predictive and so like, you know, my sense here is that technically we can bring this together through the cloud. And I think the use case is that we can apply some real time sort of predictive analytics on these streams and feed this into the transactions so that when we make a decision in terms of what to do as a result of a transaction, we have this real-time input. >> Sanjeev, did you have a comment? >> Yeah, I was just going to say that to Dave's point, you know, we have to think of streaming very different because in the historical databases, we used to bring the data and store the data and then we used to run rules on top, aggregations and all. But in case of streaming, the mindset changes because the rules are normally the inference, all of that is fixed, but the data is constantly changing. So it's a completely reversed way of thinking and building applications on top of that. >> So Dave Menninger, there seem to be some disagreement about the default. What kind of timeframe are you thinking about? Is this end of decade it becomes the default? What would you pin? >> I think around, you know, between five to 10 years, I think this becomes the reality. >> I think its... >> It'll be more and more common between now and then, but it becomes the default. And I also want Sanjeev at some point, maybe in one of our subsequent conversations, we need to talk about governing streaming data. 'Cause that's a whole nother set of challenges. >> We've also talked about it rather in two dimensions, historical and streaming, and there's lots of low latency, micro batch, sub-second, that's not quite streaming, but in many cases its fast enough and we're seeing a lot of adoption of near real time, not quite real-time as good enough for many applications. (indistinct cross talk from panelists) >> Because nobody's really taking the hardware dimension (mumbles). >> That'll just happened, Carl. (panelists laughing) >> So near real time. But maybe before you lose the customer, however we define that, right? Okay, let's move on to Brad. Brad, you want to talk about automation, AI, the pipeline people feel like, hey, we can just automate everything. What's your prediction? >> Yeah I'm an AI aficionados so apologies in advance for that. But, you know, I think that we've been seeing automation play within AI for some time now. And it's helped us do a lot of things especially for practitioners that are building AI outcomes in the enterprise. It's helped them to fill skills gaps, it's helped them to speed development and it's helped them to actually make AI better. 'Cause it, you know, in some ways provide some swim lanes and for example, with technologies like AutoML can auto document and create that sort of transparency that we talked about a little bit earlier. But I think there's an interesting kind of conversion happening with this idea of automation. And that is that we've had the automation that started happening for practitioners, it's trying to move out side of the traditional bounds of things like I'm just trying to get my features, I'm just trying to pick the right algorithm, I'm just trying to build the right model and it's expanding across that full life cycle, building an AI outcome, to start at the very beginning of data and to then continue on to the end, which is this continuous delivery and continuous automation of that outcome to make sure it's right and it hasn't drifted and stuff like that. And because of that, because it's become kind of powerful, we're starting to actually see this weird thing happen where the practitioners are starting to converge with the users. And that is to say that, okay, if I'm in Tableau right now, I can stand up Salesforce Einstein Discovery, and it will automatically create a nice predictive algorithm for me given the data that I pull in. But what's starting to happen and we're seeing this from the companies that create business software, so Salesforce, Oracle, SAP, and others is that they're starting to actually use these same ideals and a lot of deep learning (chuckles) to basically stand up these out of the box flip-a-switch, and you've got an AI outcome at the ready for business users. And I am very much, you know, I think that's the way that it's going to go and what it means is that AI is slowly disappearing. And I don't think that's a bad thing. I think if anything, what we're going to see in 2022 and maybe into 2023 is this sort of rush to put this idea of disappearing AI into practice and have as many of these solutions in the enterprise as possible. You can see, like for example, SAP is going to roll out this quarter, this thing called adaptive recommendation services, which basically is a cold start AI outcome that can work across a whole bunch of different vertical markets and use cases. It's just a recommendation engine for whatever you needed to do in the line of business. So basically, you're an SAP user, you look up to turn on your software one day, you're a sales professional let's say, and suddenly you have a recommendation for customer churn. Boom! It's going, that's great. Well, I don't know, I think that's terrifying. In some ways I think it is the future that AI is going to disappear like that, but I'm absolutely terrified of it because I think that what it really does is it calls attention to a lot of the issues that we already see around AI, specific to this idea of what we like to call at Omdia, responsible AI. Which is, you know, how do you build an AI outcome that is free of bias, that is inclusive, that is fair, that is safe, that is secure, that its audible, et cetera, et cetera, et cetera, et cetera. I'd take a lot of work to do. And so if you imagine a customer that's just a Salesforce customer let's say, and they're turning on Einstein Discovery within their sales software, you need some guidance to make sure that when you flip that switch, that the outcome you're going to get is correct. And that's going to take some work. And so, I think we're going to see this move, let's roll this out and suddenly there's going to be a lot of problems, a lot of pushback that we're going to see. And some of that's going to come from GDPR and others that Sanjeev was mentioning earlier. A lot of it is going to come from internal CSR requirements within companies that are saying, "Hey, hey, whoa, hold up, we can't do this all at once. "Let's take the slow route, "let's make AI automated in a smart way." And that's going to take time. >> Yeah, so a couple of predictions there that I heard. AI simply disappear, it becomes invisible. Maybe if I can restate that. And then if I understand it correctly, Brad you're saying there's a backlash in the near term. You'd be able to say, oh, slow down. Let's automate what we can. Those attributes that you talked about are non trivial to achieve, is that why you're a bit of a skeptic? >> Yeah. I think that we don't have any sort of standards that companies can look to and understand. And we certainly, within these companies, especially those that haven't already stood up an internal data science team, they don't have the knowledge to understand when they flip that switch for an automated AI outcome that it's going to do what they think it's going to do. And so we need some sort of standard methodology and practice, best practices that every company that's going to consume this invisible AI can make use of them. And one of the things that you know, is sort of started that Google kicked off a few years back that's picking up some momentum and the companies I just mentioned are starting to use it is this idea of model cards where at least you have some transparency about what these things are doing. You know, so like for the SAP example, we know, for example, if it's convolutional neural network with a long, short term memory model that it's using, we know that it only works on Roman English and therefore me as a consumer can say, "Oh, well I know that I need to do this internationally. "So I should not just turn this on today." >> Thank you. Carl could you add anything, any context here? >> Yeah, we've talked about some of the things Brad mentioned here at IDC and our future of intelligence group regarding in particular, the moral and legal implications of having a fully automated, you know, AI driven system. Because we already know, and we've seen that AI systems are biased by the data that they get, right? So if they get data that pushes them in a certain direction, I think there was a story last week about an HR system that was recommending promotions for White people over Black people, because in the past, you know, White people were promoted and more productive than Black people, but it had no context as to why which is, you know, because they were being historically discriminated, Black people were being historically discriminated against, but the system doesn't know that. So, you know, you have to be aware of that. And I think that at the very least, there should be controls when a decision has either a moral or legal implication. When you really need a human judgment, it could lay out the options for you. But a person actually needs to authorize that action. And I also think that we always will have to be vigilant regarding the kind of data we use to train our systems to make sure that it doesn't introduce unintended biases. In some extent, they always will. So we'll always be chasing after them. But that's (indistinct). >> Absolutely Carl, yeah. I think that what you have to bear in mind as a consumer of AI is that it is a reflection of us and we are a very flawed species. And so if you look at all of the really fantastic, magical looking supermodels we see like GPT-3 and four, that's coming out, they're xenophobic and hateful because the people that the data that's built upon them and the algorithms and the people that build them are us. So AI is a reflection of us. We need to keep that in mind. >> Yeah, where the AI is biased 'cause humans are biased. All right, great. All right let's move on. Doug you mentioned mentioned, you know, lot of people that said that data lake, that term is not going to live on but here's to be, have some lakes here. You want to talk about lake house, bring it on. >> Yes, I do. My prediction is that lake house and this idea of a combined data warehouse and data lake platform is going to emerge as the dominant data management offering. I say offering that doesn't mean it's going to be the dominant thing that organizations have out there, but it's going to be the pro dominant vendor offering in 2022. Now heading into 2021, we already had Cloudera, Databricks, Microsoft, Snowflake as proponents, in 2021, SAP, Oracle, and several of all of these fabric virtualization/mesh vendors joined the bandwagon. The promise is that you have one platform that manages your structured, unstructured and semi-structured information. And it addresses both the BI analytics needs and the data science needs. The real promise there is simplicity and lower cost. But I think end users have to answer a few questions. The first is, does your organization really have a center of data gravity or is the data highly distributed? Multiple data warehouses, multiple data lakes, on premises, cloud. If it's very distributed and you'd have difficulty consolidating and that's not really a goal for you, then maybe that single platform is unrealistic and not likely to add value to you. You know, also the fabric and virtualization vendors, the mesh idea, that's where if you have this highly distributed situation, that might be a better path forward. The second question, if you are looking at one of these lake house offerings, you are looking at consolidating, simplifying, bringing together to a single platform. You have to make sure that it meets both the warehouse need and the data lake need. So you have vendors like Databricks, Microsoft with Azure Synapse. New really to the data warehouse space and they're having to prove that these data warehouse capabilities on their platforms can meet the scaling requirements, can meet the user and query concurrency requirements. Meet those tight SLS. And then on the other hand, you have the Oracle, SAP, Snowflake, the data warehouse folks coming into the data science world, and they have to prove that they can manage the unstructured information and meet the needs of the data scientists. I'm seeing a lot of the lake house offerings from the warehouse crowd, managing that unstructured information in columns and rows. And some of these vendors, Snowflake a particular is really relying on partners for the data science needs. So you really got to look at a lake house offering and make sure that it meets both the warehouse and the data lake requirement. >> Thank you Doug. Well Tony, if those two worlds are going to come together, as Doug was saying, the analytics and the data science world, does it need to be some kind of semantic layer in between? I don't know. Where are you in on this topic? >> (chuckles) Oh, didn't we talk about data fabrics before? Common metadata layer (chuckles). Actually, I'm almost tempted to say let's declare victory and go home. And that this has actually been going on for a while. I actually agree with, you know, much of what Doug is saying there. Which is that, I mean I remember as far back as I think it was like 2014, I was doing a study. I was still at Ovum, (indistinct) Omdia, looking at all these specialized databases that were coming up and seeing that, you know, there's overlap at the edges. But yet, there was still going to be a reason at the time that you would have, let's say a document database for JSON, you'd have a relational database for transactions and for data warehouse and you had basically something at that time that resembles a dupe for what we consider your data life. Fast forward and the thing is what I was seeing at the time is that you were saying they sort of blending at the edges. That was saying like about five to six years ago. And the lake house is essentially on the current manifestation of that idea. There is a dichotomy in terms of, you know, it's the old argument, do we centralize this all you know in a single place or do we virtualize? And I think it's always going to be a union yeah and there's never going to be a single silver bullet. I do see that there are also going to be questions and these are points that Doug raised. That you know, what do you need for your performance there, or for your free performance characteristics? Do you need for instance high concurrency? You need the ability to do some very sophisticated joins, or is your requirement more to be able to distribute and distribute our processing is, you know, as far as possible to get, you know, to essentially do a kind of a brute force approach. All these approaches are valid based on the use case. I just see that essentially that the lake house is the culmination of it's nothing. It's a relatively new term introduced by Databricks a couple of years ago. This is the culmination of basically what's been a long time trend. And what we see in the cloud is that as we start seeing data warehouses as a check box items say, "Hey, we can basically source data in cloud storage, in S3, "Azure Blob Store, you know, whatever, "as long as it's in certain formats, "like, you know parquet or CSP or something like that." I see that as becoming kind of a checkbox item. So to that extent, I think that the lake house, depending on how you define is already reality. And in some cases, maybe new terminology, but not a whole heck of a lot new under the sun. >> Yeah. And Dave Menninger, I mean a lot of these, thank you Tony, but a lot of this is going to come down to, you know, vendor marketing, right? Some people just kind of co-op the term, we talked about you know, data mesh washing, what are your thoughts on this? (laughing) >> Yeah, so I used the term data platform earlier. And part of the reason I use that term is that it's more vendor neutral. We've tried to sort of stay out of the vendor terminology patenting world, right? Whether the term lake houses, what sticks or not, the concept is certainly going to stick. And we have some data to back it up. About a quarter of organizations that are using data lakes today, already incorporate data warehouse functionality into it. So they consider their data lake house and data warehouse one in the same, about a quarter of organizations, a little less, but about a quarter of organizations feed the data lake from the data warehouse and about a quarter of organizations feed the data warehouse from the data lake. So it's pretty obvious that three quarters of organizations need to bring this stuff together, right? The need is there, the need is apparent. The technology is going to continue to converge. I like to talk about it, you know, you've got data lakes over here at one end, and I'm not going to talk about why people thought data lakes were a bad idea because they thought you just throw stuff in a server and you ignore it, right? That's not what a data lake is. So you've got data lake people over here and you've got database people over here, data warehouse people over here, database vendors are adding data lake capabilities and data lake vendors are adding data warehouse capabilities. So it's obvious that they're going to meet in the middle. I mean, I think it's like Tony says, I think we should declare victory and go home. >> As hell. So just a follow-up on that, so are you saying the specialized lake and the specialized warehouse, do they go away? I mean, Tony data mesh practitioners would say or advocates would say, well, they could all live. It's just a node on the mesh. But based on what Dave just said, are we gona see those all morphed together? >> Well, number one, as I was saying before, there's always going to be this sort of, you know, centrifugal force or this tug of war between do we centralize the data, do we virtualize? And the fact is I don't think that there's ever going to be any single answer. I think in terms of data mesh, data mesh has nothing to do with how you're physically implement the data. You could have a data mesh basically on a data warehouse. It's just that, you know, the difference being is that if we use the same physical data store, but everybody's logically you know, basically governing it differently, you know? Data mesh in space, it's not a technology, it's processes, it's governance process. So essentially, you know, I basically see that, you know, as I was saying before that this is basically the culmination of a long time trend we're essentially seeing a lot of blurring, but there are going to be cases where, for instance, if I need, let's say like, Upserve, I need like high concurrency or something like that. There are certain things that I'm not going to be able to get efficiently get out of a data lake. And, you know, I'm doing a system where I'm just doing really brute forcing very fast file scanning and that type of thing. So I think there always will be some delineations, but I would agree with Dave and with Doug, that we are seeing basically a confluence of requirements that we need to essentially have basically either the element, you know, the ability of a data lake and the data warehouse, these need to come together, so I think. >> I think what we're likely to see is organizations look for a converge platform that can handle both sides for their center of data gravity, the mesh and the fabric virtualization vendors, they're all on board with the idea of this converged platform and they're saying, "Hey, we'll handle all the edge cases "of the stuff that isn't in that center of data gravity "but that is off distributed in a cloud "or at a remote location." So you can have that single platform for the center of your data and then bring in virtualization, mesh, what have you, for reaching out to the distributed data. >> As Dave basically said, people are happy when they virtualized data. >> I think we have at this point, but to Dave Menninger's point, they are converging, Snowflake has introduced support for unstructured data. So obviously literally splitting here. Now what Databricks is saying is that "aha, but it's easy to go from data lake to data warehouse "than it is from databases to data lake." So I think we're getting into semantics, but we're already seeing these two converge. >> So take somebody like AWS has got what? 15 data stores. Are they're going to 15 converge data stores? This is going to be interesting to watch. All right, guys, I'm going to go down and list do like a one, I'm going to one word each and you guys, each of the analyst, if you would just add a very brief sort of course correction for me. So Sanjeev, I mean, governance is going to to be... Maybe it's the dog that wags the tail now. I mean, it's coming to the fore, all this ransomware stuff, which you really didn't talk much about security, but what's the one word in your prediction that you would leave us with on governance? >> It's going to be mainstream. >> Mainstream. Okay. Tony Baer, mesh washing is what I wrote down. That's what we're going to see in 2022, a little reality check, you want to add to that? >> Reality check, 'cause I hope that no vendor jumps the shark and close they're offering a data niche product. >> Yeah, let's hope that doesn't happen. If they do, we're going to call them out. Carl, I mean, graph databases, thank you for sharing some high growth metrics. I know it's early days, but magic is what I took away from that, so magic database. >> Yeah, I would actually, I've said this to people too. I kind of look at it as a Swiss Army knife of data because you can pretty much do anything you want with it. That doesn't mean you should. I mean, there's definitely the case that if you're managing things that are in fixed schematic relationship, probably a relation database is a better choice. There are times when the document database is a better choice. It can handle those things, but maybe not. It may not be the best choice for that use case. But for a great many, especially with the new emerging use cases I listed, it's the best choice. >> Thank you. And Dave Menninger, thank you by the way, for bringing the data in, I like how you supported all your comments with some data points. But streaming data becomes the sort of default paradigm, if you will, what would you add? >> Yeah, I would say think fast, right? That's the world we live in, you got to think fast. >> Think fast, love it. And Brad Shimmin, love it. I mean, on the one hand I was saying, okay, great. I'm afraid I might get disrupted by one of these internet giants who are AI experts. I'm going to be able to buy instead of build AI. But then again, you know, I've got some real issues. There's a potential backlash there. So give us your bumper sticker. >> I'm would say, going with Dave, think fast and also think slow to talk about the book that everyone talks about. I would say really that this is all about trust, trust in the idea of automation and a transparent and visible AI across the enterprise. And verify, verify before you do anything. >> And then Doug Henschen, I mean, I think the trend is your friend here on this prediction with lake house is really becoming dominant. I liked the way you set up that notion of, you know, the data warehouse folks coming at it from the analytics perspective and then you get the data science worlds coming together. I still feel as though there's this piece in the middle that we're missing, but your, your final thoughts will give you the (indistinct). >> I think the idea of consolidation and simplification always prevails. That's why the appeal of a single platform is going to be there. We've already seen that with, you know, DoOP platforms and moving toward cloud, moving toward object storage and object storage, becoming really the common storage point for whether it's a lake or a warehouse. And that second point, I think ESG mandates are going to come in alongside GDPR and things like that to up the ante for good governance. >> Yeah, thank you for calling that out. Okay folks, hey that's all the time that we have here, your experience and depth of understanding on these key issues on data and data management really on point and they were on display today. I want to thank you for your contributions. Really appreciate your time. >> Enjoyed it. >> Thank you. >> Thanks for having me. >> In addition to this video, we're going to be making available transcripts of the discussion. We're going to do clips of this as well we're going to put them out on social media. I'll write this up and publish the discussion on wikibon.com and siliconangle.com. No doubt, several of the analysts on the panel will take the opportunity to publish written content, social commentary or both. I want to thank the power panelists and thanks for watching this special CUBE presentation. This is Dave Vellante, be well and we'll see you next time. (bright music)
SUMMARY :
and I'd like to welcome you to I as moderator, I'm going to and that is the journey to weigh in on there, and it's going to demand more solid data. Brad, I wonder if you that are specific to individual use cases in the past is because we I like the fact that you the data from, you know, Dave Menninger, I mean, one of the things that all need to be managed collectively. Oh thank you Dave. and to the community I think we could have a after the fact to say, okay, is this incremental to the market? the magic it does and to do it and that slows the system down. I know the story, but And that is a problem that the languages move on to Dave Menninger. So in the next say three to five years, the guy who has followed that people still need to do their taxes, And I agree 100% with you and the streaming data as the I mean, when you think about, you know, and because of basically the all of that is fixed, but the it becomes the default? I think around, you know, but it becomes the default. and we're seeing a lot of taking the hardware dimension That'll just happened, Carl. Okay, let's move on to Brad. And that is to say that, Those attributes that you And one of the things that you know, Carl could you add in the past, you know, I think that what you have to bear in mind that term is not going to and the data science needs. and the data science world, You need the ability to do lot of these, thank you Tony, I like to talk about it, you know, It's just a node on the mesh. basically either the element, you know, So you can have that single they virtualized data. "aha, but it's easy to go from I mean, it's coming to the you want to add to that? I hope that no vendor Yeah, let's hope that doesn't happen. I've said this to people too. I like how you supported That's the world we live I mean, on the one hand I And verify, verify before you do anything. I liked the way you set up We've already seen that with, you know, the time that we have here, We're going to do clips of this as well
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Menninger | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Doug Henschen | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Brad Shimmin | PERSON | 0.99+ |
Doug | PERSON | 0.99+ |
Tony Baer | PERSON | 0.99+ |
Dave Velannte | PERSON | 0.99+ |
Tony | PERSON | 0.99+ |
Carl | PERSON | 0.99+ |
Brad | PERSON | 0.99+ |
Carl Olofson | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
2014 | DATE | 0.99+ |
Sanjeev Mohan | PERSON | 0.99+ |
Ventana Research | ORGANIZATION | 0.99+ |
2022 | DATE | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
January of 2022 | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
381 databases | QUANTITY | 0.99+ |
IDC | ORGANIZATION | 0.99+ |
Informatica | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
two | QUANTITY | 0.99+ |
Sanjeev | PERSON | 0.99+ |
2021 | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
Omdia | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
SanjMo | ORGANIZATION | 0.99+ |
79% | QUANTITY | 0.99+ |
second question | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
15 data stores | QUANTITY | 0.99+ |
100% | QUANTITY | 0.99+ |
SAP | ORGANIZATION | 0.99+ |
Som Shahapurkar & Adam Williams, Iron Mountain | AWS re:Invent 2021
(upbeat music) >> We're back at AWS re:Invent 2021. You're watching theCUBE and we're really excited to have Adam Williams on, he's a senior director of engineering at Iron Mountain. Som Shahapurkar, who's the product engineering of vertical solutions at Iron Mountain. Guys, great to see you. Thanks for coming on. >> Thank you >> Thank you. All right Adam, we know Iron Mountain trucks, tapes, what's new? >> What's new. So we've developed a SaaS platform for digitizing, classifying and bringing out and unlocking the value of our customer's data and putting their data to work. The content services platform that we've developed, goes together with an IDP that we call an intelligent document processing capability to do basic content management, but also to do data extraction and to increase workflow capabilities for our customers. >> Yeah, so I was kind of joking before Iron Mountain, the legacy business of course, everybody's seeing the trucks, but $4 billion company, $13 billion market cap, the stock's been on fire. The pandemic obviously has been a tailwind for you guys, but Som, if you had to describe it to like my mother, what's the sound bite that you'd give. >> Well the sound bite, as everyone knows data is gold today, right? And we are sitting figuratively and literally on a mountain of data. And now we have the technology to take that data partner with AWS, the heavy machinery to convert that into value, into value that people can use to complete the human story of healthcare, of mortgage, finance. A lot of this sits in systems, but it also sits in paper. And we are bridging that paper to digital divide, the physical and digital divide to create one story. >> This has been a journey for you guys. I mean, I recall that when you kind of laid this vision out a number of years ago, I think he made some acquisitions. And so maybe take us through that amazing transformation that Iron Mountain has made, but help the audience understand that. >> Transformations really been going from the physical records management that we've built our business around to evolving with our customers, to be able to work with all of the digital documents and not just be a transportation and records management storage company, but to actually work with them, to put their data to work, allowing them to be able to digitize a lot of their content, but also to bring in already digitized content and rich media. >> One of the problems that always existed, especially if you go back to back of my brain, 2006, the federal rules of civil procedure, which said that emails could now be evidence in a case and everyone like, oh, I don't like, how do I find email. So one of the real problems was classifying the information for retention policies. The lawyers wanted to throw everything out after whatever six or seven years, the business people wanted to keep everything forever. Neither of those strategies work, so classification and you couldn't do it manually. So have you guys solved that problem? How do you solve that problem? Does the machine intelligence help? It used to be, I'll use support vector machines or math or probabilistic, latent, semantic, indexing, all kinds of funky stuff. And now we enter this cloud world, have you guys been able to solve that problem and how? >> So our customers already have 20 plus years of retention rules and guidelines that are built within our systems. And we've helped them define those over the years. So we're able to take those records, retention schedules that they have, and then apply them to the documents. But instead of doing that manually, we're able to do that using our classification capabilities with AI ML and that Som's expertise. >> Awesome, so lay it on me. How do you guys do that? It's a lot of math. >> Yeah, so it can get complicated real fast, but at a simple level, what's changed really from support beta machines of 2006 to today is the scale at which we can do it, right? The scale at which we are bringing those technologies. Plus the latest technologies of deep learning, your conventional neural networks going from a bag of characters and words to really the way humans look at it. You look at a document and you know this is an invoice or this is a prescription, you don't have to even know to read to know that, machines are now capable of having that vision, the computer vision to say prescription, invoice. So we train those models and have them do it at industrial scale. >> Yeah, because humans are actually pretty bad at classifying at scale. >> At scale like their back. >> You remember, we used to try to do, oh, it was just tag it, oh, what a nightmare. And then when something changes and so now machines and the cloud and Jane said, how about, I mean, I presume highly regulated industries are the target, but maybe you could talk about the industry solutions a little bit. >> Right. Regulated industries are a challenge, right. Especially when you talk about black box methodologies like AI, where we don't know, okay, why does it classify this as this and that is that? But that's where I think a combined approach of what we are trying to say, composite AI. So the human knowledge, plus AI knowledge combined together to say, okay, we know about these regulations and hey, AI, be cognizant of this regulations while you do our stuff, don't go blindly. So we keep the AI in the guardrails and guided to be within those lines. >> And other part of that is we know our customers really well. We spent a lot of time with them. And so now we're able to take a lot of the challenges they have and go meet those needs with the document classification. But we also go beyond that, allowing them to implement their own workflows within the system, allowing them to be able to define their own capabilities and to be able to take those records into the future and to use our content management system as a true content services platform. >> Okay, take me through the before and the after. So the workflow used to be, I'd ring you up, or maybe you come in and every week grab a box of records, put them in the truck and then stick them in the Iron Mountain. And that was the workflow. And you wanted them back, you'd go get it back and it take awhile. So you've digitized that whole and when you say I'm inferring that the customer can define their own workflow because it's now software defined, right. So that's what you guys have engineered. Some serious engineering work. So what's the tech behind that. Can you paint a picture? >> So the tech behind it is we've run all of our cloud systems and Kubernetes. So using Kubernetes, we can scale really, really large. All of our capabilities are obviously cloud-based, which allows us to be able to scale rapidly. With that we run elastic search is our search engine and MongoDB is our no SQL database. And that allows us to be able to run millions of documents per minute through our system. We have customers that we're doing eight million documents a day for the reel over the process. And they're able to do that with a known level of accuracy. And they can go look at the documents that have had any exceptions. And we can go back to what Som was talking about to go through and retrain models and relabel documents so that we can catch that extra percentage and get it as close to 100% accuracy as we would like, or they would like. >> So what happens? So take me through the customer experience. What is that like? I mean, do they still... we you know the joke, the paperless bathroom will occur before the paperless office, right? So there's still paper in the office, but so what's the workload? I presume a lot of this is digitized at the office, but there's still paper, so help us understand that. >> Customers can take a couple of different paths. One is that we already have the physical documents that they'd like us to scan. We call that backfile scanning. So we already have the documents, they're in a box they're in a record center. We can move them between different records centers and get them imaged in our high volume scanning operation centers. From there-- >> Sorry to interrupt. And at that point, you're auto classifying, right? It's not already classified, I mean, it kind of is manually, but you're going to reclassify it on creation. >> Correct. >> Is that electronic document? >> For some of our customers, we have base metadata that gives us some clues as to what documents may be. But for other documents, we're able to train the models to know if their invoices or if their contracts commonly formatted documents, but customers can also bring in their already digitized content. They can bring in basic PDFs or Word documents or Google Docs for instance, but they can also bring in rich media, such as video and audio. And from there, we also do a speech to text for video and audio, in addition to just basic OCR for documents. >> Public sector, financial services, health care, insurance, I got to imagine that those have got to be the sweet spots. >> Another sweet spot for us is the federal space in public sector. We achieved FedRAMP, which is a major certification to be able to work with, with the federal government. >> Now, how would he work with AWS? What's your relationship with them? How do you use the cloud? Maybe you could describe that a little bit. >> Well, yeah, at multiple levels, right? So of course we use their cloud infrastructure to run our computing because with the AI and machine learning, you need a lot of computing power, right. And AWS is the one who can reliably provide it, space to store the digital data, computing the processes, extract all the information, train our models, and then process these, like he's talking about, we are talking about eight, 12, 16 million documents a day. So now you need seconds and sub second processing times, right? So at different levels, at the company infrastructure level, also the AI and machine learning algorithms levels, AWS has great, like Tesseract is one the ones that everyone knows but there is others purpose-built model APIs that we utilize. And then we'll put our secret sauce on top of that to build that pathway up and make it really compelling. >> And the secret sauce is obviously there's a workflow and the flexibility of the workflow, there's the classification and the machine learning and intelligence and all the engineering that makes the cloud work you manage. What else is there? >> Knowledge graphs, like he was saying, right, the domain. So mortgage is not that a document that looks very similar in mortgage versus a bank stated mortgage and bank statement in healthcare have different meanings. You're looking at different things. So you have something called a knowledge graph that maintains the knowledge of a person working in that field. And then we have those created for different fields and within those fields, different applications and use cases. So that's unique and that's powerful. >> That provides the ability to prior to hierarchy for our customers, so they can trace a document back to the original box that was given to us some many years ago. >> You got that providence and that lineage, I know you're not go to market guys, but conceptually, how do you price? Is it that, it's SaaS? Is it licensed? Is it term? Is it is a consumption based, based on how much I ingest? >> We have varying different pricing models. So we first off we're in six major markets from EU, Latin America, North America and others that we serve. So within those markets, we offer different capabilities. We have an essentials offering on AWS that we've launched in the last two weeks that allows you to be able to bring in base content. And that has a per object pricing. And then from there, we go into our standard edition that has ability to bring in additional workflows and have some custom pricing. And then we have what we call the enterprise. And for enterprise, we look at the customer's problem. We look at custom AI and ML models who might be developing and the solution that we're having to build for them and we provide a custom price and capability for what they need. >> And then the nativists this week announced a new glacier tier. So you guys are all over that. That's where you use it, right? The cheapest and the deepest, right? >> Yeah, one of the major things that AWS provides us as well is the compliance capabilities for our customers. So our customers really require us to have highly secure, highly trusted environments in the cloud. And then the ability to do that with data sovereignty is really important. And so we're able to meet that with AWS as well. >> What do you do in situations where AWS might not have a region? Do you have to find your own data center to do that stuff or? >> Well, so data privacy laws can be really complex. When you work with the customer, we can often find that the nearest data center in their region works, but we also do, we've explored the ability to run cloud capabilities within data centers, within the region that allows us to be able to bridge that. We also do have offerings where we can run on-premise, but obviously our focus here is on the cloud. >> Awesome business. Does Iron Mountain have any competitors? I mean like... >> Yeah. >> You don't have to name them, but I mean, this is awesome business. You've been around for a long time. >> And we found that we have new competitors now that we're in a new business. >> They are trying to disrupt and okay. So you guys are transforming as an incumbent. You're the incumbent disruptor. >> Yes. >> Yes, it's self disruption to some extent, right. Saying, hey, let's broaden our horizon perspective offering value. But I think the key thing is, I want to focus more on the competitive advantage rather than the competitors is that we have the end to end flow, right? From the high volume scanning operations, trucking, the physical world, then up and about into the digital world, right? So you extract it, it's not just PDFs. And then you go into database, machine learnings, unstructured to structured extraction. And then about that value added models. It's not just about classification. Well, now that you have classified and you have all this documents and you have all this data, what can you glean from it? What can you learn about your customers, the customers, customers, and provide them better services. So we are adding value all throughout this chain. And think we are the only ones that can do that full stack. >> That's the real competitive advantage. Guys, really super exciting. Congratulations on getting there. I know it's been a lot of hard work and engineering and way to go. >> Thank you. >> It's fun. >> Dave: It's good, suppose to have you back. >> Thanks. >> All right and thank you for watching. This is Dave Vellante for theCUBE, the leader in live tech coverage. (upbeat music)
SUMMARY :
the product engineering All right Adam, we know and to increase workflow describe it to like my mother, And now we have the I mean, I recall that when you of the digital documents So have you guys solved that problem? and then apply them to the documents. How do you guys do that? of having that vision, Yeah, because humans but maybe you could talk about and guided to be within those lines. and to be able to take those inferring that the customer and get it as close to 100% we you know the joke, One is that we already And at that point, you're And from there, we also have got to be the sweet spots. to be able to work with, How do you use the cloud? And AWS is the one who that makes the cloud work you manage. that maintains the knowledge to prior to hierarchy and others that we serve. So you guys are all over that. And then the ability to do here is on the cloud. Does Iron Mountain have any competitors? You don't have to And we found that we So you guys are transforming Well, now that you have classified That's the real competitive advantage. suppose to have you back. the leader in live tech coverage.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jane | PERSON | 0.99+ |
Adam Williams | PERSON | 0.99+ |
Som Shahapurkar | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
2006 | DATE | 0.99+ |
Dave | PERSON | 0.99+ |
$4 billion | QUANTITY | 0.99+ |
Adam | PERSON | 0.99+ |
$13 billion | QUANTITY | 0.99+ |
20 plus years | QUANTITY | 0.99+ |
six | QUANTITY | 0.99+ |
Iron Mountain | ORGANIZATION | 0.99+ |
Word | TITLE | 0.99+ |
100% | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
North America | LOCATION | 0.99+ |
one | QUANTITY | 0.99+ |
seven years | QUANTITY | 0.99+ |
today | DATE | 0.98+ |
Kubernetes | TITLE | 0.98+ |
Som | PERSON | 0.98+ |
Latin America | LOCATION | 0.98+ |
Google Docs | TITLE | 0.98+ |
EU | LOCATION | 0.98+ |
this week | DATE | 0.98+ |
MongoDB | TITLE | 0.97+ |
first | QUANTITY | 0.96+ |
Iron Mountain | LOCATION | 0.96+ |
six major markets | QUANTITY | 0.95+ |
one story | QUANTITY | 0.93+ |
pandemic | EVENT | 0.9+ |
12, 16 million documents a day | QUANTITY | 0.9+ |
millions of documents per minute | QUANTITY | 0.89+ |
second | QUANTITY | 0.87+ |
eight million documents a day | QUANTITY | 0.87+ |
last two weeks | DATE | 0.85+ |
SQL | TITLE | 0.82+ |
seconds | QUANTITY | 0.77+ |
about | QUANTITY | 0.76+ |
Invent 2021 | EVENT | 0.72+ |
Tesseract | TITLE | 0.72+ |
of years ago | DATE | 0.71+ |
re: | EVENT | 0.7+ |
many years ago | DATE | 0.68+ |
Invent | EVENT | 0.67+ |
rules of civil procedure | TITLE | 0.66+ |
eight | QUANTITY | 0.63+ |
theCUBE | ORGANIZATION | 0.51+ |
2021 | DATE | 0.48+ |
Ryan Mac Ban, UiPath & Michael Engel, PwC | UiPath FORWARD IV
(upbeat music) >> From the Bellagio Hotel in Las Vegas, It's theCUBE. Covering UiPath FORWARD IV. Brought to you by UiPath. >> Welcome back to theCUBE's coverage of UiPath FORWARD IV. Live from the Bellagio, in Las Vegas. I'm Lisa Martin with Dave Vellante. We're here all day today and tomorrow. We're going to talk about process mining next. We've got two guests here. Mike Engel is here, intelligent automation and process intelligence leader at PWC. And Ryan McMahon, the SVP of growth at UiPath. Gentlemen, welcome to the program. >> Thank you, Lisa. >> Thank you. >> So Ryan, I'm going to start with you. Talk to us about process mining. How does UiPath do it differently and what are some of the things being unveiled at this event? >> So look, I would tell you it's actually more than process mining and hopefully, not only you but others saw this this morning with Param. It's really about the full capabilities of that discovery suite. In which, obviously, process mining is part of. But it starts with task capture. So, going out and actually working with subject matter experts on a process. Accounts payable, accounts receivable, order to cash, digitally capturing that process or how they believe it should work or execute across one's environment. Right Mike? And then from there, actually validating or verifying with things or capabilities like process mining. Giving you a full digital x-ray of actually how that process is being executed in the enterprise. Showing you process bottlenecks. For things like accounts payable, showing you days outstanding, maverick buying, so you can actually pin point and do a few things. Fix your process, right? Where process should be fixed. Fix your application because it's probably not doing what you think it is, and then third, and where the value comes, is in our platform of which process mining is a capability, our PA platform. Really moving directly to automations, right? And then, having the ability with even task mining to drill into a specific bottleneck. Capturing keystrokes, clicks, and then moving to, with both of those, process mining and task mining, into Automation Hub, as part of our discovery platform as well. Being able to crowdsource, prioritize, all of those potential, if you will, just capabilities of automations, and saying, "Okay, let's go and prioritize these. These deliver to the greatest value," and executing across them. So, as much as it is about process mining, it's actually the whole entire discovery suite of capabilities that differentiates UiPath from other RPA vendors, as the only RPA vendor that delivers process mining, task mining and this discovery suite as part of our enterprise automation platform. >> Such a critical point, Ryan. I mean, it's multi-dimensional. It's not just one component. It's not just process mining or task mining, it's the combination that's really impactful. Agree with you a hundred percent. >> So, one of the things that people who watch our shows know, I'm like a broken record on this, the early days of RPA, I called it paving the cow path. And that was good because somebody knew the process, they just repeat it. But the problem was, the process wasn't necessarily the best process. As you just described. So, when you guys made the acquisition of ProcessGold, I said, "Okay, now I'm starting to connect the dots," and now a couple years on, we're starting to see that come together. This is what I think is most misunderstood about UiPath, and I wonder, from a practitioner's perspective, if you can sort of fill in some of those gaps. It's that, it's different from a point tool, it's different from a productivity tool. Like Power Automate, I'll just say it, that's running in Azure Cloud, that's cool or a vertically integrated part of some ERP Stack. This is a horizontal play that is end to end. Which is a bigger automation agenda, it's bold but it's potentially huge. $60 billion dollar TAM, I think that's understated. Maybe you could, from a practitioner's perspective, share with us the old way, >> Yeah. >> And kind of, the new way. >> Well obviously, we all made a lot of investments in this space, early on, to determine what should we be automating in the first place? We even went so far as, we have platforms that will transcribe these kind of surveys and discussions that we're having with our clients, right. But at the end of the day all we're learning is what they know about the process. What they as individuals know about the process. And that's problematic. Once we get into the next phase of actually developing something, we miss something, right? Because we're trying to do this rapidly. So, I think what we have now is really this opportunity to have data driven insights and our clients are really grabbing onto that idea, that it's good to have a sense of what they think they do but it's more important to have a sense of what they actually do. >> Are you seeing, in the last year in a half we've seen the acceleration of a lot of things, there's some silver linings but we've also seen the acceleration in automation as a mandate. Where is it? In terms of a priority, that you're seeing with customers, and are there any industries that you're seeing that are really leading the edge here? >> Well I do see it as a priority and of course, in the role that I have, obviously everybody I talk to, it's a priority for them. But I think it's kind of changing. People are understanding that it's not just a sense of, as Ryan was pointing out, it's not just a sense of getting an understanding of what we do today, it's really driving it to that next step of actually getting something impactful out the other end. Clients are starting to understand that. I like to categorize them, there's three types of clients, there's starters, there's stall-ers and those that want to scale. >> Right? So we're seeing a lot more on the other ends of this now, where clients are really getting started and they're getting a good sense that this is important for them because they know that identifying the opportunities in the first place is the most difficult part of automation. That's what's stalling the programs. Then on the other end of the spectrum, we've got these clients that are saying, "Hey, I want to do this really at scale, can you help us do that?" >> (Ryan) Right. >> And it's quite a challenge. >> How do I build a pipeline of automations? So I've had success in finance and accounting, fantastic. How do I take this to operations? How do I take this this to supply chain? How do I take this to HR? And when I do that, it all starts with, as Wendy Batchelder, Chief Data Officer at VMware, would say and as a customer, "It starts with data but more importantly, process." So focusing on process and where we can actually deliver automation. So it's not just about those insights, it's about moving from insights to actionable next steps. >> Right. >> And that is where we're seeing this convergence, if you will, take place. As we've seen it many times before. I mentioned I worked at Cisco in the past, we saw this with Voice Over IP converging on the network. We saw this at VMware, who I know you guys have spoken to multiple times. When a move from a hypervisor to including NSX with the network, to including cloud management and also VSAN for storage, and converging in software. We're seeing it too with process, really. Instead of kids and clipboards, as they used to call it, and many Six Sigma and Lean workshops, with whiteboards and sticky papers, to actually showing people within, really, days how a process is being executed within their organization. And then, suggesting here's where there's automation capabilities, go execute against them. >> So Ryan, this is why sometimes I scoff at the TAM analysis. I get you've got to do the TAM analysis, you've got to communicate to Wall Street. But basically what you do is you pull out IDC or Gartner data, which is very stovepipe, and you kind of say, "Okay we're in this market." It's the convergence of these markets. It's cloud, it's containers, it's IS, it's PaaS, it's Saas, it's blockchain, it's automation. They're all coming together to form this, it sound like a buzzword but this digital matrix, if you will. And it's how well you leverage that digital matrix, which defines your digital business. So, talk about the role that automation, generally, RPA specifically, process mining specifically, play in a digital business. >> Do you want to take that Mike or do you want me to take it? >> We can both do it? How about that? >> Yeah, perfect. >> So I'll start with it. I mean all this is about convergence at this point, right? There are a number of platform providers out there, including UiPath, that are kind of teaching us that. Often times led by the software vendors in terms of how we think of it but what we know is that there's no one solution. We went down the RPA path, lots of clients and got a lot of excitement and a lot of impact but if you really want to drive it broader, what clients are looking at now, is what is the ecosystem of tools that we need to have in place to make that happen? And from our perspective, it's got to start with really, process intelligence. >> What I would say too, if you look at digital transformation, it was usually driven from an application. Right? Really. And what I think customers found was that, "Hey," I'm going to name some folks here, "Put everything in SAP and we'll solve all your problems." Larry Ellison will tell you, "Put everything into Oracle and we'll solve all your problems." Salesforce, now, I'm a salesperson, I've never used an out of the box Salesforce dashboard in my life, to run my business because I want to run it the way I want to run it. Having said that though, they would say the same thing, "Put everything into our platform and we'll make sure that we can access it and you can use it everywhere and we'll solve all of your problems." I think what customers found is that that's not the case. So they said, "Okay, where are there other ways. Yes, I've got my application doing what it's doing, I've improved my process but hang on. There's things that are repeatable here that I can remove to actually focus on higher level orders." And that's where UiPath comes in. We've kind of had a bottom up swell but I would tell you that as we deliver ROI within days or weeks, versus potentially years and with a heavy, heavy investment up front. We're able to do it. We're able to then work with our partners like PWC, to then demonstrate with business process modeling, the ability to do it across all those, as I call, Silo's of excellence in an organization, to deliver true value, in a timeline, with integrated services from our partner, to execute and deliver on ROI. >> You mentioned some of the great software companies that have been created over the years. One you didn't mention but I want you to comment on it is Service Now. Because essentially McDermott's trying to create the platform of platforms. All about workflow and service management. They bought an RPA company, "Hey we got this too." But it's still a walled garden. It's still the same concept is put everything in here. My question is, how are you different? Yeah look, we're going to integrate with customers who want to integrate because we're an open platform and that's the right approach. We believe there will be some overlap and there'll be some choices to be made. Instead of that top down different approach, which may be a little bit heavy and a large investment up front, with varied results, as far as what that looks like, ours is really a bottoms up. I would tell you too, if you look at our community, which is a million and a half, I believe, strong now and growing, it's really about that practitioner and those people that have embraced it from the bottom up that really change how it gets implemented. And you don't have what I used to call the white blood cells, pushing back when you're trying to say, "Hey, let's take it from this finance and accounting to HR, to the supply chain, to the other sides of the organization," saying, "Hey look, be part of this," instead of, "No, you will do." >> Yeah, there's no, at least that I know of, there's no SAP or Salesforce freemium. You can't try it before you buy. And the entry price is way higher. I mean generally. I guess Salesforce not necessarily but I could taste automation for well under $100,000. I could get in for, I bet you most of your customers started at 25 of $50,000 departmental deployments. >> It's a bottoms up ground swell, that's exactly right. And it's really that approach. Which is much more like an Atlassian, I will tell you and it's really getting to the point where we obviously, and I'm saying this, I work at UiPath, we make really good software. And so, out of the box, it's getting easier and easier to use. It all integrates. Which makes it seamless. The reason people move to RPA first was because they got tired of bouncing between applications to do a task. Now we deliver this enterprise automation platform where you can go from process discovery to crowd sourcing and prioritizing your automations with your pipeline of automations, into Studio, into creating those automations, into testing them and back again, right? We give you the opportunity not to leave the platform and extract the most value out of our, what we call enterprise automation platform. Inclusive of process mining. Inclusive of testing and all those capabilities, document understanding, which is also mine, and it's fantastic. It's very differentiated from others that are out there. >> Well it's about having the right framework in place. >> That's it. From an automation perspective. I think that's a little bit different from what you would expect from the SAP's of the world. Mike, where are you seeing, in the large organizations that you work with, we think of what you describe as the automation pipeline, where are some of the key priorities that you're finding in large organizations? What's in that pipeline and in what order? >> It's interesting because every time we have a conversation whether it's internal or with our clients, we come up with another use case for this type of technology. Obviously, when we're having the initial conversations, what we're talking about is really automation. How do we stuff that pipe with automation. But you know, we have clients that are saying, "Hey listen, I'm trying to carve out of a parent company and what I need to do is document all of my processes in a meaningful way, that I can, at some point, take action on, so there's meaningful outcomes." Whether it be a shared services organization that's looking to outsource, all different types of use cases. So, prioritizing is, I think, it's about impact and the quickest way to impact seems to be automation. >> Is it fair to say, can I look at you UiPath as automation infrastructure? Is that okay or do you guys want to say, "Oh, we're an application." The reason I ask, so then you can answer, is if you look at the great infrastructure plays, they all had a role. The DBA, the CCIE from Cisco, the Cloud Architect, the VMware admin, you've been at all of them, Ryan. So, is there a role emerging here and if it's not plumbing or infrastructure, I know, okay that's cool but course correct me on the infrastructure comment and then, is there a role emerging? >> You know, I think the difference between UiPath and some of the infrastructure companies is, it used to take, Dave, years to give an ROI, really. You'd invest in infrastructure and it's like, if we build it they will come. In fact, we've seen this with Cloud, where we kind of started doing some of that on prem, right? We can do this but then you had Amazon, Azure and others kind of take it and say, "Look, we can do it better, faster and cheaper." It's that simple. So, I would say that we are an application and that we reference it as an enterprise automation platform. It's more than infrastructure. Now, are we going to, as I mentioned, integrate to an open platform, to other capabilities? Absolutely. I think, as you see with our investments and as we continue to build this out, starting in core RPA, buying ProcessGold and getting into our discovery suite of capabilities I covered, getting into, what I see next is, as you start launching many bots into your organization, you're touching multiple applications, so you got to test it. Any time you would launch an application you're going to test it before you go live, right? We see another convergence with testing and I know you had Garrett on and Matt, earlier, with testing, application testing, which has been a legacy, kind of dinosaur market, converging with RPA, where you can deliver automations to do it better, faster and cheaper. >> Thank you for that clarification but now Mike, is that role, I know roles are emerging in RPA and automation but is there, I mean, we're seeing centers of excellence pop up, is there an analogy there or sort of a similar- >> Yeah, I think the new role, if you will, it's not super new but it's really that sense of an automation solution architect. It's a whole different thing. We're talking about now more about recombinant innovation. >> Mike: Yeah. >> Than we are about build it from scratch. Because of the convergence of these low-code, no-code types of solutions. It's a different skill set. >> And we see it at PWC. You have somebody who is potentially a process expert but then also somebody who understands automations. It's the convergences of those two, as well, that's a different skill set. It really is. And it's actually bringing those together to get the most value. And we see this across multiple organizations. It starts with a COE. We've done great with our community, so we have that upswell going and then people are saying, "Hang on, I understand process but I also understand automations. let me put the two together," and that's where we get our true value. >> Bringing in the education and training. >> No question. >> That's a huge thing. >> The traditional components of it still need to exist but I think there are new roles that are emerging, for sure. >> It's a big cultural shift. >> Oh absolutely, yeah. >> How do you guys, how does PWC and UiPath, and maybe you each can answer this in the last minute or so, how do you help facilitate that cultural shift in a business that's growing at warp speed, in a market that is very tumultuous? How do you do that? >> Want to go first or I can go? >> I'll go ahead and go first. It's working with great partners like Mike because they see it and they're converging two different practices within their organization to actually bring this value to customers and also that executive relevance. But even on our side, when we're meeting with customers, just in general, we're actually talking about, how do we deal with, there's what? 13 and a half million job openings, I guess, right now and there's 8500 people that are unemployed, is the last number that I heard. We couldn't even fill all of those jobs if we wanted to. So it's like, okay, what is it that we could potentially automate so maybe we don't need all those jobs. And that's not a negative, it's just saying, we couldn't fill them anyway. So let's focus on where we can and where, there again, can extract the most value in working with our partners but create this new domain that's not networking or virtualization but it's actually, potentially, process and automation. It's testing and automation. It might even be security and automation. Which, I will tell you, is probably coming next, having come out of the security space. You know, I sit there and listen to all these threats and I see these people chasing, really, automated threats. It's like, guys a threat hunter that's really good goes through the same 15 steps that they would when they're chasing a false positive, as if a bot would do that for them. >> I mean, I've written about the productivity declines over the past several decades in western countries, it's not universal around the world and maybe we have a productivity boost because of Covid but it's like this perpetual workday now. That's not sustainable. So we're not going to be able to solve the worlds great problems. Whether it's climate change, diversity, massive deaths, on and on and on, unless we deal with that labor gap. >> That's right. >> And the only way to do that is automation. It's so clear to me that that's the answer. Part of the answer. >> It is part of the answer and I think, to your point Lisa, it's a cultural shift that's going to happen whether we want it to or not. When you think about people that are coming into the work force, it's an expectation now. So if you want to retain or you know, attract and retain the right people, you'd better be prepared for it as an organization. >> Yeah, remember the old, proficient in Word and Excel. Makes it almost trivial. It's trivial compared to that. I think if you don't have automation chops, going forward, it's going to be an issue. Hey, we have whatever, 5000 bots running at our company, how could you help? Huh? What's a bot? >> That's right. You're right. We see this too. I'll give you an example at Cisco. One of their financial analysts, junior starter, he says, "Part of our training program, is creating automations. Why? Because it's not just about finance anymore. It's about what can I automate in my role to actually focus on higher level orders and this for me, is just amazing." And you know, it's Rajiv Ramaswamy's son who's over there at Cisco now as a financial analyst. I was sitting on my couch on a Saturday, no kidding, right Dave? And I get a text from Rajiv, who's now CEO at Nutanix, and he says, "I can't believe I just created a bot." And I said, "I'm at the right place." Really. >> That's cool, I mean hey, you're right too. You want to work for Amazon, you got to know how to provision a EC2 instance or you don't get the job. >> Yeah. >> You got to train for that. And these are the types of skills that are expected- >> That's right. >> For the future. >> Awesome. Guys- >> I'm glad I'm older. >> Are you no longer proficient in Word is the question. >> Guys, thanks for joining us, talking about what you guys are doing together, how you're really facilitating this massive growth trajectory. It's great to be back in person and we look forward to hearing from some of your customers later today. >> Terrific. >> Great. >> Thank you for the opportunity. >> Thank you for having us. >> Thank you guys. >> Our pleasure. For Dave Vellante, I'm Lisa Martin, you're watching theCUBE live from the Bellagio in Las Vegas, at UiPath FORWARD IV. Stick around. We'll be back after a short break. (upbeat music)
SUMMARY :
Brought to you by UiPath. And Ryan McMahon, the So Ryan, I'm going to start with you. It's really about the full capabilities it's the combination play that is end to end. idea, that it's good to have that are really leading the edge here? it's really driving it to that next step on the other ends of this now, How do I take this this to supply chain? to including NSX with the network, And it's how well you it's got to start with is that that's not the case. and that's the right approach. I could get in for, I bet you and it's really getting to the right framework in place. we think of what you describe and the quickest way to Is that okay or do you guys want to say, and that we reference it as it's really that sense of Because of the convergence It's the convergences of it still need to exist is the last number that I heard. and maybe we have a productivity that that's the answer. that are coming into the work force, I think if you don't have And I said, "I'm at the or you don't get the job. You got to train for that. in Word is the question. talking about what you from the Bellagio in Las Vegas,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Rajiv | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Ryan | PERSON | 0.99+ |
Ryan McMahon | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Mike Engel | PERSON | 0.99+ |
Larry Ellison | PERSON | 0.99+ |
Mike | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
$60 billion | QUANTITY | 0.99+ |
PWC | ORGANIZATION | 0.99+ |
Wendy Batchelder | PERSON | 0.99+ |
25 | QUANTITY | 0.99+ |
8500 people | QUANTITY | 0.99+ |
5000 bots | QUANTITY | 0.99+ |
15 steps | QUANTITY | 0.99+ |
Word | TITLE | 0.99+ |
UiPath | ORGANIZATION | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Dave | PERSON | 0.99+ |
Rajiv Ramaswamy | PERSON | 0.99+ |
two guests | QUANTITY | 0.99+ |
Matt | PERSON | 0.99+ |
tomorrow | DATE | 0.99+ |
Excel | TITLE | 0.99+ |
two | QUANTITY | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
UiPath | TITLE | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
one component | QUANTITY | 0.98+ |
$50,000 | QUANTITY | 0.98+ |
ProcessGold | ORGANIZATION | 0.98+ |
a million and a half | QUANTITY | 0.98+ |
third | QUANTITY | 0.98+ |
Azure | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
Ryan Mac Ban | PERSON | 0.98+ |
One | QUANTITY | 0.98+ |
Michael Engel | PERSON | 0.98+ |
Atlassian | ORGANIZATION | 0.98+ |
three types | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
Saturday | DATE | 0.97+ |
under $100,000 | QUANTITY | 0.96+ |
Azure Cloud | TITLE | 0.96+ |
EC2 | TITLE | 0.96+ |
each | QUANTITY | 0.95+ |
Salesforce | ORGANIZATION | 0.95+ |
Silo | ORGANIZATION | 0.94+ |
Wall Street | LOCATION | 0.93+ |
Cloud Architect | ORGANIZATION | 0.93+ |
TAM | ORGANIZATION | 0.91+ |
hundred percent | QUANTITY | 0.9+ |
Dr. Eng Lim Goh, HPE | HPE Discover 2021
>>Please >>welcome back to HPD discovered 2021. The cubes virtual coverage, continuous coverage of H P. S H. P. S. Annual customer event. My name is Dave Volonte and we're going to dive into the intersection of high performance computing data and AI with DR Eng limb go who is the senior vice president and CTO for AI Hewlett Packard enterprise Doctor go great to see you again. Welcome back to the cube. >>Hello Dave, Great to talk to you again. >>You might remember last year we talked a lot about swarm intelligence and how AI is evolving. Of course you hosted the day two keynotes here at discover and you talked about thriving in the age of insights and how to craft a data centric strategy. And you addressed you know some of the biggest problems I think organizations face with data that's You got a data is plentiful but insights they're harder to come by. And you really dug into some great examples in retail banking and medicine and health care and media. But stepping back a little bit with zoom out on discovered 21, what do you make of the events so far? And some of your big takeaways? >>Mm Well you started with the insightful question, Right? Yeah, data is everywhere then. But we like the insight. Right? That's also part of the reason why that's the main reason why you know Antonio on day one focused and talked about that. The fact that we are now in the age of insight, right? Uh and uh and and how to thrive thrive in that in this new age. What I then did on the day to kino following Antonio is to talk about the challenges that we need to overcome in order in order to thrive in this new asia. >>So maybe we could talk a little bit about some of the things that you took away in terms I'm specifically interested in some of the barriers to achieving insights when customers are drowning in data. What do you hear from customers? What we take away from some of the ones you talked about today? >>Oh, very pertinent question. Dave You know the two challenges I spoke about right now that we need to overcome in order to thrive in this new age. The first one is is the current challenge and that current challenge is uh you know stated is no barriers to insight. You know when we are awash with data. So that's a statement. Right? How to overcome those barriers. What are the barriers of these two insight when we are awash in data? Um I in the data keynote I spoke about three main things. Three main areas that received from customers. The first one, the first barrier is in many with many of our customers. A data is siloed. All right. You know, like in a big corporation you've got data siloed by sales, finance, engineering, manufacturing, and so on, uh supply chain and so on. And uh there's a major effort ongoing in many corporations to build a federation layer above all those silos so that when you build applications above they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the that was the first barrier. We spoke about barriers to incite when we are washed with data. The second barrier is uh that we see amongst our customers is that uh data is raw and dispersed when they are stored and and uh and you know, it's tough to get tough to to get value out of them. Right? And I in that case I I used the example of uh you know the May 6 2010 event where the stock market dropped a trillion dollars in in tens of minutes. You know, we we all know those who are financially attuned with know about this uh incident, But this is not the only incident. There are many of them out there and for for that particular May six event, uh you know, it took a long time to get insight months. Yeah, before we for months we had no insight as to what happened, why it happened, right. Um, and and there were many other incidences like this and the regulators were looking for that one rule that could, that could mitigate many of these incidences. Um, one of our customers decided to take the hard road to go with the tough data right? Because data is rolling dispersed. So they went into all the different feeds of financial transaction information, took the took the tough took the tough road and analyze that data took a long time to assemble. And they discovered that there was quote stuffing right? That uh people were sending a lot of traits in and then cancelling them almost immediately. You have to manipulate the market. Um And why why why didn't we see it immediately? Well, the reason is the process reports that everybody sees the rule in there that says all trades, less than 100 shares don't need to report in there. And so what people did was sending a lot of less than 103 100 100 shares trades uh to fly under the radar to do this manipulation. So here is here the second barrier right? Data could be raw and dispersed. Um Sometimes you just have to take the hard road and um and to get insight And this is 1 1 great example. And then the last barrier is uh is has to do with sometimes when you start a project to to get insight to get uh to get answers and insight. You you realize that all the datas around you but you don't you don't seem to find the right ones to get what you need. You don't you don't seem to get the right ones. Yeah. Um here we have three quick examples of customers. 111 was it was a great example right? Where uh they were trying to build a language translator, a machine language translator between two languages. Right? By not do that. They need to get hundreds of millions of word pairs, you know, of one language compared uh with a corresponding other hundreds of millions of them. They say, well I'm going to get all these word pairs. Someone creative thought of a willing source. And you thought it was the United Nations, you see. So sometimes you think you don't have the right data with you, but there might be another source. And the willing one that could give you that data Right? The 2nd 1 has to do with uh there was uh the uh sometimes you you may just have to generate that data, interesting one. We had an autonomous car customer that collects all these data from their cars, right? Massive amounts of data, loss of sensors, collect loss of data. And uh, you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car uh in in um in fine weather and collected the car driving on this highway in rain and also in stone, but never had the opportunity to collect the car in hill because that's a rare occurrence. So instead of waiting for a time where the car can dr inhale, they build a simulation you by having the car collector in snow and simulated him. So, these are some of the examples where we have customers working to overcome barriers, right? You have barriers that is associated the fact that data silo the Federated it various associated with data. That's tough to get that. They just took the hard road, right? And, and sometimes, thirdly, you just have to be creative to get the right data. You need, >>wow, I I'll tell you, I have about 100 questions based on what you just said. Uh, there's a great example, the flash crash. In fact, Michael Lewis wrote about this in his book The Flash Boys and essentially right. It was high frequency traders trying to front run the market and sending in small block trades trying to get on the front end it. So that's and they, and they chalked it up to a glitch like you said, for months. Nobody really knew what it was. So technology got us into this problem. I guess my question is, can technology help us get out of the problem? And that maybe is where AI fits in. >>Yes, yes. Uh, in fact, a lot of analytics, we went in to go back to the raw data that is highly dispersed from different sources, right, assemble them to see if you can find a material trend, right? You can see lots of trends, right? Like, uh, you know, we if if humans look at things right, we tend to see patterns in clouds, right? So sometimes you need to apply statistical analysis, um math to to be sure that what the model is seeing is is real. Right? And and that required work. That's one area. The second area is uh you know, when um uh there are times when you you just need to to go through that uh that tough approach to to find the answer. Now, the issue comes to mind now is is that humans put in the rules to decide what goes into a report that everybody sees. And in this case uh before the change in the rules. Right? But by the way, after the discovery, uh authorities change the rules and all all shares, all traits of different any sizes. It has to be reported. No. Yeah. Right. But the rule was applied uh you know, to say earlier that shares under 100 trades under 100 shares need not be reported. So sometimes you just have to understand that reports were decided by humans and and under for understandable reasons. I mean they probably didn't want that for various reasons not to put everything in there so that people could still read it uh in a reasonable amount of time. But uh we need to understand that rules were being put in by humans for the reports we read. And as such there are times you just need to go back to the raw data. >>I want to ask, >>it's gonna be tough. >>Yeah. So I want to ask a question about AI is obviously it's in your title and it's something you know a lot about but and I want to make a statement, you tell me if it's on point or off point. So it seems that most of the Ai going on in the enterprise is modeling data science applied to troves of data but but there's also a lot of ai going on in consumer whether it's you know, fingerprint technology or facial recognition or natural language processing will a two part question will the consumer market as has so often in the enterprise sort of inform us uh the first part and then will there be a shift from sort of modeling if you will to more you mentioned autonomous vehicles more ai influencing in real time. Especially with the edge you can help us understand that better. >>Yeah, it's a great question. Right. Uh there are three stages to just simplify, I mean, you know, it's probably more sophisticated than that but let's simplify three stages. All right. To to building an Ai system that ultimately can predict, make a prediction right or to to assist you in decision making, have an outcome. So you start with the data massive amounts of data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data and the machine uh starts to evolve a model based on all the data is seeing. It starts to evolve right to the point that using a test set of data that you have separately kept a site that you know the answer for. Then you test the model uh you know after you trained it with all that data to see whether it's prediction accuracy is high enough and once you are satisfied with it, you you then deploy the model to make the decision and that's the influence. Right? So a lot of times depend on what what we are focusing on. We we um in data science are we working hard on assembling the right data to feed the machine with, That's the data preparation organization work. And then after which you build your models, you have to pick the right models for the decisions and prediction you wanted to make. You pick the right models and then you start feeding the data with it. Sometimes you you pick one model and the prediction isn't that robust, it is good but then it is not consistent right now. What you do is uh you try another model so sometimes it's just keep trying different models until you get the right kind. Yeah, that gives you a good robust decision making and prediction after which It is tested well Q eight. You would then take that model and deploy it at the edge. Yeah. And then at the edges is essentially just looking at new data, applying it to the model that you have trained and then that model will give you a prediction decision. Right? So uh it is these three stages. Yeah, but more and more uh your question reminds me that more and more people are thinking as the edge become more and more powerful. Can you also do learning at the edge? Right. That's the reason why we spoke about swarm learning the last time, learning at the edge as a swamp, right? Because maybe individually they may not have enough power to do so. But as a swamp they made >>is that learning from the edge? You're learning at the edge? In other words? >>Yes. >>Yeah, I understand the question. Yeah. >>That's a great question. That's a great question. Right? So uh the quick answer is learning at the edge, right? Uh and and also from the edge, but the main goal, right? The goal is to learn at the edge so that you don't have to move the data that the edge sees first back to the cloud or the core to do the learning because that would be the reason. One of the main reasons why you want to learn at the edge, right? Uh So so that you don't need to have to send all that data back and assemble it back from all the different Edge devices, assemble it back to the cloud side to to do the learning right. With someone you can learn it and keep the data at the edge and learn at that point. >>And then maybe only selectively send the autonomous vehicle example you gave us great because maybe there, you know, there may be only persisting, they're not persisting data that is inclement weather or when a deer runs across the front. And then maybe they they do that and then they send that smaller data set back and maybe that's where it's modelling done. But the rest can be done at the edges. It's a new world that's coming down. Let me ask you a question, is there a limit to what data should be collected and how it should be collected? >>That's a great question again, you know uh wow today, full of these uh insightful questions that actually touches on the second challenge. Right? How do we uh in order to thrive in this new age of insight? The second challenge is are you know the is our future challenge, right? What do we do for our future? And and in there is uh the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that I talk about what to collect right? When to organize it when you collect and where will your data be, you know, going forward that you are collecting from? So what, when and where for the what data for the what data to collect? That? That was the question you ask. Um it's it's a question that different industries have to ask themselves because it will vary, right? Um Let me give you the, you use the autonomous car example, let me use that. And We have this customer collecting massive amounts of data. You know, we're talking about 10 petabytes a day from the fleet of their cars. And these are not production autonomous cars, right? These are training autonomous cars, collecting data so they can train and eventually deploy commercial cars. Right? Um, so this data collection cars they collect as a fleet of them collect 10 petabytes a day and when it came to us uh building a storage system yeah, to store all of that data, they realized they don't want to afford to store all of it. Now here comes the dilemma, right? Should what should I after I spent so much effort building all these cars and sensors and collecting data, I've now decide what to delete. That's a dilemma right now in working with them on this process of trimming down what they collected. You know, I'm constantly reminded of the sixties and seventies, right? To remind myself 16 seventies we call a large part of our D. N. A junk DNA. Today we realize that a large part of that what we call john has function as valuable function. They are not jeans, but they regulate the function of jeans, you know? So, so what's jumped in the yesterday could be valuable today or what's junk today could be valuable tomorrow. Right? So, so there's this tension going on right between you decided not wanting to afford to store everything that you can get your hands on. But on the other hand, you you know, you worry you you you ignore the wrong ones, right? You can see this tension in our customers, right? And it depends on industry here. Right? In health care, they say I have no choice. I I want it. All right. One very insightful point brought up by one health care provider that really touched me was, you know, we are not we don't only care. Of course we care a lot. We care a lot about the people we are caring for, right? But you also care for the people were not caring for. How do we find them? Mhm. Right. And that therefore they did not just need to collect data that is uh that they have with from their patients. They also need to reach out right to outside data so that they can figure out who they are not caring for. Right? So they want it all. So I tell us them. So what do you do with funding if you want it all? They say they have no choice but to figure out a way to fund it and perhaps monetization of what they have now is the way to come around and find out. Of course they also come back to us rightfully that, you know, we have to then work out a way to help them build that system, you know, so that health care, right? And and if you go to other industries like banking, they say they can't afford to keep them on, but they are regulated. Seems like healthcare, they are regulated as to uh privacy and such. Like so many examples different industries having different needs but different approaches to how what they collect. But there is this constant tension between um you perhaps deciding not wanting to fund all of that uh all that you can stall right on the other hand, you know, if you if you kind of don't want to afford it and decide not to store some uh if he does some become highly valuable in the future right? Don't worry. >>We can make some assumptions about the future, can't we? I mean, we know there's gonna be a lot more data than than we've ever seen before. We know that we know. Well notwithstanding supply constraints on things like nand, we know the prices of storage is gonna continue to decline. We also know and not a lot of people are really talking about this but the processing power but he says moore's law is dead. Okay, it's waning. But the processing power when you combine the Cpus and N. P. U. S. And Gpus and accelerators and and so forth actually is is increasing. And so when you think about these use cases at the edge, you're going to have much more processing power, you're going to have cheaper storage and it's going to be less expensive processing. And so as an ai practitioner, what can you do with that? >>So the amount of data that's gonna come in, it's gonna we exceed right? Our drop in storage costs are increasing computer power. Right? So what's the answer? Right? So so the the answer must be knowing that we don't and and even the drop in price and increase in bandwidth, it will overwhelm the increased five G will overwhelm five G. Right? Given amount 55 billion of them collecting. Right? So the answer must be that there might need to be a balance between you needing to bring all that data from the 55 billion devices data back to a central as a bunch of central. Cause because you may not be able to afford to do that firstly band with even with five G. M and and SD when you'll still be too expensive given the number of devices out there, Were you given storage costs dropping? You'll still be too expensive to try and store them all. So the answer must be to start at least to mitigate the problem to some leave both a lot of the data out there. Right? And only send back the pertinent ones as you said before. But then if you did that, then how are we gonna do machine learning at the core and the cloud side? If you don't have all the data, you want rich data to train with. Right? Some sometimes you wanna mix of the uh positive type data and the negative type data so you can train the machine in a more balanced way. So the answer must be eventually right. As we move forward with these huge number of devices out of the edge to do machine learning at the edge today, we don't have enough power. Right? The edge typically is characterized by a lower uh energy capability and therefore lower compute power. But soon, you know, even with lower energy they can do more with compute power, improving in energy efficiency, Right? Uh So learning at the edge today we do influence at the edge. So we data model deploy and you do in France at the age, that's what we do today. But more and more I believe given a massive amount of data at the edge, you, you have to have to start doing machine learning at the edge and, and if when you don't have enough power then you aggregate multiple devices, compute power into a swamp and learn as a swan. >>Oh, interesting. So now of course, if, if I were sitting and fly, fly on the wall in hp board meeting, I said okay. HB is as a leading provider of compute how do you take advantage of that? I mean we're going, we're, I know its future, but you must be thinking about that and participating in those markets. I know today you are, you have, you know, edge line and other products. But there's, it seems to me that it's, it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that >>opportunity for the customers? The world will have to have a balance right? Where today the default? Well, the more common mode is to collect the data from the edge and train at uh at some centralized location or a number of centralized location um going forward. Given the proliferation of the edge devices, we'll need a balance. We need both. We need capability at the cloud side. Right? And it has to be hybrid and then we need capability on the edge side. Yeah. That they want to build systems that that on one hand, uh is uh edge adapted, right? Meaning the environmentally adapted because the edge different. They are on a lot of times. On the outside. Uh They need to be packaging adapted and also power adapted, right? Because typically many of these devices are battery power. Right? Um, so you have to build systems that adapt to it. But at the same time they must not be custom. That's my belief. They must be using standard processes and standard operating system so that they can run a rich set of applications. So yes. Um that's that's also the insightful for that Antonio announced in 2018 Uh the next four years from 2018, right $4 billion dollars invested to strengthen our edge portfolio. Edge product lines, Right. Edge solutions. >>I can doctor go, I could go on for hours with you. You're you're just such a great guest. Let's close. What are you most excited about in the future? Of of of it. Certainly H. P. E. But the industry in general. >>Yeah. I think the excitement is uh the customers, right? The diversity of customers and and the diversity in a way they have approached their different problems with data strategy. So the excitement is around data strategy, right? Just like you know uh you know, the the statement made was was so was profound, right? Um And Antonio said we are in the age of insight powered by data. That's the first line, right. Uh The line that comes after that is as such were becoming more and more data centric with data, the currency. Now the next step is even more profound. That is um You know, we are going as far as saying that you know um data should not be treated as cost anymore. No. Right. But instead as an investment in a new asset class called data with value on our balance sheet, this is a this is a step change right? In thinking that is going to change the way we look at data, the way we value it. So that's a statement that this is the exciting thing because because for for me, a city of Ai right uh machine is only as intelligent as the data you feed it with data is a source of the machine learning to be intelligent. So, so that's that's why when when people start to value data, right? And and and say that it is an investment when we collect it, it is very positive for AI because an AI system gets intelligent, get more intelligence because it has a huge amounts of data and the diversity of data. So it would be great if the community values values data. Well, >>you certainly see it in the valuations of many companies these days. Um and I think increasingly you see it on the income statement, you know, data products and people monetizing data services and maybe eventually you'll see it in the in the balance. You know, Doug Laney, when he was a gardener group wrote a book about this and a lot of people are thinking about it. That's a big change, isn't it? Dr >>yeah. Question is is the process and methods evaluation right. But I believe we'll get there, we need to get started and then we'll get there. Believe >>doctor goes on >>pleasure. And yeah. And then the Yeah, I will well benefit greatly from it. >>Oh yeah, no doubt people will better understand how to align you know, some of these technology investments, Doctor goes great to see you again. Thanks so much for coming back in the cube. It's been a real pleasure. >>Yes. A system. It's only as smart as the data you feed it with. >>Excellent. We'll leave it there, thank you for spending some time with us and keep it right there for more great interviews from HP discover 21 this is Dave Volonte for the cube. The leader in enterprise tech coverage right back
SUMMARY :
Hewlett Packard enterprise Doctor go great to see you again. And you addressed you That's also part of the reason why that's the main reason why you know Antonio on day one So maybe we could talk a little bit about some of the things that you The first one is is the current challenge and that current challenge is uh you know stated So that's and they, and they chalked it up to a glitch like you said, is is that humans put in the rules to decide what goes into So it seems that most of the Ai going on in the enterprise is modeling It starts to evolve right to the point that using a test set of data that you have Yeah. The goal is to learn at the edge so that you don't have to move And then maybe only selectively send the autonomous vehicle example you gave us great because But on the other hand, you you know, you worry you you you But the processing power when you combine the Cpus and N. that there might need to be a balance between you needing to bring all that data from the I know today you are, you have, you know, edge line and other products. Um, so you have to build systems that adapt to it. What are you most excited about in the future? machine is only as intelligent as the data you feed it with data Um and I think increasingly you see it on the income statement, you know, data products and people Question is is the process and methods evaluation right. And then the Yeah, I will well benefit greatly from it. Doctor goes great to see you again. It's only as smart as the data you feed it with. We'll leave it there, thank you for spending some time with us and keep it right there for more great interviews
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Michael Lewis | PERSON | 0.99+ |
Dave Volonte | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Doug Laney | PERSON | 0.99+ |
France | LOCATION | 0.99+ |
two languages | QUANTITY | 0.99+ |
The Flash Boys | TITLE | 0.99+ |
55 billion | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
Today | DATE | 0.99+ |
two challenges | QUANTITY | 0.99+ |
second challenge | QUANTITY | 0.99+ |
one language | QUANTITY | 0.99+ |
second area | QUANTITY | 0.99+ |
2021 | DATE | 0.99+ |
May 6 2010 | DATE | 0.99+ |
last year | DATE | 0.99+ |
tomorrow | DATE | 0.99+ |
tens of minutes | QUANTITY | 0.99+ |
HPD | ORGANIZATION | 0.99+ |
less than 100 shares | QUANTITY | 0.99+ |
second barrier | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
first part | QUANTITY | 0.99+ |
Eng Lim Goh | PERSON | 0.99+ |
One | QUANTITY | 0.99+ |
HP | ORGANIZATION | 0.99+ |
first barrier | QUANTITY | 0.99+ |
both | QUANTITY | 0.98+ |
three stages | QUANTITY | 0.98+ |
Hewlett Packard | ORGANIZATION | 0.98+ |
$4 billion dollars | QUANTITY | 0.98+ |
one model | QUANTITY | 0.98+ |
two part | QUANTITY | 0.98+ |
first line | QUANTITY | 0.98+ |
United Nations | ORGANIZATION | 0.98+ |
one area | QUANTITY | 0.98+ |
Antonio | PERSON | 0.98+ |
first one | QUANTITY | 0.98+ |
one rule | QUANTITY | 0.98+ |
hundreds of millions | QUANTITY | 0.98+ |
HPE | ORGANIZATION | 0.97+ |
May six | DATE | 0.97+ |
about 100 questions | QUANTITY | 0.97+ |
john | PERSON | 0.96+ |
two insight | QUANTITY | 0.96+ |
10 petabytes a day | QUANTITY | 0.95+ |
yesterday | DATE | 0.95+ |
asia | LOCATION | 0.92+ |
HB | ORGANIZATION | 0.92+ |
under 100 shares | QUANTITY | 0.92+ |
Three main areas | QUANTITY | 0.92+ |
first | QUANTITY | 0.92+ |
hundreds of millions of word pairs | QUANTITY | 0.91+ |
under 100 trades | QUANTITY | 0.91+ |
less than 103 100 100 shares | QUANTITY | 0.91+ |
Q eight | OTHER | 0.9+ |
three quick examples | QUANTITY | 0.9+ |
two keynotes | QUANTITY | 0.9+ |
55 billion devices | QUANTITY | 0.89+ |
firstly | QUANTITY | 0.88+ |
three main things | QUANTITY | 0.88+ |
Dr. | PERSON | 0.87+ |
day one | QUANTITY | 0.86+ |
H P. S H. P. S. Annual customer | EVENT | 0.85+ |
2nd 1 | QUANTITY | 0.84+ |
Eng limb | PERSON | 0.81+ |
one | QUANTITY | 0.8+ |
16 seventies | QUANTITY | 0.77+ |
a trillion dollars | QUANTITY | 0.74+ |
one health care provider | QUANTITY | 0.73+ |
one of | QUANTITY | 0.72+ |
sixties | QUANTITY | 0.69+ |
DR | PERSON | 0.69+ |
customers | QUANTITY | 0.68+ |
Dr Eng Lim Goh, High Performance Computing & AI | HPE Discover 2021
>>Welcome back to HPD discovered 2021 the cubes virtual coverage, continuous coverage of H P. S H. P. S. Annual customer event. My name is Dave Volonte and we're going to dive into the intersection of high performance computing data and AI with DR Eng limb go who is the senior vice president and CTO for AI at Hewlett Packard enterprise Doctor go great to see you again. Welcome back to the cube. >>Hello Dave, Great to talk to you again. >>You might remember last year we talked a lot about swarm intelligence and how AI is evolving. Of course you hosted the day two keynotes here at discover you talked about thriving in the age of insights and how to craft a data centric strategy and you addressed you know some of the biggest problems I think organizations face with data that's You got a data is plentiful but insights they're harder to come by. And you really dug into some great examples in retail banking and medicine and health care and media. But stepping back a little bit with zoom out on discovered 21, what do you make of the events so far? And some of your big takeaways? >>Mm Well you started with the insightful question, right? Yeah. Data is everywhere then. But we like the insight. Right? That's also part of the reason why that's the main reason why you know Antonio on day one focused and talked about that. The fact that we are now in the age of insight. Right? Uh and and uh and and how to thrive thrive in that in this new age. What I then did on the day to kino following Antonio is to talk about the challenges that we need to overcome in order in order to thrive in this new age. >>So maybe we could talk a little bit about some of the things that you took away in terms I'm specifically interested in some of the barriers to achieving insights when you know customers are drowning in data. What do you hear from customers? What we take away from some of the ones you talked about today? >>Oh, very pertinent question. Dave you know the two challenges I spoke about right now that we need to overcome in order to thrive in this new age. The first one is is the current challenge and that current challenge is uh you know stated is you know, barriers to insight, you know when we are awash with data. So that's a statement right? How to overcome those barriers. What are the barriers of these two insight when we are awash in data? Um I in the data keynote I spoke about three main things. Three main areas that received from customers. The first one, the first barrier is in many with many of our customers. A data is siloed. All right. You know, like in a big corporation you've got data siloed by sales, finance, engineering, manufacturing, and so on, uh supply chain and so on. And uh, there's a major effort ongoing in many corporations to build a federation layer above all those silos so that when you build applications above they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the that was the first barrier we spoke about barriers to incite when we are washed with data. The second barrier is uh, that we see amongst our customers is that uh data is raw and dispersed when they are stored and and uh and you know, it's tough to get tough to to get value out of them. Right? And I in that case I I used the example of uh you know the May 6 2010 event where the stock market dropped a trillion dollars in in tens of ministerial. We we all know those who are financially attuned with know about this uh incident But this is not the only incident. There are many of them out there and for for that particular May six event uh you know, it took a long time to get insight months. Yeah before we for months we had no insight as to what happened, why it happened, right. Um and and there were many other incidences like this. And the regulators were looking for that one rule that could, that could mitigate many of these incidences. Um one of our customers decided to take the hard road go with the tough data right? Because data is rolling dispersed. So they went into all the different feeds of financial transaction information. Uh took the took the tough uh took the tough road and analyze that data took a long time to assemble and they discovered that there was court stuffing right? That uh people were sending a lot of traits in and then cancelling them almost immediately. You have to manipulate the market. Um And why why why didn't we see it immediately? Well the reason is the process reports that everybody sees uh rule in there that says all trades. Less than 100 shares don't need to report in there. And so what people did was sending a lot of less than 103 100 100 shares trades uh to fly under the radar to do this manipulation. So here is here the second barrier right? Data could be raw and dispersed. Um Sometimes you just have to take the hard road and um and to get insight And this is 1 1 great example. And then the last barrier is uh is has to do with sometimes when you start a project to to get insight to get uh to get answers and insight. You you realize that all the datas around you but you don't you don't seem to find the right ones To get what you need. You don't you don't seem to get the right ones. Yeah. Um here we have three quick examples of customers. 111 was it was a great example right? Where uh they were trying to build a language translator, a machine language translator between two languages. Right? But not do that. They need to get hundreds of millions of word pairs, you know, of one language compared uh with the corresponding other hundreds of millions of them. They say we are going to get all these word pairs. Someone creative thought of a willing source and a huge, so it was a United Nations you see. So sometimes you think you don't have the right data with you, but there might be another source and a willing one that could give you that data right. The second one has to do with uh there was uh the uh sometimes you you may just have to generate that data, interesting one. We had an autonomous car customer that collects all these data from their cars, right, massive amounts of data, loss of senses, collect loss of data. And uh you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car uh in in um in fine weather and collected the car driving on this highway in rain and also in stone, but never had the opportunity to collect the car in hale because that's a rare occurrence. So instead of waiting for a time where the car can dr inhale, they build a simulation you by having the car collector in snow and simulated him. So these are some of the examples where we have customers working to overcome barriers, right? You have barriers that is associated the fact that data is silo Federated, it various associated with data. That's tough to get that. They just took the hard road, right? And sometimes, thirdly, you just have to be creative to get the right data you need, >>wow, I tell you, I have about 100 questions based on what you just said. Uh, there's a great example, the flash crash. In fact, Michael Lewis wrote about this in his book, The Flash Boys and essentially right. It was high frequency traders trying to front run the market and sending in small block trades trying to get on the front end it. So that's and they, and they chalked it up to a glitch like you said, for months, nobody really knew what it was. So technology got us into this problem. I guess my question is, can technology help us get out of the problem? And that maybe is where AI fits in. >>Yes, yes. Uh, in fact, a lot of analytics, we went in, uh, to go back to the raw data that is highly dispersed from different sources, right, assemble them to see if you can find a material trend, right? You can see lots of trends right? Like, uh, you know, we, if if humans look at things right, we tend to see patterns in clouds, right? So sometimes you need to apply statistical analysis, um math to be sure that what the model is seeing is is real. Right? And and that required work. That's one area. The second area is uh you know, when um uh there are times when you you just need to to go through that uh that tough approach to to find the answer. Now, the issue comes to mind now is is that humans put in the rules to decide what goes into a report that everybody sees in this case uh before the change in the rules. Right? But by the way, after the discovery, the authorities change the rules and all all shares, all traits of different any sizes. It has to be reported. No. Yeah. Right. But the rule was applied uh you know, to say earlier that shares under 100 trades under 100 shares need not be reported. So sometimes you just have to understand that reports were decided by humans and and under for understandable reasons. I mean they probably didn't want that for various reasons not to put everything in there so that people could still read it uh in a reasonable amount of time. But uh we need to understand that rules were being put in by humans for the reports we read. And as such, there are times you just need to go back to the raw data. >>I want to ask, >>albeit that it's gonna be tough. >>Yeah. So I want to ask a question about AI is obviously it's in your title and it's something you know a lot about but and I want to make a statement, you tell me if it's on point or off point. So it seems that most of the Ai going on in the enterprise is modeling data science applied to troves of data >>but >>but there's also a lot of ai going on in consumer whether it's you know, fingerprint technology or facial recognition or natural language processing. Will a two part question will the consumer market has so often in the enterprise sort of inform us uh the first part and then will there be a shift from sort of modeling if you will to more you mentioned autonomous vehicles more ai influencing in real time. Especially with the edge. She can help us understand that better. >>Yeah, it's a great question. Right. Uh there are three stages to just simplify, I mean, you know, it's probably more sophisticated than that but let's simplify three stages. All right. To to building an Ai system that ultimately can predict, make a prediction right or to to assist you in decision making, have an outcome. So you start with the data massive amounts data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data and the machine uh starts to evolve a model based on all the data is seeing. It starts to evolve right to the point that using a test set of data that you have separately campus site that you know the answer for. Then you test the model uh you know after you trained it with all that data to see whether it's prediction accuracy is high enough and once you are satisfied with it, you you then deploy the model to make the decision and that's the influence. Right? So a lot of times depend on what what we are focusing on. We we um in data science are we working hard on assembling the right data to feed the machine with, That's the data preparation organization work. And then after which you build your models, you have to pick the right models for the decisions and prediction you wanted to make. You pick the right models and then you start feeding the data with it. Sometimes you you pick one model and the prediction isn't that robust, it is good but then it is not consistent right now what you do is uh you try another model so sometimes it's just keep trying different models until you get the right kind. Yeah, that gives you a good robust decision making and prediction after which It is tested well Q eight. You would then take that model and deploy it at the edge. Yeah. And then at the edges is essentially just looking at new data, applying it to the model, you're you're trained and then that model will give you a prediction decision. Right? So uh it is these three stages. Yeah, but more and more uh you know, your question reminds me that more and more people are thinking as the edge become more and more powerful. Can you also do learning at the edge? Right. That's the reason why we spoke about swarm learning the last time, learning at the edge as a swamp, right? Because maybe individually they may not have enough power to do so. But as a swampy me, >>is that learning from the edge or learning at the edge? In other words? Yes. Yeah. Question Yeah. >>That's a great question. That's a great question. Right? So uh the quick answer is learning at the edge, right? Uh and also from the edge, but the main goal, right? The goal is to learn at the edge so that you don't have to move the data that the Edge sees first back to the cloud or the core to do the learning because that would be the reason. One of the main reasons why you want to learn at the edge, right? Uh So so that you don't need to have to send all that data back and assemble it back from all the different edge devices, assemble it back to the cloud side to to do the learning right? With swampland. You can learn it and keep the data at the edge and learn at that point. >>And then maybe only selectively send the autonomous vehicle example you gave us. Great because maybe there, you know, there may be only persisting, they're not persisting data that is inclement weather or when a deer runs across the front and then maybe they they do that and then they send that smaller data set back and maybe that's where it's modelling done. But the rest can be done at the edges. It's a new world that's coming down. Let me ask you a question, is there a limit to what data should be collected and how it should be collected? >>That's a great question again. You know uh wow today, full of these uh insightful questions that actually touches on the second challenge. Right? How do we uh in order to thrive in this new age of inside? The second challenge is are you know the is our future challenge, right? What do we do for our future? And and in there is uh the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that I talk about what to collect right? When to organize it when you collect and then where will your data be, you know going forward that you are collecting from? So what, when and where for the what data for the what data to collect? That? That was the question you ask. Um it's it's a question that different industries have to ask themselves because it will vary, right? Um let me give you the you use the autonomous car example, let me use that. And you have this customer collecting massive amounts of data. You know, we're talking about 10 petabytes a day from the fleet of their cars. And these are not production autonomous cars, right? These are training autonomous cars collecting data so they can train and eventually deploy commercial cars, right? Um so this data collection cars they collect as a fleet of them collect temporal bikes a day. And when it came to us building a storage system to store all of that data, they realized they don't want to afford to store all of it. Now, here comes the dilemma, right? What should I after I spent so much effort building all these cars and sensors and collecting data, I've now decide what to delete. That's a dilemma right now in working with them on this process of trimming down what they collected. You know, I'm constantly reminded of the sixties and seventies, right? To remind myself 60 and seventies, we call a large part of our D. N. A junk DNA. Today. We realize that a large part of that what we call john has function as valuable function. They are not jeans, but they regulate the function of jeans, you know, So, so what's jump in the yesterday could be valuable today or what's junk today could be valuable tomorrow, Right? So, so there's this tension going on right between you decided not wanting to afford to store everything that you can get your hands on. But on the other hand, you you know, you worry you you you ignore the wrong ones, right? You can see this tension in our customers, right? And it depends on industry here, right? In health care, they say I have no choice. I I want it. All right. One very insightful point brought up by one health care provider that really touched me was, you know, we are not we don't only care. Of course we care a lot. We care a lot about the people we are caring for, right? But you also care for the people were not caring for. How do we find them? Mhm. Right. And that therefore, they did not just need to collect data. That is that they have with from their patients. They also need to reach out right to outside data so that they can figure out who they are not caring for, right? So they want it all. So I tell us them, so what do you do with funding if you want it all? They say they have no choice but to figure out a way to fund it and perhaps monetization of what they have now is the way to come around and find that. Of course they also come back to us rightfully that you know, we have to then work out a way to help them build that system, you know? So that's health care, right? And and if you go to other industries like banking, they say they can't afford to keep them off, but they are regulated, seems like healthcare, they are regulated as to uh privacy and such. Like so many examples different industries having different needs, but different approaches to how what they collect. But there is this constant tension between um you perhaps deciding not wanting to fund all of that uh all that you can store, right? But on the other hand, you know, if you if you kind of don't want to afford it and decide not to store some uh if he does some become highly valuable in the future, right? Yeah. >>We can make some assumptions about the future, can't we? I mean, we know there's gonna be a lot more data than than we've ever seen before. We know that we know well notwithstanding supply constraints on things like nand. We know the prices of storage is going to continue to decline. We also know, and not a lot of people are really talking about this but the processing power but he says moore's law is dead okay. It's waning. But the processing power when you combine the Cpus and NP US and GPUS and accelerators and and so forth actually is is increasing. And so when you think about these use cases at the edge, you're going to have much more processing power, you're gonna have cheaper storage and it's going to be less expensive processing And so as an ai practitioner, what can you do with that? >>Yeah, it's highly again, another insightful questions that we touched on our keynote and that that goes up to the why I do the where? Right, When will your data be? Right. We have one estimate that says that by next year there will be 55 billion connected devices out there. Right. 55 billion. Right. What's the population of the world? Of the other? Of 10 billion? But this thing is 55 billion. Right? Uh and many of them, most of them can collect data. So what do you what do you do? Right. Um So the amount of data that's gonna come in, it's gonna weigh exceed right? Our drop in storage costs are increasing computer power. Right? So what's the answer? Right. So, so the the answer must be knowing that we don't and and even the drop in price and increase in bandwidth, it will overwhelm the increased five G will overwhelm five G. Right? Given amount 55 billion of them collecting. Right? So, the answer must be that there might need to be a balance between you needing to bring all that data from the 55 billion devices of data back to a central as a bunch of central Cause because you may not be able to afford to do that firstly band with even with five G. M and and SD when you'll still be too expensive given the number of devices out there. Were you given storage cause dropping will still be too expensive to try and store them all. So the answer must be to start at least to mitigate the problem to some leave both a lot of the data out there. Right? And only send back the pertinent ones as you said before. But then if you did that, then how are we gonna do machine learning at the core and the cloud side? If you don't have all the data you want rich data to train with. Right? Some sometimes you want a mix of the uh positive type data and the negative type data so you can train the machine in a more balanced way. So the answer must be eventually right. As we move forward with these huge number of devices out of the edge to do machine learning at the edge. Today, we don't have enough power. Right? The edge typically is characterized by a lower uh, energy capability and therefore lower compute power. But soon, you know, even with lower energy, they can do more with compute power improving in energy efficiency, Right? Uh, so learning at the edge today, we do influence at the edge. So we data model deploy and you do influence at the age, that's what we do today. But more and more, I believe, given a massive amount of data at the edge, you you have to have to start doing machine learning at the edge. And and if when you don't have enough power, then you aggregate multiple devices, compute power into a swamp and learn as a swan, >>interesting. So now, of course, if I were sitting and fly on the wall in HP board meeting, I said, okay, HP is as a leading provider of compute, how do you take advantage of that? I mean, we're going, I know it's future, but you must be thinking about that and participating in those markets. I know today you are you have, you know, edge line and other products. But there's it seems to me that it's it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that >>opportunity for your customers? Uh the world will have to have a balance right? Where today the default, Well, the more common mode is to collect the data from the edge and train at uh at some centralized location or a number of centralized location um going forward. Given the proliferation of the edge devices, we'll need a balance. We need both. We need capability at the cloud side. Right. And it has to be hybrid. And then we need capability on the edge side. Yeah. That they want to build systems that that on one hand, uh is uh edge adapted, right? Meaning the environmentally adapted because the edge different they are on a lot of times on the outside. Uh They need to be packaging adapted and also power adapted, right? Because typically many of these devices are battery powered. Right? Um so you have to build systems that adapt to it, but at the same time they must not be custom. That's my belief. They must be using standard processes and standard operating system so that they can run rich a set of applications. So yes. Um that's that's also the insightful for that Antonio announced in 2018, Uh the next four years from 2018, right, $4 billion dollars invested to strengthen our edge portfolio, edge product lines, right Edge solutions. >>I get a doctor go. I could go on for hours with you. You're you're just such a great guest. Let's close what are you most excited about in the future of of of it? Certainly H. P. E. But the industry in general. >>Yeah I think the excitement is uh the customers right? The diversity of customers and and the diversity in a way they have approached their different problems with data strategy. So the excitement is around data strategy right? Just like you know uh you know the the statement made was was so was profound. Right? Um And Antonio said we are in the age of insight powered by data. That's the first line right? The line that comes after that is as such were becoming more and more data centric with data the currency. Now the next step is even more profound. That is um you know we are going as far as saying that you know um data should not be treated as cost anymore. No right. But instead as an investment in a new asset class called data with value on our balance sheet, this is a this is a step change right in thinking that is going to change the way we look at data the way we value it. So that's a statement that this is the exciting thing because because for for me a city of AI right uh machine is only as intelligent as the data you feed it with. Data is a source of the machine learning to be intelligent. So so that's that's why when when people start to value data right? And and and say that it is an investment when we collect it. It is very positive for ai because an Ai system gets intelligent, more intelligence because it has a huge amounts of data and the diversity of data. So it'd be great if the community values values data. Well >>you certainly see it in the valuations of many companies these days. Um and I think increasingly you see it on the income statement, you know data products and people monetizing data services and maybe eventually you'll see it in the in the balance. You know Doug Laney when he was a gardener group wrote a book about this and a lot of people are thinking about it. That's a big change isn't it? Dr >>yeah. Question is is the process and methods evaluation. Right. But uh I believe we'll get there, we need to get started then we'll get their belief >>doctor goes on and >>pleasure. And yeah and then the yeah I will will will will benefit greatly from it. >>Oh yeah, no doubt people will better understand how to align you know, some of these technology investments, Doctor goes great to see you again. Thanks so much for coming back in the cube. It's been a real pleasure. >>Yes. A system. It's only as smart as the data you feed it with. >>Excellent. We'll leave it there. Thank you for spending some time with us and keep it right there for more great interviews from HP discover 21. This is dave a lot for the cube. The leader in enterprise tech coverage right back.
SUMMARY :
at Hewlett Packard enterprise Doctor go great to see you again. the age of insights and how to craft a data centric strategy and you addressed you know That's also part of the reason why that's the main reason why you know Antonio on day one So maybe we could talk a little bit about some of the things that you The first one is is the current challenge and that current challenge is uh you know stated So that's and they, and they chalked it up to a glitch like you said, is is that humans put in the rules to decide what goes into So it seems that most of the Ai going on in the enterprise is modeling be a shift from sort of modeling if you will to more you mentioned autonomous It starts to evolve right to the point that using a test set of data that you have is that learning from the edge or learning at the edge? The goal is to learn at the edge so that you don't have to move the data that the And then maybe only selectively send the autonomous vehicle example you gave us. But on the other hand, you know, if you if you kind of don't want to afford it and But the processing power when you combine the Cpus and NP that there might need to be a balance between you needing to bring all that data from the I know today you are you have, you know, edge line and other products. Um so you have to build systems that adapt to it, but at the same time they must not Let's close what are you most excited about in the future of machine is only as intelligent as the data you feed it with. Um and I think increasingly you see it on the income statement, you know data products and Question is is the process and methods evaluation. And yeah and then the yeah I will will will will benefit greatly from it. Doctor goes great to see you again. It's only as smart as the data you feed it with. Thank you for spending some time with us and keep it right there for more great
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Michael Lewis | PERSON | 0.99+ |
Dave Volonte | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
HP | ORGANIZATION | 0.99+ |
two languages | QUANTITY | 0.99+ |
The Flash Boys | TITLE | 0.99+ |
55 billion | QUANTITY | 0.99+ |
10 billion | QUANTITY | 0.99+ |
second challenge | QUANTITY | 0.99+ |
Hewlett Packard | ORGANIZATION | 0.99+ |
two challenges | QUANTITY | 0.99+ |
second area | QUANTITY | 0.99+ |
one language | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
last year | DATE | 0.99+ |
Doug Laney | PERSON | 0.99+ |
tomorrow | DATE | 0.99+ |
next year | DATE | 0.99+ |
both | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
first line | QUANTITY | 0.99+ |
first part | QUANTITY | 0.99+ |
May 6 2010 | DATE | 0.99+ |
$4 billion dollars | QUANTITY | 0.99+ |
two part | QUANTITY | 0.99+ |
Less than 100 shares | QUANTITY | 0.99+ |
HPD | ORGANIZATION | 0.99+ |
one model | QUANTITY | 0.98+ |
one rule | QUANTITY | 0.98+ |
one area | QUANTITY | 0.98+ |
second barrier | QUANTITY | 0.98+ |
60 | QUANTITY | 0.98+ |
55 billion devices | QUANTITY | 0.98+ |
Antonio | PERSON | 0.98+ |
john | PERSON | 0.98+ |
three stages | QUANTITY | 0.98+ |
hundreds of millions | QUANTITY | 0.97+ |
about 100 questions | QUANTITY | 0.97+ |
Eng Lim Goh | PERSON | 0.97+ |
HPE | ORGANIZATION | 0.97+ |
first barrier | QUANTITY | 0.97+ |
first one | QUANTITY | 0.97+ |
Three main areas | QUANTITY | 0.97+ |
yesterday | DATE | 0.96+ |
tens of ministerial | QUANTITY | 0.96+ |
two insight | QUANTITY | 0.96+ |
Q eight | OTHER | 0.95+ |
2021 | DATE | 0.94+ |
seventies | QUANTITY | 0.94+ |
two keynotes | QUANTITY | 0.93+ |
a day | QUANTITY | 0.93+ |
first | QUANTITY | 0.92+ |
H P. S H. P. S. Annual customer | EVENT | 0.91+ |
United Nations | ORGANIZATION | 0.91+ |
less than 103 100 100 shares | QUANTITY | 0.91+ |
under 100 trades | QUANTITY | 0.9+ |
under 100 shares | QUANTITY | 0.9+ |
day one | QUANTITY | 0.88+ |
about 10 petabytes a day | QUANTITY | 0.88+ |
three quick examples | QUANTITY | 0.85+ |
one health care provider | QUANTITY | 0.85+ |
one estimate | QUANTITY | 0.84+ |
three main things | QUANTITY | 0.83+ |
hundreds of millions of word pairs | QUANTITY | 0.82+ |
Antonio | ORGANIZATION | 0.81+ |
sixties | QUANTITY | 0.78+ |
one | QUANTITY | 0.77+ |
May six | DATE | 0.75+ |
firstly | QUANTITY | 0.74+ |
trillion dollars | QUANTITY | 0.73+ |
second one | QUANTITY | 0.71+ |
HP discover 21 | ORGANIZATION | 0.69+ |
DR Eng limb | PERSON | 0.69+ |
one of our customers | QUANTITY | 0.66+ |
Dr Eng Lim Goh, Vice President, CTO, High Performance Computing & AI
(upbeat music) >> Welcome back to HPE Discover 2021, theCube's virtual coverage, continuous coverage of HPE's annual customer event. My name is Dave Vellante and we're going to dive into the intersection of high-performance computing, data and AI with Dr. Eng Lim Goh who's a Senior Vice President and CTO for AI at Hewlett Packard Enterprise. Dr. Goh, great to see you again. Welcome back to theCube. >> Hey, hello, Dave. Great to talk to you again. >> You might remember last year we talked a lot about swarm intelligence and how AI is evolving. Of course you hosted the Day 2 keynotes here at Discover. And you talked about thriving in the age of insights and how to craft a data-centric strategy and you addressed some of the biggest problems I think organizations face with data. And that's, you got to look, data is plentiful, but insights, they're harder to come by and you really dug into some great examples in retail, banking, and medicine and healthcare and media. But stepping back a little bit we'll zoom out on Discover '21, you know, what do you make of the events so far and some of your big takeaways? >> Hmm, well, you started with the insightful question. Data is everywhere then but we lack the insight. That's also part of the reason why that's a main reason why, Antonio on Day 1 focused and talked about that, the fact that we are in the now in the age of insight and how to thrive in this new age. What I then did on the Day 2 keynote following Antonio is to talk about the challenges that we need to overcome in order to thrive in this new age. >> So maybe we could talk a little bit about some of the things that you took away in terms of, I'm specifically interested in some of the barriers to achieving insights when you know customers are drowning in data. What do you hear from customers? What were your takeaway from some of the ones you talked about today? >> Very pertinent question, Dave. You know, the two challenges I spoke about how to, that we need to overcome in order to thrive in this new age, the first one is the current challenge. And that current challenge is, you know state of this, you know, barriers to insight, when we are awash with data. So that's a statement. How to overcome those barriers. One of the barriers to insight when we are awash in data, in the Day 2 keynote, I spoke about three main things, three main areas that receive from customers. The first one, the first barrier is with many of our customers, data is siloed. You know, like in a big corporation, you've got data siloed by sales, finance, engineering, manufacturing, and so on supply chain and so on. And there's a major effort ongoing in many corporations to build a Federation layer above all those silos so that when you build applications above they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the first barrier we spoke about, you know, barriers to insight when we are awash with data. The second barrier is that we see amongst our customers is that data is raw and disperse when they are stored. And it's tough to get to value out of them. In that case I use the example of the May 6, 2010 event where the stock market dropped a trillion dollars in tens of minutes. We all know those who are financially attuned with, know about this incident. But that this is not the only incident. There are many of them out there. And for that particular May 6, event, you know it took a long time to get insight, months, yeah, before we, for months we had no insight as to what happened, why it happened. And there were many other incidences like this and the regulators were looking for that one rule that could mitigate many of these incidences. One of our customers decided to take the hard road to go with the tough data. Because data is raw and dispersed. So they went into all the different feeds of financial transaction information, took the tough, you know, took a tough road and analyze that data took a long time to assemble. And he discovered that there was quote stuffing. That people were sending a lot of trades in and then canceling them almost immediately. You have to manipulate the market. And why didn't we see it immediately? Well, the reason is the process reports that everybody sees had the rule in there that says all trades less than 100 shares don't need to report in there. And so what people did was sending a lot of less than 100 shares trades to fly under the radar to do this manipulation. So here is, here the second barrier. Data could be raw and disperse. Sometimes it's just have to take the hard road and to get insight. And this is one great example. And then the last barrier has to do with sometimes when you start a project to get insight, to get answers and insight, you realize that all the data's around you, but you don't seem to find the right ones to get what you need. You don't seem to get the right ones, yeah. Here we have three quick examples of customers. One was a great example where they were trying to build a language translator a machine language translator between two languages. But in order to do that they need to get hundreds of millions of word pairs of one language compare with the corresponding other hundreds of millions of them. They say, "Where I'm going to get all these word pairs?" Someone creative thought of a willing source and huge source, it was a United Nations. You see, so sometimes you think you don't have the right data with you, but there might be another source and a willing one that could give you that data. The second one has to do with, there was the, sometimes you may just have to generate that data. Interesting one. We had an autonomous car customer that collects all these data from their cars. Massive amounts of data, lots of sensors, collect lots of data. And, you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car in fine weather and collected the car driving on this highway in rain and also in snow. But never had the opportunity to collect the car in hail because that's a rare occurrence. So instead of waiting for a time where the car can drive in hail, they build a simulation by having the car collected in snow and simulated hail. So these are some of the examples where we have customers working to overcome barriers. You have barriers that is associated with the fact, that data silo, if federated barriers associated with data that's tough to get at. They just took the hard road. And sometimes thirdly, you just have to be creative to get the right data you need. >> Wow, I tell you, I have about 100 questions based on what you just said. And as a great example, the flash crash in fact Michael Lewis wrote about this in his book, the "Flash Boys" and essentially. It was high frequency traders trying to front run the market and sending in small block trades trying to get sort of front ended. So that's, and they chalked it up to a glitch. Like you said, for months, nobody really knew what it was. So technology got us into this problem. Can I guess my question is can technology help us get get out of the problem? And that maybe is where AI fits in. >> Yes. Yes. In fact, a lot of analytics work went in to go back to the raw data that is highly dispersed from different sources, assemble them to see if you can find a material trend. You can see lots of trends. Like, no, we, if humans at things we tend to see patterns in clouds. So sometimes you need to apply statistical analysis, math to be sure that what the model is seeing is real. And that required work. That's one area. The second area is, you know, when this, there are times when you just need to go through that tough approach to find the answer. Now, the issue comes to mind now is that humans put in the rules to decide what goes into a report that everybody sees. And in this case before the change in the rules. By the way, after the discovery, the authorities changed the rules and all shares all trades of different, any sizes it has to be reported. Not, yeah. But the rule was applied to to say earlier that shares under 100, trades under 100 shares need not be reported. So sometimes you just have to understand that reports were decided by humans and for understandable reasons. I mean, they probably didn't, wanted for various reasons not to put everything in there so that people could still read it in a reasonable amount of time. But we need to understand that rules were being put in by humans for the reports we read. And as such there are times we just need to go back to the raw data. >> I want to ask you-- Or be it that it's going to be tough there. >> Yeah, so I want to ask you a question about AI as obviously it's in your title and it's something you know a lot about and I'm going to make a statement. You tell me if it's on point or off point. Seems that most of the AI going on in the enterprise is modeling data science applied to troves of data. But there's also a lot of AI going on in consumer, whether it's fingerprint technology or facial recognition or natural language processing. Will, to two-part question, will the consumer market, let's say as it has so often in the enterprise sort of inform us is sort of first part. And then will there be a shift from sort of modeling, if you will, to more, you mentioned autonomous vehicles more AI inferencing in real-time, especially with the Edge. I think you can help us understand that better. >> Yeah, this is a great question. There are three stages to just simplify, I mean, you know, it's probably more sophisticated than that, but let's just simplify there're three stages to building an AI system that ultimately can predict, make a prediction. Or to assist you in decision-making, have an outcome. So you start with the data, massive amounts of data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data. And the machine starts to evolve a model based on all the data is seeing it starts to evolve. To a point that using a test set of data that you have separately kept a site that you know the answer for. Then you test the model, you know after you're trained it with all that data to see whether his prediction accuracy is high enough. And once you are satisfied with it, you then deploy the model to make the decision and that's the inference. So a lot of times depending on what we are focusing on. We in data science are we working hard on assembling the right data to feed the machine with? That's the data preparation organization work. And then after which you build your models you have to pick the right models for the decisions and prediction you wanted to make. You pick the right models and then you start feeding the data with it. Sometimes you pick one model and a prediction isn't that a robust, it is good, but then it is not consistent. Now what you do is you try another model. So sometimes you just keep trying different models until you get the right kind, yeah, that gives you a good robust decision-making and prediction. Now, after which, if it's tested well, Q8 you will then take that model and deploy it at the Edge, yeah. And then at the Edge is essentially just looking at new data applying it to the model that you have trained and then that model will give you a prediction or a decision. So it is these three stages, yeah. But more and more, your question reminds me that more and more people are thinking as the Edge become more and more powerful, can you also do learning at the Edge? That's the reason why we spoke about swarm learning the last time, learning at the Edge as a swarm. Because maybe individually they may not have enough power to do so, but as a swarm, they may. >> Is that learning from the Edge or learning at the Edge. In other words, is it-- >> Yes. >> Yeah, you don't understand my question, yeah. >> That's a great question. That's a great question. So answer is learning at the Edge, and also from the Edge, but the main goal, the goal is to learn at the Edge so that you don't have to move the data that Edge sees first back to the Cloud or the call to do the learning. Because that would be the reason, one of the main reasons why you want to learn at the Edge. So that you don't need to have to send all that data back and assemble it back from all the different Edge devices assemble it back to the Cloud side to do the learning. With swarm learning, you can learn it and keep the data at the Edge and learn at that point, yeah. >> And then maybe only selectively send the autonomous vehicle example you gave is great 'cause maybe they're, you know, there may be only persisting. They're not persisting data that is an inclement weather, or when a deer runs across the front and then maybe they do that and then they send that smaller data set back and maybe that's where it's modeling done but the rest can be done at the Edge. It's a new world that's coming to, let me ask you a question. Is there a limit to what data should be collected and how it should be collected? >> That's a great question again, yeah, well, today full of these insightful questions that actually touches on the second challenge. How do we, to in order to thrive in this new age of insight. The second challenge is our future challenge. What do we do for our future? And in there is the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that, I talk about what to collect, and when to organize it when you collect, and then where will your data be going forward that you are collecting from? So what, when, and where. For the what data, for what data to collect that was the question you asked. It's a question that different industries have to ask themselves because it will vary. Let me give you the, you use the autonomous car example. Let me use that and you have this customer collecting massive amounts of data. You know, we talking about 10 petabytes a day from a fleet of their cars and these are not production autonomous cars. These are training autonomous cars, collecting data so they can train and eventually deploy a commercial cars. Also these data collection cars, they collect 10 as a fleet of them collect 10 petabytes a day. And then when it came to us, building a storage system to store all of that data they realize they don't want to afford to store all of it. Now here comes the dilemma. What should I, after I spent so much effort building all this cars and sensors and collecting data, I've now decide what to delete. That's a dilemma. Now in working with them on this process of trimming down what they collected. I'm constantly reminded of the 60s and 70s. To remind myself 60s and 70s, we call a large part of our DNA, junk DNA. Today we realized that a large part of that, what we call junk has function has valuable function. They are not genes but they regulate the function of genes. So what's junk in yesterday could be valuable today, or what's junk today could be valuable tomorrow. So there's this tension going on between you deciding not wanting to afford to store everything that you can get your hands on. But on the other hand, you know you worry, you ignore the wrong ones. You can see this tension in our customers. And then it depends on industry here. In healthcare they say, I have no choice. I want it all, why? One very insightful point brought up by one healthcare provider that really touched me was you know, we are not, we don't only care. Of course we care a lot. We care a lot about the people we are caring for. But we also care for the people we are not caring for. How do we find them? And therefore, they did not just need to collect data that they have with, from their patients they also need to reach out to outside data so that they can figure out who they are not caring for. So they want it all. So I asked them, "So what do you do with funding if you want it all?" They say they have no choice but they'll figure out a way to fund it and perhaps monetization of what they have now is the way to come around and fund that. Of course, they also come back to us, rightfully that you know, we have to then work out a way to to help them build a system. So that healthcare. And if you go to other industries like banking, they say they can afford to keep them all. But they are regulated same like healthcare. They are regulated as to privacy and such like. So many examples, different industries having different needs but different approaches to how, what they collect. But there is this constant tension between you perhaps deciding not wanting to fund all of that, all that you can store. But on the other hand you know, if you kind of don't want to afford it and decide not to store some, maybe those some become highly valuable in the future. You worry. >> Well, we can make some assumptions about the future, can't we? I mean we know there's going to be a lot more data than we've ever seen before, we know that. We know, well not withstanding supply constraints and things like NAND. We know the price of storage is going to continue to decline. We also know and not a lot of people are really talking about this but the processing power, everybody says, Moore's Law is dead. Okay, it's waning but the processing power when you combine the CPUs and NPUs, and GPUs and accelerators and so forth, actually is increasing. And so when you think about these use cases at the Edge you're going to have much more processing power. You're going to have cheaper storage and it's going to be less expensive processing. And so as an AI practitioner, what can you do with that? >> Yeah, it's a highly, again another insightful question that we touched on, on our keynote and that goes up to the why, I'll do the where. Where will your data be? We have one estimate that says that by next year, there will be 55 billion connected devices out there. 55 billion. What's the population of the world? Well, off the order of 10 billion, but this thing is 55 billion. And many of them, most of them can collect data. So what do you do? So the amount of data that's going to come in is going to way exceed our drop in storage costs our increasing compute power. So what's the answer? The answer must be knowing that we don't and even a drop in price and increase in bandwidth, it will overwhelm the 5G, it'll will overwhelm 5G, given the amount of 55 billion of them collecting. So the answer must be that there needs to be a balance between you needing to bring all that data from the 55 billion devices of the data back out to a central, as a bunch of central cost because you may not be able to afford to do that. Firstly bandwidth, even with 5G and as the, when you still be too expensive given the number of devices out there. You know given storage costs dropping it'll still be too expensive to try and install them all. So the answer must be to start at least to mitigate the problem to some leave most a lot of the data out there. And only send back the pertinent ones, as you said before. But then if you did that then, how are we going to do machine learning at the core and the Cloud side, if you don't have all the data you want rich data to train with. Sometimes you want to a mix of the positive type data, and the negative type data. So you can train the machine in a more balanced way. So the answer must be you eventually, as we move forward with these huge number of devices are at the Edge to do machine learning at the Edge. Today we don't even have power. The Edge typically is characterized by a lower energy capability and therefore, lower compute power. But soon, you know, even with low energy, they can do more with compute power, improving in energy efficiency. So learning at the Edge today we do inference at the Edge. So we data, model, deploy and you do inference at age. That's what we do today. But more and more, I believe given a massive amount of data at the Edge you have to have to start doing machine learning at the Edge. And if when you don't have enough power then you aggregate multiple devices' compute power into a swarm and learn as a swarm. >> Oh, interesting, so now of course, if I were sitting in a flyer flying the wall on HPE Board meeting I said, "Okay, HPE is a leading provider of compute." How do you take advantage that? I mean, we're going, I know it's future but you must be thinking about that and participating in those markets. I know today you are, you have, you know, Edge line and other products, but there's, it seems to me that it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that opportunity for your customers? >> The wall will have to have a balance. Where today the default, well, the more common mode is to collect the data from the Edge and train at some centralized location or number of centralized location. Going forward, given the proliferation of the Edge devices, we'll need a balance, we need both. We need capability at the Cloud side. And it has to be hybrid. And then we need capability on the Edge side. Yeah that we need to build systems that on one hand is Edge-adapted. Meaning they environmentally-adapted because the Edge differently are on it. A lot of times on the outside, they need to be packaging-adapted and also power-adapted. Because typically many of these devices are battery-powered. So you have to build systems that adapts to it. But at the same time, they must not be custom. That's my belief. They must be using standard processes and standard operating system so that they can run a rich set of applications. So yes, that's also the insightful for that. Antonio announced in 2018 for the next four years from 2018, $4 billion invested to strengthen our Edge portfolio our Edge product lines, Edge solutions. >> Dr. Goh, I could go on for hours with you. You're just such a great guest. Let's close. What are you most excited about in the future of certainly HPE, but the industry in general? >> Yeah, I think the excitement is the customers. The diversity of customers and the diversity in the way they have approached their different problems with data strategy. So the excitement is around data strategy. Just like, you know, the statement made for us was so, was profound. And Antonio said we are in the age of insight powered by data. That's the first line. The line that comes after that is as such we are becoming more and more data-centric with data the currency. Now the next step is even more profound. That is, you know, we are going as far as saying that data should not be treated as cost anymore, no. But instead, as an investment in a new asset class called data with value on our balance sheet. This is a step change in thinking that is going to change the way we look at data, the way we value it. So that's a statement. So this is the exciting thing, because for me a CTO of AI, a machine is only as intelligent as the data you feed it with. Data is a source of the machine learning to be intelligent. So that's why when the people start to value data and say that it is an investment when we collect it it is very positive for AI because an AI system gets intelligent, get more intelligence because it has huge amounts of data and a diversity of data. So it'd be great if the community values data. >> Well, are you certainly see it in the valuations of many companies these days? And I think increasingly you see it on the income statement, you know data products and people monetizing data services, and yeah, maybe eventually you'll see it in the balance sheet, I know. Doug Laney when he was at Gartner Group wrote a book about this and a lot of people are thinking about it. That's a big change, isn't it? Dr. Goh. >> Yeah, yeah, yeah. Your question is the process and methods in valuation. But I believe we'll get there. We need to get started and then we'll get there, I believe, yeah. >> Dr. Goh it's always my pleasure. >> And then the AI will benefit greatly from it. >> Oh yeah, no doubt. People will better understand how to align some of these technology investments. Dr. Goh, great to see you again. Thanks so much for coming back in theCube. It's been a real pleasure. >> Yes, a system is only as smart as the data you feed it with. (both chuckling) >> Well, excellent, we'll leave it there. Thank you for spending some time with us so keep it right there for more great interviews from HPE Discover '21. This is Dave Vellante for theCube, the leader in enterprise tech coverage. We'll be right back (upbeat music)
SUMMARY :
Dr. Goh, great to see you again. Great to talk to you again. and you addressed some and how to thrive in this new age. of the ones you talked about today? One of the barriers to insight And as a great example, the flash crash is that humans put in the rules to decide that it's going to be tough there. and it's something you know a lot about And the machine starts to evolve a model Is that learning from the Yeah, you don't So that you don't need to have but the rest can be done at the Edge. But on the other hand you know, And so when you think about and the Cloud side, if you I know today you are, you So you have to build about in the future as the data you feed it with. And I think increasingly you Your question is the process And then the AI will Dr. Goh, great to see you again. as the data you feed it with. Thank you for spending some time with us
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Michael Lewis | PERSON | 0.99+ |
Doug Laney | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Antonio | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
10 billion | QUANTITY | 0.99+ |
$4 billion | QUANTITY | 0.99+ |
second challenge | QUANTITY | 0.99+ |
55 billion | QUANTITY | 0.99+ |
two languages | QUANTITY | 0.99+ |
two challenges | QUANTITY | 0.99+ |
May 6 | DATE | 0.99+ |
Flash Boys | TITLE | 0.99+ |
two-part | QUANTITY | 0.99+ |
55 billion | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
Gartner Group | ORGANIZATION | 0.99+ |
second area | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
last year | DATE | 0.99+ |
less than 100 shares | QUANTITY | 0.99+ |
hundreds of millions | QUANTITY | 0.99+ |
first line | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
second barrier | QUANTITY | 0.99+ |
May 6, 2010 | DATE | 0.99+ |
10 | QUANTITY | 0.99+ |
first barrier | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
less than 100 share | QUANTITY | 0.99+ |
Dr. | PERSON | 0.99+ |
one model | QUANTITY | 0.99+ |
tens of minutes | QUANTITY | 0.98+ |
one area | QUANTITY | 0.98+ |
one language | QUANTITY | 0.98+ |
Edge | ORGANIZATION | 0.98+ |
three stages | QUANTITY | 0.98+ |
yesterday | DATE | 0.98+ |
first part | QUANTITY | 0.98+ |
one rule | QUANTITY | 0.98+ |
Goh | PERSON | 0.98+ |
Firstly | QUANTITY | 0.98+ |
first one | QUANTITY | 0.97+ |
United Nations | ORGANIZATION | 0.97+ |
first | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
first barrier | QUANTITY | 0.97+ |
Hewlett Packard Enterprise | ORGANIZATION | 0.96+ |
about 100 questions | QUANTITY | 0.96+ |
10 petabytes a day | QUANTITY | 0.95+ |
Day 2 | QUANTITY | 0.94+ |
Eng Lim Goh | PERSON | 0.94+ |
Day 1 | QUANTITY | 0.93+ |
under 100 | QUANTITY | 0.92+ |
Dr | PERSON | 0.92+ |
one estimate | QUANTITY | 0.91+ |
Dr Eng Lim Goh, Vice President, CTO, High Performance Computing & AI
(upbeat music) >> Welcome back to HPE Discover 2021, theCUBE's virtual coverage, continuous coverage of HPE's Annual Customer Event. My name is Dave Vellante, and we're going to dive into the intersection of high-performance computing, data and AI with Doctor Eng Lim Goh, who's a Senior Vice President and CTO for AI at Hewlett Packard Enterprise. Doctor Goh, great to see you again. Welcome back to theCUBE. >> Hello, Dave, great to talk to you again. >> You might remember last year we talked a lot about Swarm intelligence and how AI is evolving. Of course, you hosted the Day 2 Keynotes here at Discover. And you talked about thriving in the age of insights, and how to craft a data-centric strategy. And you addressed some of the biggest problems, I think organizations face with data. That's, you've got a, data is plentiful, but insights, they're harder to come by. >> Yeah. >> And you really dug into some great examples in retail, banking, in medicine, healthcare and media. But stepping back a little bit we zoomed out on Discover '21. What do you make of the events so far and some of your big takeaways? >> Hmm, well, we started with the insightful question, right, yeah? Data is everywhere then, but we lack the insight. That's also part of the reason why, that's a main reason why Antonio on day one focused and talked about the fact that we are in the now in the age of insight, right? And how to try thrive in that age, in this new age? What I then did on a Day 2 Keynote following Antonio is to talk about the challenges that we need to overcome in order to thrive in this new age. >> So, maybe we could talk a little bit about some of the things that you took away in terms of, I'm specifically interested in some of the barriers to achieving insights. You know customers are drowning in data. What do you hear from customers? What were your takeaway from some of the ones you talked about today? >> Oh, very pertinent question, Dave. You know the two challenges I spoke about, that we need to overcome in order to thrive in this new age. The first one is the current challenge. And that current challenge is, you know, stated is now barriers to insight, when we are awash with data. So that's a statement on how do you overcome those barriers? What are the barriers to insight when we are awash in data? In the Day 2 Keynote, I spoke about three main things. Three main areas that we receive from customers. The first one, the first barrier is in many, with many of our customers, data is siloed, all right. You know, like in a big corporation, you've got data siloed by sales, finance, engineering, manufacturing and so on supply chain and so on. And there's a major effort ongoing in many corporations to build a federation layer above all those silos so that when you build applications above, they can be more intelligent. They can have access to all the different silos of data to get better intelligence and more intelligent applications built. So that was the first barrier we spoke about, you know? Barriers to insight when we are awash with data. The second barrier is that we see amongst our customers is that data is raw and disperse when they are stored. And you know, it's tough to get at, to tough to get a value out of them, right? And in that case, I use the example of, you know, the May 6, 2010 event where the stock market dropped a trillion dollars in terms of minutes. We all know those who are financially attuned with know about this incident but that this is not the only incident. There are many of them out there. And for that particular May 6 event, you know, it took a long time to get insight. Months, yeah, before we, for months we had no insight as to what happened. Why it happened? Right, and there were many other incidences like this and the regulators were looking for that one rule that could mitigate many of these incidences. One of our customers decided to take the hard road they go with the tough data, right? Because data is raw and dispersed. So they went into all the different feeds of financial transaction information, took the tough, you know, took a tough road. And analyze that data took a long time to assemble. And they discovered that there was caught stuffing, right? That people were sending a lot of trades in and then canceling them almost immediately. You have to manipulate the market. And why didn't we see it immediately? Well, the reason is the process reports that everybody sees, the rule in there that says, all trades less than a hundred shares don't need to report in there. And so what people did was sending a lot of less than a hundred shares trades to fly under the radar to do this manipulation. So here is the second barrier, right? Data could be raw and dispersed. Sometimes it's just have to take the hard road and to get insight. And this is one great example. And then the last barrier has to do with sometimes when you start a project to get insight, to get answers and insight, you realize that all the data's around you, but you don't seem to find the right ones to get what you need. You don't seem to get the right ones, yeah? Here we have three quick examples of customers. One was a great example, right? Where they were trying to build a language translator or machine language translator between two languages, right? By not do that, they need to get hundreds of millions of word pairs. You know of one language compare with the corresponding other. Hundreds of millions of them. They say, well, I'm going to get all these word pairs. Someone creative thought of a willing source and a huge, it was a United Nations. You see? So sometimes you think you don't have the right data with you, but there might be another source and a willing one that could give you that data, right? The second one has to do with, there was the sometimes you may just have to generate that data. Interesting one, we had an autonomous car customer that collects all these data from their their cars, right? Massive amounts of data, lots of sensors, collect lots of data. And, you know, but sometimes they don't have the data they need even after collection. For example, they may have collected the data with a car in fine weather and collected the car driving on this highway in rain and also in snow. But never had the opportunity to collect the car in hill because that's a rare occurrence. So instead of waiting for a time where the car can drive in hill, they build a simulation by having the car collected in snow and simulated him. So these are some of the examples where we have customers working to overcome barriers, right? You have barriers that is associated. In fact, that data silo, they federated it. Virus associated with data, that's tough to get at. They just took the hard road, right? And sometimes thirdly, you just have to be creative to get the right data you need. >> Wow! I tell you, I have about a hundred questions based on what you just said, you know? (Dave chuckles) And as a great example, the Flash Crash. In fact, Michael Lewis, wrote about this in his book, the Flash Boys. And essentially, right, it was high frequency traders trying to front run the market and sending into small block trades (Dave chuckles) trying to get sort of front ended. So that's, and they chalked it up to a glitch. Like you said, for months, nobody really knew what it was. So technology got us into this problem. (Dave chuckles) I guess my question is can technology help us get out of the problem? And that maybe is where AI fits in? >> Yes, yes. In fact, a lot of analytics work went in to go back to the raw data that is highly dispersed from different sources, right? Assembled them to see if you can find a material trend, right? You can see lots of trends, right? Like, no, we, if humans look at things that we tend to see patterns in Clouds, right? So sometimes you need to apply statistical analysis math to be sure that what the model is seeing is real, right? And that required, well, that's one area. The second area is you know, when this, there are times when you just need to go through that tough approach to find the answer. Now, the issue comes to mind now is that humans put in the rules to decide what goes into a report that everybody sees. Now, in this case, before the change in the rules, right? But by the way, after the discovery, the authorities changed the rules and all shares, all trades of different any sizes it has to be reported. >> Right. >> Right, yeah? But the rule was applied, you know, I say earlier that shares under a hundred, trades under a hundred shares need not be reported. So, sometimes you just have to understand that reports were decided by humans and for understandable reasons. I mean, they probably didn't wanted a various reasons not to put everything in there. So that people could still read it in a reasonable amount of time. But we need to understand that rules were being put in by humans for the reports we read. And as such, there are times we just need to go back to the raw data. >> I want to ask you... >> Oh, it could be, that it's going to be tough, yeah. >> Yeah, I want to ask you a question about AI as obviously it's in your title and it's something you know a lot about but. And I'm going to make a statement, you tell me if it's on point or off point. So seems that most of the AI going on in the enterprise is modeling data science applied to, you know, troves of data. But there's also a lot of AI going on in consumer. Whether it's, you know, fingerprint technology or facial recognition or natural language processing. Well, two part question will the consumer market, as it has so often in the enterprise sort of inform us is sort of first part. And then, there'll be a shift from sort of modeling if you will to more, you mentioned the autonomous vehicles, more AI inferencing in real time, especially with the Edge. Could you help us understand that better? >> Yeah, this is a great question, right? There are three stages to just simplify. I mean, you know, it's probably more sophisticated than that. But let's just simplify that three stages, right? To building an AI system that ultimately can predict, make a prediction, right? Or to assist you in decision-making. I have an outcome. So you start with the data, massive amounts of data that you have to decide what to feed the machine with. So you feed the machine with this massive chunk of data, and the machine starts to evolve a model based on all the data it's seeing. It starts to evolve, right? To a point that using a test set of data that you have separately kept aside that you know the answer for. Then you test the model, you know? After you've trained it with all that data to see whether its prediction accuracy is high enough. And once you are satisfied with it, you then deploy the model to make the decision. And that's the inference, right? So a lot of times, depending on what we are focusing on, we in data science are, are we working hard on assembling the right data to feed the machine with? That's the data preparation organization work. And then after which you build your models you have to pick the right models for the decisions and prediction you need to make. You pick the right models. And then you start feeding the data with it. Sometimes you pick one model and a prediction isn't that robust. It is good, but then it is not consistent, right? Now what you do is you try another model. So sometimes it gets keep trying different models until you get the right kind, yeah? That gives you a good robust decision-making and prediction. Now, after which, if it's tested well, QA, you will then take that model and deploy it at the Edge. Yeah, and then at the Edge is essentially just looking at new data, applying it to the model that you have trained. And then that model will give you a prediction or a decision, right? So it is these three stages, yeah. But more and more, your question reminds me that more and more people are thinking as the Edge become more and more powerful. Can you also do learning at the Edge? >> Right. >> That's the reason why we spoke about Swarm Learning the last time. Learning at the Edge as a Swarm, right? Because maybe individually, they may not have enough power to do so. But as a Swarm, they may. >> Is that learning from the Edge or learning at the Edge? In other words, is that... >> Yes. >> Yeah. You do understand my question. >> Yes. >> Yeah. (Dave chuckles) >> That's a great question. That's a great question, right? So the quick answer is learning at the Edge, right? And also from the Edge, but the main goal, right? The goal is to learn at the Edge so that you don't have to move the data that Edge sees first back to the Cloud or the Call to do the learning. Because that would be the reason, one of the main reasons why you want to learn at the Edge. Right? So that you don't need to have to send all that data back and assemble it back from all the different Edge devices. Assemble it back to the Cloud Site to do the learning, right? Some on you can learn it and keep the data at the Edge and learn at that point, yeah. >> And then maybe only selectively send. >> Yeah. >> The autonomous vehicle, example you gave is great. 'Cause maybe they're, you know, there may be only persisting. They're not persisting data that is an inclement weather, or when a deer runs across the front. And then maybe they do that and then they send that smaller data setback and maybe that's where it's modeling done but the rest can be done at the Edge. It's a new world that's coming through. Let me ask you a question. Is there a limit to what data should be collected and how it should be collected? >> That's a great question again, yeah. Well, today full of these insightful questions. (Dr. Eng chuckles) That actually touches on the the second challenge, right? How do we, in order to thrive in this new age of insight? The second challenge is our future challenge, right? What do we do for our future? And in there is the statement we make is we have to focus on collecting data strategically for the future of our enterprise. And within that, I talked about what to collect, right? When to organize it when you collect? And then where will your data be going forward that you are collecting from? So what, when, and where? For what data to collect? That was the question you asked, it's a question that different industries have to ask themselves because it will vary, right? Let me give you the, you use the autonomous car example. Let me use that. And we do have this customer collecting massive amounts of data. You know, we're talking about 10 petabytes a day from a fleet of their cars. And these are not production autonomous cars, right? These are training autonomous cars, collecting data so they can train and eventually deploy commercial cars, right? Also this data collection cars, they collect 10, as a fleet of them collect 10 petabytes a day. And then when they came to us, building a storage system you know, to store all of that data, they realized they don't want to afford to store all of it. Now here comes the dilemma, right? What should I, after I spent so much effort building all this cars and sensors and collecting data, I've now decide what to delete. That's a dilemma, right? Now in working with them on this process of trimming down what they collected, you know, I'm constantly reminded of the 60s and 70s, right? To remind myself 60s and 70s, we called a large part of our DNA, junk DNA. >> Yeah. (Dave chuckles) >> Ah! Today, we realized that a large part of that what we call junk has function as valuable function. They are not genes but they regulate the function of genes. You know? So what's junk in yesterday could be valuable today. Or what's junk today could be valuable tomorrow, right? So, there's this tension going on, right? Between you deciding not wanting to afford to store everything that you can get your hands on. But on the other hand, you worry, you ignore the wrong ones, right? You can see this tension in our customers, right? And then it depends on industry here, right? In healthcare they say, I have no choice. I want it all, right? Oh, one very insightful point brought up by one healthcare provider that really touched me was you know, we don't only care. Of course we care a lot. We care a lot about the people we are caring for, right? But who also care for the people we are not caring for? How do we find them? >> Uh-huh. >> Right, and that definitely, they did not just need to collect data that they have with from their patients. They also need to reach out, right? To outside data so that they can figure out who they are not caring for, right? So they want it all. So I asked them, so what do you do with funding if you want it all? They say they have no choice but to figure out a way to fund it and perhaps monetization of what they have now is the way to come around and fund that. Of course, they also come back to us rightfully, that you know we have to then work out a way to help them build a system, you know? So that's healthcare, right? And if you go to other industries like banking, they say they can afford to keep them all. >> Yeah. >> But they are regulated, seemed like healthcare, they are regulated as to privacy and such like. So many examples different industries having different needs but different approaches to what they collect. But there is this constant tension between you perhaps deciding not wanting to fund all of that, all that you can install, right? But on the other hand, you know if you kind of don't want to afford it and decide not to start some. Maybe those some become highly valuable in the future, right? (Dr. Eng chuckles) You worry. >> Well, we can make some assumptions about the future. Can't we? I mean, we know there's going to be a lot more data than we've ever seen before. We know that. We know, well, not withstanding supply constraints and things like NAND. We know the prices of storage is going to continue to decline. We also know and not a lot of people are really talking about this, but the processing power, but the says, Moore's law is dead. Okay, it's waning, but the processing power when you combine the CPUs and NPUs, and GPUs and accelerators and so forth actually is increasing. And so when you think about these use cases at the Edge you're going to have much more processing power. You're going to have cheaper storage and it's going to be less expensive processing. And so as an AI practitioner, what can you do with that? >> Yeah, it's a highly, again, another insightful question that we touched on our Keynote. And that goes up to the why, uh, to the where? Where will your data be? Right? We have one estimate that says that by next year there will be 55 billion connected devices out there, right? 55 billion, right? What's the population of the world? Well, of the other 10 billion? But this thing is 55 billion. (Dave chuckles) Right? And many of them, most of them can collect data. So what do you do? Right? So the amount of data that's going to come in, it's going to way exceed, right? Drop in storage costs are increasing compute power. >> Right. >> Right. So what's the answer, right? So the answer must be knowing that we don't, and even a drop in price and increase in bandwidth, it will overwhelm the, 5G, it will overwhelm 5G, right? Given the amount of 55 billion of them collecting. So the answer must be that there needs to be a balance between you needing to bring all of that data from the 55 billion devices of the data back to a central, as a bunch of central cost. Because you may not be able to afford to do that. Firstly bandwidth, even with 5G and as the, when you'll still be too expensive given the number of devices out there. You know given storage costs dropping is still be too expensive to try and install them all. So the answer must be to start, at least to mitigate from to, some leave most a lot of the data out there, right? And only send back the pertinent ones, as you said before. But then if you did that then how are we going to do machine learning at the Core and the Cloud Site, if you don't have all the data? You want rich data to train with, right? Sometimes you want to mix up the positive type data and the negative type data. So you can train the machine in a more balanced way. So the answer must be eventually, right? As we move forward with these huge number of devices all at the Edge to do machine learning at the Edge. Today we don't even have power, right? The Edge typically is characterized by a lower energy capability and therefore lower compute power. But soon, you know? Even with low energy, they can do more with compute power improving in energy efficiency, right? So learning at the Edge, today we do inference at the Edge. So we data, model, deploy and you do inference there is. That's what we do today. But more and more, I believe given a massive amount of data at the Edge, you have to start doing machine learning at the Edge. And when you don't have enough power then you aggregate multiple devices, compute power into a Swarm and learn as a Swarm, yeah. >> Oh, interesting. So now of course, if I were sitting and fly on the wall and the HPE board meeting I said, okay, HPE is a leading provider of compute. How do you take advantage of that? I mean, we're going, I know it's future but you must be thinking about that and participating in those markets. I know today you are, you have, you know, Edge line and other products. But there's, it seems to me that it's not the general purpose that we've known in the past. It's a new type of specialized computing. How are you thinking about participating in that opportunity for the customers? >> Hmm, the wall will have to have a balance, right? Where today the default, well, the more common mode is to collect the data from the Edge and train at some centralized location or number of centralized location. Going forward, given the proliferation of the Edge devices, we'll need a balance, we need both. We need capability at the Cloud Site, right? And it has to be hybrid. And then we need capability on the Edge side that we need to build systems that on one hand is an Edge adapter, right? Meaning they environmentally adapted because the Edge differently are on it, a lot of times on the outside. They need to be packaging adapted and also power adapted, right? Because typically many of these devices are battery powered. Right? So you have to build systems that adapts to it. But at the same time, they must not be custom. That's my belief. It must be using standard processes and standard operating system so that they can run a rich set of applications. So yes, that's also the insight for that Antonio announced in 2018. For the next four years from 2018, right? $4 billion invested to strengthen our Edge portfolio. >> Uh-huh. >> Edge product lines. >> Right. >> Uh-huh, Edge solutions. >> I could, Doctor Goh, I could go on for hours with you. You're just such a great guest. Let's close. What are you most excited about in the future of, certainly HPE, but the industry in general? >> Yeah, I think the excitement is the customers, right? The diversity of customers and the diversity in the way they have approached different problems of data strategy. So the excitement is around data strategy, right? Just like, you know, the statement made for us was so was profound, right? And Antonio said, we are in the age of insight powered by data. That's the first line, right? The line that comes after that is as such we are becoming more and more data centric with data that currency. Now the next step is even more profound. That is, you know, we are going as far as saying that, you know, data should not be treated as cost anymore. No, right? But instead as an investment in a new asset class called data with value on our balance sheet. This is a step change, right? Right, in thinking that is going to change the way we look at data, the way we value it. So that's a statement. (Dr. Eng chuckles) This is the exciting thing, because for me a CTO of AI, right? A machine is only as intelligent as the data you feed it with. Data is a source of the machine learning to be intelligent. Right? (Dr. Eng chuckles) So, that's why when the people start to value data, right? And say that it is an investment when we collect it it is very positive for AI. Because an AI system gets intelligent, get more intelligence because it has huge amounts of data and a diversity of data. >> Yeah. >> So it'd be great, if the community values data. >> Well, you certainly see it in the valuations of many companies these days. And I think increasingly you see it on the income statement. You know data products and people monetizing data services. And yeah, maybe eventually you'll see it in the balance sheet. I know Doug Laney, when he was at Gartner Group, wrote a book about this and a lot of people are thinking about it. That's a big change, isn't it? >> Yeah, yeah. >> Dr. Goh... (Dave chuckles) >> The question is the process and methods in valuation. Right? >> Yeah, right. >> But I believe we will get there. We need to get started. And then we'll get there. I believe, yeah. >> Doctor Goh, it's always my pleasure. >> And then the AI will benefit greatly from it. >> Oh, yeah, no doubt. People will better understand how to align, you know some of these technology investments. Dr. Goh, great to see you again. Thanks so much for coming back in theCUBE. It's been a real pleasure. >> Yes, a system is only as smart as the data you feed it with. (Dave chuckles) (Dr. Eng laughs) >> Excellent. We'll leave it there. Thank you for spending some time with us and keep it right there for more great interviews from HPE Discover 21. This is Dave Vellante for theCUBE, the leader in Enterprise Tech Coverage. We'll be right back. (upbeat music)
SUMMARY :
Doctor Goh, great to see you again. great to talk to you again. And you talked about thriving And you really dug in the age of insight, right? of the ones you talked about today? to get what you need. And as a great example, the Flash Crash. is that humans put in the rules to decide But the rule was applied, you know, that it's going to be tough, yeah. So seems that most of the AI and the machine starts to evolve a model they may not have enough power to do so. Is that learning from the Edge You do understand my question. or the Call to do the learning. but the rest can be done at the Edge. When to organize it when you collect? But on the other hand, to help them build a system, you know? all that you can install, right? And so when you think about So what do you do? of the data back to a central, in that opportunity for the customers? And it has to be hybrid. about in the future of, as the data you feed it with. if the community values data. And I think increasingly you The question is the process We need to get started. And then the AI will Dr. Goh, great to see you again. as smart as the data Thank you for spending some time with us
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Michael Lewis | PERSON | 0.99+ |
Doug Laney | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
$4 billion | QUANTITY | 0.99+ |
Antonio | PERSON | 0.99+ |
two languages | QUANTITY | 0.99+ |
10 billion | QUANTITY | 0.99+ |
55 billion | QUANTITY | 0.99+ |
two challenges | QUANTITY | 0.99+ |
second challenge | QUANTITY | 0.99+ |
55 billion | QUANTITY | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
Gartner Group | ORGANIZATION | 0.99+ |
first line | QUANTITY | 0.99+ |
10 | QUANTITY | 0.99+ |
second area | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
Hundreds of millions | QUANTITY | 0.99+ |
Today | DATE | 0.99+ |
today | DATE | 0.99+ |
second barrier | QUANTITY | 0.99+ |
two part | QUANTITY | 0.99+ |
May 6, 2010 | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
Edge | ORGANIZATION | 0.99+ |
first barrier | QUANTITY | 0.99+ |
less than a hundred shares | QUANTITY | 0.99+ |
next year | DATE | 0.98+ |
Eng | PERSON | 0.98+ |
yesterday | DATE | 0.98+ |
first part | QUANTITY | 0.98+ |
May 6 | DATE | 0.98+ |
United Nations | ORGANIZATION | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
one area | QUANTITY | 0.98+ |
one model | QUANTITY | 0.98+ |
first one | QUANTITY | 0.98+ |
Hewlett Packard Enterprise | ORGANIZATION | 0.98+ |
Dr. | PERSON | 0.97+ |
less than a hundred shares | QUANTITY | 0.97+ |
three stages | QUANTITY | 0.97+ |
one rule | QUANTITY | 0.97+ |
Three main areas | QUANTITY | 0.97+ |
Flash Boys | TITLE | 0.97+ |
one language | QUANTITY | 0.97+ |
one | QUANTITY | 0.96+ |
10 petabytes a day | QUANTITY | 0.96+ |
Flash Crash | TITLE | 0.95+ |
under a hundred | QUANTITY | 0.95+ |
Firstly | QUANTITY | 0.95+ |
LIVE Panel: "Easy CI With Docker"
>>Hey, welcome to the live panel. My name is Brett. I am your host, and indeed we are live. In fact, if you're curious about that, if you don't believe us, um, let's just show a little bit of the browser real quick to see. Yup. There you go. We're live. So, all right. So how this is going to work is I'm going to bring in some guests and, uh, in one second, and we're going to basically take your questions on the topic designer of the day, that continuous integration testing. Uh, thank you so much to my guests welcoming into the panel. I've got Carlos, Nico and Mandy. Hello everyone. >>Hello? All right, >>Let's go. Let's go around the room and all pretend we don't know each other and that the internet didn't read below the video who we are. Uh, hi, my name is Brett. I am a Docker captain, which means I'm supposed to know something about Docker. I'm coming from Virginia Beach. I'm streaming here from Virginia Beach, Virginia, and, uh, I make videos on the internet and courses on you to me, Carlos. Hey, >>Hey, what's up? I'm Carlos Nunez. I am a solutions architect, VMware. I do solution things with computers. It's fun. I live in Dallas when I'm moving to Houston in a month, which is where I'm currently streaming. I've been all over the Northeast this whole week. So, um, it's been fun and I'm excited to meet with all of you and talk about CIA and Docker. Sure. >>Yeah. Hey everyone. Uh, Nico, Khobar here. I'm a solution engineer at HashiCorp. Uh, I am streaming to you from, uh, the beautiful Austin, Texas. Uh, ignore, ignore the golden gate bridge here. This is from my old apartment in San Francisco. Uh, just, uh, you know, keeping that, to remember all the good days, um, that that lived at. But, uh, anyway, I work at Patrick Corp and I work on all things, automation, um, and cloud and dev ops. Um, and I'm excited to be here and Mandy, >>Hi. Yeah, Mandy Hubbard. I am streaming from Austin, Texas. I am, uh, currently a DX engineer at ship engine. Um, I've worked in QA and that's kind of where I got my, uh, my Docker experience and, um, uh, moving into DX to try and help developers better understand and use our products and be an advocate for them. >>Nice. Well, thank you all for joining me. Uh, I really appreciate you taking the time out of your busy schedule to be here. And so for those of you in chat, the reason we're doing this live, because it's always harder to do things live. The reason we're here is to answer a question. So we didn't come with a bunch of slides and demos or anything like that. We're here to talk amongst ourselves about ideas and really here for you. So we've, we obviously, this is about easy CII, so we're, we're going to try to keep the conversation around testing and continuous integration and all the things that that entails with containers. But we may, we may go down rabbit holes. We may go veer off and start talking about other things, and that's totally fine if it's in the realm of dev ops and containers and developer and ops workflows, like, Hey, it's, it's kinda game. >>And, uh, these people have a wide variety of expertise. They haven't done just testing, right? We, we live in a world where you all kind of have to wear many hats. So feel free to, um, ask what you think is on the top of your mind. And we'll do our best to answer. It may, might not be the best answer or the correct answer, but we're going to do our best. Um, well, let's get it start off. Uh, let's, let's get a couple of topics to start off with. Uh, th the, the easy CGI was my, one of my three ideas. Cause he's the, one of the things that I'm most excited about is the innovation we're seeing around easier testing, faster testing, automated testing, uh, because as much as we've all been doing this stuff for, you know, 15 years, since 20 years since the sort of Jenkins early days, um, it it's, it seems like it's still really hard and it's still a lot of work. >>So, um, let's go around the room real quick, and everybody can just kind of talk for a minute about like your experience with testing and maybe some of your pain points, like what you don't like about our testing world. Um, and we can talk about some pains, cause I think that will lead us to kind of talk about what, what are the things we're seeing now that might be better, uh, ideas about how to do this. I know for me, uh, testing, obviously there's the code part, but just getting it automated, but mostly getting it in the hands of developers so that they can control their own testing. And don't have to go talk to a person to run that test again, or the mysterious Jenkins platform somewhere. I keep mentioning Jenkins cause it's, it is still the dominant player out there. Um, so for me, I'm, I'm, I, I don't like it when I'm walking into a room and there's, there's only one or two people that know how the testing works or know how to make the new tests go into the testing platform and stuff like that. So I'm always trying to free those things so that any of the developers are enabled and empowered to do that stuff. So someone else, Carlos, anybody, um, >>Oh, I have a lot of opinions on that. Having been a QA engineer for most of my career. Um, the shift that we're saying is everyone is dev ops and everyone is QA. Th the issue I see is no one asked developers if they wanted to be QA. Um, and so being the former QA on the team, when there's a problem, even though I'm a developer and we're all running QA, they always tend to come to the one of the former QA engineers. And they're not really owning that responsibility and, um, and digging in. So that's kind of what I'm saying is that we're all expected to test now. And some people, well, some people don't know how it's, uh, for me it was kind of an intuitive skill. It just kind of fit with my personality, but not knowing what to look for, not knowing what to automate, not even understanding how your API end points are used by your front end to know what to test when a change is made. It's really overwhelming for developers. And, um, we're going to need to streamline that and, and hold their hands a little bit until they get their feet wet with also being QA. >>Right. Right. So, um, uh, Carlos, >>Yeah, uh, testing is like, Tesla is one of my favorite subjects to talk about when I'm baring with developers. And a lot of it is because of what Mandy said, right? Like a lot of developers now who used to write a test and say, Hey, QA, go. Um, I wrote my unit tests. Now write the rest of the test. Essentially. Now developers are expected to be able to understand how testing, uh, testing methodologies work, um, in their local environments, right? Like they're supposed to understand how to write an integration tasks federate into and tasks, a component test. And of course, how to write unit tests that aren't just, you know, assert true is true, right? Like more comprehensive, more comprehensive, um, more high touch unit tests, which include things like mocking and stubbing and spine and all that stuff. And, you know, it's not so much getting those tests. Well, I've had a lot of challenges with developers getting those tests to run in Docker because of usually because of dependency hell, but, um, getting developers to understand how to write tests that matter and mean something. Um, it's, it's, it can be difficult, but it's also where I find a lot of the enjoyment of my work comes into play. So yeah. I mean, that's the difficulty I've seen around testing. Um, big subject though. Lots to talk about there. >>Yeah. We've got, we've already got so many questions coming in. You already got an hour's worth of stuff. So, uh, Nico 81st thoughts on that? >>Yeah, I think I definitely agree with, with other folks here on the panel, I think from a, um, the shift from a skillset perspective that's needed to adopt the new technologies, but I think from even from, uh, aside from the organizational, um, and kind of key responsibilities that, that the new developers have to kinda adapt to and, and kind of inherit now, um, there's also from a technical perspective as there's, you know, um, more developers are owning the full stack, including the infrastructure piece. So that adds a lot more to the plate in Tim's oaf, also testing that component that they were not even, uh, responsible for before. Um, and, um, also the second challenge that, you know, I'm seeing is that on, you know, the long list of added, um, uh, tooling and, you know, there's new tool every other day. Um, and, um, that kind of requires more customization to the testing, uh, that each individual team, um, any individual developer Y by extension has to learn. Uh, so the customization, uh, as well as the, kind of the scope that had, uh, you know, now in conferences, the infrastructure piece, um, uh, both of act to the, to the challenges that we're seeing right now for, um, for CGI and overall testing, um, uh, the developers are saying, uh, in, in the market today. >>Yeah. We've got a lot of questions, um, about all the, all the different parts of this. So, uh, let me just go straight to them. Cause that's why we're here is for the people, uh, a lot of people asking about your favorite tools and in one of this is one of the challenges with integration, right? Is, um, there is no, there are dominant players, but there, there is such a variety. I mean, every one of my customers seems like they're using a different workflow and a different set of tools. So, and Hey, we're all here to just talk about what we're, what we're using, uh, you know, whether your favorite tools. So like a lot of the repeated questions are, what are your favorite tools? Like if you could create it from scratch, uh, what would you use? Pierre's asking, you know, GitHub actions sounds like they're a fan of GitHub actions, uh, w you know, mentioning, pushing the ECR and Docker hub and, uh, using vs code pipeline, I guess there may be talking about Azure pipelines. Um, what, what's your preferred way? So, does anyone have any, uh, thoughts on that anyone want to throw out there? Their preferred pipeline of tooling? >>Well, I have to throw out mine. I might as Jenkins, um, like kind of a honorary cloud be at this point, having spoken a couple of times there, um, all of the plugins just make the functionality. I don't love the UI, but I love that it's been around so long. It has so much community support, and there are so many plugins so that if you want to do something, you don't have to write the code it's already been tested. Um, unfortunately I haven't been able to use Jenkins in, uh, since I joined ship engine, we, most of our, um, our, our monolithic core application is, is team city. It's a dotnet application and TeamCity plays really well with.net. Um, didn't love it, uh, Ms. Jenkins. And I'm just, we're just starting some new initiatives that are using GitHub actions, and I'm really excited to learn, to learn those. I think they have a lot of the same functionality that you're looking for, but, um, much more simplified in is right there and get hubs. So, um, the integration is a lot more seamless, but I do have to go on record that my favorite CICT tools Jenkins. >>All right. You heard it here first people. All right. Anyone else? You're muted? I'm muted. Carlin says muted. Oh, Carla says, guest has muted themselves to Carlos. You got to unmute. >>Yes. I did mute myself because I was typing a lot, trying to, you know, try to answer stuff in the chat. And there's a lot of really dark stuff in there. That's okay. Two more times today. So yeah, it's fine. Yeah, no problem. So totally. And it's the best way to start a play more. So I'm just going to go ahead and light it up. Um, for enterprise environments, I actually am a huge fan of Jenkins. Um, it's a tool that people really understand. Um, it has stood the test of time, right? I mean, people were using Hudson, but 15 years ago, maybe longer. And, you know, the way it works, hasn't really changed very much. I mean, Jenkins X is a little different, but, um, the UI and the way it works internally is pretty familiar to a lot of enterprise environments, which is great. >>And also in me, the plugin ecosystem is amazing. There's so many plugins for everything, and you can make your own if you know, Java groovy. I'm sure there's a perfect Kotlin in there, but I haven't tried myself, but it's really great. It's also really easy to write, um, CIS code, which is something I'm a big fan of. So Jenkins files have been, have worked really well for me. I, I know that I can get a little bit more complex as you start to build your own models and such, but, you know, for enterprise enterprise CIO CD, if you want, especially if you want to roll your own or own it yourself, um, Jenkins is the bellwether and for very good reason now for my personal projects. And I see a lot on the chat here, I think y'all, y'all been agreed with me get hub actions 100%, my favorite tool right now. >>Um, I love GitHub actions. It's, it's customizable, it's modular. There's a lot of plugins already. I started using getting that back maybe a week after when GA and there was no documentation or anything. And I still, it was still my favorite CIA tool even then. Um, and you know, the API is really great. There's a lot to love about GitHub actions and, um, and I, and I use it as much as I can from my personal project. So I still have a soft spot for Travis CAI. Um, you know, they got acquired and they're a little different now trying to see, I, I can't, I can't let it go. I just love it. But, um, yeah, I mean, when it comes to Seattle, those are my tools. So light me up in the comments I will respond. Yeah. >>I mean, I, I feel with you on the Travis, the, I think, cause I think that was my first time experiencing, you know, early days get hub open source and like a free CIA tool that I could describe. I think it was the ammo back then. I don't actually remember, but yeah, it was kind of an exciting time from my experience. There was like, oh, this is, this is just there as a service. And I could just use it. It doesn't, it's like get hub it's free from my open source stuff. And so it does have a soft spot in my heart too. So yeah. >>All right. We've got questions around, um, cam, so I'm going to ask some questions. We don't have to have these answers because sometimes they're going to be specific, but I want to call them out because people in chat may have missed that question. And there's probably, you know, that we have smart people in chat too. So there's probably someone that knows the answer to these things. If, if it's not us, um, they're asking about building Docker images in Kubernetes, which to me is always a sore spot because it's Kubernetes does not build images by default. It's not meant for that out of the gate. And, uh, what is the best way to do this without having to use privileged containers, which privileged containers just implying that yeah, you, you, it probably has more privileges than by default as a container in Kubernetes. And that is a hard thing because, uh, I don't, I think Docker doesn't lie to do that out of the gate. So I don't know if anyone has an immediate answer to that. That's a pretty technical one, but if you, if you know the answer to that in chat, call it out. >>Um, >>I had done this, uh, but I'm pretty sure I had to use a privileged, um, container and install the Docker Damon on the Kubernetes cluster. And I CA I can't give you a better solution. Um, I've done the same. So, >>Yeah, uh, Chavonne asks, um, back to the Jenkins thing, what's the easiest way to integrate Docker into a Jenkins CICB pipeline. And that's one of the challenges I find with Jenkins because I don't claim to be the expert on Jenkins. Is there are so many plugins because of this, of this such a huge ecosystem. Um, when you go searching for Docker, there's a lot that comes back, right. So I, I don't actually have a preferred way because every team I find uses it differently. Um, I don't know, is there a, do you know if there's a Jenkins preferred, a default plugin? I don't even know for Docker. Oh, go ahead. Yeah. Sorry for Docker. And jacon sorry, Docker plugins for Jenkins. Uh, as someone's asking like the preferred or easy way to do that. Um, and I don't, I don't know the back into Jenkins that well, so, >>Well, th the new, the new way that they're doing, uh, Docker builds with the pipeline, which is more declarative versus the groovy. It's really simple, and their documentation is really good. They, um, they make it really easy to say, run this in this image. So you can pull down, you know, public images and add your own layers. Um, so I don't know the name of that plugin, uh, but I can certainly take a minute after this session and going and get that. Um, but if you really are overwhelmed by the plugins, you can just write your, you know, your shell command in Jenkins. You could just by, you know, doing everything in bash, calling the Docker, um, Damon directly, and then getting it working just to see that end to end, and then start browsing for plugins to see if you even want to use those. >>The plugins will allow more integration from end to end. Some of the things that you input might be available later on in the process for having to manage that yourself. But, you know, you don't have to use any of the plugins. You can literally just, you know, do a block where you write your shell command and get it working, and then decide if, for plugins for you. Um, I think it's always under important to understand what is going on under the hood before you, before you adopt the magic of a plugin, because, um, once you have a problem, if you're, if it's all a lockbox to you, it's going to be more difficult to troubleshoot. It's kind of like learning, get command line versus like get cracking or something. Once, once you get in a bind, if you don't understand the underlying steps, it's really hard to get yourself out of a bind, versus if you understand what the plugin or the app is doing, then, um, you can get out of situations a lot easier. That's a good place. That's, that's where I'd start. >>Yeah. Thank you. Um, Camden asks better to build test environment images, every commit in CII. So this is like one of those opinions of we're all gonna have some different, uh, or build on build images on every commit, leveraging the cash, or build them once outside the test pile pipeline. Um, what say you people? >>Uh, well, I I've seen both and generally speaking, my preference is, um, I guess the ant, the it's a consultant answer, right? I think it depends on what you're trying to do, right. So if you have a lot of small changes that are being made and you're creating images for each of those commits, you're going to have a lot of images in your, in your registry, right? And on top of that, if you're building those images, uh, through CAI frequently, if you're using Docker hub or something like that, you might run into rate limiting issues because of Docker's new rate, limiting, uh, rate limits that they put in place. Um, but that might be beneficial if the, if being able to roll back between those small changes while you're testing is important to you. Uh, however, if all you care about is being able to use Docker images, um, or being able to correlate versions to your Docker images, or if you're the type of team that doesn't even use him, uh, does he even use, uh, virgins in your image tags? Then I would think that that might be a little, much you might want to just have in your CIO. You might want to have a stage that builds your Docker images and Docker image and pushes it into your registry, being done first particular branches instead of having to be done on every commit regardless of branch. But again, it really depends on the team. It really depends on what you're building. It really depends on your workflow. It can depend on a number of things like a curse sometimes too. Yeah. Yeah. >>Once had two points here, you know, I've seen, you know, the pattern has been at every, with every, uh, uh, commit, assuming that you have the right set of tests that would kind of, uh, you would benefit from actually seeing, um, the, the, the, the testing workflow go through and can detect any issue within, within the build or whatever you're trying to test against. But if you're just a building without the appropriate set of tests, then you're just basically consuming almond, adding time, as well as all the, the image, uh, stories associated with it without treaty reaping the benefit of, of, of this pattern. Uh, and the second point is, again, I think if you're, if you're going to end up doing a per commit, uh, definitely recommend having some type of, uh, uh, image purging, um, uh, and, and, and garbage collection process to ensure that you're not just wasting, um, all the stories needed and also, um, uh, optimizing your, your bill process, because that will end up being the most time-consuming, um, um, you know, within, within your pipeline. So this is my 2 cents on this. >>Yeah, that's good stuff. I mean, those are both of those are conversations that could lead us into the rabbit hole for the rest of the day on storage management, uh, you know, CP CPU minutes for, uh, you know, your build stuff. I mean, if you're in any size team, more than one or two people, you immediately run into headaches with cost of CIA, because we have now the problem of tools, right? We have so many tools. We can have the CIS system burning CPU cycles all day, every day, if we really wanted to. And so you re very quickly, I think, especially if you're on every commit on every branch, like that gets you into a world of cost mitigation, and you probably are going to have to settle somewhere in the middle on, uh, between the budget, people that are saying you're spending way too much money on the CII platform, uh, because of all these CPU cycles, and then the developers who would love to have everything now, you know, as fast as possible and the biggest, biggest CPU's, and the biggest servers, and have the bills, because the bills can never go fast enough, right. >>There's no end to optimizing your build workflow. Um, we have another question on that. This is another topic that we'll all probably have different takes on is, uh, basically, uh, version tags, right? So on images, we, we have a very established workflow in get for how we make commits. We have commit shots. We have, uh, you know, we know get tags and there's all these things there. And then we go into images and it's just this whole new world that's opened up. Like there's no real consensus. Um, so what, what are your thoughts on the strategy for teams in their image tag? Again, another, another culture thing. Um, commander, >>I mean, I'm a fan of silver when we have no other option. Um, it's just clean and I like the timestamp, you know, exactly when it was built. Um, I don't really see any reason to use another, uh, there's just normal, incremental, um, you know, numbering, but I love the fact that you can pull any tag and know exactly when it was created. So I'm a big fan of bar, if you can make that work for your organization. >>Yep. People are mentioned that in chat, >>So I like as well. Uh, I'm a big fan of it. I think it's easy to be able to just be as easy to be able to signify what a major changes versus a minor change versus just a hot fix or, you know, some or some kind of a bad fix. The problem that I've found with having teams adopt San Bernardo becomes answering these questions and being able to really define what is a major change, what is a minor change? What is a patch, right? And this becomes a bit of an overhead or not so much of an overhead, but, uh, uh, uh, a large concern for teams who have never done versioning before, or they never been responsible for their own versioning. Um, in fact, you know, I'm running into that right now, uh, with, with a client that I'm working with, where a lot, I'm working with a lot of teams, helping them move their applications from a legacy production environment into a new one. >>And in doing so, uh, versioning comes up because Docker images, uh, have tags and usually the tax correlate to versions, but some teams over there, some teams that I'm working with are only maintaining a script and others are maintaining a fully fledged JAK, three tier application, you know, with lots of dependencies. So telling the script, telling the team that maintains a script, Hey, you know, you should use somber and you should start thinking about, you know, what's major, what's my number what's patch. That might be a lot for them. And for someone or a team like that, I might just suggest using commit shots as your versions until you figure that out, or maybe using, um, dates as your version, but for the more for the team, with the larger application, they probably already know the answers to those questions. In which case they're either already using Sember or they, um, or they may be using some other version of the strategy and might be in December, might suit them better. So, um, you're going to hear me say, it depends a lot, and I'm just going to say here, it depends. Cause it really does. Carlos. >>I think you hit on something interesting beyond just how to version, but, um, when to consider it a major release and who makes those decisions, and if you leave it to engineers to version, you're kind of pushing business decisions down the pipe. Um, I think when it's a minor or a major should be a business decision and someone else needs to make that call someone closer to the business should be making that call as to when we want to call it major. >>That's a really good point. And I add some, I actually agree. Um, I absolutely agree with that. And again, it really depends on the team that on the team and the scope of it, it depends on the scope that they're maintaining, right? And so it's a business application. Of course, you're going to have a product manager and you're going to have, you're going to have a product manager who's going to want to make that call because that version is going to be out in marketing. People are going to use it. They're going to refer to and support calls. They're going to need to make those decisions. Sember again, works really, really well for that. Um, but for a team that's maintaining the scripts, you know, I don't know, having them say, okay, you must tell me what a major version is. It's >>A lot, but >>If they want it to use some birds great too, which is why I think going back to what you originally said, Sember in the absence of other options. I think that's a good strategy. >>Yeah. There's a, there's a, um, catching up on chat. I'm not sure if I'm ever going to catch up, but there's a lot of people commenting on their favorite CII systems and it's, and it, it just goes to show for the, the testing and deployment community. Like how many tools there are out there, how many tools there are to support the tools that you're using. Like, uh, it can be a crazy wilderness. And I think that's, that's part of the art of it, uh, is that these things are allowing us to build our workflows to the team's culture. Um, and, uh, but I do think that, you know, getting into like maybe what we hope to be at what's next is I do hope that we get to, to try to figure out some of these harder problems of consistency. Uh, one of the things that led me to Docker at the beginning to begin with was the fact that it wa it created a consistent packaging solution for me to get my code, you know, off of, off of my site of my local system, really, and into the server. >>And that whole workflow would at least the thing that I was making at each step was going to be the same thing used. Right. And that, that was huge. Uh, it was also, it also took us a long time to get there. Right. We all had to, like Docker was one of those ones that decade kind of ideas of let's solidify the, enter, get the consensus of the community around this idea. And we, and it's not perfect. Uh, you know, the Docker Docker file is not the most perfect way to describe how to make your app, but it is there and we're all using it. And now I'm looking for that next piece, right. Then hopefully the next step in that, um, that where we can all arrive at a consensus so that once you hop teams, you know, okay. We all knew Docker. We now, now we're all starting to get to know the manifests, but then there's this big gap in the middle where it's like, it might be one of a dozen things. Um, you know, so >>Yeah, yeah. To that, to that, Brett, um, you know, uh, just maybe more of a shameless plug here and wanting to kind of talk about one of the things that I'm on. So excited, but I work, I work at Tasha Corp. I don't know anyone, or I don't know if many people have heard of, um, you know, we tend to focus a lot on workflows versus technologies, right. Because, you know, as you can see, even just looking at the chat, there's, you know, ton of opinions on the different tooling, right. And, uh, imagine having, you know, I'm working with clients that have 10,000 developers. So imagine taking the folks in the chat and being partnered with one organization or one company and having to make decisions on how to build software. Um, but there's no way you can conversion one or, or one way or one tool, uh, and that's where we're facing in the industry. >>So one of the things that, uh, I'm pretty excited about, and I don't know if it's getting as much traction as you know, we've been focused on it. This is way point, which is a project, an open source project. I believe we got at least, uh, last year, um, which is, it's more of, uh, it's, it is aim to address that really, uh, uh, Brad set on, you know, to come to tool to, uh, make it extremely easy and simple. And, you know, to describe how you want to build, uh, deploy or release your application, uh, in, in a consistent way, regardless of the tools. So similar to how you can think of Terraform and having that pluggability to say Terraform apply or plan against any cloud infrastructure, uh, without really having to know exactly the details of how to do it, uh, this is what wave one is doing. Um, and it can be applied with, you know, for the CIA, uh, framework. So, you know, task plugability into, uh, you know, circle CEI tests to Docker helm, uh, Kubernetes. So that's the, you know, it's, it's a hard problem to solve, but, um, I'm hopeful that that's the path that we're, you know, we'll, we'll eventually get to. So, um, hope, you know, you can, you can, uh, see some of the, you know, information, data on it, on, on HashiCorp site, but I mean, I'm personally excited about it. >>Yeah. Uh I'm to gonna have to check that out. And, um, I told you on my live show, man, we'll talk about it, but talk about it for a whole hour. Uh, so there's another question here around, uh, this, this is actually a little bit more detailed, but it is one that I think a lot of people deal with and I deal with a lot too, is essentially the question is from Cameron, uh, D essentially, do you use compose in your CIO or not Docker compose? Uh, because yes I do. Yeah. Cause it, it, it, it solves so many problems am and not every CGI can, I don't know, there's some problems with a CIO is trying to do it for me. So there are pros and cons and I feel like I'm still on the fence about it because I use it all the time, but also it's not perfect. It's not always meant for CIA. And CIA sometimes tries to do things for you, like starting things up before you start other parts and having that whole order, uh, ordering problem of things anyway. W thoughts and when have thoughts. >>Yes. I love compose. It's one of my favorite tools of all time. Um, and the reason why it's, because what I often find I'm working with teams trying to actually let me walk that back, because Jack on the chat asked a really interesting question about what, what, what the hardest thing about CIS for a lot of teams. And in my experience, the hardest thing is getting teams to build an app that is the same app as what's built in production. A lot of CGI does things that are totally different than what you would do in your local, in your local dev. And as a result of that, you get, you got this application that either doesn't work locally, or it does work, but it's a completely different animal than what you would get in production. Right? So what I've found in trying to get teams to bridge that gap by basically taking their CGI, shifting the CII left, I hate the shift left turn, but I'll use it. >>I'm shifting the CIO left to your local development is trying to say, okay, how do we build an app? How do we, how do we build mot dependencies of that app so that we can build so that we can test our app? How do we run tests, right? How do we build, how do we get test data? And what I found is that trying to get teams to do all this in Docker, which is normally a first for a lot of teams that I'm working with, trying to get them all to do all of this. And Docker means you're running Docker, build a lot running Docker, run a lot. You're running Docker, RM a lot. You ran a lot of Docker, disparate Docker commands. And then on top of that, trying to bridge all of those containers together into a single network can be challenging without compose. >>So I like using a, to be able to really easily categorize and compartmentalize a lot of the things that are going to be done in CII, like building a Docker image, running tests, which is you're, you're going to do it in CII anyway. So running tests, building the image, pushing it to the registry. Well, I wouldn't say pushing it to the registry, but doing all the things that you would do in local dev, but in the same network that you might have a mock database or a mock S3 instance or some of something else. Um, so it's just easy to take all those Docker compose commands and move them into your Yammel file using the hub actions or your dankest Bob using Jenkins, or what have you. Right. It's really, it's really portable that way, but it doesn't work for every team. You know, for example, if you're just a team that, you know, going back to my script example, if it's a really simple script that does one thing on a somewhat routine basis, then that might be a lot of overhead. Um, in that case, you know, you can get away with just Docker commands. It's not a big deal, but the way I looked at it is if I'm, if I'm building, if I build something that's similar to a make bile or rate file, or what have you, then I'm probably gonna want to use Docker compose. If I'm working with Docker, that's, that's a philosophy of values, right? >>So I'm also a fan of Docker compose. And, um, you know, to your point, Carlos, the whole, I mean, I'm also a fan of shifting CEI lift and testing lift, but if you put all that logic in your CTI, um, it changes the L the local development experience from the CGI experience. Versus if you put everything in a compose file so that what you build locally is the same as what you build in CGI. Um, you're going to have a better experience because you're going to be testing something more, that's closer to what you're going to be releasing. And it's also very easy to look at a compose file and kind of, um, understand what the dependencies are and what's happening is very readable. And once you move that stuff to CGI, I think a lot of developers, you know, they're going to be intimidated by the CGI, um, whatever the scripting language is, it's going to be something they're going to have to wrap their head around. >>Um, but they're not gonna be able to use it locally. You're going to have to have another local solution. So I love the idea of a composed file use locally, um, especially if he can Mount the local workspace so that they can do real time development and see their changes in the exact same way as it's going to be built and tested in CGI. It gives developers a high level of confidence. And then, you know, you're less likely to have issues because of discrepancies between how it was built in your local test environment versus how it's built in NCI. And so Docker compose really lets you do all of that in a way that makes your solution more portable, portable between local dev and CGI and reduces the number of CGI cycles to get, you know, the test, the test data that you need. So that's why I like it for really, for local dev. >>It'll be interesting. Um, I don't know if you all were able to see the keynote, but there was a, there was a little bit, not a whole lot, but a little bit talk of the Docker, compose V two, which has now built into the Docker command line. And so now we're shifting from the Python built compose, which was a separate package. You could that one of the challenges was getting it into your CA solution because if you don't have PIP and you got down on the binary and the binary wasn't available for every platform and, uh, it was a PI installer. It gets a little nerdy into how that works, but, uh, and the team is now getting, be able to get unified with it. Now that it's in Golang and it's, and it's plugged right into the Docker command line, it hopefully will be easier to distribute, easier to, to use. >>And you won't have to necessarily have dependencies inside of where you're running it because there'll be a statically compiled binary. Um, so I've been playing with that, uh, this year. And so like training myself to do Docker going from Docker dash compose to Docker space, compose. It is a thing I I'm almost to the point of having to write a shell replacement. Yeah. Alias that thing. Um, but, um, I'm excited to see what that's going, cause there's already new features in it. And it, these built kit by default, like there's all these things. And I, I love build kit. We could make a whole session on build kit. Um, in fact there's actually, um, maybe going on right now, or right around this time, there is a session on, uh, from Solomon hikes, the seat, uh, co-founder of Docker, former CTO, uh, on build kit using, uh, using some other tool on top of build kit or whatever. >>So that, that would be interesting for those of you that are not watching that one. Cause you're here, uh, to do a check that one out later. Um, all right. So another good question was caching. So another one, another area where there is no wrong answers probably, and everyone has a different story. So the question is, what are your thoughts on CII build caching? There's often a debate between security. This is from Quentin. Thank you for this great question. There's often a debate between security reproducibility and build speeds. I haven't found a good answer so far. I will just throw my hat in the ring and say that the more times you want to build, like if you're trying to build every commit or every commit, if you're building many times a day, the more caching you need. So like the more times you're building, the more caching you're gonna likely want. And in most cases caching doesn't bite you in the butt, but that could be, yeah, we, can we get the bit about that? So, yeah. Yeah. >>I'm going to quote Carlos again and say, it depends on, on, you know, how you're talking, you know, what you're trying to build and I'm quoting your colors. Um, yeah, it's, it's got, it's gonna depend because, you know, there are some instances where you definitely want to use, you know, depends on the frequency that you're building and how you're building. Um, it's you would want to actually take advantage of cashing functionalities, um, for the build, uh, itself. Um, but if, um, you know, as you mentioned, there could be some instances where you would want to disable, um, any caching because you actually want to either pull a new packages or, um, you know, there could be some security, um, uh, disadvantages related to security aspects that would, you know, you know, using a cache version of, uh, image layer, for example, could be a problem. And you, you know, if you have a fleet of build, uh, engines, you don't have a good grasp of where they're being cashed. We would have to, um, disable caching in that, in that, um, in those instances. So it, it would depend. >>Yeah, it's, it's funny you have that problem on both sides of cashing. Like there are things that, especially in Docker world, they will cash automatically. And, and then, and then you maybe don't realize that some of that caching could be bad. It's, it's actually using old, uh, old assets, old artifacts, and then there's times where you would expect it to cash, that it doesn't cash. And then you have to do something extra to enable that caching, especially when you're dealing with that cluster of, of CIS servers. Right. And the cloud, the whole clustering problem with caching is even more complex, but yeah, >>But that's, that's when, >>Uh, you know, ever since I asked you to start using build kits and able to build kit, you know, between it's it's it's reader of Boston in, in detecting word, you know, where in, in the bill process needs to cash, as well as, uh, the, the, um, you know, the process. I don't think I've seen any other, uh, approach there that comes close to how efficient, uh, that process can become how much time it can actually save. Uh, but again, I think, I think that's, for me that had been my default approach, unless I actually need something that I would intentionally to disable caching for that purpose, but the benefits, at least for me, the benefits of, um, how bill kit actually been processing my bills, um, from the builds as well as, you know, using the cash up until, you know, how it detects the, the difference in, in, in the assets within the Docker file had been, um, you know, uh, pretty, you know, outweigh the disadvantages that it brings in. So it, you know, take it each case by case. And based on that, determine if you want to use it, but definitely recommend those enabling >>In the absence of a reason not to, um, I definitely think that it's a good approach in terms of speed. Um, yeah, I say you cash until you have a good reason not to personally >>Catch by default. There you go. I think you catch by default. Yeah. Yeah. And, uh, the trick is, well, one, it's not always enabled by default, especially when you're talking about cross server. So that's a, that's a complexity for your SIS admins, or if you're on the cloud, you know, it's usually just an option. Um, I think it also is this, this veers into a little bit of, uh, the more you cash the in a lot of cases with Docker, like the, from like, if you're from images and checked every single time, if you're not pinning every single thing, if you're not painting your app version, you're at your MPN versions to the exact lock file definition. Like there's a lot of these things where I'm I get, I get sort of, I get very grouchy with teams that sort of let it, just let it all be like, yeah, we'll just build two images and they're totally going to have different dependencies because someone happened to update that thing and after whatever or MPM or, or, and so I get grouchy about that, cause I want to lock it all down, but I also know that that's going to create administrative burden. >>Like the team is now going to have to manage versions in a very much more granular way. Like, do we need to version two? Do we need to care about curl? You know, all that stuff. Um, so that's, that's kind of tricky, but when you get to, when you get to certain version problems, uh, sorry, uh, cashing problems, you, you, you don't want those set those caches to happen because it, if you're from image changes and you're not constantly checking for a new image, and if you're not pinning that V that version, then now you, you don't know whether you're getting the latest version of Davion or whatever. Um, so I think that there's, there's an art form to the more you pen, the less you have, the less, you have to be worried about things changing, but the more you pen, the, uh, all your versions of everything all the way down the stack, the more administrative stuff, because you're gonna have to manually change every one of those. >>So I think it's a balancing act for teams. And as you mature, I to find teams, they tend to pin more until they get to a point of being more comfortable with their testing. So the other side of this argument is if you trust your testing, then you, and you have better testing to me, the less likely to the subtle little differences in versions have to be penned because you can get away with those minor or patch level version changes. If you're thoroughly testing your app, because you're trusting your testing. And this gets us into a whole nother rant, but, uh, yeah, but talking >>About penny versions, if you've got a lot of dependencies isn't that when you would want to use the cash the most and not have to rebuild all those layers. Yeah. >>But if you're not, but if you're not painting to the exact patch version and you are caching, then you're not technically getting the latest versions because it's not checking for all the time. It's a weird, there's a lot of this subtle nuance that people don't realize until it's a problem. And that's part of the, the tricky part of allow this stuff, is it, sometimes the Docker can be almost so much magic out of the box that you, you, you get this all and it all works. And then day two happens and you built it a second time and you've got a new version of open SSL in there and suddenly it doesn't work. Um, so anyway, uh, that was a great question. I've done the question on this, on, uh, from heavy. What do you put, where do you put testing in your pipeline? Like, so testing the code cause there's lots of types of testing, uh, because this pipeline gets longer and longer and Docker building images as part of it. And so he says, um, before staging or after staging, but before production, where do you put it? >>Oh man. Okay. So, um, my, my main thought on this is, and of course this is kind of religious flame bait, so sure. You know, people are going to go into the compensation wrong. Carlos, the boy is how I like to think about it. So pretty much in every stage or every environment that you're going to be deploying your app into, or that your application is going to touch. My idea is that there should be a build of a Docker image that has all your applications coded in, along with its dependencies, there's testing that tests your application, and then there's a deployment that happens into whatever infrastructure there is. Right. So the testing, they can get tricky though. And the type of testing you do, I think depends on the environment that you're in. So if you're, let's say for example, your team and you have, you have a main branch and then you have feature branches that merged into the main branch. >>You don't have like a pre-production branch or anything like that. So in those feature branches, whenever I'm doing CGI that way, I know when I freak, when I cut my poll request, that I'm going to merge into main and everything's going to work in my feature branches, I'm going to want to probably just run unit tests and maybe some component tests, which really, which are just, you know, testing that your app can talk to another component or another part, another dependency, like maybe a database doing tests like that, that don't take a lot of time that are fascinating and right. A lot of would be done at the beach branch level and in my opinion, but when you're going to merge that beach branch into main, as part of a release in that activity, you're going to want to be able to do an integration tasks, to make sure that your app can actually talk to all the other dependencies that it talked to. >>You're going to want to do an end to end test or a smoke test, just to make sure that, you know, someone that actually touches the application, if it's like a website can actually use the website as intended and it meets the business cases and all that, and you might even have testing like performance testing, low performance load testing, or security testing, compliance testing that would want to happen in my opinion, when you're about to go into production with a release, because those are gonna take a long time. Those are very expensive. You're going to have to cut new infrastructure, run those tests, and it can become quite arduous. And you're not going to want to run those all the time. You'll have the resources, uh, builds will be slower. Uh, release will be slower. It will just become a mess. So I would want to save those for when I'm about to go into production. Instead of doing those every time I make a commit or every time I'm merging a feature ranch into a non main branch, that's the way I look at it, but everything does a different, um, there's other philosophies around it. Yeah. >>Well, I don't disagree with your build test deploy. I think if you're going to deploy the code, it needs to be tested. Um, at some level, I mean less the same. You've got, I hate the term smoke tests, cause it gives a false sense of security, but you have some mental minimum minimal amount of tests. And I would expect the developer on the feature branch to add new tests that tested that feature. And that would be part of the PR why those tests would need to pass before you can merge it, merge it to master. So I agree that there are tests that you, you want to run at different stages, but the earlier you can run the test before going to production. Um, the fewer issues you have, the easier it is to troubleshoot it. And I kind of agree with what you said, Carlos, about the longer running tests like performance tests and things like that, waiting to the end. >>The only problem is when you wait until the end to run those performance tests, you kind of end up deploying with whatever performance you have. It's, it's almost just an information gathering. So if you don't run your performance test early on, um, and I don't want to go down a rabbit hole, but performance tests can be really useless if you don't have a goal where it's just information gap, uh, this is, this is the performance. Well, what did you expect it to be? Is it good? Is it bad? They can get really nebulous. So if performance is really important, um, you you're gonna need to come up with some expectations, preferably, you know, set up the business level, like what our SLA is, what our response times and have something to shoot for. And then before you're getting to production. If you have targets, you can test before staging and you can tweak the code before staging and move that performance initiative. Sorry, Carlos, a little to the left. Um, but if you don't have a performance targets, then it's just a check box. So those are my thoughts. I like to test before every deployment. Right? >>Yeah. And you know what, I'm glad that you, I'm glad that you brought, I'm glad that you brought up Escalades and performance because, and you know, the definition of performance says to me, because one of the things that I've seen when I work with teams is that oftentimes another team runs a P and L tests and they ended, and the development team doesn't really have too much insight into what's going on there. And usually when I go to the performance team and say, Hey, how do you run your performance test? It's usually just a generic solution for every single application that they support, which may or may not be applicable to the application team that I'm working with specifically. So I think it's a good, I'm not going to dig into it. I'm not going to dig into the rabbit hole SRE, but it is a good bridge into SRE when you start trying to define what does reliability mean, right? >>Because the reason why you test performance, it's test reliability to make sure that when you cut that release, that customers would go to your site or use your application. Aren't going to see regressions in performance and are not going to either go to another website or, you know, lodge in SLA violation or something like that. Um, it does, it does bridge really well with defining reliability and what SRE means. And when you have, when you start talking about that, that's when you started talking about how often do I run? How often do I test my reliability, the reliability of my application, right? Like, do I have nightly tasks in CGI that ensure that my main branch or, you know, some important branch I does not mean is meeting SLA is meeting SLR. So service level objectives, um, or, you know, do I run tasks that ensure that my SLA is being met in production? >>Like whenever, like do I use, do I do things like game days where I test, Hey, if I turn something off or, you know, if I deploy this small broken code to production and like what happens to my performance? What happens to my security and compliance? Um, you can, that you can go really deep into and take creating, um, into creating really robust tests that cover a lot of different domains. But I liked just using build test deploy is the overall answer to that because I find that you're going to have to build your application first. You're going to have to test it out there and build it, and then you're going to want to deploy it after you test it. And that order generally ensures that you're releasing software. That works. >>Right. Right. Um, I was going to ask one last question. Um, it's going to have to be like a sentence answer though, for each one of you. Uh, this is, uh, do you lint? And if you lint, do you lent all the things, if you do, do you fail the linters during your testing? Yes or no? I think it's going to depend on the culture. I really do. Sorry about it. If we >>Have a, you know, a hook, uh, you know, on the get commit, then theoretically the developer can't get code there without running Melinta anyway, >>So, right, right. True. Anyone else? Anyone thoughts on that? Linting >>Nice. I saw an additional question online thing. And in the chat, if you would introduce it in a multi-stage build, um, you know, I was wondering also what others think about that, like typically I've seen, you know, with multi-stage it's the most common use case is just to produce the final, like to minimize the, the, the, the, the, the image size and produce a final, you know, thin, uh, layout or thin, uh, image. Uh, so if it's not for that, like, I, I don't, I haven't seen a lot of, you know, um, teams or individuals who are actually within a multi-stage build. There's nothing really against that, but they think the number one purpose of doing multi-stage had been just producing the minimalist image. Um, so just wanted to kind of combine those two answers in one, uh, for sure. >>Yeah, yeah, sure. Um, and with that, um, thank you all for the great questions. We are going to have to wrap this up and we could go for another hour if we all had the time. And if Dr. Khan was a 24 hour long event and it didn't sadly, it's not. So we've got to make room for the next live panel, which will be Peter coming on and talking about security with some developer ex security experts. And I wanted to thank again, thank you all three of you for being here real quick, go around the room. Um, uh, where can people reach out to you? I am, uh, at Bret Fisher on Twitter. You can find me there. Carlos. >>I'm at dev Mandy with a Y D E N D Y that's me, um, >>Easiest name ever on Twitter, Carlos and DFW on LinkedIn. And I also have a LinkedIn learning course. So if you check me out on my LinkedIn learning, >>Yeah. I'm at Nicola Quebec. Um, one word, I'll put it in the chat as well on, on LinkedIn, as well as, uh, uh, as well as Twitter. Thanks for having us, Brett. Yeah. Thanks for being here. >>Um, and, and you all stay around. So if you're in the room with us chatting, you're gonna, you're gonna, if you want to go to see the next live panel, I've got to go back to the beginning and do that whole thing, uh, and find the next, because this one will end, but we'll still be in chat for a few minutes. I think the chat keeps going. I don't actually know. I haven't tried it yet. So we'll find out here in a minute. Um, but thanks you all for being here, I will be back a little bit later, but, uh, coming up next on the live stuff is Peter Wood security. Ciao. Bye.
SUMMARY :
Uh, thank you so much to my guests welcoming into the panel. Virginia, and, uh, I make videos on the internet and courses on you to me, So, um, it's been fun and I'm excited to meet with all of you and talk Uh, just, uh, you know, keeping that, to remember all the good days, um, uh, moving into DX to try and help developers better understand and use our products And so for those of you in chat, the reason we're doing this So feel free to, um, ask what you think is on the top of your And don't have to go talk to a person to run that Um, and so being the former QA on the team, So, um, uh, Carlos, And, you know, So, uh, Nico 81st thoughts on that? kind of the scope that had, uh, you know, now in conferences, what we're using, uh, you know, whether your favorite tools. if you want to do something, you don't have to write the code it's already been tested. You got to unmute. And, you know, the way it works, enterprise CIO CD, if you want, especially if you want to roll your own or own it yourself, um, Um, and you know, the API is really great. I mean, I, I feel with you on the Travis, the, I think, cause I think that was my first time experiencing, And there's probably, you know, And I CA I can't give you a better solution. Um, when you go searching for Docker, and then start browsing for plugins to see if you even want to use those. Some of the things that you input might be available later what say you people? So if you have a lot of small changes that are being made and time-consuming, um, um, you know, within, within your pipeline. hole for the rest of the day on storage management, uh, you know, CP CPU We have, uh, you know, we know get tags and there's Um, it's just clean and I like the timestamp, you know, exactly when it was built. Um, in fact, you know, I'm running into that right now, telling the script, telling the team that maintains a script, Hey, you know, you should use somber and you should start thinking I think you hit on something interesting beyond just how to version, but, um, when to you know, I don't know, having them say, okay, you must tell me what a major version is. If they want it to use some birds great too, which is why I think going back to what you originally said, a consistent packaging solution for me to get my code, you know, Uh, you know, the Docker Docker file is not the most perfect way to describe how to make your app, To that, to that, Brett, um, you know, uh, just maybe more of So similar to how you can think of Terraform and having that pluggability to say Terraform uh, D essentially, do you use compose in your CIO or not Docker compose? different than what you would do in your local, in your local dev. I'm shifting the CIO left to your local development is trying to say, you know, you can get away with just Docker commands. And, um, you know, to your point, the number of CGI cycles to get, you know, the test, the test data that you need. Um, I don't know if you all were able to see the keynote, but there was a, there was a little bit, And you won't have to necessarily have dependencies inside of where you're running it because So that, that would be interesting for those of you that are not watching that one. I'm going to quote Carlos again and say, it depends on, on, you know, how you're talking, you know, And then you have to do something extra to enable that caching, in, in the assets within the Docker file had been, um, you know, Um, yeah, I say you cash until you have a good reason not to personally uh, the more you cash the in a lot of cases with Docker, like the, there's an art form to the more you pen, the less you have, So the other side of this argument is if you trust your testing, then you, and you have better testing to the cash the most and not have to rebuild all those layers. And then day two happens and you built it a second And the type of testing you do, which really, which are just, you know, testing that your app can talk to another component or another you know, someone that actually touches the application, if it's like a website can actually Um, the fewer issues you have, the easier it is to troubleshoot it. So if you don't run your performance test early on, um, and you know, the definition of performance says to me, because one of the things that I've seen when I work So service level objectives, um, or, you know, do I run Hey, if I turn something off or, you know, if I deploy this small broken code to production do you lent all the things, if you do, do you fail the linters during your testing? So, right, right. And in the chat, if you would introduce it in a multi-stage build, And I wanted to thank again, thank you all three of you for being here So if you check me out on my LinkedIn Um, one word, I'll put it in the chat as well on, Um, but thanks you all for being here,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Carlos Nunez | PERSON | 0.99+ |
Carla | PERSON | 0.99+ |
Carlos | PERSON | 0.99+ |
Brett | PERSON | 0.99+ |
Dallas | LOCATION | 0.99+ |
Houston | LOCATION | 0.99+ |
Nico | PERSON | 0.99+ |
Virginia Beach | LOCATION | 0.99+ |
Chavonne | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
December | DATE | 0.99+ |
Mandy | PERSON | 0.99+ |
Khobar | PERSON | 0.99+ |
Carlin | PERSON | 0.99+ |
Jack | PERSON | 0.99+ |
Seattle | LOCATION | 0.99+ |
CIA | ORGANIZATION | 0.99+ |
two points | QUANTITY | 0.99+ |
24 hour | QUANTITY | 0.99+ |
Tasha Corp. | ORGANIZATION | 0.99+ |
Pierre | PERSON | 0.99+ |
Patrick Corp | ORGANIZATION | 0.99+ |
Peter | PERSON | 0.99+ |
Jenkins X | TITLE | 0.99+ |
second point | QUANTITY | 0.99+ |
second challenge | QUANTITY | 0.99+ |
Python | TITLE | 0.99+ |
Docker | TITLE | 0.99+ |
2 cents | QUANTITY | 0.99+ |
10,000 developers | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
both | QUANTITY | 0.99+ |
Austin, Texas | LOCATION | 0.99+ |
Cameron | PERSON | 0.99+ |
two images | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
15 years | QUANTITY | 0.99+ |
Jenkins | TITLE | 0.99+ |
Khan | PERSON | 0.99+ |
HashiCorp | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
each case | QUANTITY | 0.99+ |
Brad | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
three ideas | QUANTITY | 0.99+ |
this year | DATE | 0.99+ |
Quentin | PERSON | 0.98+ |
both sides | QUANTITY | 0.98+ |
Tim | PERSON | 0.98+ |
last year | DATE | 0.98+ |
20 years | QUANTITY | 0.98+ |
Camden | PERSON | 0.98+ |
each step | QUANTITY | 0.98+ |
Two more times | QUANTITY | 0.98+ |
Avi Shua, Orca Security | CUBE Conversation May 2021
(calm music)- Hello, and welcome to this CUBE conversation here in Palo Alto, California in theCUBE Studios, I'm John Furrier, host of theCUBE. We are here with the hot startup really working on some real, super important security technology for the cloud, great company, Orca Security, Avi Shua, CEO, and co founder. Avi, thank you for coming on theCUBE and share your story >> Thanks for having me. >> So one of the biggest problems that enterprises and large scale, people who are going to the cloud and are in the cloud and are evolving with cloud native, have realized that the pace of change and the scale is a benefit to the organizations for the security teams, and getting that security equation, right, is always challenging, and it's changing. You guys have a solution for that, I really want to hear what you guys are doing. I like what you're talking about. I like what you're thinking about, and you have some potentially new technologies. Let's get into it. So before we get started, talk about what is Orca Security, what do you guys do? What problem do you solve? >> So what we invented in Orca, is a unique technology called site scanning, that essentially enables us to connect to any cloud environment in a way which is as simple as installing a smartphone application and getting a full stack visibility of your security posture, meaning seeing all of the risk, whether it's vulnerability, misconfiguration, lateral movement risk, work that already been compromised, and more and more, literally in minutes without deploying any agent, without running any network scanners, literally with no change. And while it sounds to many of us like it can't happen, it's snake oil, it's simply because we are so used to on premise environment where it simply wasn't possible in physical server, but it is possible in the cloud. >> Yeah, and you know, we've had many (indistinct) on theCUBE over the years. One (indistinct) told us that, and this is a direct quote, I'll find the clip and share it on Twitter, but he said, "The cloud is more secure than on premise, because it's more changes going on." And I asked him, "Okay, how'd you do?" He says, "It's hard, you got to stay on top of it." A lot of people go to the cloud, and they see some security benefits with the scale. But there are gaps. You guys are building something that solves those gaps, those blind spots, because of things are always changing, you're adding more services, sometimes you're integrating, you now have containers that could have, for instance, you know, malware on it, gets introduced into a cluster, all kinds of things can go on in a cloud environment, that was fine yesterday, you could have a production cluster that's infected. So you have all of these new things. How do you figure out the gaps and the blind spots? That's what you guys do, I believe, what are the gaps in cloud security? Share with us. >> So definitely, you're completely correct. You know, I totally agree the cloud can be dramatically more secluded on-prem. At the end of the day, unlike an on-prem data center, where someone can can plug a new firewall, plug a new switch, change things. And if you don't instrument, it won't see what's inside. This is not possible in the cloud. In the cloud it's all code. It's all running on one infrastructure that can be used for the instrumentation. On the other hand, the cloud enabled businesses to act dramatically faster, by say dramatically, we're talking about order of magnitude faster, you can create new networks in matter of minutes, workloads can come and go within seconds. And this creates a lot of changes that simply haven't happened before. And it involves a lot of challenges, also from security instrumentation point of view. And you cannot use the same methodologies that you used for the on-prem because if you use them, you're going to lose, they were a compromise, that worked for certain physics, certain set of constraints that no longer apply. And our thesis is that essentially, you need to use the capabilities of the cloud itself, for the instrumentation of everything that can runs on the cloud. And when you do that, by definition, you have full coverage, because if it's run on the cloud, it can be instrumented on cloud, this essentially what Docker does. And you're able to have this full visibility for all of the risks and the importance because all of them, essentially filter workload, which we're able to analyze. >> What are some of the blind spots in the public cloud, for instance. I mean, that you guys are seeing that you guys point out or see with the software and the services that you guys have. >> So the most common ones are the things that we have seen in the last decades. I don't think they are materially different simply on steroids. We see things, services that are launched, nobody maintained for years, using things like improper segmentation, that everyone have permission to access everything. And therefore if one environment is breached, everything is breached. We see organization where something goes dramatically hardened. So people find a way to a very common thing is that, and now ever talks about CIM and the tightening their permission and making sure that every workload have only the capabilities that they need. But sometimes developers are a bit lazy. So they'll walk by that, but also have keys that are stored that can bypass the entire mechanism that, again, everyone can do everything on any environment. So at the end of the day, I think that the most common thing is the standard aging issues, making sure that your environment is patched, it's finger tightened, there is no alternative ways to go to the environment, at scale, because the end of the day, they are destined for security professional, you need to secure everything that they can just need to find one thing that was missed. >> And you guys provide that visibility into the cloud. So to identify those. >> Exactly. I think one of the top reasons that we implemented Orca using (indistinct) technology that I've invented, is essentially because it guarantees coverage. For the first time, we can guarantee you that if you scan it, that way, we'll see every instance, every workload, every container, because of its running, is a native workload, whether it's a Kubernetes, whether it's a service function, we see it all because we don't rely on any (indistinct) integration, we don't rely on friction within the organization. So many times in my career, I've been in discussion with customer that has been breached. And when we get to the core of the issue, it was, you couldn't, you haven't installed that agent, you haven't configured that firewall, the IPS was not up to date. So the protections weren't applied. So this is technically true, but it doesn't solve the customer problem, which is, I need the security to be applied to all of my environment, and I can't rely on people to do manual processes, because they will fail. >> Yeah, yeah. I mean, it's you can't get everything now and the velocity, the volume of activity. So let me just get this right, you guys are scanning container. So the risk I hear a lot is, you know, with Kubernetes, in containers is, a fully secure cluster could have a container come in with malware, and penetrate. And even if it's air gapped, it's still there. So problematic, you would scan that? Is that how it would work? >> So yes, but so for nothing but we are not scanning only containers, the essence of Orca is scanning the cloud environment holistically. We scan your cloud configuration, we scan your Kubernetes configuration, we scan your Dockers, the containers that run on top of them, we scan the images that are installed and we scan the permission that these images are one, and most importantly, we combined these data points. So it's not like you buy one solution that look to AWS configuration, is different solution that locate your virtual machines at one cluster, another one that looks at your cluster configuration. Another one that look at a web server and one that look at identity. And then you have resolved from five different tools that each one of them claims that this is the most important issue. But in fact, you need to infuse the data and understand yourself what is the most important items or they're correlated. We do it in an holistic way. And at the end of the day, security is more about thinking case graphs is vectors, rather than list. So it is to tell you something like this is a container, which is vulnerable, it has permission to access your sensitive data, it's running on a pod that is indirectly connected to the internet to this load balancer, which is exposed. So this is an attack vector that can be utilized, which is just a tool that to say you have a vulnerable containers, but you might have hundreds, where 99% of them are not exposed. >> Got it, so it's really more logical, common sense vectoring versus the old way, which was based on perimeter based control points, right? So is that what I get? is that right is that you're looking at it like okay, a whole new view of it. Not necessarily old way. Is that right? >> Yes, it is right, we are looking at as one problem that is entered in one tool that have one unified data model. And on top of that, one scanning technology that can provide all the necessary data. We are not a tool that say install vulnerability scanner, install identity access management tools and infuse all of the data to Orca will make sense, and if you haven't installed the tools to you, it's not our problem. We are scanning your environment, all of your containers, virtual machine serverless function, cloud configuration using guard technology. When standard risk we put them in a graph and essentially what is the attack vectors that matter for you? >> The sounds like a very promising value proposition. if I've workloads, production workloads, certainly in the cloud and someone comes to me and says you could have essentially a holistic view of your security posture at any given point in that state of operations. I'm going to look at it. So I'm compelled by it. Now tell me how it works. Is there overhead involved? What's the cost to, (indistinct) Australian dollars, but you can (indistinct) share the price to would be great. But like, I'm more thinking of me as a customer. What do I have to do? What operational things, what set up? What's my cost operationally, and is there overhead to performance? >> You won't believe me, but it's almost zero. Deploying Orca is literally three clicks, you just go log into the application, you give it the permission to read only permission to the environment. And it does the rest, it doesn't run a single awkward in the environment, it doesn't send a single packet. It doesn't create any overhead we have within our public customer list companies with a very critical workloads, which are time sensitive, I can quote some names companies like Databricks, Robinhood, Unity, SiteSense, Lemonade, and many others that have critical workloads that have deployed it for all of the environment in a very quick manner with zero interruption to the business continuity. And then focusing on that, because at the end of the day, in large organization, friction is the number one thing that kills security. You want to deploy your security tool, you need to talk with the team, the team says, okay, we need to check it doesn't affect the environment, let's schedule it in six months, in six months is something more urgent then times flybys and think of security team in a large enterprise that needs to coordinate with 500 teams, and make sure it's deployed, it can't work, Because we can guarantee, we do it because we leverage the native cloud capabilities, there will be zero impact. This allows to have the coverage and find these really weak spot nobody's been looking at. >> Yeah, I mean, this having the technology you have is also good, but the security teams are burning out. And this is brings up the cultural issue we were talking before we came on camera around the cultural impact of the security assessment kind of roles and responsibilities inside companies. Could you share your thoughts on this because this is a real dynamic, the people involved as a people process technology, the classic, you know, things that are impacted with digital transformation. But really the cultural impact of how developers push code, the business drivers, how the security teams get involved. And sometimes it's about the security teams are not under the CIO or under these different groups, all kinds of impacts to how the security team behaves in context to how code gets shipped. What's your vision and view on the cultural impact of security in the cloud. >> So, in fact, many times when people say that the cloud is not secure, I say that the culture that came with the cloud, sometimes drive us to non secure processes, or less secure processes. If you think about that, only a decade ago, if an organization could deliver a new service in a year, it would be an amazing achievement, from design to deliver. Now, if an organization cannot ship it, within weeks, it's considered a failure. And this is natural, something that was enabled by the cloud and by the technologies that came with the cloud. But it also created a situation where security teams that used to be some kind of a checkpoint in the way are no longer in that position. They're in one end responsible to audit and make sure that things are acting as they should. But on the other end, things happen without involvement. And this is a very, very tough place to be, nobody wants to be the one that tells the business you can't move as fast as you want. Because the business want to move fast. So this is essentially the friction that exists whether can we move fast? And how can we move fast without breaking things, and without breaking critical security requirements. So I believe that security is always about a triode, of educate, there's nothing better than educate about putting the guardrails to make sure that people cannot make mistakes, but also verify an audit because there will be failures in even if you educate, even if you put guardrails, things won't work as needed. And essentially, our position within this, triode is to audit, to verify to empower the security teams to see exactly what's happening, and this is an enabler for a discussion. Because if you see what are the risks, the fact that you have, you know, you have this environment that hasn't been patched for a decade with the password one to six, it's a different case, then I need you to look at this environment because I'm concerned that I haven't reviewed it in a year. >> That's exactly a great comment. You mentioned friction kills innovation earlier. This is one friction point that mismatch off cadence between ownership of process, business owners goals of shipping fast, security teams wanting to be secure. And developers just want to write code faster too. So productivity, burnout, innovation all are a factor in cloud security. What can a company do to get involved? You mentioned easy to deploy. How do I work with Orca? You guys are just, is it a freemium? What is the business model? How do I engage with you if I'm interested in deploying? >> So one thing that I really love about the way that we work is that you don't need to trust a single word I said, you can get a free trial of Orca at website orca.security, one a scan on your cloud environment, and see for yourself, whether there are critical ways that were overlooked, whether everything is said and there is no need for a tool or whether they some areas that are neglected and can be acted at any given moment (indistinct) been breached. We are not a freemium but we offer free trials. And I'm also a big believer in simplicity and pricing, we just price by the average number workload that you have, you don't need to read a long formula to understand the pricing. >> Reducing friction, it's a very ethos sounds like you guys have a good vision on making things easy and frictionless and sets that what we want. So maybe I should ask you a question. So I want to get your thoughts because a lot of conversations in the industry around shifting left. And that's certainly makes a lot of sense. Which controls insecurity do you want to shift left and which ones you want to shift right? >> So let me put it at, I've been in this industry for more than two decades. And like any industry every one's involved, there is a trend and of something which is super valuable. But some people believe that this is the only thing that you need to do. And if you know Gartner Hype Cycle, at the beginning, every technology is (indistinct) of that. And we believe that this can do everything and then it reaches (indistinct) productivity of the area of the value that it provides. Now, I believe that shifting left is similar to that, of course, you want to shift left as much as possible, you want things to be secure as they go out of the production line. This doesn't mean that you don't need to audit what's actually warning, because everything you know, I can quote, Amazon CTO, Werner Vogels about everything that can take will break, everything fails all the time. You need to assume that everything will fail all the time, including all of the controls that you baked in. So you need to bake as much as possible early on, and audit what's actually happening in your environment to find the gaps, because this is the responsibility of security teams. Now, just checking everything after the fact, of course, it's a bad idea. But only investing in shifting left and education have no controls of what's actually happening is a bad idea as well. >> A lot of people, first of all, great call out there. I totally agree, shift left as much as possible, but also get the infrastructure and your foundational data strategies, right and when you're watching and auditing. I have to ask you the next question on the context of the data, right, because you could audit all day long, all night long. But you're going to have a pile of needles looking for haystack of needles, as they say, and you got to have context. And you got to understand when things can be jumped on. You can have alert fatigue, for instance, you don't know what to look at, you can have too much data. So how do you manage the difference between making the developers productive in the shift left more with the shift right auditing? What's the context and (indistinct)? How do you guys talk about that? Because I can imagine, yeah, it makes sense. But I want to get the right alert at the right time when it matters the most. >> We look at risk as a combination of three things. Risk is not only how pickable the lock is. If I'll come to your office and will tell you that you have security issue, is that they cleaning, (indistinct) that lock can be easily picked. You'll laugh at me, technically, it might be the most pickable lock in your environment. But you don't care because the exposure is limited, you need to get to the office, and there's nothing valuable inside. So I believe that we always need to take, to look at risk as the exposure, who can reach that lock, how easily pickable this lock is, and what's inside, is at your critical plan tools, is it keys that can open another lock that includes this plan tools or just nothing. And when you take this into context, and the one wonderful thing about the cloud, is that for the first time in the history of computing, the data that is necessary to understand the exposure and the impact is in the same place where you can understand also the risk of the locks. You can make a very concise decision of easily (indistinct) that makes sense. That is a critical attack vector, that is a (indistinct) critical vulnerability that is exposed, it is an exposed service and the service have keys that can download all of my data, or maybe it's an internal service, but the port is blocked, and it just have a default web server behind it. And when you take that, you can literally quantize 0.1% of the alert, even less than that, that can be actually exploited versus device that might have the same severity scores or sound is critical, but don't have a risk in terms of exposure or business impact. >> So this is why context matters. I want to just connect what you said earlier and see if I get this right. What you just said about the lock being picked, what's behind the door can be more keys. I mean, they're all there and the thieves know, (indistinct) bad guys know exactly what these vectors are. And they're attacking them. But the context is critical. But now that's what you were getting at before by saying there's no friction or overhead, because the old way was, you know, send probes out there, send people out in the network, send packers to go look at things which actually will clutter the traffic up or, you know, look for patterns, that's reliant on footsteps or whatever metaphor you want to use. You don't do that, because you just wire up the map. And then you put context to things that have weights, I'm imagining graph technologies involved or machine learning. Is that right? Am I getting that kind of conceptually, right, that you guys are laying it out holistically and saying, that's a lock that can be picked, but no one really cares. So no one's going to pick and if they do, there's no consequence, therefore move on and focus energy. Is that kind of getting it right? Can you correct me where I got that off or wrong? >> So you got it completely right. On one end, we do the agentless deep assessment to understand your workloads, your virtual machine or container, your apps and service that exists with them. And using the site scanning technology that some people you know, call the MRI for the cloud. And we build the map to understand what are connected to the security groups, the load balancer, the keys that they hold, what these keys open, and we use this graph to essentially understand the risk. Now we have a graph that includes risk and exposure and trust. And we use this graph to prioritize detect vectors that matters to you. So you might have thousands upon thousands of vulnerabilities on servers that are simply internal and these cannot be manifested, that will be (indistinct) and 0.1% of them, that can be exploited indirectly to a load balancer, and we'll be able to highlight these one. And this is the way to solve alert fatigue. We've been in large organizations that use other tools that they had million critical alerts, using the tools before Orca. We ran our scanner, we found 30. And you can manage 30 alerts if you're a large organization, no one can manage a million alerts. >> Well, I got to say, I love the value proposition. I think you're bringing a smart view of this. I see you have the experience there, Avi and team, congratulations, and it makes sense of the cloud is a benefit, it can be leveraged. And I think security being rethought this way, is smart. And I think it's being validated. Now, I did check the news, you guys have raised significant traction as valuation certainly raised around the funding of (indistinct) 10 million, I believe, a (indistinct) Funding over a billion dollar valuation, pushes a unicorn status. I'm sure that's a reflection of your customer interaction. Could you share customer success that you're having? What's the adoption look like? What are some of the things customers are saying? Why do they like your product? Why is this happening? I mean, I can connect the dots myself, but I want to hear what your customers think. >> So definitely, we're seeing huge traction. We grew by thousands of percent year over year, literally where times during late last year, where our sales team, literally you had to wait two or three weeks till you managed to speak to a seller to work with Orca. And we see the reasons as organization have the same problems that we were in, and that we are focusing. They have cloud environments, they don't know their security posture, they need to own it. And they need to own it now in a way which guarantees coverage guarantees that they'll see the important items and there was no other solution that could do that before Orca. And this is the fact. We literally reduce deployment (indistinct) it takes months to minutes. And this makes it something that can happen rather than being on the roadmap and waiting for the next guy to come and do that. So this is what we hear from our customers and the basic value proposition for Orca haven't changed. We're providing literally Cloud security that actually works that is providing full coverage, comprehensive and contextual, in a seamless manner. >> So talk about the benefits to customers, I'll give you an example. Let's just say theCUBE, we have our own cloud. It's growing like crazy. And we have a DevOps team, very small team, and we start working with big companies, they all want to know what our security posture is. I have to go hire a bunch of security people, do I just work with Orca, because that's the more the trend is integration. I just was talking to another CEO of a hot startup and the platform engineering conversations about people are integrating in the cloud and across clouds and on premises. So integration is all about posture, as well, too I want to know, people want to know who they're working with. How does that, does that factor into anything? Because I think, that's a table stakes for companies to have almost a posture report, almost like an MRI you said, or a clean (indistinct) health. >> So definitely, we are both providing the prioritized risk assessment. So let's say that your cloud team want to check their security, the cloud security risk, they'll will connect Orca, they'll see the (indistinct) in a very, very clear way, what's been compromised (indistinct) zero, what's in an imminent compromise meaning the attacker can utilize today. And you probably want to fix it as soon as possible and things that are hazardous in terms that they are very risky, but there is no clear attack vectors that can utilize them today, there might be things that combining other changes will become imminent compromise. But on top of that, when standard people also have compliance requirements, people are subject to a regulation like PCI CCPA (indistinct) and others. So we also show the results in the lens of these compliance frameworks. So you can essentially export a report showing, okay, we were scanned by Orca, and we comply with all of these requirements of SOC 2, etc. And this is another value proposition of essentially not only showing it in a risk lens, but also from the compliance lens. >> You got to be always on with security and cloud. Avi, great conversation. Thank you for sharing nice knowledge and going deep on some of the solution and appreciate your conversation. Thanks for coming on. >> Thanks for having me. >> Obviously, you are CEO and co founder of Orca Security, hot startup, taking on security in the cloud and getting it right. I'm John Furrier with theCUBE. Thanks for watching. (calm music)
SUMMARY :
technology for the cloud, and are in the cloud and are but it is possible in the cloud. And I asked him, "Okay, how'd you do?" of everything that can runs on the cloud. I mean, that you guys are seeing So at the end of the day, And you guys provide that For the first time, we can guarantee you So the risk I hear a lot is, So it is to tell you something like So is that what I get? and infuse all of the data the price to would be great. And it does the rest, the classic, you know, I say that the culture What is the business model? about the way that we work is that and which ones you want to shift right? that you need to do. I have to ask you the next question is that for the first time that you guys are laying it out that some people you know, What are some of the things and the basic value proposition So talk about the in the lens of these and going deep on some of the solution taking on security in the
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Orca Security | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Orca | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
Avi Shua | PERSON | 0.99+ |
500 teams | QUANTITY | 0.99+ |
May 2021 | DATE | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
30 alerts | QUANTITY | 0.99+ |
99% | QUANTITY | 0.99+ |
Robinhood | ORGANIZATION | 0.99+ |
SiteSense | ORGANIZATION | 0.99+ |
hundreds | QUANTITY | 0.99+ |
0.1% | QUANTITY | 0.99+ |
thousands | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Avi | PERSON | 0.99+ |
SOC 2 | TITLE | 0.99+ |
Lemonade | ORGANIZATION | 0.99+ |
six months | QUANTITY | 0.99+ |
five different tools | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
first time | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
Werner Vogels | PERSON | 0.99+ |
Unity | ORGANIZATION | 0.99+ |
three weeks | QUANTITY | 0.99+ |
three clicks | QUANTITY | 0.99+ |
one tool | QUANTITY | 0.99+ |
single packet | QUANTITY | 0.98+ |
one problem | QUANTITY | 0.98+ |
10 million | QUANTITY | 0.98+ |
a decade ago | DATE | 0.98+ |
late last year | DATE | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.97+ |
CUBE | ORGANIZATION | 0.97+ |
six | QUANTITY | 0.97+ |
a year | QUANTITY | 0.97+ |
30 | QUANTITY | 0.97+ |
more than two decades | QUANTITY | 0.97+ |
each one | QUANTITY | 0.96+ |
one thing | QUANTITY | 0.96+ |
one cluster | QUANTITY | 0.96+ |
one environment | QUANTITY | 0.96+ |
last decades | DATE | 0.95+ |
Kubernetes | TITLE | 0.95+ |
single word | QUANTITY | 0.95+ |
single | QUANTITY | 0.95+ |
thousands of percent | QUANTITY | 0.95+ |
today | DATE | 0.94+ |
orca.security | ORGANIZATION | 0.94+ |
three things | QUANTITY | 0.93+ |
one solution | QUANTITY | 0.92+ |
Gartner Hype Cycle | ORGANIZATION | 0.92+ |
ORGANIZATION | 0.91+ | |
one end | QUANTITY | 0.91+ |
million critical alerts | QUANTITY | 0.91+ |
One | QUANTITY | 0.9+ |
a decade | QUANTITY | 0.89+ |
over a billion dollar | QUANTITY | 0.87+ |
zero impact | QUANTITY | 0.83+ |
million alerts | QUANTITY | 0.8+ |
DevOps | ORGANIZATION | 0.77+ |
theCUBE Studios | ORGANIZATION | 0.77+ |
Kiernan Taylor, Kevin Surace and Issac Sacolick | BizOps Chaos to Clarity 2021
(upbeat music) >> Welcome to this BizOps Manifesto Power Panel, Data Lake or Data Landfill. We're going to be talking about that today. I've got three guests joining me. We're going to dive through that. Kieran Taylor is here the CMO of Broadcom's Enterprise Software Division. Kieran, great to have you on the program. >> Thank you, Lisa. >> Kevin Surace is here as well. Chairman and CTO of Appvance, hey Kevin. >> Hey Lisa. >> And Isaac Sacolick Author and CEO of StarCIO. Isaac, welcome. >> Hi Lisa, thanks for having me. >> So we're going to spend the next 25 to 30 minutes talking about the challenges and the opportunities that data brings to organizations. You guys are going to share some of your best practices for how organizations can actually sort through all this data to make data-driven decisions. We're also going to be citing some statistics from the Inaugural BizOps Industry Survey of the State of Digital Business in which 519 business and technology folks were surveyed across five nations. Let's go ahead and jump right in and the first one in that server that I just mentioned 97% of organizations say we've got data related challenges, limiting the amount of information that we actually have available to the business. Big conundrum there. How do organizations get out of that conundrum? Kieran, we're going to start with you. >> Thanks Lisa. You know, I think, I don't know if it's so much limiting information as it is limiting answers. There's no real shortage of data I don't think being captured, recently met with a unnamed auto manufacturer Who's collecting petabytes of data from their connected cars and they're doing that because they don't really yet know what questions they have of the data. So I think you get out of this Data Landfill conundrum by first understanding what questions to ask. It's not algorithms, it's not analytics. It's not, you know, math that's going to solve this problem. It's really, really understanding your customer's issues and what questions to ask of the data >> Understanding what questions to ask of the data. Kevin, what are your thoughts? >> Yeah, look, I think it gets down to what questions you want to ask and what you want out of it, right? So there's questions you want to ask but what are the business outcomes you're looking for, which is the core of BizOps anyway, right? What are the business outcomes and what business outcomes can I act upon? So there are so many business outcomes you can get from data and you go, well, I can't legally act upon that. I can't practically act upon that. I can't, whether it's lay off people or hire people or whatever it is, right? So what are the actionable items? There is plenty of data. We would argue too much data. Now we could say, is the data good? Is the data bad? Is it poorly organized? Is it, noisy? There's all other problems, right? There's plenty of data. What do I do with it? What can I do that's actionable? If I was an automaker and I had lots of sensors on the road, I had petabytes, as Kieran says and I'd probably bringing in petabytes potentially every day. Well, I could make myself driving systems better. That's an obvious place to start or that's what I would do but I could also potentially use that to change people's insurance and say, if you drive in a certain way something we've never been able to do. If you drive in a certain way, based on the sensors you get a lower insurance rate, then nobody's done that. But now there's interesting business opportunities for that data that you didn't have one minute ago and I just gave away. So, (laughs) it's all about the actionable items in the data. How do you drive something to the top line and the bottom line? 'Cause in the end, that's how we're all measured. >> And Isaac, I know you say data is the lifeblood. What are your thoughts on this conundrum? >> Well, I think, you know, they gave you the start and the end of the equation, start with a question. What are you really trying to answer? What you don't understand that you want to learn about your business connect it to an outcome that is valuable to you. And really what most organizations struggle with is a process that goes through discovery, learning what's in the data, addressing data, quality issues, loading new data sources if required and really doing that iteratively and we're all agile people here at BizOps, right? So doing it iteratively, getting some answers out and understanding what the issues are with the underlying data and then going back and revisiting and reprioritizing what you want to do next. Do you want to go look at another question? Is the answer heading down a path that you can drive outcomes? Do you got to go cleanse some data? So it's really that, how do you put it together so that you can peel the onion back and start looking at data and getting insights out of it. >> Great advice, another challenge though, that the survey identified was that nearly 70% of the respondents and again, 519 business and technology professionals from five countries said, we are struggling to create business metrics from our data with so much data, so much that we can't access. Can you guys share best practices for how organizations would sort through and identify the best data sources from which they can identify the ideal business metrics? Kieran, take it away. >> Sure thing, I guess I'll build on Isaac's statements. Every company has some gap in data, right? And so when you do that, that data gap analysis I think you really, I don't know. It's like Alice in Wonderland, begin at the beginning, right? You start with that question like Isaac said, And I think the best questions are really born from an understanding of what your customers value. And if you dig into that, you understand what the customers value, you build it off of actual customer feedback, market research then you know what questions to ask and then from that, hey, what inputs do I need to really understand how to solve that particular business issue or problem. >> Kevin, what are your thoughts? >> Yeah, I'm going to add to that, completely agree but, look, let's start with sales data, right? So sales data is something, everybody watching this understands, even if they're not in sales, they go well, okay, I understand sales data. What's interesting there is we know who our customers are. We could probably figure out if we have enough data, why they buy, are they buying because of a certain sales person? Are they buying because it's a certain region? Are they buying because of some demographic that we don't understand, but AI can pull out, right? So I would like to know, who's buying and why they're buying. Because if I know that I might make more of what more of those people want whatever that is, certain fundamental sales changes or product changes or whatever it is. So if you could certainly start there, if you start nowhere else, say I sell X today. I'd like to sell X times 1.2 by next year. Okay, great. Can I learn from the last five years of sales, millions of units or million or whatever it is, how to do that better and the answer is for sure yes and yes there's problems with the data and there's holes in the data as Kieran said and there's missing data. It doesn't matter, there's a lot of data around sales. So you can just start there and probably drive some top line growth, just doing what you're already doing but doing it better and learning how to do it better. >> Learning how to do it better. Isaac, talk to us about what your thoughts are here with respect to this challenge. >> Well, when you look at that percentage 70% struggling with business metrics, you know what I see is some companies struggling when they have too few metrics and you know, their KPIs, it really doesn't translate well to people doing work for a customer for an application, responding to an issue. So when you have too few in there too disconnected from the work, people don't understand how to use them and then on the flip side I see other organizations trying to create metrics around every single part of the operation, you know, dozens of different ways of measuring user experience and so forth. And that doesn't work because now we don't know what to prioritize. So I think the art of this is management coming back and saying, what are the metrics? Do we want to see impact and changes over in a short amount of time, over the next quarter, over the next six months and to pick a couple in each category, certainly starting with the customer, certainly looking at sales but then also looking at operations and looking at quality and looking at risk and say to the organization, these are the two or three we're going to focus on in the next six months and then I think that's what simplifies it for organizations. >> Thanks, Isaac. So something that I found interesting, it's not surprising in that the survey found that too much data is one of the biggest challenges that organizations have followed by the limitations that we just talked about in terms of identifying what are the ideal business metrics, but a whopping 74% of survey respondents said we failed to have key data available in real time, which is a big inhibitor for data-driven decision-making. Can you guys offer some advice to organizations? How can they harness this data and glean insights from it faster, Kieran, take it away. >> Yeah, I think there are probably five steps to establishing business KPIs and Lisa your first two questions and these gentleman's answers laid out the first two that is define the questions that you want answers for and then identify what those data inputs would be. You know, if you've got a formula in mind, what data inputs do do you need? The remaining three steps. One is, you know, to evaluate the data you've got and then identify what's missing, you know what do you need to then fetch? And then that fetching, you need to think about the measurement method, the frequency I think Isaac mentioned, you know this concept of tools for all. We have too many tools to collect data. So, the measurement method and frequency is important standardizing on tools and automating that collection wherever possible. And then the last step, this is really the people component of the formula. You need to identify stakeholders that will own those business KPIs and even communicate them within the organization. That human element is sometimes forgotten and it's really important. >> It is important, it's one of the challenges as well. Kevin, talk to us about your thoughts here. >> Yeah, again I mean, for sure you've got in the end you've got the human element. You can give people all kinds of KPIs as Isaac said, often it's too many. You have now KPI the business to death and nobody can get out and do anything that doesn't work. Obviously you can't improve things until you measure them. So you have to measure, we get that. But this question of live data is interesting. My personal view is only certain kinds of data are interesting, absolutely live in the moment. So I think people get in their mind, oh, well if I could deploy IOT everywhere and get instantaneous access within one second to the amalgam of that data, I'm making up words too. That would be interesting. Are you sure that'd be interesting? I might rather analyze the last week of real, real data, really deep analysis, right? Build you know, a real model around that and say, okay for the next week, you ought to do the following. Now I get that if you're in the high-frequency stock trading business you know, every millisecond counts, okay? But most of our businesses do not run by the millisecond and we're not going to make a business decision especially humans involved in a millisecond anyway. We make business decisions based on a fair bit of data, days and weeks. So this is just my own personal opinion. I think people get hung up on this. I've got to have all this live data. No, you want great data analysis using AI and machine learning to evaluate as much data as you can get over whatever period of time that is a week, a month a year and start making some rational decisions off of that information. I think that is how you run a business that's going to crush your competition. >> Good advice, Isaac what are your thoughts on these comments? >> Yeah, I'm going to pair off of Kevin's comments. You know, how do you chip away at this problem at getting more real time data? And I'll share two insights first, from the top down, you know, when StarCIO works with a group of CEO and their executive group, you know how are they getting their data? Well, they're getting it in a boardroom with PowerPoints with spreadsheets behind those PowerPoints, with analysts doing a lot of number crunching and behind all that are all the systems of record around the CRM and the ERP and all the other systems that are telling them how they're performing. And I suggest to them for a month, leave the world of PowerPoint and Excel and bring your analysts in to show you the data live in the systems, ask questions and see what it's like to work with real time data. That first changes the perspective in terms of all the manual work that goes into homogenizing that data for them. But then they start getting used to looking at the tools where the data is actually living. So that's an exercise from the top down from the bottom up when we talk to the it groups, you know so much of our data technologies were built at a time when batch processing in our data centers was the only way to go. We ran these things overnight to move data from point a to point B and with the Cloud, with data streaming technologies it's really a new game in town. And so it's really time for many organizations to modernize and thinking about how they're streaming data. Doesn't necessarily have to be real time. It's not really IOT but it's really saying, I need to have my data updated on a regular basis with an SLA against it so that my teams and my businesses can make good decisions around things. >> So let's talk now about digital transformation. We've been talking about that for years. We talked a lot about in 2020, the acceleration of digital transformation for obvious reasons. But when organizations are facing this data conundrum that we talked about, this sort of data disconnect too much can't get what we need right away. Do we need it right away? How did they flip the script on that so that it doesn't become an impediment to digital transformation but it becomes an accelerant. Kieran >> You know, a lot of times you'll hear vendors talk about technology as being the answer, right? So MI, ML, my math is better than your math, et cetera. And technology is important but it's only effective to the point that which people can actually interpret understand and use the data. And so I would put forth this notion of having data at all levels throughout an organization too often. What you'll see is that I think Isaac mentioned it, you know the data is delivered to the C-suite via PowerPoint and it's been sanitized and scrubbed, et cetera. But heck, by the time it gets to the C-suite it's three weeks old. Data at all levels is making sure that throughout organization, the right people have real-time access to data and can make actionable decisions based upon that. So I think that's a real vital ingredient to successful digital transformation. >> Kevin. >> Well, I like to think of digital transformation as looking at all of your relatively manual or paper-based or other processes whatever they are throughout the organization and saying is this something that can now be done for lack of a better word by a machine, right? And that machine could be algorithms. It could be computers, it could be humans it could be Cloud, it could be AI it could be IOT doesn't really matter. (clears throat) And so there's a reason to do that and of course, the basis of that is the data. You've got to collect data to say, this is how we've been performing. This is what we've been doing. So an example, a simple example of digitalization is people doing RPA around customer support. Now you collect a lot of data on how customer support has been supporting customers. You break that into tiers and you say, here's the easiest, lowest tier. I had farmed that out to probably some other country 20 years ago or 10 years ago. Can I even with the systems in place, can I automate that with a set of processes, Robotic Process Automation that digitizes that process now, Now there still might be, you know 20 different screens that click on all different kinds of things, whatever it is, but can I do that? Can I do it with some Chatbots? Can I do it with it? No, I'm not going to do all the customer support that way but I could probably do a fair bit. Can I digitize that process? Can I digitize the process? Great example we all know is insurance companies taking claims. Okay, I have a phone. Can, I take a picture of my car that just got smashed send it in, let AI analyze it and frankly, do an ACH transfer within the hour, because if it costs them insurance company on average 300 to $500 depending on who they are to process a claim, it's cheaper to just send me the $500 then even question it. And if I did it two or three times, well then I'm trying to steal their money and I should go to jail, right? So these are just, I'm giving these as examples 'cause they're examples that everyone who is watching this would go, oh I understand you're digitizing a process. So now when we get to much more complex processes that we're digitizing in data or hiring or whatever, those are a little harder to understand but I just tried to give those as like everyone understands yes, you should digitize those. Those are obvious, right? >> Now those are great examples, you're right. They're relatable across the board here. Isaac, talk to me about what your thoughts are about. Okay, let's do the conundrum. How do we flip the script and leverage data, access to it insights to drive and facilitate digital transformation rather than impede it. >> Well remember, you know, digital transformation is really about changing the business model, changing how you're working with customers and what markets you're going after. You're being forced to do that because of the pace digital technologies are enabling competitors to outpace you. And so we really like starting digital transformations with a vision. What does this business need to do better, differently more of what markets are we going to go after? What types of technologies are important? And we're going to create that vision but we know long-term planning, doesn't work. We know multi-year planning, doesn't work. So we're going to send our teams out on an agile journey over the next sprint, over the next quarter and we're going to use data to give us information about whether we're heading in the right direction. Should we do more of something? Is this feature higher priority? Is there a certain customer segment that we need to pay attention to more? Is there a set of defects happening in our technology that we have to address? Is there a new competitor stealing market share all that kind of data is what the organization needs to be looking at on a very regular basis to say, do we need to pivot, what we're doing? Do we need to accelerate something? Are we heading in the right direction? Should we give ourselves high fives and celebrate a quick win? Because we've accomplished something 'cause so much of transformation is what we're doing today. We're going to change what we're doing over the next three years, and then guess what? There's going to be a new set of technologies. There's going to be another disruption that we can't anticipate and we want our teams sitting on their toes waiting to look at data and saying, what should we do next? >> That's a great segue Isaac into our last question, which is around culture that's always one of those elephants in the room, right? Because so much cultural transformation is necessary but it's incredibly difficult. So question for you guys, Kieran we'll start with you is, should you advise leadership, should really create a culture, a company-wide culture around data? What do you think? >> Absolutely. I mean, this reminds me of DevOps in many ways and you know, the data has to be shared at all levels and has to empower people to make decisions at their respective levels so that we're not, you know kind of siloed in our knowledge or our decision-making, it's through that collective intelligence that I think organizations can move forward more quickly but they do have to change the culture and they've got to have everyone in the room. Everyone's got a stake in driving business success from the C-suite down to the individual contributor >> Right, Kevin, your thoughts >> You know what? Kieran's right. Data silos, one of the biggest brick walls in all of our way, all the time, you know SecOps says there is no way I'm going to share that database because it's got PII. Okay, well, how about if we strip the PII? Well, then that won't be good for something else and you're getting these huge arguments and if you're not driving it from the top, certainly the CIO, maybe the CFO, maybe the CEO I would argue the CEO, drives it from the top. 'Cause the CEO drives company culture and you know, we talk BizOps and the first word of that is Biz. It's the business, right? It's Ops being driven by business goals and the CEO has to set the business goals. It's not really up to the CIO to set business goals. They're setting operational goals, it's up to the CEO. So when the CEO comes out and says our business goals are to drive up sales by this drive down cost by this drive up speed of product development, whatever it is and we're going to digitize all of our processes to do that. We're going to set in KPIs. We're going to measure everything that we do and everybody's going to work around this table. By the way just like we did with DevOps a decade ago, right? And said, Dev, you actually have to work with Ops now and they go, those dangerous guys way over in that other building, we don't even know who they are but in time people realize that we're all on the same team and that if developers develop something that operations can't host and support and keep alive, it's junk right? And we used to do that and now we're much better at it. And whether it's Dev, SecOps or Dev two-way Ops, whatever all those teams working together. Now we're going to spread that out and make it a bigger pyre on the company and it starts with the CEO. And when the CEO makes it a directive for the company I think we're all going to be successful. >> Isaac, what are your thoughts? >> I think we're really talking about a culture of transformation and a culture of collaboration. I mean, again, everything that we're doing now we're going to build, we're going to learn. We're going to use data to pivot what we're doing. We're going to release a product to customers. We're going to get feedback. We're going to continue to iterate over those things. Same thing when it comes to sales, same things that you know, the experiments that we do for marketing, what we're doing today, we're constantly learning. We're constantly challenging our assumptions. We're trying to throw out the sacred cows with status quo, 'cause we know there's going to be another Island that we have to go after and that's the transformation part. The collaboration part is really you know, what you're hearing. Multiple teams, not just Dev and Ops and not just data and Dev, but really the spectrum of business of product, of stakeholders, of marketing and sales, working with technologists and saying, look this is the things that we need to go after over these time periods and work collaboratively and iteratively around them. And again, the data is the foundation for this, right? And we talk about a learning culture as part of that, the data is a big part of that learning, learning new skills and what new skills to learn is as part of that. But when I think about culture, you know the things that slow down organizations is when they're not transforming fast enough, or they're going in five or six different directions, they're not collaborative enough and the data is the element in there that is an equalizer. It's what you show everybody to say, look what we're doing today is not going to make us survive over the next three years. >> The data equalizer, that sounds like it could be movie coming out in 2021. (laughing) Gentlemen, thank you for walking us through some of those interesting metrics coming out of the BizOps Inaugural Survey. Yes, there are challenges with data. Many of them aren't surprising but there's also a lot of tremendous opportunity and I liked how you kind of brought it around to from a cultural perspective. It's got to start from that C-suite to Kieran's point all the way down. I know we could keep talking, we're out of time, but we'll have to keep following, this as a very interesting topic. One that is certainly pervasive across industries. Thanks guys for sharing your insights. >> Than you. >> Thank you, Lisa. >> Thank you, Lisa. >> For Kieran Taylor, Kevin Surace and Isaac Sacolick. I'm Lisa Martin. Thanks for watching. (upbeat music)
SUMMARY :
Kieran, great to have you on the program. Chairman and CTO of Appvance, hey Kevin. Author and CEO of StarCIO. and the first one in that So I think you get out of questions to ask of the data. and what you want out of it, right? And Isaac, I know you and the end of the equation, and identify the best data sources And so when you do that, but doing it better and learning how to do it better. Learning how to do it better. the operation, you know, dozens in that the survey found and then identify what's missing, you know of the challenges as well. You have now KPI the business to death and behind all that are all the systems to digital transformation it gets to the C-suite and of course, the basis Isaac, talk to me about what We're going to change what we're doing elephants in the room, right? from the C-suite down to and the CEO has to set the business goals. and Dev, but really the and I liked how you kind Surace and Isaac Sacolick.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Kevin | PERSON | 0.99+ |
Isaac Sacolick | PERSON | 0.99+ |
Isaac | PERSON | 0.99+ |
Kieran | PERSON | 0.99+ |
Kevin Surace | PERSON | 0.99+ |
Lisa | PERSON | 0.99+ |
Kieran Taylor | PERSON | 0.99+ |
Kevin Surace | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Issac Sacolick | PERSON | 0.99+ |
Kiernan Taylor | PERSON | 0.99+ |
$500 | QUANTITY | 0.99+ |
2021 | DATE | 0.99+ |
five | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
74% | QUANTITY | 0.99+ |
97% | QUANTITY | 0.99+ |
70% | QUANTITY | 0.99+ |
Excel | TITLE | 0.99+ |
three | QUANTITY | 0.99+ |
PowerPoint | TITLE | 0.99+ |
today | DATE | 0.99+ |
five countries | QUANTITY | 0.99+ |
next year | DATE | 0.99+ |
Broadcom | ORGANIZATION | 0.99+ |
three times | QUANTITY | 0.99+ |
a week | QUANTITY | 0.99+ |
three weeks | QUANTITY | 0.99+ |
five nations | QUANTITY | 0.99+ |
20 different screens | QUANTITY | 0.99+ |
next week | DATE | 0.99+ |
five steps | QUANTITY | 0.99+ |
one second | QUANTITY | 0.99+ |
StarCIO | ORGANIZATION | 0.98+ |
Alice in Wonderland | TITLE | 0.98+ |
first two questions | QUANTITY | 0.98+ |
each category | QUANTITY | 0.98+ |
three guests | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
first word | QUANTITY | 0.98+ |
PowerPoints | TITLE | 0.98+ |
Appvance | ORGANIZATION | 0.98+ |
first one | QUANTITY | 0.98+ |
first two | QUANTITY | 0.98+ |
10 years ago | DATE | 0.98+ |
20 years ago | DATE | 0.98+ |
dozens | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
BizOps | ORGANIZATION | 0.97+ |
a decade ago | DATE | 0.97+ |
last week | DATE | 0.97+ |
Inaugural BizOps Industry Survey | EVENT | 0.96+ |
a month a year | QUANTITY | 0.96+ |
nearly 70% | QUANTITY | 0.96+ |