Image Title

Search Results for about 100 analytics:

Peter Fetterolf, ACG Business Analytics & Charles Tsai, Dell Technologies | MWC Barcelona 2023


 

>> Narrator: TheCUBE's live coverage is made possible by funding from Dell Technologies. Creating technologies that drive human progress. (light airy music) >> Hi, everybody, welcome back to the Fira in Barcelona. My name is Dave Vellante. I'm here with my co-host Dave Nicholson. Lisa Martin is in the house. John Furrier is pounding the news from our Palo Alto studio. We are super excited to be talking about cloud at the edge, what that means. Charles Tsai is here. He's the Senior Director of product management at Dell Technologies and Peter Fetterolf is the Chief Technology Officer at ACG Business Analytics, a firm that goes deep into the TCO and the telco space, among other things. Gents, welcome to theCUBE. Thanks for coming on. Thank you. >> Good to be here. >> Yeah, good to be here. >> So I've been in search all week of the elusive next wave of monetization for the telcos. We know they make great money on connectivity, they're really good at that. But they're all talking about how they can't let this happen again. Meaning we can't let the over the top vendors yet again, basically steal our cookies. So we're going to not mess it up this time. We're going to win in the monetization. Charles, where are those monetization opportunities? Obviously at the edge, the telco cloud at the edge. What is that all about and where's the money? >> Well, Dave, I think from a Dell's perspective, what we want to be able to enable operators is a solution that enable them to roll out services much quicker, right? We know there's a lot of innovation around IoT, MEG and so on and so forth, but they continue to rely on traditional technology and way of operations is going to take them years to enable new services. So what Dell is doing is now, creating the entire vertical stack from the hardware through CAST and automation that enable them, not only to push out services very quickly, but operating them using cloud principles. >> So it's when you say the entire vertical stack, it's the integrated hardware components with like, for example, Red Hat on top- >> Right. >> Or a Wind River? >> That's correct. >> Okay, and then open API, so the developers can create workloads, I presume data companies. We just had a data conversation 'cause that was part of the original stack- >> That's correct. >> So through an open ecosystem, you can actually sort of recreate that value, correct? >> That's correct. >> Okay. >> So one thing Dell is doing, is we are offering an infrastructure block where we are taking over the overhead of certifying every release coming from the Red Hat or the Wind River of the world, right? We want telcos to spend their resources on what is going to generate them revenue. Not the overhead of creating this cloud stack. >> Dave, I remember when we went through this in the enterprise and you had companies like, you know, IBM with the AS400 and the mainframe saying it's easier to manage, which it was, but it's still, you know, it was subsumed by the open systems trend. >> Yeah, yeah. And I think that's an important thing to probe on, is this idea of what is, what exactly does it mean to be cloud at the edge in the telecom space? Because it's a much used term. >> Yeah. >> When we talk about cloud and edge, in sort of generalized IT, but what specifically does it mean? >> Yeah, so when we talk about telco cloud, first of all it's kind of different from what you're thinking about public cloud today. And there's a couple differences. One, if you look at the big hyperscaler public cloud today, they tend to be centralized in huge data centers. Okay, telco cloud, there are big data centers, but then there's also regional data centers. There are edge data centers, which are your typical like access central offices that have turned data centers, and then now even cell sites are becoming mini data centers. So it's distributed. I mean like you could have like, even in a country like say Germany, you'd have 30,000 soul sites, each one of them being a data center. So it's a very different model. Now the other thing I want to go back to the question of monetization, okay? So how do you do monetization? The only way to do that, is to be able to offer new services, like Charles said. How do you offer new services? You have to have an open ecosystem that's going to be very, very flexible. And if we look at where telcos are coming from today, they tend to be very inflexible 'cause they're all kind of single vendor solutions. And even as we've moved to virtualization, you know, if you look at packet core for instance, a lot of them are these vertical stacks of say a Nokia or Ericson or Huawei where you know, you can't really put any other vendors or any other solutions into that. So basically the idea is this kind of horizontal architecture, right? Where now across, not just my central data centers, but across my edge data centers, which would be traditionally my access COs, as well as my cell sites. I have an open environment. And we're kind of starting with, you know, packet core obviously with, and UPFs being distributed, but now open ran or virtual ran, where I can have CUs and DUs and I can split CUs, they could be at the soul site, they could be in edge data centers. But then moving forward, we're going to have like MEG, which are, you know, which are new kinds of services, you know, could be, you know, remote cars it could be gaming, it could be the Metaverse. And these are going to be a multi-vendor environment. So one of the things you need to do is you need to have you know, this cloud layer, and that's what Charles was talking about with the infrastructure blocks is helping the service providers do that, but they still own their infrastructure. >> Yeah, so it's still not clear to me how the service providers win that game but we can maybe come back to that because I want to dig into TCO a little bit. >> Sure. >> Because I have a lot of friends at Dell. I don't have a lot of friends at HPE. I've always been critical when they take an X86 server put a name on it that implies edge and they throw it over the fence to the edge, that's not going to work, okay? We're now seeing, you know we were just at the Dell booth yesterday, you did the booth crawl, which was awesome. Purpose-built servers for this environment. >> Charles: That's right. >> So there's two factors here that I want to explore in TCO. One is, how those next gen servers compare to the previous gen, especially in terms of power consumption but other factors and then how these sort of open ran, open ecosystem stacks compared to proprietary stacks. Peter, can you help us understand those? >> Yeah, sure. And Charles can comment on this as well. But I mean there, there's a couple areas. One is just moving the next generation. So especially on the Intel side, moving from Ice Lake to the Sapphire Rapids is a big deal, especially when it comes to the DU. And you know, with the radios, right? There's the radio unit, the RU, and then there's the DU the distributed unit, and the CU. The DU is really like part of the radio, but it's virtualized. When we moved from Ice lake to Sapphire Rapids, which is third generation intel to fourth generation intel, we're literally almost doubling the performance in the DU. And that's really important 'cause it means like almost half the number of servers and we're talking like 30, 40, 50,000 servers in some cases. So, you know, being able to divide that by two, that's really big, right? In terms of not only the the cost but all the TCO and the OpEx. Now another area that's really important, when I was talking moving from these vertical silos to the horizontal, the issue with the vertical silos is, you can't place any other workloads into those silos. So it's kind of inefficient, right? Whereas when we have the horizontal architecture, now you can place workloads wherever you want, which basically also means less servers but also more flexibility, more service agility. And then, you know, I think Charles can comment more, specifically on the XR8000, some things Dell's doing, 'cause it's really exciting relative to- >> Sure. >> What's happening in there. >> So, you know, when we start looking at putting compute at the edge, right? We recognize the first thing we have to do is understand the environment we are going into. So we spend with a lot of time with telcos going to the south side, going to the edge data center, looking at operation, how do the engineer today deal with maintenance replacement at those locations? Then based on understanding the operation constraints at those sites, we create innovation and take a traditional server, remodel it to make sure that we minimize the disruption to the operations, right? Just because we are helping them going from appliances to open compute, we do not want to disrupt what is have been a very efficient operation on the remote sites. So we created a lot of new ideas and develop them on general compute, where we believe we can save a lot of headache and disruptions and still provide the same level of availability, resiliency, and redundancy on an open compute platform. >> So when we talk about open, we don't mean generic? Fair? See what I mean? >> Open is more from the software workload perspective, right? A Dell server can run any type of workload that customer intend. >> But it's engineered for this? >> Environment. >> Environment. >> That's correct. >> And so what are some of the environmental issues that are dealt with in the telecom space that are different than the average data center? >> The most basic one, is in most of the traditional cell tower, they are deployed within cabinets instead of racks. So they are depth constraints that you just have no access to the rear of the chassis. So that means on a server, is everything you need to access, need to be in the front, nothing should be in the back. Then you need to consider how labor union come into play, right? There's a lot of constraint on who can go to a cell tower and touch power, who can go there and touch compute, right? So we minimize all that disruption through a modular design and make it very efficient. >> So when we took a look at XR8000, literally right here, sitting on the desk. >> Uh-huh. >> Took it apart, don't panic, just pulled out some sleds and things. >> Right, right. >> One of the interesting demonstrations was how it compared to the size of a shoe. Now apparently you hired someone at Dell specifically because they wear a size 14 shoe, (Charles laughs) so it was even more dramatic. >> That's right. >> But when you see it, and I would suggest that viewers go back and take a look at that segment, specifically on the hardware. You can see exactly what you just referenced. This idea that everything is accessible from the front. Yeah. >> So I want to dig in a couple things. So I want to push back a little bit on what you were saying about the horizontal 'cause there's the benefit, if you've got the horizontal infrastructure, you can run a lot more workloads. But I compare it to the enterprise 'cause I, that was the argument, I've made that argument with converged infrastructure versus say an Oracle vertical stack, but it turned out that actually Oracle ran Oracle better, okay? Is there an analog in telco or is this new open architecture going to be able to not only service the wide range of emerging apps but also be as resilient as the proprietary infrastructure? >> Yeah and you know, before I answer that, I also want to say that we've been writing a number of white papers. So we have actually three white papers we've just done with Dell looking at infrastructure blocks and looking at vertical versus horizontal and also looking at moving from the previous generation hardware to the next generation hardware. So all those details, you can find the white papers, and you can find them either in the Dell website or at the ACG research website >> ACGresearch.com? >> ACG research. Yeah, if you just search ACG research, you'll find- >> Yeah. >> Lots of white papers on TCO. So you know, what I want to say, relative to the vertical versus horizontal. Yeah, obviously in the vertical side, some of those things will run well, I mean it won't have issues. However, that being said, as we move to cloud native, you know, it's very high performance, okay? In terms of the stack, whether it be a Red Hat or a VMware or other cloud layers, that's really become much more mature. It now it's all CNF base, which is really containerized, very high performance. And so I don't think really performance is an issue. However, my feeling is that, if you want to offer new services and generate new revenue, you're not going to do it in vertical stacks, period. You're going to be able to do a packet core, you'll be able to do a ran over here. But now what if I want to offer a gaming service? What if I want to do metaverse? What if I want to do, you have to have an environment that's a multi-vendor environment that supports an ecosystem. Even in the RAN, when we look at the RIC, and the xApps and the rApps, these are multi-vendor environments that's going to create a lot of flexibility and you can't do that if you're restricted to, I can only have one vendor running on this hardware. >> Yeah, we're seeing these vendors work together and create RICs. That's obviously a key point, but what I'm hearing is that there may be trade offs, but the incremental value is going to overwhelm that. Second question I have, Peter is, TCO, I've been hearing a lot about 30%, you know, where's that 30% come from? Is it Op, is it from an OpEx standpoint? Is it labor, is it power? Is it, you mentioned, you know, cutting the number of servers in half. If I can unpack the granularity of that TCO, where's the benefit coming from? >> Yeah, the answer is yes. (Peter and Charles laugh) >> Okay, we'll do. >> Yeah, so- >> One side that, in terms of, where is the big bang for the bucks? >> So I mean, so you really need to look at the white paper to see details, but definitely power, definitely labor, definitely reducing the number of servers, you know, reducing the CapEx. The other thing is, is as you move to this really next generation horizontal telco cloud, there's the whole automation and orchestration, that is a key component as well. And it's enabled by what Dell is doing. It's enabled by the, because the thing is you're not going to have end-to-end automation if you have all this legacy stuff there or if you have these vertical stacks where you can't integrate. I mean you can automate that part and then you have separate automation here, you separate. you need to have integrated automation and orchestration across the whole thing. >> One other point I would add also, right, on the hardware perspective, right? With the customized hardware, what we allow operator to do is, take out the existing appliance and push a edge optimized server without reworking the entire infrastructure. There is a significant saving where you don't have to rethink about what is my power infrastructure, right? What is my security infrastructure? The server is designed to leverage the existing, what is already there. >> How should telco, Charles, plan for this transformation? Are there specific best practices that you would recommend in terms of the operational model? >> Great question. I think first thing is do an inventory of what you have. Understand what your constraints are and then come to Dell, we will love to consult with you, based on our experience on the best practices. We know how to minimize additional changes. We know how to help your support engineer, understand how to shift appliance based operation to a cloud-based operation. >> Is that a service you offer? Is that a pre-sales freebie? What is maybe both? >> It's both. >> Yeah. >> It's both. >> Yeah. >> Guys- >> Just really quickly. >> We're going to wrap. >> The, yeah. Dave loves the TCO discussion. I'm always thinking in terms of, well how do you measure TCO when you're comparing something where you can't do something to an environment where you're going to be able to do something new? And I know that that's always the challenge in any kind of emerging market where things are changing, any? >> Well, I mean we also look at, not only TCO, but we look at overall business case. So there's basically service at GLD and revenue and then there's faster time to revenues. Well, and actually ACG, we actually have a platform called the BAE or Business Analytics Engine that's a very sophisticated simulation cloud-based platform, where we can actually look at revenue month by month. And we look at what's the impact of accelerating revenue by three months. By four months. >> So you're looking into- >> By six months- >> So you're forward looking. You're just not consistently- >> So we're not just looking at TCO, we're looking at the overall business case benefit. >> Yeah, exactly right. There's the TCO, which is the hard dollars. >> Right. >> CFO wants to see that, he or she needs to see that. But you got to, you can convince that individual, that there's a business case around it. >> Peter: Yeah. >> And then you're going to sign up for that number. >> Peter: Yeah. >> And they're going to be held to it. That's the story the world wants. >> At the end of the day, telcos have to be offered new services 'cause look at all the money that's been spent. >> Dave: Yeah, that's right. >> On investment on 5G and everything else. >> 0.5 trillion over the next seven years. All right, guys, we got to go. Sorry to cut you off. >> Okay, thank you very much. >> But we're wall to wall here. All right, thanks so much for coming on. >> Dave: Fantastic. >> All right, Dave Vellante, for Dave Nicholson. Lisa Martin's in the house. John Furrier in Palo Alto Studios. Keep it right there. MWC 23 live from the Fira in Barcelona. (light airy music)

Published Date : Mar 1 2023

SUMMARY :

that drive human progress. and Peter Fetterolf is the of the elusive next wave of creating the entire vertical of the original stack- or the Wind River of the world, right? AS400 and the mainframe in the telecom space? So one of the things you need to do how the service providers win that game the fence to the edge, to the previous gen, So especially on the Intel side, We recognize the first thing we have to do from the software workload is in most of the traditional cell tower, sitting on the desk. Took it apart, don't panic, One of the interesting demonstrations accessible from the front. But I compare it to the Yeah and you know, Yeah, if you just search ACG research, and the xApps and the rApps, but the incremental value Yeah, the answer is yes. and then you have on the hardware perspective, right? inventory of what you have. Dave loves the TCO discussion. and then there's faster time to revenues. So you're forward looking. So we're not just There's the TCO, But you got to, you can And then you're going to That's the story the world wants. At the end of the day, and everything else. Sorry to cut you off. But we're wall to wall here. Lisa Martin's in the house.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave NicholsonPERSON

0.99+

Lisa MartinPERSON

0.99+

Dave VellantePERSON

0.99+

DavePERSON

0.99+

IBMORGANIZATION

0.99+

CharlesPERSON

0.99+

Charles TsaiPERSON

0.99+

Peter FetterolfPERSON

0.99+

NokiaORGANIZATION

0.99+

EricsonORGANIZATION

0.99+

HuaweiORGANIZATION

0.99+

John FurrierPERSON

0.99+

PeterPERSON

0.99+

DellORGANIZATION

0.99+

Dell TechnologiesORGANIZATION

0.99+

30QUANTITY

0.99+

telcoORGANIZATION

0.99+

bothQUANTITY

0.99+

ACG Business AnalyticsORGANIZATION

0.99+

30%QUANTITY

0.99+

three monthsQUANTITY

0.99+

ACGORGANIZATION

0.99+

TCOORGANIZATION

0.99+

four monthsQUANTITY

0.99+

Palo AltoLOCATION

0.99+

BarcelonaLOCATION

0.99+

IntelORGANIZATION

0.99+

SecondQUANTITY

0.99+

twoQUANTITY

0.99+

0.5 trillionQUANTITY

0.99+

yesterdayDATE

0.99+

two factorsQUANTITY

0.99+

six monthsQUANTITY

0.98+

todayDATE

0.98+

OracleORGANIZATION

0.98+

MWC 23EVENT

0.98+

GermanyLOCATION

0.98+

Red HatORGANIZATION

0.98+

oneQUANTITY

0.98+

XR8000COMMERCIAL_ITEM

0.98+

Ice LakeCOMMERCIAL_ITEM

0.98+

OneQUANTITY

0.97+

one vendorQUANTITY

0.97+

Palo Alto StudiosLOCATION

0.97+

third generationQUANTITY

0.97+

fourth generationQUANTITY

0.96+

40, 50,000 serversQUANTITY

0.96+

theCUBEORGANIZATION

0.96+

telcosORGANIZATION

0.95+

telco cloudORGANIZATION

0.95+

each oneQUANTITY

0.95+

theCUBE's New Analyst Talks Cloud & DevOps


 

(light music) >> Hi everybody. Welcome to this Cube Conversation. I'm really pleased to announce a collaboration with Rob Strechay. He's a guest cube analyst, and we'll be working together to extract the signal from the noise. Rob is a long-time product pro, working at a number of firms including AWS, HP, HPE, NetApp, Snowplow. I did a stint as an analyst at Enterprise Strategy Group. Rob, good to see you. Thanks for coming into our Marlboro Studios. >> Well, thank you for having me. It's always great to be here. >> I'm really excited about working with you. We've known each other for a long time. You've been in the Cube a bunch. You know, you're in between gigs, and I think we can have a lot of fun together. Covering events, covering trends. So. let's get into it. What's happening out there? We're sort of exited the isolation economy. Things were booming. Now, everybody's tapping the brakes. From your standpoint, what are you seeing out there? >> Yeah. I'm seeing that people are really looking how to get more out of their data. How they're bringing things together, how they're looking at the costs of Cloud, and understanding how are they building out their SaaS applications. And understanding that when they go in and actually start to use Cloud, it's not only just using the base services anymore. They're looking at, how do I use these platforms as a service? Some are easier than others, and they're trying to understand, how do I get more value out of that relationship with the Cloud? They're also consolidating the number of Clouds that they have, I would say to try to better optimize their spend, and getting better pricing for that matter. >> Are you seeing people unhook Clouds, or just reduce maybe certain Cloud activities and going maybe instead of 60/40 going 90/10? >> Correct. It's more like the 90/10 type of rule where they're starting to say, Hey I'm not going to get rid of Azure or AWS or Google. I'm going to move a portion of this over that I was using on this one service. Maybe I got a great two-year contract to start with on this platform as a service or a database as a service. I'm going to unhook from that and maybe go with an independent. Maybe with something like a Snowflake or a Databricks on top of another Cloud, so that I can consolidate down. But it also gives them more flexibility as well. >> In our last breaking analysis, Rob, we identified six factors that were reducing Cloud consumption. There were factors and customer tactics. And I want to get your take on this. So, some of the factors really, you got fewer mortgage originations. FinTech, obviously big Cloud user. Crypto, not as much activity there. Lower ad spending means less Cloud. And then one of 'em, which you kind of disagreed with was less, less analytics, you know, fewer... Less frequency of calculations. I'll come back to that. But then optimizing compute using Graviton or AMD instances moving to cheaper storage tiers. That of course makes sense. And then optimize pricing plans. Maybe going from On Demand, you know, to, you know, instead of pay by the drink, buy in volume. Okay. So, first of all, do those make sense to you with the exception? We'll come back and talk about the analytics piece. Is that what you're seeing from customers? >> Yeah, I think so. I think that was pretty much dead on with what I'm seeing from customers and the ones that I go out and talk to. A lot of times they're trying to really monetize their, you know, understand how their business utilizes these Clouds. And, where their spend is going in those Clouds. Can they use, you know, lower tiers of storage? Do they really need the best processors? Do they need to be using Intel or can they get away with AMD or Graviton 2 or 3? Or do they need to move in? And, I think when you look at all of these Clouds, they always have pricing curves that are arcs from the newest to the oldest stuff. And you can play games with that. And understanding how you can actually lower your costs by looking at maybe some of the older generation. Maybe your application was written 10 years ago. You don't necessarily have to be on the best, newest processor for that application per se. >> So last, I want to come back to this whole analytics piece. Last June, I think it was June, Dev Ittycheria, who's the-- I call him Dev. Spelled Dev, pronounced Dave. (chuckles softly) Same pronunciation, different spelling. Dev Ittycheria, CEO of Mongo, on the earnings call. He was getting, you know, hit. Things were starting to get a little less visible in terms of, you know, the outlook. And people were pushing him like... Because you're in the Cloud, is it easier to dial down? And he said, because we're the document database, we support transaction applications. We're less discretionary than say, analytics. Well on the Snowflake earnings call, that same month or the month after, they were all over Slootman and Scarpelli. Oh, the Mongo CEO said that they're less discretionary than analytics. And Snowflake was an interesting comment. They basically said, look, we're the Cloud. You can dial it up, you can dial it down, but the area under the curve over a period of time is going to be the same, because they get their customers to commit. What do you say? You disagreed with the notion that people are running their calculations less frequently. Is that because they're trying to do a better job of targeting customers in near real time? What are you seeing out there? >> Yeah, I think they're moving away from using people and more expensive marketing. Or, they're trying to figure out what's my Google ad spend, what's my Meta ad spend? And what they're trying to do is optimize that spend. So, what is the return on advertising, or the ROAS as they would say. And what they're looking to do is understand, okay, I have to collect these analytics that better understand where are these people coming from? How do they get to my site, to my store, to my whatever? And when they're using it, how do they they better move through that? What you're also seeing is that analytics is not only just for kind of the retail or financial services or things like that, but then they're also, you know, using that to make offers in those categories. When you move back to more, you know, take other companies that are building products and SaaS delivered products. They may actually go and use this analytics for making the product better. And one of the big reasons for that is maybe they're dialing back how many product managers they have. And they're looking to be more data driven about how they actually go and build the product out or enhance the product. So maybe they're, you know, an online video service and they want to understand why people are either using or not using the whiteboard inside the product. And they're collecting a lot of that product analytics in a big way so that they can go through that. And they're doing it in a constant manner. This first party type tracking within applications is growing rapidly by customers. >> So, let's talk about who wins in that. So, obviously the Cloud guys, AWS, Google and Azure. I want to come back and unpack that a little bit. Databricks and Snowflake, we reported on our last breaking analysis, it kind of on a collision course. You know, a couple years ago we were thinking, okay, AWS, Snowflake and Databricks, like perfect sandwich. And then of course they started to become more competitive. My sense is they still, you know, compliment each other in the field, right? But, you know, publicly, they've got bigger aspirations, they get big TAMs that they're going after. But it's interesting, the data shows that-- So, Snowflake was off the charts in terms of spending momentum and our EPR surveys. Our partner down in New York, they kind of came into line. They're both growing in terms of market presence. Databricks couldn't get to IPO. So, we don't have as much, you know, visibility on their financials. You know, Snowflake obviously highly transparent cause they're a public company. And then you got AWS, Google and Azure. And it seems like AWS appears to be more partner friendly. Microsoft, you know, depends on what market you're in. And Google wants to sell BigQuery. >> Yeah. >> So, what are you seeing in the public Cloud from a data platform perspective? >> Yeah. I think that was pretty astute in what you were talking about there, because I think of the three, Google is definitely I think a little bit behind in how they go to market with their partners. Azure's done a fantastic job of partnering with these companies to understand and even though they may have Synapse as their go-to and where they want people to go to do AI and ML. What they're looking at is, Hey, we're going to also be friendly with Snowflake. We're also going to be friendly with a Databricks. And I think that, Amazon has always been there because that's where the market has been for these developers. So, many, like Databricks' and the Snowflake's have gone there first because, you know, Databricks' case, they built out on top of S3 first. And going and using somebody's object layer other than AWS, was not as simple as you would think it would be. Moving between those. >> So, one of the financial meetups I said meetup, but the... It was either the CEO or the CFO. It was either Slootman or Scarpelli talking at, I don't know, Merrill Lynch or one of the other financial conferences said, I think it was probably their Q3 call. Snowflake said 80% of our business goes through Amazon. And he said to this audience, the next day we got a call from Microsoft. Hey, we got to do more. And, we know just from reading the financial statements that Snowflake is getting concessions from Amazon, they're buying in volume, they're renegotiating their contracts. Amazon gets it. You know, lower the price, people buy more. Long term, we're all going to make more money. Microsoft obviously wants to get into that game with Snowflake. They understand the momentum. They said Google, not so much. And I've had customers tell me that they wanted to use Google's AI with Snowflake, but they can't, they got to go to to BigQuery. So, honestly, I haven't like vetted that so. But, I think it's true. But nonetheless, it seems like Google's a little less friendly with the data platform providers. What do you think? >> Yeah, I would say so. I think this is a place that Google looks and wants to own. Is that now, are they doing the right things long term? I mean again, you know, you look at Google Analytics being you know, basically outlawed in five countries in the EU because of GDPR concerns, and compliance and governance of data. And I think people are looking at Google and BigQuery in general and saying, is it the best place for me to go? Is it going to be in the right places where I need it? Still, it's still one of the largest used databases out there just because it underpins a number of the Google services. So you almost get, like you were saying, forced into BigQuery sometimes, if you want to use the tech on top. >> You do strategy. >> Yeah. >> Right? You do strategy, you do messaging. Is it the right call by Google? I mean, it's not a-- I criticize Google sometimes. But, I'm not sure it's the wrong call to say, Hey, this is our ace in the hole. >> Yeah. >> We got to get people into BigQuery. Cause, first of all, BigQuery is a solid product. I mean it's Cloud native and it's, you know, by all, it gets high marks. So, why give the competition an advantage? Let's try to force people essentially into what is we think a great product and it is a great product. The flip side of that is, they're giving up some potential partner TAM and not treating the ecosystem as well as one of their major competitors. What do you do if you're in that position? >> Yeah, I think that that's a fantastic question. And the question I pose back to the companies I've worked with and worked for is, are you really looking to have vendor lock-in as your key differentiator to your service? And I think when you start to look at these companies that are moving away from BigQuery, moving to even, Databricks on top of GCS in Google, they're looking to say, okay, I can go there if I have to evacuate from GCP and go to another Cloud, I can stay on Databricks as a platform, for instance. So I think it's, people are looking at what platform as a service, database as a service they go and use. Because from a strategic perspective, they don't want that vendor locking. >> That's where Supercloud becomes interesting, right? Because, if I can run on Snowflake or Databricks, you know, across Clouds. Even Oracle, you know, they're getting into business with Microsoft. Let's talk about some of the Cloud players. So, the big three have reported. >> Right. >> We saw AWSs Cloud growth decelerated down to 20%, which is I think the lowest growth rate since they started to disclose public numbers. And they said they exited, sorry, they said January they grew at 15%. >> Yeah. >> Year on year. Now, they had some pretty tough compares. But nonetheless, 15%, wow. Azure, kind of mid thirties, and then Google, we had kind of low thirties. But, well behind in terms of size. And Google's losing probably almost $3 billion annually. But, that's not necessarily a bad thing by advocating and investing. What's happening with the Cloud? Is AWS just running into the law, large numbers? Do you think we can actually see a re-acceleration like we have in the past with AWS Cloud? Azure, we predicted is going to be 75% of AWS IAS revenues. You know, we try to estimate IAS. >> Yeah. >> Even though they don't share that with us. That's a huge milestone. You'd think-- There's some people who have, I think, Bob Evans predicted a while ago that Microsoft would surpass AWS in terms of size. You know, what do you think? >> Yeah, I think that Azure's going to keep to-- Keep growing at a pretty good clip. I think that for Azure, they still have really great account control, even though people like to hate Microsoft. The Microsoft sellers that are out there making those companies successful day after day have really done a good job of being in those accounts and helping people. I was recently over in the UK. And the UK market between AWS and Azure is pretty amazing, how much Azure there is. And it's growing within Europe in general. In the states, it's, you know, I think it's growing well. I think it's still growing, probably not as fast as it is outside the U.S. But, you go down to someplace like Australia, it's also Azure. You hear about Azure all the time. >> Why? Is that just because of the Microsoft's software state? It's just so convenient. >> I think it has to do with, you know, and you can go with the reasoning they don't break out, you know, Office 365 and all of that out of their numbers is because they have-- They're in all of these accounts because the office suite is so pervasive in there. So, they always have reasons to go back in and, oh by the way, you're on these old SQL licenses. Let us move you up here and we'll be able to-- We'll support you on the old version, you know, with security and all of these things. And be able to move you forward. So, they have a lot of, I guess you could say, levers to stay in those accounts and be interesting. At least as part of the Cloud estate. I think Amazon, you know, is hitting, you know, the large number. Laws of large numbers. But I think that they're also going through, and I think this was seen in the layoffs that they were making, that they're looking to understand and have profitability in more of those services that they have. You know, over 350 odd services that they have. And you know, as somebody who went there and helped to start yet a new one, while I was there. And finally, it went to beta back in September, you start to look at the fact that, that number of services, people, their own sellers don't even know all of their services. It's impossible to comprehend and sell that many things. So, I think what they're going through is really looking to rationalize a lot of what they're doing from a services perspective going forward. They're looking to focus on more profitable services and bringing those in. Because right now it's built like a layer cake where you have, you know, S3 EBS and EC2 on the bottom of the layer cake. And then maybe you have, you're using IAM, the authorization and authentication in there and you have all these different services. And then they call it EMR on top. And so, EMR has to pay for that entire layer cake just to go and compete against somebody like Mongo or something like that. So, you start to unwind the costs of that. Whereas Azure, went and they build basically ground up services for the most part. And Google kind of falls somewhere in between in how they build their-- They're a sort of layer cake type effect, but not as many layers I guess you could say. >> I feel like, you know, Amazon's trying to be a platform for the ecosystem. Yes, they have their own products and they're going to sell. And that's going to drive their profitability cause they don't have to split the pie. But, they're taking a piece of-- They're spinning the meter, as Ziyas Caravalo likes to say on every time Snowflake or Databricks or Mongo or Atlas is, you know, running on their system. They take a piece of the action. Now, Microsoft does that as well. But, you look at Microsoft and security, head-to-head competitors, for example, with a CrowdStrike or an Okta in identity. Whereas, it seems like at least for now, AWS is a more friendly place for the ecosystem. At the same time, you do a lot of business in Microsoft. >> Yeah. And I think that a lot of companies have always feared that Amazon would just throw, you know, bodies at it. And I think that people have come to the realization that a two pizza team, as Amazon would call it, is eight people. I think that's, you know, two slices per person. I'm a little bit fat, so I don't know if that's enough. But, you start to look at it and go, okay, if they're going to start out with eight engineers, if I'm a startup and they're part of my ecosystem, do I really fear them or should I really embrace them and try to partner closer with them? And I think the smart people and the smart companies are partnering with them because they're realizing, Amazon, unless they can see it to, you know, a hundred million, $500 million market, they're not going to throw eight to 16 people at a problem. I think when, you know, you could say, you could look at the elastic with OpenSearch and what they did there. And the licensing terms and the battle they went through. But they knew that Elastic had a huge market. Also, you had a number of ecosystem companies building on top of now OpenSearch, that are now domain on top of Amazon as well. So, I think Amazon's being pretty strategic in how they're doing it. I think some of the-- It'll be interesting. I think this year is a payout year for the cuts that they're making to some of the services internally to kind of, you know, how do we take the fat off some of those services that-- You know, you look at Alexa. I don't know how much revenue Alexa really generates for them. But it's a means to an end for a number of different other services and partners. >> What do you make of this ChatGPT? I mean, Microsoft obviously is playing that card. You want to, you want ChatGPT in the Cloud, come to Azure. Seems like AWS has to respond. And we know Google is, you know, sharpening its knives to come up with its response. >> Yeah, I mean Google just went and talked about Bard for the first time this week and they're in private preview or I guess they call it beta, but. Right at the moment to select, select AI users, which I have no idea what that means. But that's a very interesting way that they're marketing it out there. But, I think that Amazon will have to respond. I think they'll be more measured than say, what Google's doing with Bard and just throwing it out there to, hey, we're going into beta now. I think they'll look at it and see where do we go and how do we actually integrate this in? Because they do have a lot of components of AI and ML underneath the hood that other services use. And I think that, you know, they've learned from that. And I think that they've already done a good job. Especially for media and entertainment when you start to look at some of the ways that they use it for helping do graphics and helping to do drones. I think part of their buy of iRobot was the fact that iRobot was a big user of RoboMaker, which is using different models to train those robots to go around objects and things like that, so. >> Quick touch on Kubernetes, the whole DevOps World we just covered. The Cloud Native Foundation Security, CNCF. The security conference up in Seattle last week. First time they spun that out kind of like reinforced, you know, AWS spins out, reinforced from reinvent. Amsterdam's coming up soon, the CubeCon. What should we expect? What's hot in Cubeland? >> Yeah, I think, you know, Kubes, you're going to be looking at how OpenShift keeps growing and I think to that respect you get to see the momentum with people like Red Hat. You see others coming up and realizing how OpenShift has gone to market as being, like you were saying, partnering with those Clouds and really making it simple. I think the simplicity and the manageability of Kubernetes is going to be at the forefront. I think a lot of the investment is still going into, how do I bring observability and DevOps and AIOps and MLOps all together. And I think that's going to be a big place where people are going to be looking to see what comes out of CubeCon in Amsterdam. I think it's that manageability ease of use. >> Well Rob, I look forward to working with you on behalf of the whole Cube team. We're going to do more of these and go out to some shows extract the signal from the noise. Really appreciate you coming into our studio. >> Well, thank you for having me on. Really appreciate it. >> You're really welcome. All right, keep it right there, or thanks for watching. This is Dave Vellante for the Cube. And we'll see you next time. (light music)

Published Date : Feb 7 2023

SUMMARY :

I'm really pleased to It's always great to be here. and I think we can have the number of Clouds that they have, contract to start with those make sense to you And, I think when you look in terms of, you know, the outlook. And they're looking to My sense is they still, you know, in how they go to market And he said to this audience, is it the best place for me to go? You do strategy, you do messaging. and it's, you know, And I think when you start Even Oracle, you know, since they started to to be 75% of AWS IAS revenues. You know, what do you think? it's, you know, I think it's growing well. Is that just because of the And be able to move you forward. I feel like, you know, I think when, you know, you could say, And we know Google is, you know, And I think that, you know, you know, AWS spins out, and I think to that respect forward to working with you Well, thank you for having me on. And we'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AmazonORGANIZATION

0.99+

Dave VellantePERSON

0.99+

Bob EvansPERSON

0.99+

MicrosoftORGANIZATION

0.99+

HPORGANIZATION

0.99+

AWSORGANIZATION

0.99+

RobPERSON

0.99+

GoogleORGANIZATION

0.99+

OracleORGANIZATION

0.99+

Rob StrechayPERSON

0.99+

New YorkLOCATION

0.99+

SeptemberDATE

0.99+

SeattleLOCATION

0.99+

JanuaryDATE

0.99+

Dev IttycheriaPERSON

0.99+

HPEORGANIZATION

0.99+

NetAppORGANIZATION

0.99+

AmsterdamLOCATION

0.99+

75%QUANTITY

0.99+

UKLOCATION

0.99+

AWSsORGANIZATION

0.99+

JuneDATE

0.99+

SnowplowORGANIZATION

0.99+

eightQUANTITY

0.99+

80%QUANTITY

0.99+

ScarpelliPERSON

0.99+

15%QUANTITY

0.99+

AustraliaLOCATION

0.99+

MongoORGANIZATION

0.99+

SlootmanPERSON

0.99+

two-yearQUANTITY

0.99+

AMDORGANIZATION

0.99+

EuropeLOCATION

0.99+

DatabricksORGANIZATION

0.99+

six factorsQUANTITY

0.99+

threeQUANTITY

0.99+

Merrill LynchORGANIZATION

0.99+

Last JuneDATE

0.99+

five countriesQUANTITY

0.99+

eight peopleQUANTITY

0.99+

U.S.LOCATION

0.99+

last weekDATE

0.99+

16 peopleQUANTITY

0.99+

Databricks'ORGANIZATION

0.99+

Jesse Cugliotta & Nicholas Taylor | The Future of Cloud & Data in Healthcare


 

(upbeat music) >> Welcome back to Supercloud 2. This is Dave Vellante. We're here exploring the intersection of data and analytics in the future of cloud and data. In this segment, we're going to look deeper into the life sciences business with Jesse Cugliotta, who leads the Healthcare and Life Sciences industry practice at Snowflake. And Nicholas Nick Taylor, who's the executive director of Informatics at Ionis Pharmaceuticals. Gentlemen, thanks for coming in theCUBE and participating in the program. Really appreciate it. >> Thank you for having us- >> Thanks for having me. >> You're very welcome, okay, we're go really try to look at data sharing as a use case and try to understand what's happening in the healthcare industry generally and specifically, how Nick thinks about sharing data in a governed fashion whether tapping the capabilities of multiple clouds is advantageous long term or presents more challenges than the effort is worth. And to start, Jesse, you lead this industry practice for Snowflake and it's a challenging and vibrant area. It's one that's hyper-focused on data privacy. So the first question is, you know there was a time when healthcare and other regulated industries wouldn't go near the cloud. What are you seeing today in the industry around cloud adoption and specifically multi-cloud adoption? >> Yeah, for years I've heard that healthcare and life sciences has been cloud diverse, but in spite of all of that if you look at a lot of aspects of this industry today, they've been running in the cloud for over 10 years now. Particularly when you look at CRM technologies or HR or HCM, even clinical technologies like EDC or ETMF. And it's interesting that you mentioned multi-cloud as well because this has always been an underlying reality especially within life sciences. This industry grows through acquisition where companies are looking to boost their future development pipeline either by buying up smaller biotechs, they may have like a late or a mid-stage promising candidate. And what typically happens is the larger pharma could then use their commercial muscle and their regulatory experience to move it to approvals and into the market. And I think the last few decades of cheap capital certainly accelerated that trend over the last couple of years. But this typically means that these new combined institutions may have technologies that are running on multiple clouds or multiple cloud strategies in various different regions to your point. And what we've often found is that they're not planning to standardize everything onto a single cloud provider. They're often looking for technologies that embrace this multi-cloud approach and work seamlessly across them. And I think this is a big reason why we, here at Snowflake, we've seen such strong momentum and growth across this industry because healthcare and life science has actually been one of our fastest growing sectors over the last couple of years. And a big part of that is in fact that we run on not only all three major cloud providers, but individual accounts within each and any one of them, they had the ability to communicate and interoperate with one another, like a globally interconnected database. >> Great, thank you for that setup. And so Nick, tell us more about your role and Ionis Pharma please. >> Sure. So I've been at Ionis for around five years now. You know, when when I joined it was, the IT department was pretty small. There wasn't a lot of warehousing, there wasn't a lot of kind of big data there. We saw an opportunity with Snowflake pretty early on as a provider that would be a lot of benefit for us, you know, 'cause we're small, wanted something that was fairly hands off. You know, I remember the days where you had to get a lot of DBAs in to fine tune your databases, make sure everything was running really, really well. The notion that there's, you know, no indexes to tune, right? There's very few knobs and dials, you can turn on Snowflake. That was appealing that, you know, it just kind of worked. So we found a use case to bring the platform in. We basically used it as a logging replacement as a Splunk kind of replacement with a platform called Elysium Analytics as a way to just get it in the door and give us the opportunity to solve a real world use case, but also to help us start to experiment using Snowflake as a platform. It took us a while to A, get the funding to bring it in, but B, build the momentum behind it. But, you know, as we experimented we added more data in there, we ran a few more experiments, we piloted in few more applications, we really saw the power of the platform and now, we are becoming a commercial organization. And with that comes a lot of major datasets. And so, you know, we really see Snowflake as being a very important part of our ecology going forward to help us build out our infrastructure. >> Okay, and you are running, your group runs on Azure, it's kind of mono cloud, single cloud, but others within Ionis are using other clouds, but you're not currently, you know, collaborating in terms of data sharing. And I wonder if you could talk about how your data needs have evolved over the past decade. I know you came from another highly regulated industry in financial services. So what's changed? You sort of touched on this before, you had these, you know, very specialized individuals who were, you know, DBAs, and, you know, could tune databases and the like, so that's evolved, but how has generally your needs evolved? Just kind of make an observation over the last, you know, five or seven years. What have you seen? >> Well, we, I wasn't in a group that did a lot of warehousing. It was more like online trade capture, but, you know, it was very much on-prem. You know, being in the cloud is very much a dirty word back then. I know that's changed since I've left. But in, you know, we had major, major teams of everyone who could do everything, right. As I mentioned in the pharma organization, there's a lot fewer of us. So the data needs there are very different, right? It's, we have a lot of SaaS applications. One of the difficulties with bringing a lot of SaaS applications on board is obviously data integration. So making sure the data is the same between them. But one of the big problems is joining the data across those SaaS applications. So one of the benefits, one of the things that we use Snowflake for is to basically take data out of these SaaS applications and load them into a warehouse so we can do those joins. So we use technologies like Boomi, we use technologies like Fivetran, like DBT to bring this data all into one place and start to kind of join that basically, allow us to do, run experiments, do analysis, basically take better, find better use for our data that was siloed in the past. You mentioned- >> Yeah. And just to add on to Nick's point there. >> Go ahead. >> That's actually something very common that we're seeing across the industry is because a lot of these SaaS applications that you mentioned, Nick, they're with from vendors that are trying to build their own ecosystem in walled garden. And by definition, many of them do not want to integrate with one another. So from a, you know, from a data platform vendor's perspective, we see this as a huge opportunity to help organizations like Ionis and others kind of deal with the challenges that Nick is speaking about because if the individual platform vendors are never going to make that part of their strategy, we see it as a great way to add additional value to these customers. >> Well, this data sharing thing is interesting. There's a lot of walled gardens out there. Oracle is a walled garden, AWS in many ways is a walled garden. You know, Microsoft has its walled garden. You could argue Snowflake is a walled garden. But the, what we're seeing and the whole reason behind the notion of super-cloud is we're creating an abstraction layer where you actually, in this case for this use case, can share data in a governed manner. Let's forget about the cross-cloud for a moment. I'll come back to that, but I wonder, Nick, if you could talk about how you are sharing data, again, Snowflake sort of, it's, I look at Snowflake like the app store, Apple, we're going to control everything, we're going to guarantee with data clean rooms and governance and the standards that we've created within that platform, we're going to make sure that it's safe for you to share data in this highly regulated industry. Are you doing that today? And take us through, you know, the considerations that you have in that regard. >> So it's kind of early days for us in Snowflake in general, but certainly in data sharing, we have a couple of examples. So data marketplace, you know, that's a great invention. It's, I've been a small IT shop again, right? The fact that we are able to just bring down terabyte size datasets straight into our Snowflake and run analytics directly on that is huge, right? The fact that we don't have to FTP these massive files around run jobs that may break, being able to just have that on tap is huge for us. We've recently been talking to one of our CRO feeds- CRO organizations about getting their data feeds in. Historically, this clinical trial data that comes in on an FTP file, we have to process it, take it through the platforms, put it into the warehouse. But one of the CROs that we talked to recently when we were reinvestigate in what data opportunities they have, they were a Snowflake customer and we are, I think, the first production customer they have, have taken that feed. So they're basically exposing their tables of data that historically came in these FTP files directly into our Snowflake instance now. We haven't taken advantage of that. It only actually flipped the switch about three or four weeks ago. But that's pretty big for us again, right? We don't have to worry about maintaining those jobs that take those files in. We don't have to worry about the jobs that take those and shove them on the warehouse. We now have a feed that's directly there that we can use a tool like DBT to push through directly into our model. And then the third avenue that's came up, actually fairly recently as well was genetics data. So genetics data that's highly, highly regulated. We had to be very careful with that. And we had a conversation with Snowflake about the data white rooms practice, and we see that as a pretty interesting opportunity. We are having one organization run genetic analysis being able to send us those genetic datasets, but then there's another organization that's actually has the in quotes "metadata" around that, so age, ethnicity, location, et cetera. And being able to join those two datasets through some kind of mechanism would be really beneficial to the organization. Being able to build a data white room so we can put that genetic data in a secure place, anonymize it, and then share the amalgamated data back out in a way that's able to be joined to the anonymized metadata, that could be pretty huge for us as well. >> Okay, so this is interesting. So you talk about FTP, which was the common way to share data. And so you basically, it's so, I got it now you take it and do whatever you want with it. Now we're talking, Jesse, about sharing the same copy of live data. How common is that use case in your industry? >> It's become very common over the last couple of years. And I think a big part of it is having the right technology to do it effectively. You know, as Nick mentioned, historically, this was done by people sending files around. And the challenge with that approach, of course, while there are multiple challenges, one, every time you send a file around your, by definition creating a copy of the data because you have to pull it out of your system of record, put it into a file, put it on some server where somebody else picks it up. And by definition at that point you've lost governance. So this creates challenges in general hesitation to doing so. It's not that it hasn't happened, but the other challenge with it is that the data's no longer real time. You know, you're working with a copy of data that was as fresh as at the time at that when that was actually extracted. And that creates limitations in terms of how effective this can be. What we're starting to see now with some of our customers is live sharing of information. And there's two aspects of that that are important. One is that you're not actually physically creating the copy and sending it to someone else, you're actually exposing it from where it exists and allowing another consumer to interact with it from their own account that could be in another region, some are running in another cloud. So this concept of super-cloud or cross-cloud could becoming realized here. But the other important aspect of it is that when that other- when that other entity is querying your data, they're seeing it in a real time state. And this is particularly important when you think about use cases like supply chain planning, where you're leveraging data across various different enterprises. If I'm a manufacturer or if I'm a contract manufacturer and I can see the actual inventory positions of my clients, of my distributors, of the levels of consumption at the pharmacy or the hospital that gives me a lot of indication as to how my demand profile is changing over time versus working with a static picture that may have been from three weeks ago. And this has become incredibly important as supply chains are becoming more constrained and the ability to plan accurately has never been more important. >> Yeah. So the race is on to solve these problems. So it start, we started with, hey, okay, cloud, Dave, we're going to simplify database, we're going to put it in the cloud, give virtually infinite resources, separate compute from storage. Okay, check, we got that. Now we've moved into sort of data clean rooms and governance and you've got an ecosystem that's forming around this to make it safer to share data. And then, you know, nirvana, at least near term nirvana is we're going to build data applications and we're going to be able to share live data and then you start to get into monetization. Do you see, Nick, in the near future where I know you've got relationships with, for instance, big pharma like AstraZeneca, do you see a situation where you start sharing data with them? Is that in the near term? Is that more long term? What are the considerations in that regard? >> I mean, it's something we've been thinking about. We haven't actually addressed that yet. Yeah, I could see situations where, you know, some of these big relationships where we do need to share a lot of data, it would be very nice to be able to just flick a switch and share our data assets across to those organizations. But, you know, that's a ways off for us now. We're mainly looking at bringing data in at the moment. >> One of the things that we've seen in financial services in particular, and Jesse, I'd love to get your thoughts on this, is companies like Goldman or Capital One or Nasdaq taking their stack, their software, their tooling actually putting it on the cloud and facing it to their customers and selling that as a new monetization vector as part of their digital or business transformation. Are you seeing that Jesse at all in healthcare or is it happening today or do you see a day when that happens or is healthier or just too scary to do that? >> No, we're seeing the early stages of this as well. And I think it's for some of the reasons we talked about earlier. You know, it's a much more secure way to work with a colleague if you don't have to copy your data and potentially expose it. And some of the reasons that people have historically copied that data is that they needed to leverage some sort of algorithm or application that a third party was providing. So maybe someone was predicting the ideal location and run a clinical trial for this particular rare disease category where there are only so many patients around the world that may actually be candidates for this disease. So you have to pick the ideal location. Well, sending the dataset to do so, you know, would involve a fairly complicated process similar to what Nick was mentioning earlier. If the company who was providing the logic or the algorithm to determine that location could bring that algorithm to you and you run it against your own data, that's a much more ideal and a much safer and more secure way for this industry to actually start to work with some of these partners and vendors. And that's one of the things that we're looking to enable going into this year is that, you know, the whole concept should be bring the logic to your data versus your data to the logic and the underlying sharing mechanisms that we've spoken about are actually what are powering that today. >> And so thank you for that, Jesse. >> Yes, Dave. >> And so Nick- Go ahead please. >> Yeah, if I could add, yeah, if I could add to that, that's something certainly we've been thinking about. In fact, we'd started talking to Snowflake about that a couple of years ago. We saw the power there again of the platform to be able to say, well, could we, we were thinking in more of a data share, but could we share our data out to say an AI/ML vendor, have them do the analytics and then share the data, the results back to us. Now, you know, there's more powerful mechanisms to do that within the Snowflake ecosystem now, but you know, we probably wouldn't need to have onsite AI/ML people, right? Some of that stuff's very sophisticated, expensive resources, hard to find, you know, it's much better for us to find a company that would be able to build those analytics, maintain those analytics for us. And you know, we saw an opportunity to do that a couple years ago and we're kind of excited about the opportunity there that we can just basically do it with a no op, right? We share the data route, we have the analytics done, we get the result back and it's just fairly seamless. >> I mean, I could have a whole another Cube session on this, guys, but I mean, I just did a a session with Andy Thurai, a Constellation research about how difficult it's been for organization to get ROI because they don't have the expertise in house so they want to either outsource it or rely on vendor R&D companies to inject that AI and machine intelligence directly into applications. My follow-up question to you Nick is, when you think about, 'cause Jesse was talking about, you know, let the data basically stay where it is and you know bring the compute to that data. If that data lives on different clouds, and maybe it's not your group, but maybe it's other parts of Ionis or maybe it's your partners like AstraZeneca, or you know, the AI/ML partners and they're potentially on other clouds or that data is on other clouds. Do you see that, again, coming back to super-cloud, do you see it as an advantage to be able to have a consistent experience across those clouds? Or is that just kind of get in the way and make things more complex? What's your take on that, Nick? >> Well, from the vendors, so from the client side, it's kind of seamless with Snowflake for us. So we know for a fact that one of the datasets we have at the moment, Compile, which is a, the large multi terabyte dataset I was talking about. They're on AWS on the East Coast and we are on Azure on the West Coast. And they had to do a few tweaks in the background to make sure the data was pushed over from, but from my point of view, the data just exists, right? So for me, I think it's hugely beneficial that Snowflake supports this kind of infrastructure, right? We don't have to jump through hoops to like, okay, well, we'll download it here and then re-upload it here. They already have the mechanism in the background to do these multi-cloud shares. So it's not important for us internally at the moment. I could see potentially at some point where we start linking across different groups in the organization that do have maybe Amazon or Google Cloud, but certainly within our providers. We know for a fact that they're on different services at the moment and it just works. >> Yeah, and we learned from Benoit Dageville, who came into the studio on August 9th with first Supercloud in 2022 that Snowflake uses a single global instance across regions and across clouds, yeah, whether or not you can query across you know, big regions, it just depends, right? It depends on latency. You might have to make a copy or maybe do some tweaks in the background. But guys, we got to jump, I really appreciate your time. Really thoughtful discussion on the future of data and cloud, specifically within healthcare and pharma. Thank you for your time. >> Thanks- >> Thanks for having us. >> All right, this is Dave Vellante for theCUBE team and my co-host, John Furrier. Keep it right there for more action at Supercloud 2. (upbeat music)

Published Date : Jan 3 2023

SUMMARY :

and analytics in the So the first question is, you know And it's interesting that you Great, thank you for that setup. get the funding to bring it in, over the last, you know, So one of the benefits, one of the things And just to add on to Nick's point there. that you mentioned, Nick, and the standards that we've So data marketplace, you know, And so you basically, it's so, And the challenge with Is that in the near term? bringing data in at the moment. One of the things that we've seen that algorithm to you and you And so Nick- the results back to us. Or is that just kind of get in the way in the background to do on the future of data and cloud, All right, this is Dave Vellante

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jesse CugliottaPERSON

0.99+

Dave VellantePERSON

0.99+

GoldmanORGANIZATION

0.99+

AstraZenecaORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

John FurrierPERSON

0.99+

Capital OneORGANIZATION

0.99+

JessePERSON

0.99+

Andy ThuraiPERSON

0.99+

AWSORGANIZATION

0.99+

August 9thDATE

0.99+

NickPERSON

0.99+

NasdaqORGANIZATION

0.99+

Nicholas Nick TaylorPERSON

0.99+

fiveQUANTITY

0.99+

AmazonORGANIZATION

0.99+

IonisORGANIZATION

0.99+

DavePERSON

0.99+

Ionis PharmaORGANIZATION

0.99+

Nicholas TaylorPERSON

0.99+

Ionis PharmaceuticalsORGANIZATION

0.99+

SnowflakeORGANIZATION

0.99+

first questionQUANTITY

0.99+

Benoit DagevillePERSON

0.99+

AppleORGANIZATION

0.99+

seven yearsQUANTITY

0.99+

OracleORGANIZATION

0.99+

2022DATE

0.99+

todayDATE

0.99+

over 10 yearsQUANTITY

0.98+

SnowflakeTITLE

0.98+

oneQUANTITY

0.98+

OneQUANTITY

0.98+

two aspectsQUANTITY

0.98+

firstQUANTITY

0.98+

this yearDATE

0.97+

eachQUANTITY

0.97+

two datasetsQUANTITY

0.97+

West CoastLOCATION

0.97+

four weeks agoDATE

0.97+

around five yearsQUANTITY

0.97+

threeQUANTITY

0.95+

first productionQUANTITY

0.95+

East CoastLOCATION

0.95+

third avenueQUANTITY

0.95+

one organizationQUANTITY

0.94+

theCUBEORGANIZATION

0.94+

couple years agoDATE

0.93+

single cloudQUANTITY

0.92+

single cloud providerQUANTITY

0.92+

hree weeks agoDATE

0.91+

one placeQUANTITY

0.88+

AzureTITLE

0.86+

last couple of yearsDATE

0.85+

Domenic Ravita, SingleStore | AWS re:Invent 2022


 

>>Hey guys and girls, welcome back to The Cube's Live coverage of AWS Reinvent 22 from Sin City. We've been here, this is our third day of coverage. We started Monday night first. Full day of the show was yesterday. Big news yesterday. Big news. Today we're hearing north of 50,000 people, and I'm hearing hundreds of thousands online. We've been having great conversations with AWS folks in the ecosystem, AWS customers, partners, ISVs, you name it. We're pleased to welcome back one of our alumni to the program, talking about partner ecosystem. Dominic Rav Vida joins us, the VP of Developer relations at single store. It's so great to have you on the program. Dominic. Thanks for coming. >>Thanks. Great. Great to see you >>Again. Great to see you too. We go way back. >>We do, yeah. >>So let's talk about reinvent 22. This is the 11th reinvent. Yeah. What are some of the things that you've heard this week that are exciting that are newsworthy from single stores perspective? >>I think in particular what we heard AWS announce on the zero ETL between Aurora and Redshift, I think it's, it's significant in that AWS has provided lots of services for building blocks for applications for a long time. And that's a great amount of flexibility for developers. But there are cases where, you know, it's a common thing to need to move data from transactional systems to analytics systems and making that easy with zero etl, I think it's a significant thing and in general we see in the market and especially in the data management market in the cloud, a unification of different types of workloads. So I think that's a step in the right direction for aws. And I think for the market as a whole, why it's significant for single store is, that's our specialty in particular, is to unify transactions and analytics for realtime applications and analytics. When you've got customer facing analytic applications and you need low latency data from realtime streaming data sources and you've gotta crunch and compute that. Those are diverse types of workloads over document transactional workloads as well as, you know, analytical workloads of various shapes and the data types could be diverse from geospatial time series. And then you've gotta serve that because we're all living in this digital service first world and you need that relevant, consistent, fresh data. And so that unification is what we think is like the big thing in data right >>Now. So validation for single store, >>It does feel like that. I mean, I'd say in the recent like six months, you've seen announcements from Google with Alloy db basically adding the complement to their workload types. You see it with Snowflake adding the complement to their traditional analytical workload site. You see it with Mongo and others. And yeah, we do feel it was validation cuz at single store we completed the functionality for what we call universal storage, which is, is the industry's first third type of storage after row store and column store, single store dbs, universal storage, unifies those. So on a single copy of data you can form these diverse workloads. And that was completed three years ago. So we sort of see like, you know, we're onto something >>Here. Welcome to the game guys. >>That's right. >>What's the value in that universal storage for customers, whether it's a healthcare organization, a financial institution, what's the value in it in those business outcomes that you guys are really helping to fuel? >>I think in short, if there were like a, a bumper sticker for that message, it's like, are you ready for the next interaction? The next interaction with your customer, the next interaction with your supply chain partner, the next interaction with your internal stakeholders, your operational managers being ready for that interaction means you've gotta have the historical data at the ready, accessible, efficiently accessible, and and, and queryable along with the most recent fresh data. And that's the context that's expected and be able to serve that instantaneously. So being ready for that next interaction is what single store helps companies do. >>Talk about single store helping customers. You know, every company these days has to be a data company. I always think, whether it's my grocery store that has all my information and helps keep me fed or a gas station or a car dealer or my bank. And we've also here, one of the things that John Furrier got to do, and he does this every year before aws, he gets to sit down with the CEO and gets really kind of a preview of what's gonna happen at at the show, right? And Adams Lisky said to him some interesting very poignant things. One is that that data, we talk about data democratization, but he says the role of the data analyst is gonna go away. Or that maybe that term in, in that every person within an organization, whether you're marketing, sales, ops, finance, is going to be analyzing data for their jobs to become data driven. Right? How does single store help customers really become data companies, especially powering data intensive apps like I know you do. >>Yeah, that's, there's a lot of talk about that and, and I think there's a lot of work that's been done with companies to make that easier to analyze data in all these different job functions. While we do that, it's not really our starting point because, and our starting point is like operationalizing that analytics as part of the business. So you can think of it in terms of database terms. Like is it batch analysis? Batch analytics after the fact, what happened last week? What happened last month? That's a lot of what those data teams are doing and those analysts are doing. What single store focuses more is in putting those insights into action for the business operations, which typically is more on the application side, it's the API side, you might call it a data product. If you're monetizing your data and you're transacting with that providing as an api, or you're delivering it as software as a service, and you're providing an end-to-end function for, you know, our marketing marketer, then we help power those kinds of real time data applications that have the interactivity and have that customer touchpoint or that partner touchpoint. So you can say we sort of, we put the data in action in that way. >>And that's the most, one of the most important things is putting data in action. If it's, it can be gold, it can be whatever you wanna call it, but if you can't actually put it into action, act on insights in real time, right? The value goes way down or there's liability, >>Right? And I think you have to do that with privacy in mind as well, right? And so you have to take control of that data and use it for your business strategy And the way that you can do that, there's technology like single store makes that possible in ways that weren't possible before. And I'll give you an example. So we have a, a customer named Fathom Analytics. They provide web analytics for marketers, right? So if you're in marketing, you understand this use case. Any demand gen marketer knows that they want to see what the traffic that hits their site is. What are the page views, what are the click streams, what are the sequences? Have these visitors to my website hit certain goals? So the big name in that for years of course has been Google Analytics and that's a free service. And you interact with that and you can see how your website's performing. >>So what Fathom does is a privacy first alternative to Google Analytics. And when you think about, well, how is that possible that they, and as a paid service, it's as software, as a service, how, first of all, how can you keep up with that real time deluge of clickstream data at the rate that Google Analytics can do it? That's the technical problem. But also at the data layer, how could you keep up with Google has, you know, in terms of databases And Fathom's answer to that is to use single store. Their, their prior architecture had four different types of database technologies under the hood. They were using Redis to have fast read time cash. They were using MySEQ database as the application database they were using. They were looking at last search to do full tech search. And they were using DynamoDB as part of a another kind of fast look up fast cash. They replaced all four of those with single store. And, and again, what they're doing is like sort of battling defacto giant in Google Analytics and having a great success at doing that for posting tens of thousands of websites. Some big names that you've heard of as well. >>I can imagine that's a big reduction from four to one, four x reduction in databases. The complexities that go away, the simplification that happens, I can imagine is quite huge for them. >>And we've done a study, an independent study with Giga Home Research. We published this back in June looking at total cost of ownership with benchmarks and the relevant benchmarks for transactions and analytics and databases are tpcc for transactions, TPC H for analytics, TPC DS for analytics. And we did a TCO study using those benchmark datas on a combination of transactional and analytical databases together and saw some pretty big improvements. 60% improvement over Myse Snowflake, for >>Instance. Awesome. Big business outcomes. We only have a few seconds left, so you've already given me a bumper sticker. Yeah. And I know I live in Silicon Valley, I've seen those billboards. I know single store has done some cheeky billboard marketing campaigns. But if you had a new billboard to create from your perspective about single store, what does it say? >>I, I think it's that, are you, are you ready for the next interaction? Because business is won and lost in every moment, in every location, in every digital moment passing by. And if you're not ready to, to interact and transact rather your systems on your behalf, then you're behind the curve. It's easy to be displaced people swipe left and pick your competitor. So I think that's the next bumper sticker. I may, I would say our, my favorite billboard so far of what we've run is cover your SaaS, which is what is how, what is the data layer to, to manage the next level of SaaS applications, the next generation. And we think single store is a big part >>Of that. Cover your SaaS. Love it. Dominic, thank you so much for joining me, giving us an update on single store from your perspective, what's going on there, kind of really where you are in the market. We appreciate that. We'll have to >>Have you back. Thank you. Glad to >>Be here. All right. For Dominic rta, I'm Lisa Martin. You're watching The Cube, the leader in live, emerging and enterprise tech coverage.

Published Date : Dec 1 2022

SUMMARY :

It's so great to have you on the program. Great to see you Great to see you too. What are some of the things that you've heard this week that are exciting that are newsworthy from And so that unification is what we think is like the So on a single copy of data you can form these diverse And that's the context that's expected and be able to serve that instantaneously. one of the things that John Furrier got to do, and he does this every year before aws, he gets to sit down with the CEO So you can think of it in terms of database terms. And that's the most, one of the most important things is putting data in action. And I think you have to do that with privacy in mind as well, right? But also at the data layer, how could you keep up with Google has, you know, The complexities that go away, the simplification that happens, I can imagine is quite huge for them. And we've done a study, an independent study with Giga Home Research. But if you had a new billboard to create from your perspective And if you're not ready to, to interact and transact rather your systems on Dominic, thank you so much for joining me, giving us an update on single store from your Have you back. the leader in live, emerging and enterprise tech coverage.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DominicPERSON

0.99+

Lisa MartinPERSON

0.99+

AWSORGANIZATION

0.99+

60%QUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

Sin CityLOCATION

0.99+

JuneDATE

0.99+

TodayDATE

0.99+

last weekDATE

0.99+

yesterdayDATE

0.99+

Monday nightDATE

0.99+

last monthDATE

0.99+

Giga Home ResearchORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

John FurrierPERSON

0.99+

oneQUANTITY

0.99+

third dayQUANTITY

0.99+

single storeQUANTITY

0.99+

Domenic RavitaPERSON

0.99+

Dominic Rav VidaPERSON

0.99+

Google AnalyticsTITLE

0.99+

Adams LiskyPERSON

0.99+

FathomORGANIZATION

0.98+

three years agoDATE

0.98+

singleQUANTITY

0.98+

Fathom AnalyticsORGANIZATION

0.98+

single copyQUANTITY

0.98+

hundreds of thousandsQUANTITY

0.98+

MongoORGANIZATION

0.97+

OneQUANTITY

0.97+

SingleStoreORGANIZATION

0.97+

this weekDATE

0.97+

The CubeTITLE

0.96+

first third typeQUANTITY

0.94+

fourQUANTITY

0.94+

tens of thousands of websitesQUANTITY

0.92+

MySEQTITLE

0.91+

six monthsQUANTITY

0.91+

RedisTITLE

0.9+

DynamoDBTITLE

0.9+

Dominic rtaPERSON

0.89+

Reinvent 22TITLE

0.89+

north of 50,000 peopleQUANTITY

0.86+

first alternativeQUANTITY

0.81+

firstQUANTITY

0.81+

single storesQUANTITY

0.78+

first worldQUANTITY

0.73+

AlloyTITLE

0.69+

few secondsQUANTITY

0.67+

RedshiftORGANIZATION

0.67+

SnowflakeORGANIZATION

0.67+

awsORGANIZATION

0.67+

TCOORGANIZATION

0.66+

InventEVENT

0.66+

reinvent 22EVENT

0.65+

SnowflakeTITLE

0.65+

2022DATE

0.64+

AuroraORGANIZATION

0.56+

11th reinventEVENT

0.52+

FathomPERSON

0.51+

MyseORGANIZATION

0.49+

CubeORGANIZATION

0.46+

zeroQUANTITY

0.34+

Jason Beyer & Josh Von Schaumburg | AWS Executive Summit 2022


 

(bright upbeat music) >> Well, hi everybody, John Wallace here and welcome to theCUBE, the leader in high-tech coverage. Glad to have you aboard here as we continue our coverage here at re:Invent 2022. We're out at The Venetian in Las Vegas. A lot of energy down on that exhibit floor, I promise you. We're a little bit away from the maddening crowd, but we're here with the Executive Summit sponsored by Accenture. I've got two guests I want to introduce you to. Jason Beyer who is the vice president of Data and Analytics at Bridgestone Americas. Jason, good to see you, sir. >> Hello, John. >> And Josh von Schaumburg, who is the managing director and North America lead for AWS Security at Accenture. Josh, good to see you. >> Thanks for having us. >> Yeah, first off, just quick take on the show. I know you've only been here about a day or so, but just your thoughts about what you're seeing on the floor in terms of energy, enthusiasm and, I think, turnout, right? I'm really impressed by it. We've got a lot of people down there. >> Yeah, I've been certainly impressed, John, with the turnout. But just as you say, the energy of the crowd, the excitement for the new things coming, it seems like it's a really pivotal moment for many organizations, including my own, and really excited to see what's coming over the next couple days. >> Let's jump into Bridgestone then. I kind of kidded you before we started the interview saying, all right, tires and golf balls, that's what I relate to, but you have a full array of consumer products and solution you're offering and your responsibility is managing the data and the analytics and making sure those business lines are as efficient as possible. >> Absolutely, John. So in my role, I have the privilege of being in an enterprise position. So I get to see the vast array of Bridgestone, which it is a large, highly vertically integrated company all the way from raw material sourcing of natural rubber to retail services in the automotive industry. We're at scale across those areas. The exciting thing about the company right now is we're going through this business transformation of becoming, you know, building on that heritage and that great legacy of having high quality high performance, highly focused on safety products to becoming a product and solutions company, and particular a sustainable solutions company. So what that means is we're bringing not only those great products to market, tires, golf balls, hoses, all kinds of rubber, air springs products to market, but thinking about how do we service those after they're in the market, how do we bring solutions to help fleets, vehicle owners, vehicle operators operate those in a sustainable way, in a cost effective way? So those solutions, of course, bring all new sets of data and analytics that come with it, and technology and moving to the cloud to be cloud native. So this new phase for the organization that we refer to as Bridgestone 3.0, and that business strategy is driving our cloud strategy, our technology strategy, and our data strategy and AWS and Accenture are important partners in that. >> Yeah, so we hear a lot about that these days about this transformation, this journey that people are on now. And Josh, when Bridgestone or other clients come to you and they talk about their migrations and what's their footprint going to look like and how do they get there, in the case of Bridgestone when they came to you and said, "All right, this is where we want to go with this. We're going to embark on a significant upgrade of our systems here," how do you lead 'em? How do you get 'em there? >> Yeah, I think there are a couple key cloud transformation value drivers that we've emphasized and that I've seen at Bridgestone in my time there. I mean, number one, just the rapid increase in the pace of innovation that we've seen over the last couple years. And a lot of that is also led by the scalability of all of the cloud native AWS services that we're leveraging, and in particular with the CDP platform. It really started off as a single-use case and really a single-tenant data lake. And then through the strategic vision of Jason and the leadership team, we've been able to expand that to 10 plus tenants and use cases. And a big reason behind that is the scalability of all these AWS services, right? So as we add more and more tenants, all the infrastructure just scales without any manual provisioning any tuning that we need to do. And that allows us to go really from idea, to POC, to production in really a matter of months when traditionally it might take years. >> So- >> If I can build upon that. >> Please do, yeah. >> The CDP, or central data platform, is part of a broader reference architecture that reflects that business strategy. So we looked at it and said, we could have taken a couple of different approaches to recognize the business challenges we're facing. We needed to modernize our core, our ERP, our manufacturing solutions move to smart factory and green factories, our PLM solutions. But at the same time, we're moving quickly. We have a startup mindset in our mobility solutions businesses where we're going to market on our customer and commerce solutions, and we needed to move at a different pace. And so to decouple those, we, in partnership with Accenture and AWS, built out a reference architecture that has a decoupling layer that's built around a data fabric, a data connected layer, integrated data services as well. A key part of that architecture is our central data platform built on AWS. This is a comprehensive data lake architecture using all the modern techniques within AWS to bring data together, to coalesce data, as well as recognize the multiple different modes of consumption, whether that's classic reporting, business intelligence, analytics, machine learning data science, as well as API consumption. And so we're building that out. A year ago it was a concept on a PowerPoint and just show and kind of reflect the innovation and speed. As Josh mentioned, we're up to 10 tenants, we're growing exponentially. There's high demand from the organization to leverage data at scale because of the business transformation that I mentioned and that modernization of the core ecosystem. >> That's crazy fast, right? And all of a sudden, whoa! >> Faster than I expected. >> Almost snap overnight. And you raise an interesting point too. I think when you talk about how there was a segment of your business that you wanted to get in the startup mode, whereas I don't think Bridgestone, I don't think about startup, right? I think in a much more, I wouldn't say traditional, but you've got big systems, right? And so how did you kind of inject your teams with that kind of mindset, right? That, hey, you're going to have to hit the pedal here, right? And I want you to experiment. I want you to innovate. And that might be a little bit against the grain from what they were used to. >> So just over two years ago, we built and started the organization that I have the privilege of leading, our data and analytics organization. And it's a COE. It's a center of expertise in the organization. We partner with specialized teams in product development, marketing, other places to enable data and analytics everywhere. We wanted to be pervasive, it's a team sport. But we really embraced at that moment what we refer to as a dual speed mindset. Speed one, we've got to move at the speed of the business. And that's variable. Based on the different business units and lines of lines of business and functional areas, the core modernization efforts, those are multi-year transformation programs that have multiple phases to them, and we're embedded there building the fundamentals of data governance and data management and reporting operational things. But at the same time, we needed to recognize that speed of those startup businesses where we're taking solutions and service offerings to market, doing quick minimum viable product, put it in a market, try it, learn from it adapt. Sometimes shut it down and take those learnings into the next area as well as joint ventures. We've been much more aggressive in terms of the partnerships in the marketplace, the joint ventures, the minority investments, et cetera, really to give us that edge in how we corner the market on the fleet and mobility solutions of the future. So having that dual speed approach of operating at the speed of the business, we also needed to balance that with speed two, which is building those long term capabilities and fundamentals. And that's where we've been building out those practical examples of having data governance and data management across these areas, building robust governance of how we're thinking about data science and the evolution of data science and that maturity towards machine learning. And so having that dual speed approach, it's a difficult balancing act, but it's served us well, really partnering with our key business stakeholders of where we can engage, what services they need, and where do we need to make smart choices between those two different speeds. >> Yeah, you just hit on something I want to ask Josh about, about how you said sometimes you have to shut things down, right? It's one thing to embark on I guess a new opportunity or explore, right? New avenues. And then to tell your client, "Well, might be some bumps along the way." >> Yeah. >> A lot of times people in Jason's position don't want to hear that. (laughs) It's like, I don't want to hear about bumps. >> Yeah. >> We want this to be, again, working with clients in that respect and understanding that there's going to be a learning curve and that some things might not function the way you want them to, we might have to take a right instead of a left. >> Yeah, and I think the value of AWS is you really can fail fast and try to innovate and try different use cases out. You don't have any enormous upfront capital expenditure to start building all these servers in your data center for all of your use cases. You can spin something up easily based in idea and then fail fast and move on to the next idea. And I also wanted to emphasize I think how critical top-down executive buy-in is for any cloud transformation. And you could hear it, the excitement in Jason's voice. And anytime we've seen a failed cloud transformation, the common theme is typically lack of executive buy-in and leadership and vision. And I think from day one, Bridgestone has had that buy-in from Jason throughout the whole executive team, and I think that's really evident in the success of the CDP platform. >> Absolutely. >> And what's been your experience in that regard then? Because I think that's a great point Josh raised that you might be really excited in your position, but you've got to convince the C-suite. >> Yeah. >> And there are a lot of variables there that have to be considered, that are kind of out of your sandbox, right? So for somebody else to make decisions based on a holistic approach, right? >> I could tell you, John, talking with with peers of mine, I recognize that I've probably had a little bit of privilege in that regard because the leadership at Bridgestone has recognized to move to this product and solutions organization and have sustainable solutions for the future we needed to move to the cloud. We needed to shift that technology forward. We needed to have a more data-driven approach to things. And so the selling of that was not a huge uphill a battle to be honest. It was almost more of a pull from the top, from our global group CEO, from our CEOs in our different regions, including in Bridgestone Americas. They've been pushing that forward, they've been driving it. And as Josh mentioned, that's been a really huge key to our success, is that executive alignment to move at this new pace, at this new frame of innovation, because that's what the market is demanding in the changing landscape of mobility and the movement of vehicles and things on the road. >> So how do you two work together going forward, Ben? Because you're in a great position now. You've had this tremendous acceleration in the past year, right? Talking about this tenfold increase and what the platform's enabled you to do, but as you know, you can't stand still. Right? (laughs) >> Yeah. There's so much excitement, so many use cases in the backlog now, and it's really been a snowball effect. I think one of the use cases I'm most excited about is starting to apply ML, you know, machine learning to the data sets. And I think there's an amazing IoT predictive maintenance use case there for all of the the censored data collected across all of the tires that are sold. There's an immense amount of data and ultimately we can use that data to predict failures and make our roads safer and help save lives >> Right. >> It's hard to not take a long time to explain all the things because there is a lot ahead of us. The demand curve for capabilities and the enabling things that AWS is going to support is just tremendous. As Josh mentioned, the, the AI ML use cases ahead of us, incredibly exciting. The way we're building and co-innovating things around how we make data more accessible in our data marketplace and more advanced data governance and data quality techniques. The use of, you know, creating data hubs and moving our API landscape into this environment as well is going to be incredibly empowering in terms of accessibility of data across our enterprise globally, as well as both for our internal stakeholders and our external stakeholders. So, I'll stop there because there's a lot of things in there. >> We could be here a long time. >> Yes, we could. >> But it is an exciting time and I appreciate that you're both sharing your perspectives on this because you've got a winning formula going and look forward to what's happening. And we'll see you next year right back here on the Executive Summit. >> Absolutely. >> To measure the success in 2023. How about that? >> Sounds good, thank you, Jim. >> Is that a deal? >> Awesome. >> Sounds good. >> Excellent, good deal. You've been watching AWS here at Coverage of Reinvent '22. We are the Executive Summit sponsored by Accenture and you are watching theCUBE, the leader in high tech coverage. (gentle music)

Published Date : Nov 29 2022

SUMMARY :

A lot of energy down on that Josh, good to see you. quick take on the show. and really excited to see I kind of kidded you before the cloud to be cloud native. in the case of Bridgestone And a lot of that is also because of the business in the startup mode, and mobility solutions of the future. And then to tell your client, to hear about bumps. and that some things might not function of the CDP platform. that you might be really and the movement of vehicles and what the platform's enabled you to do, for all of the the censored data and the enabling things and look forward to what's happening. To measure the success and you are watching theCUBE,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
AWSORGANIZATION

0.99+

JoshPERSON

0.99+

JohnPERSON

0.99+

Jason BeyerPERSON

0.99+

JasonPERSON

0.99+

Josh von SchaumburgPERSON

0.99+

JimPERSON

0.99+

John WallacePERSON

0.99+

BridgestoneORGANIZATION

0.99+

AccentureORGANIZATION

0.99+

2023DATE

0.99+

BenPERSON

0.99+

10 plus tenantsQUANTITY

0.99+

two guestsQUANTITY

0.99+

Josh Von SchaumburgPERSON

0.99+

PowerPointTITLE

0.99+

twoQUANTITY

0.99+

next yearDATE

0.99+

Las VegasLOCATION

0.99+

bothQUANTITY

0.99+

AWS SecurityORGANIZATION

0.98+

A year agoDATE

0.98+

singleQUANTITY

0.97+

oneQUANTITY

0.97+

one thingQUANTITY

0.96+

up to 10 tenantsQUANTITY

0.94+

reEVENT

0.86+

theCUBEORGANIZATION

0.86+

past yearDATE

0.85+

firstQUANTITY

0.85+

AWS Executive Summit 2022EVENT

0.85+

two differentQUANTITY

0.84+

over two years agoDATE

0.83+

BridgestoneLOCATION

0.82+

about a dayQUANTITY

0.81+

single-use caseQUANTITY

0.8+

Bridgestone AmericasLOCATION

0.79+

North AmericaLOCATION

0.77+

Executive SummitEVENT

0.75+

last couple yearsDATE

0.75+

Invent 2022EVENT

0.73+

Bridgestone AmericasORGANIZATION

0.71+

coupleQUANTITY

0.71+

tenfoldQUANTITY

0.68+

Bridgestone 3.0ORGANIZATION

0.67+

DataORGANIZATION

0.67+

AnalyticsORGANIZATION

0.54+

SummitEVENT

0.53+

casesQUANTITY

0.51+

Coverage of Reinvent '22EVENT

0.49+

couple daysDATE

0.48+

VenetianLOCATION

0.34+

Oracle Announces MySQL HeatWave on AWS


 

>>Oracle continues to enhance my sequel Heatwave at a very rapid pace. The company is now in its fourth major release since the original announcement in December 2020. 1 of the main criticisms of my sequel, Heatwave, is that it only runs on O. C I. Oracle Cloud Infrastructure and as a lock in to Oracle's Cloud. Oracle recently announced that heat wave is now going to be available in AWS Cloud and it announced its intent to bring my sequel Heatwave to Azure. So my secret heatwave on AWS is a significant TAM expansion move for Oracle because of the momentum AWS Cloud continues to show. And evidently the Heatwave Engineering team has taken the development effort from O. C I. And is bringing that to A W S with a number of enhancements that we're gonna dig into today is senior vice president. My sequel Heatwave at Oracle is back with me on a cube conversation to discuss the latest heatwave news, and we're eager to hear any benchmarks relative to a W S or any others. Nippon has been leading the Heatwave engineering team for over 10 years and there's over 100 and 85 patents and database technology. Welcome back to the show and good to see you. >>Thank you. Very happy to be back. >>Now for those who might not have kept up with the news, uh, to kick things off, give us an overview of my sequel, Heatwave and its evolution. So far, >>so my sequel, Heat Wave, is a fully managed my secret database service offering from Oracle. Traditionally, my secret has been designed and optimised for transaction processing. So customers of my sequel then they had to run analytics or when they had to run machine learning, they would extract the data out of my sequel into some other database for doing. Unlike processing or machine learning processing my sequel, Heat provides all these capabilities built in to a single database service, which is my sequel. He'd fake So customers of my sequel don't need to move the data out with the same database. They can run transaction processing and predicts mixed workloads, machine learning, all with a very, very good performance in very good price performance. Furthermore, one of the design points of heat wave is is a scale out architecture, so the system continues to scale and performed very well, even when customers have very large late assignments. >>So we've seen some interesting moves by Oracle lately. The collaboration with Azure we've we've covered that pretty extensively. What was the impetus here for bringing my sequel Heatwave onto the AWS cloud? What were the drivers that you considered? >>So one of the observations is that a very large percentage of users of my sequel Heatwave, our AWS users who are migrating of Aurora or so already we see that a good percentage of my secret history of customers are migrating from GWS. However, there are some AWS customers who are still not able to migrate the O. C. I to my secret heat wave. And the reason is because of, um, exorbitant cost, which was charges. So in order to migrate the workload from AWS to go see, I digress. Charges are very high fees which becomes prohibitive for the customer or the second example we have seen is that the latency of practising a database which is outside of AWS is very high. So there's a class of customers who would like to get the benefits of my secret heatwave but were unable to do so and with this support of my secret trip inside of AWS, these customers can now get all the grease of the benefits of my secret he trip without having to pay the high fees or without having to suffer with the poorly agency, which is because of the ws architecture. >>Okay, so you're basically meeting the customer's where they are. So was this a straightforward lifted shift from from Oracle Cloud Infrastructure to AWS? >>No, it is not because one of the design girls we have with my sequel, Heatwave is that we want to provide our customers with the best price performance regardless of the cloud. So when we decided to offer my sequel, he headed west. Um, we have optimised my sequel Heatwave on it as well. So one of the things to point out is that this is a service with the data plane control plane and the console are natively running on AWS. And the benefits of doing so is that now we can optimise my sequel Heatwave for the E. W s architecture. In addition to that, we have also announced a bunch of new capabilities as a part of the service which will also be available to the my secret history of customers and our CI, But we just announced them and we're offering them as a part of my secret history of offering on AWS. >>So I just want to make sure I understand that it's not like you just wrapped your stack in a container and stuck it into a W s to be hosted. You're saying you're actually taking advantage of the capabilities of the AWS cloud natively? And I think you've made some other enhancements as well that you're alluding to. Can you maybe, uh, elucidate on those? Sure. >>So for status, um, we have taken the mind sequel Heatwave code and we have optimised for the It was infrastructure with its computer network. And as a result, customers get very good performance and price performance. Uh, with my secret he trade in AWS. That's one performance. Second thing is, we have designed new interactive counsel for the service, which means that customers can now provision there instances with the council. But in addition, they can also manage their schemas. They can. Then court is directly from the council. Autopilot is integrated. The council we have introduced performance monitoring, so a lot of capabilities which we have introduced as a part of the new counsel. The third thing is that we have added a bunch of new security features, uh, expose some of the security features which were part of the My Secret Enterprise edition as a part of the service, which gives customers now a choice of using these features to build more secure applications. And finally, we have extended my secret autopilot for a number of old gpus cases. In the past, my secret autopilot had a lot of capabilities for Benedict, and now we have augmented my secret autopilot to offer capabilities for elderly people. Includes as well. >>But there was something in your press release called Auto thread. Pooling says it provides higher and sustained throughput. High concerns concerns concurrency by determining Apple number of transactions, which should be executed. Uh, what is that all about? The auto thread pool? It seems pretty interesting. How does it affect performance? Can you help us understand that? >>Yes, and this is one of the capabilities of alluding to which we have added in my secret autopilot for transaction processing. So here is the basic idea. If you have a system where there's a large number of old EP transactions coming into it at a high degrees of concurrency in many of the existing systems of my sequel based systems, it can lead to a state where there are few transactions executing, but a bunch of them can get blocked with or a pilot tried pulling. What we basically do is we do workload aware admission control and what this does is it figures out, what's the right scheduling or all of these algorithms, so that either the transactions are executing or as soon as something frees up, they can start executing, so there's no transaction which is blocked. The advantage to the customer of this capability is twofold. A get significantly better throughput compared to service like Aurora at high levels of concurrency. So at high concurrency, for instance, uh, my secret because of this capability Uh oh, thread pulling offers up to 10 times higher compared to Aurora, that's one first benefit better throughput. The second advantage is that the true part of the system never drops, even at high levels of concurrency, whereas in the case of Aurora, the trooper goes up, but then, at high concurrency is, let's say, starting, uh, level of 500 or something. It depends upon the underlying shit they're using the troopers just dropping where it's with my secret heatwave. The truth will never drops. Now, the ramification for the customer is that if the truth is not gonna drop, the user can start off with a small shape, get the performance and be a show that even the workload increases. They will never get a performance, which is worse than what they're getting with lower levels of concurrency. So this let's leads to customers provisioning a shape which is just right for them. And if they need, they can, uh, go with the largest shape. But they don't like, you know, over pay. So those are the two benefits. Better performance and sustain, uh, regardless of the level of concurrency. >>So how do we quantify that? I know you've got some benchmarks. How can you share comparisons with other cloud databases especially interested in in Amazon's own databases are obviously very popular, and and are you publishing those again and get hub, as you have done in the past? Take us through the benchmarks. >>Sure, So benchmarks are important because that gives customers a sense of what performance to expect and what price performance to expect. So we have run a number of benchmarks. And yes, all these benchmarks are available on guitar for customers to take a look at. So we have performance results on all the three castle workloads, ol DB Analytics and Machine Learning. So let's start with the Rdp for Rdp and primarily because of the auto thread pulling feature. We show that for the IPCC for attended dataset at high levels of concurrency, heatwave offers up to 10 times better throughput and this performance is sustained, whereas in the case of Aurora, the performance really drops. So that's the first thing that, uh, tend to alibi. Sorry, 10 gigabytes. B B C c. I can come and see the performance are the throughput is 10 times better than Aurora for analytics. We have done a comparison of my secret heatwave in AWS and compared with Red Ship Snowflake Googled inquiry, we find that the price performance of my secret heatwave compared to read ship is seven times better. So my sequel, Heat Wave in AWS, provides seven times better price performance than red ship. That's a very, uh, interesting results to us. Which means that customers of Red Shift are really going to take the service seriously because they're gonna get seven times better price performance. And this is all running in a W s so compared. >>Okay, carry on. >>And then I was gonna say, compared to like, Snowflake, uh, in AWS offers 10 times better price performance. And compared to Google, ubiquity offers 12 times better price performance. And this is based on a four terabyte p PCH workload. Results are available on guitar, and then the third category is machine learning and for machine learning, uh, for training, the performance of my secret heatwave is 25 times faster compared to that shit. So all the three workloads we have benchmark's results, and all of these scripts are available on YouTube. >>Okay, so you're comparing, uh, my sequel Heatwave on AWS to Red Shift and snowflake on AWS. And you're comparing my sequel Heatwave on a W s too big query. Obviously running on on Google. Um, you know, one of the things Oracle is done in the past when you get the price performance and I've always tried to call fouls you're, like, double your price for running the oracle database. Uh, not Heatwave, but Oracle Database on a W s. And then you'll show how it's it's so much cheaper on on Oracle will be like Okay, come on. But they're not doing that here. You're basically taking my sequel Heatwave on a W s. I presume you're using the same pricing for whatever you see to whatever else you're using. Storage, um, reserved instances. That's apples to apples on A W s. And you have to obviously do some kind of mapping for for Google, for big query. Can you just verify that for me, >>we are being more than fair on two dimensions. The first thing is, when I'm talking about the price performance for analytics, right for, uh, with my secret heat rape, the cost I'm talking about from my secret heat rape is the cost of running transaction processing, analytics and machine learning. So it's a fully loaded cost for the case of my secret heatwave. There has been I'm talking about red ship when I'm talking about Snowflake. I'm just talking about the cost of these databases for running, and it's only it's not, including the source database, which may be more or some other database, right? So that's the first aspect that far, uh, trip. It's the cost for running all three kinds of workloads, whereas for the competition, it's only for running analytics. The second thing is that for these are those services whether it's like shit or snowflakes, That's right. We're talking about one year, fully paid up front cost, right? So that's what most of the customers would pay for. Many of the customers would pay that they will sign a one year contract and pay all the costs ahead of time because they get a discount. So we're using that price and the case of Snowflake. The costs were using is their standard edition of price, not the Enterprise edition price. So yes, uh, more than in this competitive. >>Yeah, I think that's an important point. I saw an analysis by Marx Tamer on Wiki Bond, where he was doing the TCO comparisons. And I mean, if you have to use two separate databases in two separate licences and you have to do et yelling and all the labour associated with that, that that's that's a big deal and you're not even including that aspect in in your comparison. So that's pretty impressive. To what do you attribute that? You know, given that unlike, oh, ci within the AWS cloud, you don't have as much control over the underlying hardware. >>So look hard, but is one aspect. Okay, so there are three things which give us this advantage. The first thing is, uh, we have designed hateful foreign scale out architecture. So we came up with new algorithms we have come up with, like, uh, one of the design points for heat wave is a massively partitioned architecture, which leads to a very high degree of parallelism. So that's a lot of hype. Each were built, So that's the first part. The second thing is that although we don't have control over the hardware, but the second design point for heat wave is that it is optimised for commodity cloud and the commodity infrastructure so we can have another guys, what to say? The computer we get, how much network bandwidth do we get? How much of, like objects to a brand that we get in here? W s. And we have tuned heat for that. That's the second point And the third thing is my secret autopilot, which provides machine learning based automation. So what it does is that has the users workload is running. It learns from it, it improves, uh, various premieres in the system. So the system keeps getting better as you learn more and more questions. And this is the third thing, uh, as a result of which we get a significant edge over the competition. >>Interesting. I mean, look, any I SV can go on any cloud and take advantage of it. And that's, uh I love it. We live in a new world. How about machine learning workloads? What? What did you see there in terms of performance and benchmarks? >>Right. So machine learning. We offer three capabilities training, which is fully automated, running in France and explanations. So one of the things which many of our customers told us coming from the enterprise is that explanations are very important to them because, uh, customers want to know that. Why did the the system, uh, choose a certain prediction? So we offer explanations for all models which have been derailed by. That's the first thing. Now, one of the interesting things about training is that training is usually the most expensive phase of machine learning. So we have spent a lot of time improving the performance of training. So we have a bunch of techniques which we have developed inside of Oracle to improve the training process. For instance, we have, uh, metal and proxy models, which really give us an advantage. We use adaptive sampling. We have, uh, invented in techniques for paralysing the hyper parameter search. So as a result of a lot of this work, our training is about 25 times faster than that ship them health and all the data is, uh, inside the database. All this processing is being done inside the database, so it's much faster. It is inside the database. And I want to point out that there is no additional charge for the history of customers because we're using the same cluster. You're not working in your service. So all of these machine learning capabilities are being offered at no additional charge inside the database and as a performance, which is significantly faster than that, >>are you taking advantage of or is there any, uh, need not need, but any advantage that you can get if two by exploiting things like gravity. John, we've talked about that a little bit in the past. Or trainee. Um, you just mentioned training so custom silicon that AWS is doing, you're taking advantage of that. Do you need to? Can you give us some insight >>there? So there are two things, right? We're always evaluating What are the choices we have from hybrid perspective? Obviously, for us to leverage is right and like all the things you mention about like we have considered them. But there are two things to consider. One is he is a memory system. So he favours a big is the dominant cost. The processor is a person of the cost, but memory is the dominant cost. So what we have evaluated and found is that the current shape which we are using is going to provide our customers with the best price performance. That's the first thing. The second thing is that there are opportunities at times when we can use a specialised processor for vaccinating the world for a bit. But then it becomes a matter of the cost of the customer. Advantage of our current architecture is on the same hardware. Customers are getting very good performance. Very good, energetic performance in a very good machine learning performance. If you will go with the specialised processor, it may. Actually, it's a machine learning, but then it's an additional cost with the customers we need to pay. So we are very sensitive to the customer's request, which is usually to provide very good performance at a very low cost. And we feel is that the current design we have as providing customers very good performance and very good price performance. >>So part of that is architectural. The memory intensive nature of of heat wave. The other is A W s pricing. If AWS pricing were to flip, it might make more sense for you to take advantage of something like like cranium. Okay, great. Thank you. And welcome back to the benchmarks benchmarks. Sometimes they're artificial right there. A car can go from 0 to 60 in two seconds. But I might not be able to experience that level of performance. Do you? Do you have any real world numbers from customers that have used my sequel Heatwave on A W s. And how they look at performance? >>Yes, absolutely so the my Secret service on the AWS. This has been in Vera for, like, since November, right? So we have a lot of customers who have tried the service. And what actually we have found is that many of these customers, um, planning to migrate from Aurora to my secret heat rape. And what they find is that the performance difference is actually much more pronounced than what I was talking about. Because with Aurora, the performance is actually much poorer compared to uh, like what I've talked about. So in some of these cases, the customers found improvement from 60 times, 240 times, right? So he travels 100 for 240 times faster. It was much less expensive. And the third thing, which is you know, a noteworthy is that customers don't need to change their applications. So if you ask the top three reasons why customers are migrating, it's because of this. No change to the application much faster, and it is cheaper. So in some cases, like Johnny Bites, what they found is that the performance of their applications for the complex storeys was about 60 to 90 times faster. Then we had 60 technologies. What they found is that the performance of heat we have compared to Aurora was 100 and 39 times faster. So, yes, we do have many such examples from real workloads from customers who have tried it. And all across what we find is if it offers better performance, lower cost and a single database such that it is compatible with all existing by sequel based applications and workloads. >>Really impressive. The analysts I talked to, they're all gaga over heatwave, and I can see why. Okay, last question. Maybe maybe two and one. Uh, what's next? In terms of new capabilities that customers are going to be able to leverage and any other clouds that you're thinking about? We talked about that upfront, but >>so in terms of the capabilities you have seen, like they have been, you know, non stop attending to the feedback from the customers in reacting to it. And also, we have been in a wedding like organically. So that's something which is gonna continue. So, yes, you can fully expect that people not dressed and continue to in a way and with respect to the other clouds. Yes, we are planning to support my sequel. He tripped on a show, and this is something that will be announced in the near future. Great. >>All right, Thank you. Really appreciate the the overview. Congratulations on the work. Really exciting news that you're moving my sequel Heatwave into other clouds. It's something that we've been expecting for some time. So it's great to see you guys, uh, making that move, and as always, great to have you on the Cube. >>Thank you for the opportunity. >>All right. And thank you for watching this special cube conversation. I'm Dave Volonte, and we'll see you next time.

Published Date : Sep 14 2022

SUMMARY :

The company is now in its fourth major release since the original announcement in December 2020. Very happy to be back. Now for those who might not have kept up with the news, uh, to kick things off, give us an overview of my So customers of my sequel then they had to run analytics or when they had to run machine So we've seen some interesting moves by Oracle lately. So one of the observations is that a very large percentage So was this a straightforward lifted shift from No, it is not because one of the design girls we have with my sequel, So I just want to make sure I understand that it's not like you just wrapped your stack in So for status, um, we have taken the mind sequel Heatwave code and we have optimised Can you help us understand that? So this let's leads to customers provisioning a shape which is So how do we quantify that? So that's the first thing that, So all the three workloads we That's apples to apples on A W s. And you have to obviously do some kind of So that's the first aspect And I mean, if you have to use two So the system keeps getting better as you learn more and What did you see there in terms of performance and benchmarks? So we have a bunch of techniques which we have developed inside of Oracle to improve the training need not need, but any advantage that you can get if two by exploiting We're always evaluating What are the choices we have So part of that is architectural. And the third thing, which is you know, a noteworthy is that In terms of new capabilities that customers are going to be able so in terms of the capabilities you have seen, like they have been, you know, non stop attending So it's great to see you guys, And thank you for watching this special cube conversation.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VolontePERSON

0.99+

December 2020DATE

0.99+

AmazonORGANIZATION

0.99+

JohnPERSON

0.99+

FranceLOCATION

0.99+

AWSORGANIZATION

0.99+

10 timesQUANTITY

0.99+

two thingsQUANTITY

0.99+

OracleORGANIZATION

0.99+

HeatwaveTITLE

0.99+

100QUANTITY

0.99+

60 timesQUANTITY

0.99+

one yearQUANTITY

0.99+

12 timesQUANTITY

0.99+

GWSORGANIZATION

0.99+

60 technologiesQUANTITY

0.99+

first partQUANTITY

0.99+

240 timesQUANTITY

0.99+

two separate licencesQUANTITY

0.99+

third categoryQUANTITY

0.99+

second advantageQUANTITY

0.99+

0QUANTITY

0.99+

seven timesQUANTITY

0.99+

two secondsQUANTITY

0.99+

twoQUANTITY

0.99+

AppleORGANIZATION

0.99+

seven timesQUANTITY

0.99+

GoogleORGANIZATION

0.99+

oneQUANTITY

0.99+

25 timesQUANTITY

0.99+

second pointQUANTITY

0.99+

NovemberDATE

0.99+

85 patentsQUANTITY

0.99+

second thingQUANTITY

0.99+

AuroraTITLE

0.99+

third thingQUANTITY

0.99+

EachQUANTITY

0.99+

second exampleQUANTITY

0.99+

10 gigabytesQUANTITY

0.99+

three thingsQUANTITY

0.99+

OneQUANTITY

0.99+

two benefitsQUANTITY

0.99+

one aspectQUANTITY

0.99+

first aspectQUANTITY

0.98+

two separate databasesQUANTITY

0.98+

over 10 yearsQUANTITY

0.98+

fourth major releaseQUANTITY

0.98+

39 timesQUANTITY

0.98+

first thingQUANTITY

0.98+

Heat WaveTITLE

0.98+

Paula Hansen and Jacqui van der Leij Greyling | Democratizing Analytics Across the Enterprise


 

(light upbeat music) (mouse clicks) >> Hey, everyone. Welcome back to the program. Lisa Martin here. I've got two guests joining me. Please welcome back to The Cube, Paula Hansen, the chief revenue officer and president at Alteryx. And Jacqui Van der Leij - Greyling joins us as well, the global head of tax technology at eBay. They're going to share with you how Alteryx is helping eBay innovate with analytics. Ladies, welcome. It's great to have you both on the program. >> Thank you, Lisa. >> Thank you, Lisa. >> It's great to be here. >> Yeah, Paula. We're going to start with you. In this program, we've heard from Jason Klein, we've heard from Alan Jacobson, they talked about the need to democratize analytics across any organization to really drive innovation. With analytics as they talked about at the forefront of software investments, how's Alteryx helping its customers to develop roadmaps for success with analytics? >> Well, thank you, Lisa. It absolutely is about our customer's success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts, of course, with our innovative technology and platform but ultimately, we help our customers to create a culture of data literacy and analytics from the top of the organization, starting with the C-suite. And we partner with our customers to build their roadmaps for scaling that culture of analytics through things like enablement programs, skills assessments, hackathons, setting up centers of excellence to help their organizations scale and drive governance of this analytics capability across the enterprise. So at the end of the day, it's really about helping our customers to move up their analytics maturity curve with proven technologies and best practices so they can make better business decisions and compete in their respective industries. >> Excellent. Sounds like a very strategic program. We're going to unpack that. Jacqui let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the IDC report that showed that 93% of organizations are not utilizing the analytics skills of their employees, but then there's eBay. How, Jacqui, did eBay become one of the 7% of organizations who's really maturing and how are you using analytics across the organization at eBay? >> So I think the main thing for us is just when we started out was, is that, you know, our, especially in finance they became spreadsheet professionals, instead of the things that we really want our employees to add value to. And we realized we had to address that. And we also knew we couldn't wait for all our data to be centralized until we actually start using the data or start automating and be more effective. So ultimately, we really started very, very actively embedding analytics in our people and our data and our processes. >> Starting with people is really critical. Jacqui, continuing with you, what were some of the roadblocks to analytics adoption that you faced and how did you overcome them? >> So I think, you know, eBay is a very data driven company. We have a lot of data. I think we are 27 years around this year so we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and just finding those data sources and finding ways to connect to them to move forward. The other thing is, is that you know, people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals, right? And there was no, we're not independent. You couldn't move forward. You would've been dependent on somebody else's roadmap to get to data and to get the information you wanted. So really finding something that everybody could access analytics or access data. And finally, we have to realize is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy and that is not so daunting on somebody who's brand new to the field? And I would call those out as your major roadblocks because you always have, not always, but most of the times you have support from the top in our case, we have, but in the end of the day, it's our people that need to actually really embrace it and making that accessible for them, I would say is definitely not per se, a roadblock but basically some, a block you want to be able to move. >> It's really all about putting people first. Question for both of you, and Paula will start with you, and then Jacqui will go to you. I think the message in this program that the audience is watching with us is very clear. Analytics is for everyone, should be for everyone. Let's talk now about how both of your organizations are empowering people those in the organization that may not have technical expertise to be able to leverage data so that they can actually be data driven? Paula? >> Yes. Well, we leverage our platform across all of our business functions here at Alteryx. And just like Jacqui explained at eBay finance is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jacqui mentioned, we have this huge amount of data flowing through our enterprise and the opportunity to leverage that into insights and analytics is really endless. So our CFO, Kevin Rubin has been a key sponsor for using our own technology. We use Alteryx for forecasting, all of our key performance metrics for business planning across our audit function to help with compliance and regulatory requirements, tax and even to close our books at the end of each quarter so it's really remained across our business. And at the end of the day, it comes to how do you train users? How do you engage users to lean into this analytic opportunity to discover use cases. And so one of the other things that we've seen many companies do is to gamify that process to build a game that brings users into the experience for training and to work with each other, to problem solve, and along the way, maybe earn badges depending on the capabilities and trainings that they take. And just have a little healthy competition as an employee base around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jacqui mentioned, it's really about ensuring that people feel comfortable, that they feel supported that they have access to the training that they need. And ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >> That confidence is key. Jacqui, talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >> Yeah, I think it means to what Paula has said in terms of you know, getting people excited about it but it's also understanding that this is a journey. And everybody is the different place in their journey. You have folks that's already really advanced who has done this every day, and then you have really some folks that this is brand new and, or maybe somewhere in between. And it's about how you could get everybody in their different phases to get to the initial destination. I say initially, because I believe the journey is never really complete. What we have done is that we decided to invest in a... We build a proof of concepts and we got our CFO to sponsor a hackathon. We opened it up to everybody in finance in the middle of the pandemic. So everybody was on Zoom. And we told people, "Listen, we're going to teach you this tool, super easy. And let's just see what you can do." We ended up having 70 entries. We had only three weeks. So, and these are people that has... They do not have a background. They are not engineers, they're not data scientists. And we ended up with a 25,000 hour savings at the end of that hackathon. From the 70 entries with people that have never, ever done anything like this before and there you had the result. And then it just went from there. It was people had a proof of concept, they knew that it worked, and they overcame that initial barrier of change. And that's where we are seeing things really, really picking up now. >> That's fantastic. And the business outcome that you mentioned there, the business impact is massive helping folks get that confidence to be able to overcome sometimes the cultural barriers is key here. I think another thing that this program has really highlighted is there is a clear demand for data literacy in the job market, regardless of organization. Can each of you share more about how you're empowering the next generation of data workers? Paula will start with you. >> Absolutely. And Jacqui says it so well, which is that it really is a journey that organizations are on. And we, as people in society are on in terms of upskilling our capabilities. So one of the things that we're doing here at Alteryx to help address this skillset gap on a global level is through a program that we call SparkED, which is essentially a no-cost analytics education program that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay, and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this program's really developed to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with SparkED, we started last May, but we currently have over 850 educational institutions globally engaged across 47 countries. And we're going to continue to invest here because there's so much opportunity for people, for society and for enterprises, when we close gap and empower more people with the necessary analytics skills to solve all the problems that data can help solve. >> So SparkED just made a really big impact in such a short time period. It's going to be fun to watch the progress of that. Jacqui let's go over to you now. Talk about some of the things that eBay is doing to empower the next generation of data workers. >> So we basically wanted to make sure that we kicked that momentum from the hackathon. Like we don't lose that excitement, right? So we just launched a program called eBay Masterminds. And what it basically is, it's an inclusive innovation initiative, where we firmly believe that innovation is for upscaling for all analytics role. So it doesn't matter your background, doesn't matter which function you are in, come and participate in this, where we really focus on innovation, introducing new technologies and upscaling our people. We are... Apart from that, we also said... Well, we should just keep it to inside eBay. We have to share this innovation with the community. So we are actually working on developing an analytics high school program, which we hope to pilot by the end of this year, where we will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, but also how to use alter Alteryx. And we're working with actually, we're working with SparkED and they're helping us develop that program. And we really hope that, let us say, by the end of the year have a pilot and then also next, was hoping to roll it out in multiple locations, in multiple countries, and really, really focus on that whole concept of analytics role. >> Analytics role, sounds like Alteryx and eBay have a great synergistic relationship there, that is jointly aimed at, especially, kind of, going down the stuff and getting people when they're younger interested and understanding how they can be empowered with data across any industry. Paula let's go back to you. You were recently on The Cube's Supercloud event just a couple of weeks ago. And you talked about the challenges the companies are facing as they're navigating what is by default a multi-cloud world? How does the Alteryx Analytics Cloud platform enable CIOs to democratize analytics across their organization? >> Yes, business leaders and CIOs across all industries are realizing that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organizations. Last, I check there was 2 million data scientists in the world. So that's woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. (Paula clears throat) So what we're seeing now with CIOs, with business leaders is that they're integrating data analysis and the skillset of data analysis into virtually every job function. And that is what we think of when we think of analytics for all. And so our mission with Alteryx Analytics Cloud, is to empower all of those people in every job function regardless of their skillset. As Jacqui pointed out from people that would, you know are just getting started all the way to the most sophisticated of technical users. Every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organizations. So that's our goal with Alteryx Analytics Cloud and it operates in a multi-cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyze and report out so that we can break down data silos across the enterprise and help drive real business outcomes as a result of unlocking the potential of data. >> As well as really lessening that skills gap as you were saying, there's only 2 million data scientists. You don't need to be a data scientist. That's the beauty of what Alteryx is enabling and eBay is a great example of that. Jacqui let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where Alteryx fits in as that analytics maturity journey continues. And what are some of the things that you are most excited about as analytics truly gets democratized across eBay? >> When we started about getting excited about things when it comes to analytics, I can go on all day but I'll keep it short and sweet for you. I do think we are on the topic full of data scientists. And I really feel that that is your next step, for us anyways, it's just that, how do we get folks to not see data scientists as this big thing, like a rocket scientist, it's something completely different. And it's something that is in everybody in a certain extent. So again, partnering with Alteryx would just release the AI/ML solution, allowing, you know, folks to not have a data scientist program but actually build models and be able to solve problems that way. So we have engaged with Alteryx and we purchased the licenses quite a few. And right now, through our mastermind program we're actually running a four-months program for all skill levels. Teaching them AI/ML and machine learning and how they can build their own models. We are really excited about that. We have over 50 participants without the background from all over the organization. We have members from our customer services, we have even some of our engineers, are actually participating in the program. We just kicked it off. And I really believe that that is our next step. I want to give you a quick example of the beauty of this is where we actually just allow people to go out and think about ideas and come up with things. And one of the people in our team who doesn't have a data scientist background at all was able to develop a solution where, you know, there is a checkout feedback, checkout functionality on the eBay site, where sellers or buyers can verbatim add information. And she build a model to be able to determine what relates to tax specific, what is the type of problem, and even predict how that problem can be solved before we, as a human even step in. And now instead of us or somebody going to the bay to try to figure out what's going on there, we can focus on fixing the error versus actually just reading through things and not adding any value. And it's a beautiful tool, and I'm very impressed when you saw the demo and they've been developing that further. >> That sounds fantastic. And I think just the one word that keeps coming to mind and we've said this a number of times in the program today is, empowerment. What you're actually really doing to truly empower people across the organization with varying degrees of skill level going down to the high school level, really exciting. We'll have to stay tuned to see what some of the great things are that come from this continued partnership. Ladies, I want to thank you so much for joining me on the program today and talking about how Alteryx and eBay are really partnering together to democratize analytics and to facilitate its maturity. It's been great talking to you >> Thank you, Lisa. >> Thank you so much. (light upbeat music) >> As you heard over the course of our program, organizations where more people are using analytics who have deeper capabilities in each of the four E's that's, everyone, everything, everywhere and easy analytics. Those organizations achieve more ROI from their respective investments in analytics and automation than those who don't. We also heard a great story from eBay, great example of an enterprise that is truly democratizing analytics across its organization. It's enabling an empowering line of business users to use analytics. Not only focused on key aspects of their job, but develop new skills rather than doing the same repetitive tasks. We want to thank you so much for watching the program today. Remember you can find all of the content on thecube.net. You can find all of the news from today on siliconangle.com, and of course alteryx.com. We also want to thank Alteryx for making this program possible and for sponsoring The Cube. For all of my guests, I'm Lisa Martin. We want to thank you for watching and bye for now. (light upbeat music)

Published Date : Sep 13 2022

SUMMARY :

the global head of tax technology at eBay. going to start with you. So at the end of the day, one of the things that we talked about instead of the things that that you faced and how but most of the times you that the audience is watching and the confidence to be able to be a part Jacqui, talk about some of the ways And everybody is the different get that confidence to be able to overcome that it's difficult to find Jacqui let's go over to you now. that momentum from the hackathon. And you talked about the in the opportunity to unlock and eBay is a great example of that. example of the beauty of this is It's been great talking to you Thank you so much. in each of the four E's

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JacquiPERSON

0.99+

PaulaPERSON

0.99+

Lisa MartinPERSON

0.99+

Jason KleinPERSON

0.99+

eBayORGANIZATION

0.99+

Kevin RubinPERSON

0.99+

AlteryxORGANIZATION

0.99+

LisaPERSON

0.99+

Paula HansenPERSON

0.99+

Alan JacobsonPERSON

0.99+

70 entriesQUANTITY

0.99+

Jacqui Van der Leij - GreylingPERSON

0.99+

25,000 hourQUANTITY

0.99+

four-monthsQUANTITY

0.99+

last MayDATE

0.99+

todayDATE

0.99+

27 yearsQUANTITY

0.99+

bothQUANTITY

0.99+

93%QUANTITY

0.99+

two guestsQUANTITY

0.99+

IDCORGANIZATION

0.99+

over 50 participantsQUANTITY

0.99+

Jacqui van der Leij GreylingPERSON

0.99+

oneQUANTITY

0.98+

siliconangle.comOTHER

0.98+

eachQUANTITY

0.98+

2 million data scientistsQUANTITY

0.98+

over 850 educational institutionsQUANTITY

0.97+

47 countriesQUANTITY

0.97+

7%QUANTITY

0.97+

thecube.netOTHER

0.97+

SupercloudEVENT

0.95+

three weeksQUANTITY

0.95+

end of this yearDATE

0.93+

Jason Klein, Alteryx | Democratizing Analytics Across the Enterprise


 

>> It's no surprise that 73% of organizations indicate analytics spend will outpace other software investments in the next 12 to 18 months. After all, as we know, data is changing the world, and the world is changing with it. But is everyone's spending resulting in the same ROI? This is Lisa Martin. Welcome to the Cube's presentation of "Democratizing Analytics Across the Enterprise," made possible by Alteryx. An Alteryx-commissioned IDC InfoBrief entitled, Four Ways to Unlock Transformative Business Outcomes From Analytics Investments, found that 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. On this special Cube presentation, Jason Klein, Product Marketing Director of Alteryx, will join me to share key findings from the new Alteryx-commissioned IDC Brief, and uncover how enterprises can derive more value from their data. In our second segment, we'll hear from Alan Jacobson, Chief Data and Analytics Officer at Alteryx. He's going to discuss how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. And then, in our final segment, Paula Hansen, who is the President and Chief Revenue Officer of Alteryx, and Jacqui Van der Leij-Greyling, who is the Global Head of Tax Technology at eBay, they'll join me. They're going to share how Alteryx is helping the global eCommerce company innovate with analytics. Let's get the show started. (upbeat music) Jason Klein joins me next, Product Marketing Director at Alteryx. Jason, welcome to the program. >> Hello, nice to be here. >> Excited to talk with you. What can you tell me about the new Alteryx IDC research which spoke with about 1500 leaders? What nuggets were in there? >> Well, as the business landscape changes over the next 12 to 18 months, we're going to see that analytics is going to be a key component to navigating this change. 73% of the orgs indicated that analytics spend will outpace other software investments. But just putting more money towards technology, it isn't going to solve everything. And this is why everyone's spending is resulting in different ROIs. And one of the reasons for this gap is because 93% of organizations, they're still not fully using the analytics skills of their employees. And this widening analytics gap, it's threatening operational progress by wasting workers' time, harming business productivity, and introducing costly errors. So in this research, we developed a framework of enterprise analytics proficiency that helps organizations reap greater benefits from their investments. And we based this framework on the behaviors of organizations that saw big improvements across financial, customer, and employee metrics. And we're able to focus on the behaviors driving higher ROI. >> So the InfoBrief also revealed that nearly all organizations are planning to increase their analytics spend. And it looks like from the InfoBrief that nearly three quarters plan on spending more on analytics than any other software. And can you unpack what's driving this demand, this need for analytics across organizations? >> Sure, well, first, there's more data than ever before. The data's changing the world, and the world is changing data. Enterprises across the world, they're accelerating digital transformation to capitalize on new opportunities, to grow revenue, to increase margins, and to improve customer experiences. And analytics, along with automation and AI, is what's making digital transformation possible. They're providing the fuel to new digitally enabled lines of business. >> Yet not all analytics spending is resulting in the same ROI. So, what are some of the discrepancies that the InfoBrief uncovered with respect to ROI? >> Well, our research with IDC revealed significant roadblocks across people, processes and technologies, all preventing companies from reaping greater benefits from their investments. So on the people side, for example, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy as compared to the technology itself. And next, while data is everywhere, most organizations, 63% in our survey, are still not using the full breadth of data types available. Data has never been this prolific. It's going to continue to grow, and orgs should be using it to their advantage. And lastly, organizations, they need to provide the right analytic tools to help everyone unlock the power of data, yet instead, they're relying on outdated spreadsheet technology. Nine out of 10 survey respondents said that less than half of their knowledge workers are active users of analytics software. True analytics transformation can't happen for an organization in a few select pockets or silos. We believe everyone, regardless of skill level, should be able to participate in the data and analytics process and drive value. >> So if I look at this holistically then, what would you say organizations need to do to make sure that they're really deriving value from their investments in analytics? >> Yeah, sure. So overall, the enterprises that derive more value >> from their data and analytics and achieved more ROI, they invested more aggressively in the four dimensions of enterprise analytics proficiency. So they've invested in the comprehensiveness of analytics, across all data sources and data types, meaning they're applying analytics to everything. They've invested in the flexibility of analytics across deployment scenarios and departments, meaning they're putting analytics everywhere. They've invested in the ubiquity of analytics and insights for every skill level, meaning they're making analytics for everyone. And they've invested in the usability of analytics software, meaning they're prioritizing easy technology to accelerate analytics democratization. >> So are there any specific areas that the survey uncovered where most companies are falling short? Like any black holes organizations need to be aware of from the outset? >> It did. You need to build a data-centric culture, and this begins with people. But we found that the people aspect of analytics is most heavily skewed towards low proficiency. In order to maximize ROI, organizations need to make sure everyone has access to the data and analytics technology they need. Organizations that align their analytics investments with upskilling enjoy higher ROI than orgs that are less aligned. For example, among the high ROI achievers in our survey, 78% had good or great alignment between analytics investments and workforce upskilling, compared to only 64% among those without positive ROI. And as more enterprises adopt cloud data warehouses or cloud data lakes to manage increasingly massive data sets, analytics needs to exist everywhere, especially for those cloud environments. And what we found is organizations that use more data types and more data sources generate higher ROI from their analytics investments. Among those with improved customer metrics, 90% were good or great at utilizing all data sources compared to only 67% among the ROI laggards. >> So interesting that you mentioned people. I'm glad that you mentioned people. Data scientists, everybody talks about data scientists. They're in high demand. We know that, but there aren't enough to meet the needs of all enterprises. So given that discrepancy, how can organizations fill the gap and really maximize the investments that they're making in analytics? >> Right. So analytics democratization, it's no longer optional, but it doesn't have to be complex. So we at Alteryx, we're democratizing analytics by empowering every organization to upskill every worker into a data worker. And the data from this survey shows this is the optimal approach. Organizations with a higher percentage of knowledge workers who are actively using analytics software enjoy higher returns from their analytics investment than orgs still stuck on spreadsheets. Among those with improved financial metrics, AKA the high ROI achievers, nearly 70% say that at least a quarter of their knowledge workers are using analytics software other than spreadsheets compared to only 56% in the low ROI group. Also, among the high ROI performers, 63% said data and analytic workers collaborate well or extremely well, compared to only 51% in the low ROI group. The data from the survey shows that supporting more business domains with analytics and providing cross-functional analytics correlates with higher ROI. So to maximize ROI, orgs should be transitioning workers from spreadsheets to analytics software. They should be letting them collaborate effectively, and letting them do so cross-functionally >> Yeah, that cross-functional collaboration is essential for anyone in any organization and in any discipline. Another key thing that jumped out from the survey was around shadow IT. The business side is using more data science tools than the IT side, and is expected to spend more on analytics than other IT. What risks does this present to the overall organization? If IT and the lines of business guys and gals aren't really aligned? >> Well, there needs to be better collaboration and alignment between IT and the line of business. The data from the survey, however, shows that business managers, they're expected to spend more on analytics and use more analytics tools than IT is aware of. And this is because the lines of business have recognized the value of analytics and plan to invest accordingly. But a lack of alignment between IT and business, this will negatively impact governance, which ultimately impedes democratization and hence, ROI. >> So Jason, where can organizations that are maybe at the outset of their analytics journey, or maybe they're in environments where there's multiple analytics tools across shadow IT, where can they go to Alteryx to learn more about how they can really simplify, streamline, and dial up the value on their investment? >> Well, they can learn more, you know, on our website. I also encourage them to explore the Alteryx community, which has lots of best practices, not just in terms of how you do the analytics, but how you stand up an Alteryx environment. But also to take a look at your analytics stack, and prioritize technologies that can snap to and enhance your organization's governance posture. It doesn't have to change it, but it should be able to align to and enhance it. >> And of course, as you mentioned, it's about people, process and technologies. Jason, thank you so much for joining me today, unpacking the IDC InfoBrief and the great nuggets in there. Lots that organizations can learn, and really become empowered to maximize their analytics investments. We appreciate your time. >> Thank you. It's been a pleasure. >> In a moment, Alan Jacobson, who's the Chief Data and Analytics Officer at Alteryx, is going to join me. He's going to be here to talk about how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. You're watching the Cube, the leader in tech enterprise coverage. (gentle music)

Published Date : Sep 13 2022

SUMMARY :

in the next 12 to 18 months. Excited to talk with you. over the next 12 to 18 months, And it looks like from the InfoBrief and the world is changing data. that the InfoBrief uncovered So on the people side, for example, should be able to participate So overall, the enterprises analytics to everything. analytics needs to exist everywhere, and really maximize the investments And the data from this survey shows If IT and the lines of and plan to invest accordingly. that can snap to and really become empowered to maximize It's been a pleasure. at Alteryx, is going to join me.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jason KleinPERSON

0.99+

Paula HansenPERSON

0.99+

Alan JacobsonPERSON

0.99+

Lisa MartinPERSON

0.99+

JasonPERSON

0.99+

AlteryxORGANIZATION

0.99+

Jacqui Van der Leij-GreylingPERSON

0.99+

NineQUANTITY

0.99+

IDCORGANIZATION

0.99+

90%QUANTITY

0.99+

second segmentQUANTITY

0.99+

93%QUANTITY

0.99+

63%QUANTITY

0.99+

eBayORGANIZATION

0.99+

73%QUANTITY

0.99+

78%QUANTITY

0.99+

oneQUANTITY

0.99+

InfoBriefORGANIZATION

0.99+

firstQUANTITY

0.99+

10 survey respondentsQUANTITY

0.97+

five organizationsQUANTITY

0.97+

about 1500 leadersQUANTITY

0.97+

todayDATE

0.97+

Democratizing Analytics Across the EnterpriseTITLE

0.97+

Four Ways to Unlock Transformative Business Outcomes From Analytics InvestmentsTITLE

0.96+

Alteryx IDCORGANIZATION

0.93+

CubeORGANIZATION

0.92+

nearly 70%QUANTITY

0.92+

18 monthsQUANTITY

0.92+

CubePERSON

0.91+

less than halfQUANTITY

0.9+

56%QUANTITY

0.89+

12QUANTITY

0.86+

67%QUANTITY

0.83+

IDC InfoBriefTITLE

0.78+

64%QUANTITY

0.75+

ChiefPERSON

0.73+

51%QUANTITY

0.73+

PresidentPERSON

0.73+

three quartersQUANTITY

0.64+

Chief Revenue OfficerPERSON

0.57+

the reasonsQUANTITY

0.52+

Alan Jacobson, Alteryx | Democratizing Analytics Across the Enterprise


 

>>Hey, everyone. Welcome back to accelerating analytics, maturity. I'm your host. Lisa Martin, Alan Jacobson joins me next. The chief data and analytics officer at Altrix Ellen. It's great to have you on the program. >>Thanks Lisa. >>So Ellen, as we know, everyone knows that being data driven is very important. It's a household term these days, but 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. What's your advice, your recommendations for organizations who are just starting out with analytics >>And you're spot on many organizations really aren't leveraging the, the full capability of their knowledge workers. And really the first step is probably assessing where you are on the journey, whether that's you personally, or your organization as a whole, we just launched an assessment tool on our website that we built with the international Institute of analytics, that in a very short period of time, in about 15 minutes, you can go on and answer some questions and understand where you sit versus your peer set versus competitors and kind of where you are on the journey. >>So when people talk about data analytics, they often think, ah, this is for data science experts like people like you. So why should people in the lines of business like the finance folks, the marketing folks, why should they learn analytics? >>So domain experts are really in the best position. They, they know where the gold is buried in their companies. They know where the inefficiencies are, and it is so much easier and faster to teach a domain expert a bit about how to automate a process or how to use analytics than it is to take a data scientist and try to teach them to have the knowledge of a 20 year accounting professional or a, or a logistics expert of your company. It much harder to do that. And really, if you think about it, the world has changed dramatically in a very short period of time. If, if you were a marketing professional 30 years ago, you likely didn't need to know anything about the internet, but today, do you know what you would call that marketing professional? If they didn't know anything about the internet, probably unemployed or retired. And so knowledge workers are having to learn more and more skills to really keep up with their professions. And analytics is really no exception. Pretty much in every profession, people are needing to learn analytics, to stay current and, and be capable for their companies. And companies need people who can do that. >>Absolutely. It seems like it's table stakes. These days, let's look at different industries. Now, are there differences in how you see analytics in automation being employed in different industries? I know Altrix is being used across a lot of different types of organizations from government to retail. I also see you're now with some of the leading sports teams, any differences in industries. >>Yeah. There's an incredible actually commonality between domains industry to industry. So if you look at what an HR professional is doing, maybe attrition analysis, it's probably quite similar, whether they're in oil and gas or in a high tech software company. And so really the similarities are, are much larger than you might think. And even on the, on, on the, on the sports front, we see many of the analytics that sports teams perform are very similar. So McLaren is one of the great partners that we work with and they use TRICS across many areas of their business from finance to production, extreme sports, logistics, wind tunnel engineering, the marketing team analyzes social media data, all using Altrics. And if I take as an example, the finance team, the finance team is trying to optimize the budget to make sure that they can hit the very stringent targets that F1 sports has. And I don't see a ton of difference between the optimization that they're doing to hit their budget numbers and what I see fortune 500 finance departments doing to optimize their budget. And so really the, the commonality is very high. Even across industries. >>I bet every F fortune 500 or even every company would love to be compared to the same department within McLaren F1, just to know that wow, what they're doing is so in incre incredibly important as is what we are doing. Absolutely. So talk about lessons learned, what lessons can business leaders take from those organizations like McLaren, who are the most analytically mature >>Probably first and foremost, is that the ROI with analytics and automation is incredibly high. Companies are having a ton of success. It's becoming an existential threat to some degree, if, if your company isn't going on this journey and your competition is it, it can be a, a huge problem. IDC just did a recent study about how companies are unlocking the ROI using analytics. And the data was really clear organizations that have a higher percentage of their workforce using analytics are enjoying a much higher return from their analytic investment. And so it's not about hiring two double PhD statisticians from Oxford. It really is how widely you can bring your workforce on this journey. Can they all get 10% more capable? And that's having incredible results at businesses all over the world. An another key finding that they had is that the majority of them said that when they had many folks using analytics, they were going on the journey faster than companies they didn't. And so picking technologies, that'll help everyone do this and, and do this fast and do it easily. Having an approachable piece of software that everyone can use is really a key, >>So faster able to move faster, higher ROI. I also imagine analytics across the organization is a big competitive advantage for organizations in any industry. >>Absolutely the IDC or not. The IDC, the international Institute of analytics showed huge correlation between companies that were more analytically mature versus ones that were not. They showed correlation to growth of the company. They showed correlation to revenue and they showed correlation to shareholder values. So across really all of the, the, the key measures of business, the more analytically mature companies simply outperformed their competition. >>And that's key these days is to be able to outperform your competition. You know, one of the things that we hear so often, Alan, is people talking about democratizing data and analytics. You talked about the line of business workers, but I gotta ask you, is it really that easy for the line of business workers who aren't trained in data science, to be able to jump in, look at data, uncover and extract business insights to make decisions. >>So in, in many ways, it really is that easy. I have a 14 and 16 year old kid. Both of them have learned Altrics they're, Altrics certified. And, and it was quite easy. It took 'em about 20 hours and they were, they, they were off to the races, but there can be some hard parts. The hard parts have more to do with change management. I mean, if you're an accountant, that's been doing the best accounting work in your company for the last 20 years. And all you happen to know is a spreadsheet for those 20 years. Are you ready to learn some new skills? And, and I would suggest you probably need to, if you want, keep up with your profession. The, the big four accounting firms have trained over a hundred thousand people in Altrix just one firm has trained over a hundred thousand. >>You, you can't be an accountant or an auditor at some of these places with, without knowing Altrix. And so the hard part, really in the end, isn't the technology and learning analytics and data science. The harder part is this change management change is hard. I should probably eat better and exercise more, but it's, it's hard to always do that. And so companies are finding that that's the hard part. They need to help people go on the journey, help people with the change management to, to help them become the digitally enabled accountant of the future. The, the logistics professional that is E enabled that that's the challenge. >>That's a huge challenge. Cultural, cultural shift is a challenge. As you said, change management. How, how do you advise customers? If you might be talking with someone who might be early in their analytics journey, but really need to get up to speed and mature to be competitive, how do you guide them or give them recommendations on being able to facilitate that change management? >>Yeah, that's a great question. So, so people entering into the workforce today, many of them are starting to have these skills Altrics is used in over 800 universities around the globe to teach finance and to teach marketing and to teach logistics. And so some of this is happening naturally as new workers are entering the workforce, but for all of those who are already in the workforce have already started their careers, learning in place becomes really important. And so we work with companies to put on programmatic approaches to help their workers do this. And so it's, again, not simply putting a box of tools in the corner and saying free, take one. We put on hackathons and analytic days, and it can, it can be great fun. We, we have a great time with, with many of the customers that we work with helping them, you know, do this, helping them go on the journey and the ROI, as I said, you know, is fantastic. And not only does it sometimes affect the bottom line, it can really make societal changes. We've seen companies have breakthroughs that really make great impact to society as a whole. >>Isn't that so fantastic to see the, the difference that that can make. It sounds like you're, you guys are doing a great job of democratizing access to alter X to everybody. We talked about the line of business folks and the incredible importance of enabling them and the, the ROI, the speed, the competitive advantage. Can you share some specific examples that you think of Alter's customers that really show data breakthroughs by the lines of business using the technology? >>Yeah, absolutely. So, so many to choose from I'll I'll, I'll give you two examples. Quickly. One is armor express. They manufacture life saving equipment, defensive equipments, like armor plated vests, and they were needing to optimize their supply chain, like many companies through the pandemic. We, we see how important the supply chain is. And so adjusting supply to, to match demand is, is really vital. And so they've used all tricks to model some of their supply and demand signals and built a predictive model to optimize the supply chain. And it certainly helped out from a, a dollar standpoint, they cut over a half a million dollars of inventory in the first year, but more importantly, by matching that demand and supply signal, you're able to better meet customer customer demand. And so when people have orders and are, are looking to pick up a vest, they don't wanna wait. >>And, and it becomes really important to, to get that right. Another great example is British telecom. They're, they're a company that services the public sector. They have very strict reporting regulations that they have to meet and they had, and, and this is crazy to think about over 140 legacy spreadsheet models that they had to run to comply with these regulatory processes and, and report, and obviously running 140 legacy models that had to be done in a certain order and linked incredibly challenging. It took them over four weeks, each time that they had to go through that process. And so to, to save time and have more efficiency in doing that, they trained 50 employees over just a two week period to start using Altrix and, and, and learn Altrix. And they implemented an all new reporting process that saw a 75% reduction in the number of man hours. >>It took to run in a 60% runtime performance. And so, again, a huge improvement. I can imagine it probably had better quality as well, because now that it was automated, you don't have people copying and past data into a spreadsheet. And that was just one project that this group of, of folks were able to accomplish that had huge ROI, but now those people are moving on and automating other processes and performing analytics in, in other areas, you can imagine the impact by the end of the year that they will have on their business, you know, potentially millions upon millions of dollars. This is what we see again. And again, company after company government agency, after government agency is how analytics are really transforming the way work is being done. >>That was the word that came to mind when you were describing the all three customer examples, the transformation, this is transformative. The ability to leverage alters to, to truly democratize data and analytics, give access to the lines of business is transformative for every organization. And, and also the business outcomes. You mentioned, those are substantial metrics based business outcomes. So the ROI and leveraging a technology like alri seems to be right there, sitting in front of you. >>That's right. And, and to be honest, it's not only important for these businesses. It's important for, for the knowledge workers themselves. I mean, we, we hear it from people that they discover Alrich, they automate a process. They finally get to get home for dinner with their families, which is fantastic, but, but it leads to new career paths. And so, you know, knowledge workers that have these added skills have so much larger opportunity. And I think it's great when the needs of businesses to become more analytics and analytic and automate processes actually matches the needs of the employees. And, you know, they too wanna learn these skills and become more advanced in their capabilities, >>Huge value there for the business, for the employees themselves to expand their skillset, to, to really open up so many opportunities for not only the business to meet the demands of the demanding customer, but the employees to be able to really have that breadth and depth in their field of service. Great opportunities there. Alan, is there anywhere that you wanna point the audience to go, to learn more about how they can get started? >>Yeah. So one of the things that we're really excited about is how fast and easy it is to learn these tools. So any of the listeners who wanna experience Altrix, they can go to the website, there's a free download on the website. You can take our analytic maturity assessment, as we talked about at the beginning and, and see where you are on the journey and just reach out. You know, we'd love to work with you and your organization to see how we can help you accelerate your journey on, on analytics and automation, >>Alan, it was a pleasure talking to you about democratizing data and analytics, the power in it for organizations across every industry. We appreciate your insights and your time. >>Thank you so much >>In a moment, Paula Hanson, who is the president and chief revenue officer of ultras and Jackie Vander lay graying. Who's the global head of tax technology at eBay will join me. You're watching the cube, the leader in high tech enterprise coverage.

Published Date : Sep 13 2022

SUMMARY :

It's great to have you on the program. the analytics skills of their employees, which is creating a widening analytics gap. And really the first step is probably assessing finance folks, the marketing folks, why should they learn analytics? about the internet, but today, do you know what you would call that marketing professional? government to retail. And so really the similarities are, are much larger than you might think. to the same department within McLaren F1, just to know that wow, what they're doing is so And the data was really I also imagine analytics across the organization is a big competitive advantage for They showed correlation to revenue and they showed correlation to shareholder values. And that's key these days is to be able to outperform your competition. And all you happen to know is a spreadsheet for those 20 years. And so companies are finding that that's the hard part. their analytics journey, but really need to get up to speed and mature to be competitive, the globe to teach finance and to teach marketing and to teach logistics. job of democratizing access to alter X to everybody. So, so many to choose from I'll I'll, I'll give you two examples. models that they had to run to comply with these regulatory processes and, the end of the year that they will have on their business, you know, potentially millions upon millions So the ROI and leveraging a technology like alri seems to be right there, And so, you know, knowledge workers that have these added skills have so much larger opportunity. of the demanding customer, but the employees to be able to really have that breadth and depth in So any of the listeners who wanna experience Altrix, Alan, it was a pleasure talking to you about democratizing data and analytics, the power in it for Who's the global head of tax technology at eBay will

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

Paula HansonPERSON

0.99+

AlanPERSON

0.99+

LisaPERSON

0.99+

Alan JacobsonPERSON

0.99+

60%QUANTITY

0.99+

AltrixORGANIZATION

0.99+

14QUANTITY

0.99+

75%QUANTITY

0.99+

BothQUANTITY

0.99+

20 yearQUANTITY

0.99+

EllenPERSON

0.99+

10%QUANTITY

0.99+

50 employeesQUANTITY

0.99+

eBayORGANIZATION

0.99+

16 yearQUANTITY

0.99+

93%QUANTITY

0.99+

Jackie VanderPERSON

0.99+

McLarenORGANIZATION

0.99+

millionsQUANTITY

0.99+

20 yearsQUANTITY

0.99+

Altrix EllenORGANIZATION

0.99+

IDCORGANIZATION

0.99+

two examplesQUANTITY

0.99+

first stepQUANTITY

0.99+

over a hundred thousandQUANTITY

0.99+

over a hundred thousand peopleQUANTITY

0.98+

over 800 universitiesQUANTITY

0.98+

first yearQUANTITY

0.98+

one firmQUANTITY

0.98+

two weekQUANTITY

0.98+

AltricsORGANIZATION

0.98+

todayDATE

0.98+

each timeQUANTITY

0.98+

Institute of analyticsORGANIZATION

0.98+

AlteryxORGANIZATION

0.98+

about 20 hoursQUANTITY

0.98+

OneQUANTITY

0.97+

oneQUANTITY

0.97+

one projectQUANTITY

0.97+

over a half a million dollarsQUANTITY

0.96+

about 15 minutesQUANTITY

0.96+

over four weeksQUANTITY

0.96+

AlrichORGANIZATION

0.93+

140 legacy modelsQUANTITY

0.91+

pandemicEVENT

0.91+

fortune 500ORGANIZATION

0.9+

30 years agoDATE

0.9+

F1EVENT

0.89+

threeQUANTITY

0.87+

over 140 legacy spreadsheet modelsQUANTITY

0.84+

AlterORGANIZATION

0.84+

firsQUANTITY

0.83+

two double PhD statisticiansQUANTITY

0.83+

endDATE

0.82+

four accounting firmsQUANTITY

0.82+

OxfordORGANIZATION

0.8+

TRICSORGANIZATION

0.73+

last 20 yearsDATE

0.66+

BritishLOCATION

0.66+

F fortune 500ORGANIZATION

0.57+

ultrasORGANIZATION

0.51+

dollarsQUANTITY

0.42+

Jason Klein Alteryx


 

>> It's no surprise that 73% of organizations indicate analytics spend will outpace other software investments in the next 12 to 18 months. After all, as we know, data is changing the world, and the world is changing with it. But is everyone's spending resulting in the same ROI? This is Lisa Martin. Welcome to the Cube's presentation of "Democratizing Analytics Across the Enterprise," made possible by Alteryx. An Alteryx-commissioned IDC InfoBrief entitled, Four Ways to Unlock Transformative Business Outcomes From Analytics Investments, found that 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. On this special Cube presentation, Jason Klein, Product Marketing Director of Alteryx, will join me to share key findings from the new Alteryx-commissioned IDC Brief, and uncover how enterprises can derive more value from their data. In our second segment, we'll hear from Alan Jacobson, Chief Data and Analytics Officer at Alteryx. He's going to discuss how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. And then, in our final segment, Paula Hansen, who is the President and Chief Revenue Officer of Alteryx, and Jacqui Van der Leij-Greyling, who is the Global Head of Tax Technology at eBay, they'll join me. They're going to share how Alteryx is helping the global eCommerce company innovate with analytics. Let's get the show started. (upbeat music) Jason Klein joins me next, Product Marketing Director at Alteryx. Jason, welcome to the program. >> Hello, nice to be here. >> Excited to talk with you. What can you tell me about the new Alteryx IDC research which spoke with about 1500 leaders? What nuggets were in there? >> Well, as the business landscape changes over the next 12 to 18 months, we're going to see that analytics is going to be a key component to navigating this change. 73% of the orgs indicated that analytics spend will outpace other software investments. But just putting more money towards technology, it isn't going to solve everything. And this is why everyone's spending is resulting in different ROIs. And one of the reasons for this gap is because 93% of organizations, they're still not fully using the analytics skills of their employees. And this widening analytics gap, it's threatening operational progress by wasting workers' time, harming business productivity, and introducing costly errors. So in this research, we developed a framework of enterprise analytics proficiency that helps organizations reap greater benefits from their investments. And we based this framework on the behaviors of organizations that saw big improvements across financial, customer, and employee metrics. And we're able to focus on the behaviors driving higher ROI. >> So the InfoBrief also revealed that nearly all organizations are planning to increase their analytics spend. And it looks like from the InfoBrief that nearly three quarters plan on spending more on analytics than any other software. And can you unpack what's driving this demand, this need for analytics across organizations? >> Sure, well, first, there's more data than ever before. The data's changing the world, and the world is changing data. Enterprises across the world, they're accelerating digital transformation to capitalize on new opportunities, to grow revenue, to increase margins, and to improve customer experiences. And analytics, along with automation and AI, is what's making digital transformation possible. They're providing the fuel to new digitally enabled lines of business. >> One of the things that the study also showed was that not all analytics spending is resulting in the same ROI. What are some of the discrepancies that the InfoBrief uncovered with respect to the the changes in ROI that organizations are achieving? >> Our research with IDC revealed significant roadblocks across people, processes, and technologies. They're preventing companies from reaping greater benefits from their investments. So for example, on the people side, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy, as compared to the technology itself. And next, while data is everywhere, most organizations, 63%, from our survey, are still not using the full breadth of data types available. Yet, data's never been this prolific. It's going to continue to grow, and orgs should be using it to their advantage. And lastly, organizations, they need to provide the right analytics tools to help everyone unlock the power of data. They instead rely on outdated spreadsheet technology. In our survey, 9 out of 10 respondents said less than half of their knowledge workers are active users of analytics software beyond spreadsheets. But true analytic transformation can't happen for an organization in a few select pockets or silos. We believe everyone, regardless of skill level, should be able to participate in the data and analytics process and be driving value. >> Should we retake that, since I started talking over Jason accidentally? >> Yep, absolutely, you can do so. Yep, we'll go back to Lisa's question. Let's retake the question and the answer. >> That'll be not all analog spending results in the same ROI. What are some of the discrepancies? >> Yes, Lisa, so we'll go from your ISO, just so we can get that clean question and answer. >> Okay. >> Thank you for that. on your ISO, we're still speeding, Lisa. So give it a beat in your head, and then on you. >> Yet not all analytics spending is resulting in the same ROI. So, what are some of the discrepancies that the InfoBrief uncovered with respect to ROI? >> Well, our research with IDC revealed significant roadblocks across people, processes and technologies, all preventing companies from reaping greater benefits from their investments. So on the people side, for example, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy as compared to the technology itself. And next, while data is everywhere, most organizations, 63% in our survey, are still not using the full breadth of data types available. Data has never been this prolific. It's going to continue to grow, and orgs should be using it to their advantage. And lastly, organizations, they need to provide the right analytic tools to help everyone unlock the power of data, yet instead, they're relying on outdated spreadsheet technology. Nine out of 10 survey respondents said that less than half of their knowledge workers are active users of analytics software. True analytics transformation can't happen for an organization in a few select pockets or silos. We believe everyone, regardless of skill level, should be able to participate in the data and analytics process and drive value. >> So if I look at this holistically then, what would you say organizations need to do to make sure that they're really deriving value from their investments in analytics? >> Yeah, sure. So overall, the enterprises that derive more value from their data and analytics and achieved more ROI, they invested more aggressively in the four dimensions of enterprise analytics proficiency. So they've invested in the comprehensiveness of analytics, across all data sources and data types, meaning they're applying analytics to everything. They've invested in the flexibility of analytics across deployment scenarios and departments, meaning they're putting analytics everywhere. They've invested in the ubiquity of analytics and insights for every skill level, meaning they're making analytics for everyone. And they've invested in the usability of analytics software, meaning they're prioritizing easy technology to accelerate analytics democratization. >> So very strategic investments. Did the survey uncover any specific areas where most companies are falling short, like any black holes that organizations need to be aware of at the outset? >> It did. It did. So organizations, they need to build a data-centric culture. And this begins with people. But what the survey told us is that the people aspect of analytics is the most heavily skewed towards low proficiency. In order to maximize ROI, organizations need to make sure everyone in the organization has access to the data and analytics technology they need. And then the organizations also have to align their investments with upskilling in data literacy to enjoy that higher ROI. Companies who did so experience higher ROI than companies who underinvested in analytics literacy. So among the high ROI achievers, 78% have a good or great alignment between analytics investment and workforce upskilling compared to only 64% among those without positive ROI. And as more orgs adopt cloud data warehouses or cloud data lakes, in order to manage the massively increasing workloads. Can I start that one over? Can I redo this one? >> Sure. >> Yeah >> Of course. Stand by. >> Tongue tied. >> Yep. No worries. >> One second. >> If we could get, if we could do the same, Lisa, just have a clean break. We'll go to your question. Yep. >> Yeah. >> On you Lisa. Just give that a count and whenever you're ready, here, I'm going to give us a little break. On you Lisa. >> So are there any specific areas that the survey uncovered where most companies are falling short? Like any black holes organizations need to be aware of from the outset? >> It did. You need to build a data-centric culture, and this begins with people. But we found that the people aspect of analytics is most heavily skewed towards low proficiency. In order to maximize ROI, organizations need to make sure everyone has access to the data and analytics technology they need. Organizations that align their analytics investments with upskilling enjoy higher ROI than orgs that are less aligned. For example, among the high ROI achievers in our survey, 78% had good or great alignment between analytics investments and workforce upskilling, compared to only 64% among those without positive ROI. And as more enterprises adopt cloud data warehouses or cloud data lakes to manage increasingly massive data sets, analytics needs to exist everywhere, especially for those cloud environments. And what we found is organizations that use more data types and more data sources generate higher ROI from their analytics investments. Among those with improved customer metrics, 90% were good or great at utilizing all data sources compared to only 67% among the ROI laggards. >> So interesting that you mentioned people. I'm glad that you mentioned people. Data scientists, everybody talks about data scientists. They're in high demand. We know that, but there aren't enough to meet the needs of all enterprises. So given that discrepancy, how can organizations fill the gap and really maximize the investments that they're making in analytics? >> Right. So analytics democratization, it's no longer optional, but it doesn't have to be complex. So we at Alteryx, we're democratizing analytics by empowering every organization to upskill every worker into a data worker. And the data from this survey shows this is the optimal approach. Organizations with a higher percentage of knowledge workers who are actively using analytics software enjoy higher returns from their analytics investment than orgs still stuck on spreadsheets. Among those with improved financial metrics, AKA the high ROI achievers, nearly 70% say that at least a quarter of their knowledge workers are using analytics software other than spreadsheets compared to only 56% in the low ROI group. Also, among the high ROI performers, 63% said data and analytic workers collaborate well or extremely well, compared to only 51% in the low ROI group. The data from the survey shows that supporting more business domains with analytics and providing cross-functional analytics correlates with higher ROI. So to maximize ROI, orgs should be transitioning workers from spreadsheets to analytics software. They should be letting them collaborate effectively, and letting them do so cross-functionally >> Yeah, that cross-functional collaboration is essential for anyone in any organization and in any discipline. Another key thing that jumped out from the survey was around shadow IT. The business side is using more data science tools than the IT side, and is expected to spend more on analytics than other IT. What risks does this present to the overall organization? If IT and the lines of business guys and gals aren't really aligned? >> Well, there needs to be better collaboration and alignment between IT and the line of business. The data from the survey, however, shows that business managers, they're expected to spend more on analytics and use more analytics tools than IT is aware of. And this is because the lines of business have recognized the value of analytics and plan to invest accordingly. But a lack of alignment between IT and business, this will negatively impact governance, which ultimately impedes democratization and hence, ROI. >> So Jason, where can organizations that are maybe at the outset of their analytics journey, or maybe they're in environments where there's multiple analytics tools across shadow IT, where can they go to Alteryx to learn more about how they can really simplify, streamline, and dial up the value on their investment? >> Well, they can learn more, you know, on our website. I also encourage them to explore the Alteryx community, which has lots of best practices, not just in terms of how you do the analytics, but how you stand up an Alteryx environment. But also to take a look at your analytics stack, and prioritize technologies that can snap to and enhance your organization's governance posture. It doesn't have to change it, but it should be able to align to and enhance it. >> And of course, as you mentioned, it's about people, process and technologies. Jason, thank you so much for joining me today, unpacking the IDC InfoBrief and the great nuggets in there. Lots that organizations can learn, and really become empowered to maximize their analytics investments. We appreciate your time. >> Thank you. It's been a pleasure. >> In a moment, Alan Jacobson, who's the Chief Data and Analytics Officer at Alteryx, is going to join me. He's going to be here to talk about how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. You're watching the Cube, the leader in tech enterprise coverage. (gentle music)

Published Date : Sep 10 2022

SUMMARY :

in the next 12 to 18 months. Excited to talk with you. over the next 12 to 18 months, And it looks like from the InfoBrief and the world is changing data. that the InfoBrief uncovered So for example, on the people side, Let's retake the question and the answer. in the same ROI. just so we can get that So give it a beat in your that the InfoBrief uncovered So on the people side, for example, So overall, the enterprises organizations need to be aware of is that the people aspect We'll go to your question. here, I'm going to give us a little break. to the data and analytics and really maximize the investments And the data from this survey shows If IT and the lines of and plan to invest accordingly. that can snap to and really become empowered to maximize Thank you. at Alteryx, is going to join me.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jason KleinPERSON

0.99+

JasonPERSON

0.99+

Paula HansenPERSON

0.99+

Alan JacobsonPERSON

0.99+

LisaPERSON

0.99+

Lisa MartinPERSON

0.99+

AlteryxORGANIZATION

0.99+

Jacqui Van der Leij-GreylingPERSON

0.99+

IDCORGANIZATION

0.99+

9QUANTITY

0.99+

90%QUANTITY

0.99+

second segmentQUANTITY

0.99+

63%QUANTITY

0.99+

NineQUANTITY

0.99+

93%QUANTITY

0.99+

78%QUANTITY

0.99+

eBayORGANIZATION

0.99+

One secondQUANTITY

0.99+

73%QUANTITY

0.99+

oneQUANTITY

0.99+

InfoBriefORGANIZATION

0.99+

todayDATE

0.99+

firstQUANTITY

0.99+

five organizationsQUANTITY

0.98+

OneQUANTITY

0.98+

about 1500 leadersQUANTITY

0.97+

five organizationsQUANTITY

0.97+

nearly 70%QUANTITY

0.97+

Democratizing Analytics Across the EnterpriseTITLE

0.97+

10 respondentsQUANTITY

0.97+

10 survey respondentsQUANTITY

0.97+

Four Ways to Unlock Transformative Business Outcomes From Analytics InvestmentsTITLE

0.96+

Alteryx IDCORGANIZATION

0.93+

64%QUANTITY

0.93+

CubeORGANIZATION

0.92+

ISOORGANIZATION

0.92+

18 monthsQUANTITY

0.92+

Jason Klein AlteryxPERSON

0.9+

less than halfQUANTITY

0.87+

12QUANTITY

0.86+

four dimensionsQUANTITY

0.83+

67%QUANTITY

0.81+

51%QUANTITY

0.78+

56%QUANTITY

0.76+

CubePERSON

0.75+

ChiefPERSON

0.73+

PresidentPERSON

0.73+

three quartersQUANTITY

0.64+

thingsQUANTITY

0.63+

Chief Revenue OfficerPERSON

0.57+

the reasonsQUANTITY

0.52+

InfoBriefTITLE

0.5+

Paula Hansen Jacqui van der Leij Greyling Alteryx


 

>>Hey everyone. Welcome back to the program. Lisa Martin here, I've got two guests joining me, please. Welcome back to the cube. Paula Hansen, the chief revenue officer and president at Al alters and Jackie Vander lake grayling joins us as well. The global head of tax technology at eBay. They're gonna share with you how an alter Ricks is helping eBay innovate with analytics. Ladies. Welcome. It's great to have you both on the program. >>Thank you, Lisa. It's great to be here. >>Yeah, Paula, we're gonna start with you in this program. We've heard from Jason Klein, we've heard from Alan Jacobson, they talked about the need to democratize analytics across any organization to really drive innovation with analytics. As they talked about at the forefront of software investments, how's alters helping its customers to develop roadmaps for success with analytics. >>Well, thank you, Lisa. It absolutely is about our customer's success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts of course, with our innovative technology and platform, but ultimately we help our customers to create a culture of data literacy and analytics from the top of the organization, starting with the C-suite. And we partner with our customers to build their roadmaps for scaling that culture of analytics through things like enablement programs, skills, assessments, hackathons, setting up centers of excellence to help their organizations scale and drive governance of this analytics capability across the enterprise. So at the end of the day, it's really about helping our customers to move up their analytics, maturity curve with proven technologies and best practices so they can make better business decisions and compete in their respective industries. >>Excellent. Sounds like a very strategic program. We're gonna unpack that Jackie, let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the IDC report that showed that 93% of organizations are not utilizing the analytics skills of their employees, but then there's eBay. How Jackie did eBay become one of the 7% of organizations who's really maturing and how are you using analytics across the organization at eBay? >>So I think the main thing for us is just when we started out was is that, you know, our, especially in finance, they became spreadsheet professionals instead of the things that we really want our employees to add value to. And we realized we had to address that. And we also knew we couldn't wait for all our data to be centralized until we actually start using the data or start automating and be more effective. So ultimately we really started very, very actively embedding analytics in our people and our data and our processes, >>Starting with people is really critical. Jackie, continuing with you, what were some of the roadblocks to analytics adoption that you faced and how did you overcome them? >>So I think, you know, eBay is a very data driven company. We have a lot of data. I think we are 27 years around this year, so we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and, and just finding those data sources and finding ways to connect to them to move forward. The other thing is, is that, you know, people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals, right? And we, there was no, we're not independent. You couldn't move forward. You would've opinion on somebody else's roadmap to get to data and to get the information you wanted. So really finding something that everybody could access analytics or access data. >>And finally we have to realize is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy? And that is not so daunting on somebody who's brand new to the field. And I would, I would call those out as your, as your major roadblocks, because you always have not always, but most of the times you have support from the top in our case, we have, but in the end of the day, it's, it's our people that need to actually really embrace it and, and making that accessible for them, I would say is definitely not per se, a roadblock, but basically some, a block you wanna be able to move. >>It's really all about putting people. First question for both of you and Paula will start with you. And then Jackie will go to you. I think the message in this program that the audience is watching with us is very clear. Analytics is for everyone should be for everyone. Let's talk now about how both of your organizations are empowering people, those in the organization that may not have technical expertise to be able to leverage data so that they can actually be data driven Paula. >>Yes. Well, we leverage our platform across all of our business functions here at Altrix and just like Jackie explained it, eBay finances is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jackie mentioned, we have this huge amount of data flowing through our enterprise and the opportunity to leverage that into insights and analytics is really endless. So our CFO, Kevin Rubin has been a, a key sponsor for using our own technology. We use Altrix for forecasting, all of our key performance metrics for business planning across our audit function, to help with compliance and regulatory requirements tax, and even to close our books at the end of each quarter. So it's really remain across our business. And at the end of the day, it comes to how do you train users? How do you engage users to lean into this analytic opportunity to discover use cases? >>And so one of the other things that we've seen many companies do is to gamify that process, to build a game that brings users into the experience for training and to work with each other, to problem solve and along the way, maybe earn badges depending on the capabilities and trainings that they take. And just have a little healthy competition as an employee base around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jackie mentioned, it's really about ensuring that people feel comfortable, that they feel supported, that they have access to the training that they need. And ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >>That confidence is key. Jackie, talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >>Yeah, I think it means to what Paula has said in terms of, you know, you know, getting people excited about it, but it's also understanding that this is a journey and everybody's the different place in their journey. You have folks that's already really advanced who has done this every day. And then you have really some folks that this is brand new and, or maybe somewhere in between. And it's about how you put, get everybody in their different phases to get to the, the initial destination. I say initially, because I believe the journey is never really complete. What we have done is, is that we decided to invest in an Ebola group of concept. And we got our CFO to sponsor a hackathon. We opened it up to everybody in finance, in the middle of the pandemic. So everybody was on zoom and we had, and we told people, listen, we're gonna teach you this tool super easy. >>And let's just see what you can do. We ended up having 70 entries. We had only three weeks. So, and these are people that has N that do not have a background. They are not engineers, they're not data scientists. And we ended up with a 25,000 hour savings at the end of that hackathon from the 70 inches with people that have never, ever done anything like this before and there you had the result. And then it just went from there. It was, people had a proof of concept. They, they knew that it worked and they overcame the initial barrier of change. And that's where we are seeing things really, really picking up. Now >>That's fantastic. And the, the business outcome that you mentioned there, the business impact is massive helping folks get that confidence to be able to overcome. Sometimes the, the cultural barriers is key. I think another thing that this program has really highlighted is there is a clear demand for data literacy in the job market, regardless of organization. Can each of you share more about how you are empowering the next generation of data workers, Paula will start with you? >>Absolutely. And, and Jackie says it so well, which is that it really is a journey that organizations are on. And, and we, as people in society are on in terms of upskilling our capabilities. So one of the things that we're doing here at Altrix to help address this skillset gap on a global level is through a program that we call sparked, which is essentially a, no-cost a no cost analytics education program that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this program's really developed just to, to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with sparked. We started last may, but we currently have over 850 educational institutions globally engaged across 47 countries. And we're gonna continue to invest here because there's so much opportunity for people, for society and for enterprises, when we close gap and empower more people within necessary analytics skills to solve all the problems that data can help solve. >>So spark has made a really big impact in such a short time period. And it's gonna be fun to watch the progress of that. Jackie, let's go over to you now talk about some of the things that eBay is doing to empower the next generation of data workers. >>So we basically wanted to make sure that we keep that momentum from the hackathon that we don't lose that excitement, right? So we just launched a program called Ebo masterminds. And what it basically is, it's an inclusive innovation initiative where we firmly believe that innovation is all up scaling for all analytics for. So it doesn't matter. Your background doesn't matter which function you are in, come and participate in, in this where we really focus on innovation, introducing new technologies and upskilling our people. We are apart from that, we also say, well, we should just keep it to inside eBay. We, we have to share this innovation with the community. So we are actually working on developing an analytics high school program, which we hope to pilot by the end of this year, where we will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, but also how to use alter alter. And we're working with actually, we're working with spark and they're helping us develop that program. And we really hope that as a say, by the end of the year, have a pilot and then also make you, so we roll it out in multiple locations in multiple countries and really, really focus on, on that whole concept of analytics, role >>Analytics for all sounds like ultra and eBay have a great synergistic relationship there that is jointly aimed at, especially kind of going down the staff and getting people when they're younger, interested, and understanding how they can be empowered with data across any industry. Paula, let's go back to you. You were recently on the Cube's super cloud event just a couple of weeks ago. And you talked about the challenges the companies are facing as they're navigating. What is by default a multi-cloud world? How does the alters analytics cloud platform enable CIOs to democratize analytics across their organization? >>Yes, business leaders and CIOs across all industries are realizing that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organizations. Last I check there was 2 million data scientists in the world. So that's woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. So what we're seeing now with CIOs with business leaders is that they're integrating data analysis and the skill of data analysis into virtually every job function. And that is what we think of when we think of analytics for all. And so our mission with Altrics analytics cloud is to empower all of those people in every job function, regardless of their skillset. As Jackie pointed out from people that would, you know, are just getting started all the way to the most sophisticated of technical users. Every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organizations. So that's our goal with Altrics analytics cloud, and it operates in a multi cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyze, and report out so that we can break down data silos across the enterprise and drive real business outcomes. As a result of unlocking the potential of data, >>As well as really re lessening that skill gap. As you were saying, there's only 2 million data scientists. You don't need to be a data scientist. That's the, the beauty of what Altrics is enabling. And, and eBay is a great example of that. Jackie, let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where alters fits in on as that analytics maturity journey continues and what are some of the things that you are most excited about as analytics truly gets democratized across eBay? >>When we start about getting excited about things, when it comes to analytics, I can go on all day, but I I'll keep it short and sweet for you. I do think we are on the topic full of, of, of data scientists. And I really feel that that is your next step for us anyways, is that, how do we get folks to not see data scientists as this big thing, like a rocket scientist, it's, it's something completely different. And it's something that, that is in everybody to a certain extent. So again, partner with three X would just released the AI ML solution, allowing, you know, folks to not have a data scientist program, but actually build models and be able to solve problems that way. So we have engaged with alters and we, we purchased a license, this quite a few. And right now through our mastermind program, we're actually running a four months program for all skill levels, teaching, teaching them AI ML and machine learning and how they can build their own models. >>We are really excited about that. We have over 50 participants without the background from all over the organization. We have members from our customer services. We have even some of our engineers are actually participating in the program. We just kicked it off. And I really believe that that is our next step. I wanna give you a quick example of, of the beauty of this is where we actually just allow people to go out and think about ideas and come up with things. And one of the people in our team who doesn't have a data scientist background at all, was able to develop a solution where, you know, there is a checkout feedback checkout functionality on the eBay site where sellers or buyers can verbatim add information. And she build a model to be able to determine what relates to tax specific, what is the type of problem, and even predict how that problem can be solved before we, as a human even step in, and now instead of us or somebody going to verbatim and try to figure out what's going on there, we can focus on fixing the error versus actually just reading through things and not adding any value. >>And it's a beautiful tool and very impressed. You saw the demo and they developing that further. >>That sounds fantastic. And I think just the one word that keeps coming to mind, and we've said this a number of times in the program today is empowerment. What you're actually really doing to truly empower people across the organization with, with varying degrees of skill level, going down to the high school level, really exciting, we'll have to stay tuned to see what some of the great things are that come from this continued partnership. Ladies, I wanna thank you so much for joining me on the program today and talking about how alters and eBay are really partnering together to democratize analytics and to facilitate its maturity. It's been great talking to you. >>Thank you. >>As you heard over the course of our program organizations, where more people are using analytics who have the deeper capabilities in each of the four E's, that's, everyone, everything everywhere and easy analytics, those organizations achieve more ROI from their respective investments in analytics and automation than those who don't. We also heard a great story from eBay, great example of an enterprise that is truly democratizing analytics across its organization. It's enabling an empowering line of business users to use analytics, not only focused on key aspects of their job, but develop new skills rather than doing the same repetitive tasks. We wanna thank you so much for watching the program today. Remember you can find all of the content on the cue.net. You can find all of the news from today on Silicon angle.com and of course, alter.com. We also wanna thank alt alters for making this program possible and for sponsored in the queue for all of my guests. I'm Lisa Martin. We wanna thank you for watching and bye for now.

Published Date : Sep 10 2022

SUMMARY :

It's great to have you both on the program. Yeah, Paula, we're gonna start with you in this program. end of the day, it's really about helping our customers to move up their analytics, Speaking of analytics maturity, one of the things that we talked about in this event is the IDC instead of the things that we really want our employees to add value to. adoption that you faced and how did you overcome them? data and to get the information you wanted. And finally we have to realize is that this is uncharted territory. those in the organization that may not have technical expertise to be able to leverage data it comes to how do you train users? that people feel comfortable, that they feel supported, that they have access to the training that they need. expertise to really be data driven. And then you have really some folks that this is brand new and, And we ended up with a 25,000 folks get that confidence to be able to overcome. and colleges globally to help build the next generation of data workers. Jackie, let's go over to you now talk about some of the things that eBay is doing to empower And we really hope that as a say, by the end of the year, And you talked about the challenges the companies are facing as in terms of the opportunity for people to be a part of the analytics solution. It obviously has the right culture to adapt to that. And it's something that, that is in everybody to a certain extent. And she build a model to be able to determine what relates to tax specific, You saw the demo and they developing that skill level, going down to the high school level, really exciting, we'll have to stay tuned to see what some of We wanna thank you so much for watching the program today.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

PaulaPERSON

0.99+

Jason KleinPERSON

0.99+

Kevin RubinPERSON

0.99+

eBayORGANIZATION

0.99+

Paula HansenPERSON

0.99+

JackiePERSON

0.99+

LisaPERSON

0.99+

AltrixORGANIZATION

0.99+

Alan JacobsonPERSON

0.99+

25,000 hourQUANTITY

0.99+

70 entriesQUANTITY

0.99+

Al altersORGANIZATION

0.99+

27 yearsQUANTITY

0.99+

93%QUANTITY

0.99+

bothQUANTITY

0.99+

70 inchesQUANTITY

0.99+

First questionQUANTITY

0.99+

todayDATE

0.99+

four monthsQUANTITY

0.99+

oneQUANTITY

0.99+

over 50 participantsQUANTITY

0.99+

two guestsQUANTITY

0.99+

IDCORGANIZATION

0.98+

2 million data scientistsQUANTITY

0.97+

47 countriesQUANTITY

0.97+

7%QUANTITY

0.97+

this yearDATE

0.97+

over 850 educational institutionsQUANTITY

0.97+

eachQUANTITY

0.96+

three weeksQUANTITY

0.95+

each quarterQUANTITY

0.95+

Altrics analyticsORGANIZATION

0.95+

three XORGANIZATION

0.94+

Jackie VanderPERSON

0.93+

couple of weeks agoDATE

0.93+

alter RicksORGANIZATION

0.93+

Altrics analyticsORGANIZATION

0.93+

pandemicEVENT

0.92+

last mayDATE

0.91+

AltricsORGANIZATION

0.89+

one wordQUANTITY

0.89+

Jacqui van der Leij Greyling AlteryxPERSON

0.86+

end of this yearDATE

0.81+

ultraORGANIZATION

0.76+

four E'sQUANTITY

0.73+

cue.netORGANIZATION

0.73+

endDATE

0.71+

Silicon angle.comOTHER

0.64+

Alteryx Democratizing Analytics Across the Enterprise Full Episode V1b


 

>> It's no surprise that 73% of organizations indicate analytics spend will outpace other software investments in the next 12 to 18 months. After all as we know, data is changing the world and the world is changing with it. But is everyone's spending resulting in the same ROI? This is Lisa Martin. Welcome to "theCUBE"'s presentation of democratizing analytics across the enterprise, made possible by Alteryx. An Alteryx commissioned IDC info brief entitled, "Four Ways to Unlock Transformative Business Outcomes from Analytics Investments" found that 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. On this special "CUBE" presentation, Jason Klein, product marketing director of Alteryx, will join me to share key findings from the new Alteryx commissioned IDC brief and uncover how enterprises can derive more value from their data. In our second segment, we'll hear from Alan Jacobson, chief data and analytics officer at Alteryx. He's going to discuss how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. And then in our final segment, Paula Hansen, who is the president and chief revenue officer of Alteryx, and Jacqui Van der Leij Greyling, who is the global head of tax technology at eBay, they'll join me. They're going to share how Alteryx is helping the global eCommerce company innovate with analytics. Let's get the show started. (upbeat music) Jason Klein joins me next, product marketing director at Alteryx. Jason, welcome to the program. >> Hello, nice to be here. >> Excited to talk with you. What can you tell me about the new Alteryx IDC research, which spoke with about 1500 leaders, what nuggets were in there? >> Well, as the business landscape changes over the next 12 to 18 months, we're going to see that analytics is going to be a key component to navigating this change. 73% of the orgs indicated that analytics spend will outpace other software investments. But just putting more money towards technology, it isn't going to solve everything. And this is why everyone's spending is resulting in different ROIs. And one of the reasons for this gap is because 93% of organizations, they're still not fully using the analytics skills of their employees, and this widening analytics gap, it's threatening operational progress by wasting workers' time, harming business productivity and introducing costly errors. So in this research, we developed a framework of enterprise analytics proficiency that helps organizations reap greater benefits from their investments. And we based this framework on the behaviors of organizations that saw big improvements across financial, customer, and employee metrics, and we're able to focus on the behaviors driving higher ROI. >> So the info brief also revealed that nearly all organizations are planning to increase their analytics spend. And it looks like from the info brief that nearly three quarters plan on spending more on analytics than any other software. And can you unpack, what's driving this demand, this need for analytics across organizations? >> Sure, well first there's more data than ever before, the data's changing the world, and the world is changing data. Enterprises across the world, they're accelerating digital transformation to capitalize on new opportunities, to grow revenue, to increase margins and to improve customer experiences. And analytics along with automation and AI is what's making digital transformation possible. They're providing the fuel to new digitally enabled lines of business. >> One of the things that the study also showed was that not all analytics spending is resulting in the same ROI. What are some of the discrepancies that the info brief uncovered with respect to the changes in ROI that organizations are achieving? >> Our research with IDC revealed significant roadblocks across people, processes, and technologies. They're preventing companies from reaping greater benefits from their investments. So for example, on the people side, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy as compared to the technology itself. And next, while data is everywhere, most organizations, 63% from our survey, are still not using the full breadth of data types available. Yet data's never been this prolific, it's going to continue to grow, and orgs should be using it to their advantage. And lastly organizations, they need to provide the right analytics tools to help everyone unlock the power of data. >> So they- >> They instead rely on outdated spreadsheet technology. In our survey, nine out of 10 respondents said less than half of their knowledge workers are active users of analytics software beyond spreadsheets. But true analytic transformation can't happen for an organization in a few select pockets or silos. We believe everyone regardless of skill level should be able to participate in the data and analytics process and be driving value. >> Should we retake that, since I started talking over Jason accidentally? >> Yep, absolutely we can do so. We'll just go, yep, we'll go back to Lisa's question. Let's just, let's do the, retake the question and the answer, that'll be able to. >> It'll be not all analytics spending results in the same ROI, what are some of the discrepancies? >> Yes, Lisa, so we'll go from your ISO, just so we get that clean question and answer. >> Okay. >> Thank you for that. On your ISO, we're still speeding, Lisa, so give it a beat in your head and then on you. >> Yet not all analytics spending is resulting in the same ROI. So what are some of the discrepancies that the info brief uncovered with respect to ROI? >> Well, our research with IDC revealed significant roadblocks across people, processes, and technologies, all preventing companies from reaping greater benefits from their investments. So on the people side, for example, only one out of five organizations reported a commensurate investment in upskilling for analytics and data literacy as compared to the technology itself. And next, while data is everywhere, most organizations, 63% in our survey, are still not using the full breadth of data types available. Data has never been this prolific. It's going to continue to grow and orgs should be using it to their advantage. And lastly, organizations, they need to provide the right analytic tools to help everyone unlock the power of data, yet instead they're relying on outdated spreadsheet technology. Nine of 10 survey respondents said that less than half of their knowledge workers are active users of analytics software. True analytics transformation can't happen for an organization in a few select pockets or silos. We believe everyone regardless of skill level should be able to participate in the data and analytics process and drive value. >> So if I look at this holistically, then what would you say organizations need to do to make sure that they're really deriving value from their investments in analytics? >> Yeah, sure. So overall, the enterprises that derive more value from their data and analytics and achieve more ROI, they invested more aggressively in the four dimensions of enterprise analytics proficiency. So they've invested in the comprehensiveness of analytics across all data sources and data types, meaning they're applying analytics to everything. They've invested in the flexibility of analytics across deployment scenarios and departments, meaning they're putting analytics everywhere. They've invested in the ubiquity of analytics and insights for every skill level, meaning they're making analytics for everyone. And they've invested in the usability of analytics software, meaning they're prioritizing easy technology to accelerate analytics democratization. >> So very strategic investments. Did the survey uncover any specific areas where most companies are falling short, like any black holes that organizations need to be aware of at the outset? >> It did, it did. So organizations, they need to build a data-centric culture. And this begins with people. But what the survey told us is that the people aspect of analytics is the most heavily skewed towards low proficiency. In order to maximize ROI, organizations need to make sure everyone in the organization has access to the data and analytics technology they need. And then the organizations also have to align their investments with upskilling in data literacy to enjoy that higher ROI. Companies who did so experience higher ROI than companies who underinvested in analytics literacy. So among the high ROI achievers, 78% have a good or great alignment between analytics investment and workforce upskilling compared to only 64% among those without positive ROI. And as more orgs adopt cloud data warehouses or cloud data lakes, in order to manage the massively increasing workloads- Can I start that one over. >> Sure. >> Can I redo this one? >> Yeah. >> Of course, stand by. >> Tongue tied. >> Yep, no worries. >> One second. >> If we could do the same, Lisa, just have a clean break, we'll go your question. >> Yep, yeah. >> On you Lisa. Just give that a count and whenever you're ready. Here, I'm going to give us a little break. On you Lisa. >> So are there any specific areas that the survey uncovered where most companies are falling short? Like any black holes organizations need to be aware of from the outset? >> It did. You need to build a data-centric culture and this begins with people, but we found that the people aspect of analytics is most heavily skewed towards low proficiency. In order to maximize ROI organizations need to make sure everyone has access to the data and analytics technology they need. Organizations that align their analytics investments with upskilling enjoy higher ROI than orgs that are less aligned. For example, among the high ROI achievers in our survey, 78% had good or great alignment between analytics investments and workforce upskilling, compared to only 64% among those without positive ROI. And as more enterprises adopt cloud data warehouses or cloud data lakes to manage increasingly massive data sets, analytics needs to exist everywhere, especially for those cloud environments. And what we found is organizations that use more data types and more data sources generate higher ROI from their analytics investments. Among those with improved customer metrics, 90% were good or great at utilizing all data sources, compared to only 67% among the ROI laggards. >> So interesting that you mentioned people, I'm glad that you mentioned people. Data scientists, everybody talks about data scientists. They're in high demand, we know that, but there aren't enough to meet the needs of all enterprises. So given that discrepancy, how can organizations fill the gap and really maximize the investments that they're making in analytics? >> Right, so analytics democratization, it's no longer optional, but it doesn't have to be complex. So we at Alteryx, we're democratizing analytics by empowering every organization to upskill every worker into a data worker. And the data from this survey shows this is the optimal approach. Organizations with a higher percentage of knowledge workers who are actively using analytics software enjoy higher returns from their analytics investment than orgs still stuck on spreadsheets. Among those with improved financial metrics, AKA the high ROI achievers, nearly 70% say that at least a quarter of their knowledge workers are using analytics software other than spreadsheets compared to only 56% in the low ROI group. Also among the high ROI performers, 63% said data and analytic workers collaborate well or extremely well compared to only 51% in the low ROI group. The data from the survey shows that supporting more business domains with analytics and providing cross-functional analytics correlates with higher ROI. So to maximize ROI, orgs should be transitioning workers from spreadsheets to analytics software. They should be letting them collaborate effectively and letting them do so cross-functionally. >> Yeah, that cross-functional collaboration is essential for anyone in any organization and in any discipline. Another key thing that jumped out from the survey was around shadow IT. The business side is using more data science tools than the IT side. And it's expected to spend more on analytics than other IT. What risks does this present to the overall organization, if IT and the lines of business guys and gals aren't really aligned? >> Well, there needs to be better collaboration and alignment between IT and the line of business. The data from the survey, however, shows that business managers, they're expected to spend more on analytics and use more analytics tools than IT is aware of. And this isn't because the lines of business have recognized the value of analytics and plan to invest accordingly, but a lack of alignment between IT and business. This will negatively impact governance, which ultimately impedes democratization and hence ROI. >> So Jason, where can organizations that are maybe at the outset of their analytics journey, or maybe they're in environments where there's multiple analytics tools across shadow IT, where can they go to Alteryx to learn more about how they can really simplify, streamline, and dial up the value on their investment? >> Well, they can learn more on our website. I also encourage them to explore the Alteryx community, which has lots of best practices, not just in terms of how you do the analytics, but how you stand up in Alteryx environment, but also to take a look at your analytics stack and prioritize technologies that can snap to and enhance your organization's governance posture. It doesn't have to change it, but it should be able to align to and enhance it. >> And of course, as you mentioned, it's about people, process, and technologies. Jason, thank you so much for joining me today, unpacking the IDC info brief and the great nuggets in there. Lots that organizations can learn and really become empowered to maximize their analytics investments. We appreciate your time. >> Thank you, it's been a pleasure. >> In a moment, Alan Jacobson, who's the chief data and analytics officer at Alteryx is going to join me. He's going to be here to talk about how organizations across all industries can accelerate their analytic maturity to drive transformational business outcomes. You're watching "theCUBE", the leader in tech enterprise coverage. >> Somehow many have come to believe that data analytics is for the few, for the scientists, the PhDs, the MBAs. Well, it is for them, but that's not all. You don't have to have an advanced degree to do amazing things with data. You don't even have to be a numbers person. You can be just about anything. A titan of industry or a future titan of industry. You could be working to change the world, your neighborhood, or the course of your business. You can be saving lives or just looking to save a little time. The power of data analytics shouldn't be limited to certain job titles or industries or organizations because when more people are doing more things with data, more incredible things happen. Analytics makes us smarter and faster and better at what we do. It's practically a superpower. That's why we believe analytics is for everyone, and everything, and should be everywhere. That's why we believe in analytics for all. (upbeat music) >> Hey, everyone. Welcome back to "Accelerating Analytics Maturity". I'm your host, Lisa Martin. Alan Jacobson joins me next. The chief of data and analytics officer at Alteryx. Alan, it's great to have you on the program. >> Thanks, Lisa. >> So Alan, as we know, everyone knows that being data driven is very important. It's a household term these days, but 93% of organizations are not utilizing the analytics skills of their employees, which is creating a widening analytics gap. What's your advice, your recommendations for organizations who are just starting out with analytics? >> You're spot on, many organizations really aren't leveraging the full capability of their knowledge workers. And really the first step is probably assessing where you are on the journey, whether that's you personally, or your organization as a whole. We just launched an assessment tool on our website that we built with the International Institute of Analytics, that in a very short period of time, in about 15 minutes, you can go on and answer some questions and understand where you sit versus your peer set versus competitors and kind of where you are on the journey. >> So when people talk about data analytics, they often think, ah, this is for data science experts like people like you. So why should people in the lines of business like the finance folks, the marketing folks, why should they learn analytics? >> So domain experts are really in the best position. They know where the gold is buried in their companies. They know where the inefficiencies are. And it is so much easier and faster to teach a domain expert a bit about how to automate a process or how to use analytics than it is to take a data scientist and try to teach them to have the knowledge of a 20 year accounting professional or a logistics expert of your company. Much harder to do that. And really, if you think about it, the world has changed dramatically in a very short period of time. If you were a marketing professional 30 years ago, you likely didn't need to know anything about the internet, but today, do you know what you would call that marketing professional if they didn't know anything about the internet, probably unemployed or retired. And so knowledge workers are having to learn more and more skills to really keep up with their professions. And analytics is really no exception. Pretty much in every profession, people are needing to learn analytics to stay current and be capable for their companies. And companies need people who can do that. >> Absolutely, it seems like it's table stakes these days. Let's look at different industries now. Are there differences in how you see analytics in automation being employed in different industries? I know Alteryx is being used across a lot of different types of organizations from government to retail. I also see you're now with some of the leading sports teams. Any differences in industries? >> Yeah, there's an incredible actually commonality between the domains industry to industry. So if you look at what an HR professional is doing, maybe attrition analysis, it's probably quite similar, whether they're in oil and gas or in a high tech software company. And so really the similarities are much larger than you might think. And even on the sports front, we see many of the analytics that sports teams perform are very similar. So McLaren is one of the great partners that we work with and they use Alteryx across many areas of their business from finance to production, extreme sports, logistics, wind tunnel engineering, the marketing team analyzes social media data, all using Alteryx, and if I take as an example, the finance team, the finance team is trying to optimize the budget to make sure that they can hit the very stringent targets that F1 Sports has, and I don't see a ton of difference between the optimization that they're doing to hit their budget numbers and what I see Fortune 500 finance departments doing to optimize their budget, and so really the commonality is very high, even across industries. >> I bet every Fortune 500 or even every company would love to be compared to the same department within McLaren F1. Just to know that wow, what they're doing is so incredibly important as is what we're doing. >> So talk- >> Absolutely. >> About lessons learned, what lessons can business leaders take from those organizations like McLaren, who are the most analytically mature? >> Probably first and foremost, is that the ROI with analytics and automation is incredibly high. Companies are having a ton of success. It's becoming an existential threat to some degree, if your company isn't going on this journey and your competition is, it can be a huge problem. IDC just did a recent study about how companies are unlocking the ROI using analytics. And the data was really clear, organizations that have a higher percentage of their workforce using analytics are enjoying a much higher return from their analytic investment, and so it's not about hiring two double PhD statisticians from Oxford. It really is how widely you can bring your workforce on this journey, can they all get 10% more capable? And that's having incredible results at businesses all over the world. An another key finding that they had is that the majority of them said that when they had many folks using analytics, they were going on the journey faster than companies that didn't. And so picking technologies that'll help everyone do this and do this fast and do it easily. Having an approachable piece of software that everyone can use is really a key. >> So faster, able to move faster, higher ROI. I also imagine analytics across the organization is a big competitive advantage for organizations in any industry. >> Absolutely the IDC, or not the IDC, the International Institute of Analytics showed huge correlation between companies that were more analytically mature versus ones that were not. They showed correlation to growth of the company, they showed correlation to revenue and they showed correlation to shareholder values. So across really all of the key measures of business, the more analytically mature companies simply outperformed their competition. >> And that's key these days, is to be able to outperform your competition. You know, one of the things that we hear so often, Alan, is people talking about democratizing data and analytics. You talked about the line of business workers, but I got to ask you, is it really that easy for the line of business workers who aren't trained in data science to be able to jump in, look at data, uncover and extract business insights to make decisions? >> So in many ways, it really is that easy. I have a 14 and 16 year old kid. Both of them have learned Alteryx, they're Alteryx certified and it was quite easy. It took 'em about 20 hours and they were off to the races, but there can be some hard parts. The hard parts have more to do with change management. I mean, if you're an accountant that's been doing the best accounting work in your company for the last 20 years, and all you happen to know is a spreadsheet for those 20 years, are you ready to learn some new skills? And I would suggest you probably need to, if you want to keep up with your profession. The big four accounting firms have trained over a hundred thousand people in Alteryx. Just one firm has trained over a hundred thousand. You can't be an accountant or an auditor at some of these places without knowing Alteryx. And so the hard part, really in the end, isn't the technology and learning analytics and data science, the harder part is this change management, change is hard. I should probably eat better and exercise more, but it's hard to always do that. And so companies are finding that that's the hard part. They need to help people go on the journey, help people with the change management to help them become the digitally enabled accountant of the future, the logistics professional that is E enabled, that's the challenge. >> That's a huge challenge. Cultural shift is a challenge, as you said, change management. How do you advise customers if you might be talking with someone who might be early in their analytics journey, but really need to get up to speed and mature to be competitive, how do you guide them or give them recommendations on being able to facilitate that change management? >> Yeah, that's a great question. So people entering into the workforce today, many of them are starting to have these skills. Alteryx is used in over 800 universities around the globe to teach finance and to teach marketing and to teach logistics. And so some of this is happening naturally as new workers are entering the workforce, but for all of those who are already in the workforce, have already started their careers, learning in place becomes really important. And so we work with companies to put on programmatic approaches to help their workers do this. And so it's, again, not simply putting a box of tools in the corner and saying free, take one. We put on hackathons and analytic days, and it can be great fun. We have a great time with many of the customers that we work with, helping them do this, helping them go on the journey, and the ROI, as I said, is fantastic. And not only does it sometimes affect the bottom line, it can really make societal changes. We've seen companies have breakthroughs that have really made great impact to society as a whole. >> Isn't that so fantastic, to see the difference that that can make. It sounds like you guys are doing a great job of democratizing access to Alteryx to everybody. We talked about the line of business folks and the incredible importance of enabling them and the ROI, the speed, the competitive advantage. Can you share some specific examples that you think of Alteryx customers that really show data breakthroughs by the lines of business using the technology? >> Yeah, absolutely, so many to choose from. I'll give you two examples quickly. One is Armor Express. They manufacture life saving equipment, defensive equipments, like armor plated vests, and they were needing to optimize their supply chain, like many companies through the pandemic. We see how important the supply chain is. And so adjusting supply to match demand is really vital. And so they've used Alteryx to model some of their supply and demand signals and built a predictive model to optimize the supply chain. And it certainly helped out from a dollar standpoint. They cut over a half a million dollars of inventory in the first year, but more importantly, by matching that demand and supply signal, you're able to better meet customer demand. And so when people have orders and are looking to pick up a vest, they don't want to wait. And it becomes really important to get that right. Another great example is British Telecom. They're a company that services the public sector. They have very strict reporting regulations that they have to meet and they had, and this is crazy to think about, over 140 legacy spreadsheet models that they had to run to comply with these regulatory processes and report, and obviously running 140 legacy models that had to be done in a certain order and length, incredibly challenging. It took them over four weeks each time that they had to go through that process. And so to save time and have more efficiency in doing that, they trained 50 employees over just a two week period to start using Alteryx and learn Alteryx. And they implemented an all new reporting process that saw a 75% reduction in the number of man hours it took to run in a 60% run time performance. And so, again, a huge improvement. I can imagine it probably had better quality as well, because now that it was automated, you don't have people copying and pasting data into a spreadsheet. And that was just one project that this group of folks were able to accomplish that had huge ROI, but now those people are moving on and automating other processes and performing analytics in other areas. So you can imagine the impact by the end of the year that they will have on their business, potentially millions upon millions of dollars. And this is what we see again and again, company after company, government agency after government agency, is how analytics are really transforming the way work is being done. >> That was the word that came to mind when you were describing the all three customer examples, transformation, this is transformative. The ability to leverage Alteryx, to truly democratize data and analytics, give access to the lines of business is transformative for every organization. And also the business outcome you mentioned, those are substantial metrics based business outcomes. So the ROI in leveraging a technology like Alteryx seems to be right there, sitting in front of you. >> That's right, and to be honest, it's not only important for these businesses. It's important for the knowledge workers themselves. I mean, we hear it from people that they discover Alteryx, they automate a process, they finally get to get home for dinner with their families, which is fantastic, but it leads to new career paths. And so knowledge workers that have these added skills have so much larger opportunity. And I think it's great when the needs of businesses to become more analytic and automate processes actually matches the needs of the employees, and they too want to learn these skills and become more advanced in their capabilities. >> Huge value there for the business, for the employees themselves to expand their skillset, to really open up so many opportunities for not only the business to meet the demands of the demanding customer, but the employees to be able to really have that breadth and depth in their field of service. Great opportunities there, Alan. Is there anywhere that you want to point the audience to go to learn more about how they can get started? >> Yeah, so one of the things that we're really excited about is how fast and easy it is to learn these tools. So any of the listeners who want to experience Alteryx, they can go to the website, there's a free download on the website. You can take our analytic maturity assessment, as we talked about at the beginning, and see where you are on the journey and just reach out. We'd love to work with you and your organization to see how we can help you accelerate your journey on analytics and automation. >> Alan, it was a pleasure talking to you about democratizing data and analytics, the power in it for organizations across every industry. We appreciate your insights and your time. >> Thank you so much. >> In a moment, Paula Hansen, who is the president and chief revenue officer of Alteryx, and Jacqui Van der Leij Greyling, who's the global head of tax technology at eBay, will join me. You're watching "theCUBE", the leader in high tech enterprise coverage. >> 1200 hours of wind tunnel testing, 30 million race simulations, 2.4 second pit stops. >> Make that 2.3. >> Sector times out the wazoo. >> Way too much of this. >> Velocities, pressures, temperatures, 80,000 components generating 11.8 billion data points and one analytics platform to make sense of it all. When McLaren needs to turn complex data into winning insights, they turn to Alteryx. Alteryx, analytics automation. (upbeat music) >> Hey, everyone, welcome back to the program. Lisa Martin here, I've got two guests joining me. Please welcome back to "theCUBE" Paula Hansen, the chief revenue officer and president at Alteryx, and Jacqui Van der Leij Greyling joins us as well, the global head of tax technology at eBay. They're going to share with you how Alteryx is helping eBay innovate with analytics. Ladies, welcome, it's great to have you both on the program. >> Thank you, Lisa, it's great to be here. >> Yeah, Paula, we're going to start with you. In this program, we've heard from Jason Klein, we've heard from Alan Jacobson. They talked about the need to democratize analytics across any organization to really drive innovation. With analytics, as they talked about, at the forefront of software investments, how's Alteryx helping its customers to develop roadmaps for success with analytics? >> Well, thank you, Lisa. It absolutely is about our customers' success. And we partner really closely with our customers to develop a holistic approach to their analytics success. And it starts of course with our innovative technology and platform, but ultimately we help our customers to create a culture of data literacy and analytics from the top of the organization, starting with the C-suite. And we partner with our customers to build their roadmaps for scaling that culture of analytics, through things like enablement programs, skills assessments, hackathons, setting up centers of excellence to help their organization scale and drive governance of this analytics capability across the enterprise. So at the end of the day, it's really about helping our customers to move up their analytics maturity curve with proven technologies and best practices, so they can make better business decisions and compete in their respective industries. >> Excellent, sounds like a very strategic program, we're going to unpack that. Jacqui, let's bring you into the conversation. Speaking of analytics maturity, one of the things that we talked about in this event is the IDC report that showed that 93% of organizations are not utilizing the analytics skills of their employees, but then there's eBay. How Jacqui did eBay become one of the 7% of organizations who's really maturing and how are you using analytics across the organization at eBay? >> So I think the main thing for us is when we started out was is that, our, especially in finance, they became spreadsheet professionals instead of the things that we really want our employees to add value to. And we realized we had to address that. And we also knew we couldn't wait for all our data to be centralized until we actually start using the data or start automating and being more effective. So ultimately we really started very, very actively embedding analytics in our people and our data and our processes. >> Starting with people is really critical. Jacqui, continuing with you, what were some of the roadblocks to analytics adoption that you faced and how did you overcome them? >> So I think eBay is a very data driven company. We have a lot of data. I think we are 27 years around this year, so we have the data, but it is everywhere. And how do you use that data? How do you use it efficiently? How do you get to the data? And I believe that that is definitely one of our biggest roadblocks when we started out and just finding those data sources and finding ways to connect to them to move forward. The other thing is that people were experiencing a lot of frustration. I mentioned before about the spreadsheet professionals. And there was no, we were not independent. You couldn't move forward, you would've put it on somebody else's roadmap to get the data and to get the information if you want it. So really finding something that everybody could access analytics or access data. And finally we have to realize is that this is uncharted territory. This is not exactly something that everybody is used to working with every day. So how do you find something that is easy, and that is not so daunting on somebody who's brand new to the field? And I would call those out as your major roadblocks, because you always have, not always, but most of the times you have support from the top, and in our case we have, but at the end of the day, it's our people that need to actually really embrace it, and making that accessible for them, I would say is definitely not per se, a roadblock, but basically a block you want to be able to move. >> It's really all about putting people first. Question for both of you, and Paula we'll start with you, and then Jacqui we'll go to you. I think the message in this program that the audience is watching with us is very clear. Analytics is for everyone, should be for everyone. Let's talk now about how both of your organizations are empowering people, those in the organization that may not have technical expertise to be able to leverage data, so that they can actually be data driven. Paula. >> Yes, well, we leverage our platform across all of our business functions here at Alteryx. And just like Jacqui explained, at eBay finance is probably one of the best examples of how we leverage our own platform to improve our business performance. So just like Jacqui mentioned, we have this huge amount of data flowing through our enterprise and the opportunity to leverage that into insights and analytics is really endless. So our CFO Kevin Rubin has been a key sponsor for using our own technology. We use Alteryx for forecasting all of our key performance metrics, for business planning, across our audit function, to help with compliance and regulatory requirements, tax, and even to close our books at the end of each quarter. So it's really going to remain across our business. And at the end of the day, it comes to how do you train users? How do you engage users to lean into this analytic opportunity to discover use cases? And so one of the other things that we've seen many companies do is to gamify that process, to build a game that brings users into the experience for training and to work with each other, to problem solve and along the way, maybe earn badges depending on the capabilities and trainings that they take. And just have a little healthy competition as an employee base around who can become more sophisticated in their analytic capability. So I think there's a lot of different ways to do it. And as Jacqui mentioned, it's really about ensuring that people feel comfortable, that they feel supported, that they have access to the training that they need, and ultimately that they are given both the skills and the confidence to be able to be a part of this great opportunity of analytics. >> That confidence is key. Jacqui, talk about some of the ways that you're empowering folks without that technical expertise to really be data driven. >> Yeah, I think it means to what Paula has said in terms of getting people excited about it, but it's also understanding that this is a journey and everybody is at a different place in their journey. You have folks that's already really advanced who has done this every day. And then you have really some folks that this is brand new or maybe somewhere in between. And it's about how you get everybody in their different phases to get to the initial destination. I say initial, because I believe a journey is never really complete. What we have done is that we decided to invest, and built a proof of concept, and we got our CFO to sponsor a hackathon. We opened it up to everybody in finance in the middle of the pandemic. So everybody was on Zoom and we told people, listen, we're going to teach you this tool, it's super easy, and let's just see what you can do. We ended up having 70 entries. We had only three weeks. So and these are people that do not have a background. They are not engineers, they're not data scientists. And we ended up with a 25,000 hour savings at the end of that hackathon from the 70 entries with people that have never, ever done anything like this before. And there you have the result. And then it just went from there. People had a proof of concept. They knew that it worked and they overcame the initial barrier of change. And that's where we are seeing things really, really picking up now. >> That's fantastic. And the business outcome that you mentioned there, the business impact is massive, helping folks get that confidence to be able to overcome sometimes the cultural barriers is key here. I think another thing that this program has really highlighted is there is a clear demand for data literacy in the job market, regardless of organization. Can each of you share more about how you're empowering the next generation of data workers? Paula, we'll start with you. >> Absolutely, and Jacqui says it so well, which is that it really is a journey that organizations are on and we as people in society are on in terms of upskilling our capabilities. So one of the things that we're doing here at Alteryx to help address this skillset gap on a global level is through a program that we call SparkED, which is essentially a no-cost analytics education program that we take to universities and colleges globally to help build the next generation of data workers. When we talk to our customers like eBay and many others, they say that it's difficult to find the skills that they want when they're hiring people into the job market. And so this program's really developed just to do just that, to close that gap and to work hand in hand with students and educators to improve data literacy for the next generation. So we're just getting started with SparkED. We started last May, but we currently have over 850 educational institutions globally engaged across 47 countries, and we're going to continue to invest here because there's so much opportunity for people, for society and for enterprises, when we close the gap and empower more people with the necessary analytics skills to solve all the problems that data can help solve. >> So SparkED has made a really big impact in such a short time period. It's going to be fun to watch the progress of that. Jacqui, let's go over to you now. Talk about some of the things that eBay is doing to empower the next generation of data workers. >> So we basically wanted to make sure that we kept that momentum from the hackathon, that we don't lose that excitement. So we just launched the program called eBay Masterminds. And what it basically is, is it's an inclusive innovation in each other, where we firmly believe that innovation is for upskilling for all analytics roles. So it doesn't matter your background, doesn't matter which function you are in, come and participate in in this where we really focus on innovation, introducing new technologies and upskilling our people. We are, apart from that, we also said, well, we shouldn't just keep it to inside eBay. We have to share this innovation with the community. So we are actually working on developing an analytics high school program, which we hope to pilot by the end of this year, where we will actually have high schoolers come in and teach them data essentials, the soft skills around analytics, but also how to use Alteryx. And we're working with, actually, we're working with SparkED and they're helping us develop that program. And we really hope that at, say, by the end of the year, we have a pilot and then also next year, we want to roll it out in multiple locations in multiple countries and really, really focus on that whole concept of analytics for all. >> Analytics for all, sounds like Alteryx and eBay have a great synergistic relationship there that is jointly aimed at especially going down the stuff and getting people when they're younger interested, and understanding how they can be empowered with data across any industry. Paula, let's go back to you, you were recently on "theCUBE"'s Supercloud event just a couple of weeks ago. And you talked about the challenges the companies are facing as they're navigating what is by default a multi-cloud world. How does the Alteryx Analytics Cloud platform enable CIOs to democratize analytics across their organization? >> Yes, business leaders and CIOs across all industries are realizing that there just aren't enough data scientists in the world to be able to make sense of the massive amounts of data that are flowing through organizations. Last I checked, there was 2 million data scientists in the world, so that's woefully underrepresented in terms of the opportunity for people to be a part of the analytics solution. So what we're seeing now with CIOs, with business leaders is that they're integrating data analysis and the skillset of data analysis into virtually every job function, and that is what we think of when we think of analytics for all. And so our mission with Alteryx Analytics Cloud is to empower all of those people in every job function, regardless of their skillset, as Jacqui pointed out from people that are just getting started all the way to the most sophisticated of technical users. Every worker across that spectrum can have a meaningful role in the opportunity to unlock the potential of the data for their company and their organizations. So that's our goal with Alteryx Analytics Cloud, and it operates in a multi cloud world and really helps across all sizes of data sets to blend, cleanse, shape, analyze, and report out so that we can break down data silos across the enterprise and help drive real business outcomes as a result of unlocking the potential of data. >> As well as really lessening that skill gap. As you were saying, there's only 2 million data scientists. You don't need to be a data scientist, that's the beauty of what Alteryx is enabling and eBay is a great example of that. Jacqui, let's go ahead and wrap things with you. You talked a great deal about the analytics maturity that you have fostered at eBay. It obviously has the right culture to adapt to that. Can you talk a little bit and take us out here in terms of where Alteryx fits in as that analytics maturity journey continues and what are some of the things that you are most excited about as analytics truly gets democratized across eBay? >> When we're starting up and getting excited about things when it comes to analytics, I can go on all day, but I'll keep it short and sweet for you. I do think we are on the top of the pool of data scientists. And I really feel that that is your next step, for us anyways, is that how do we get folks to not see data scientists as this big thing, like a rocket scientist, it's something completely different. And it's something that is in everybody in a certain extent. So again, partnering with Alteryx who just released the AI ML solution, allowing folks to not have a data scientist program, but actually build models and be able to solve problems that way. So we have engaged with Alteryx and we purchased the licenses, quite a few. And right now through our Masterminds program, we're actually running a four month program for all skill levels, teaching them AI ML and machine learning and how they can build their own models. We are really excited about that. We have over 50 participants without a background from all over the organization. We have members from our customer services. We have even some of our engineers are actually participating in the program. We just kicked it off. And I really believe that that is our next step. I want to give you a quick example of the beauty of this is where we actually just allow people to go out and think about ideas and come up with things. And one of the people in our team who doesn't have a data scientist background at all, was able to develop a solution where there is a checkout feedback functionality on the eBay side where sellers or buyers can verbatim add information. And she built a model to be able to determine what relates to tax specific, what is the type of problem, and even predict how that problem can be solved before we as a human even step in, and now instead of us or somebody going to verbatim and try to figure out what's going on there, we can focus on fixing the error versus actually just reading through things and not adding any value, and it's a beautiful tool and I was very impressed when I saw the demo and definitely developing that sort of thing. >> That sounds fantastic. And I think just the one word that keeps coming to mind, and we've said this a number of times in the program today is empowerment. What you're actually really doing to truly empower people across the organization with varying degrees of skill level, going down to the high school level, really exciting. We'll have to stay tuned to see what some of the great things are that come from this continued partnership. Ladies, I want to thank you so much for joining me on the program today and talking about how Alteryx and eBay are really partnering together to democratize analytics and to facilitate its maturity. It's been great talking to you. >> Thank you, Lisa. >> Thank you so much. (cheerful electronic music) >> As you heard over the course of our program, organizations where more people are using analytics who have deeper capabilities in each of the four Es, that's everyone, everything, everywhere, and easy analytics, those organizations achieve more ROI from their respective investments in analytics and automation than those who don't. We also heard a great story from eBay, great example of an enterprise that is truly democratizing analytics across its organization. It's enabling and empowering line of business users to use analytics, not only focused on key aspects of their job, but develop new skills rather than doing the same repetitive tasks. We want to thank you so much for watching the program today. Remember you can find all of the content on thecube.net. You can find all of the news from today on siliconangle.com and of course alteryx.com. We also want to thank Alteryx for making this program possible and for sponsoring "theCUBE". For all of my guests, I'm Lisa Martin. We want to thank you for watching and bye for now. (upbeat music)

Published Date : Sep 10 2022

SUMMARY :

in the next 12 to 18 months. Excited to talk with you. over the next 12 to 18 months, And it looks like from the info brief and the world is changing data. that the info brief uncovered with respect So for example, on the people side, in the data and analytics and the answer, that'll be able to. just so we get that clean Thank you for that. that the info brief uncovered as compared to the technology itself. So overall, the enterprises to be aware of at the outset? is that the people aspect of analytics If we could do the same, Lisa, Here, I'm going to give us a little break. to the data and analytics and really maximize the investments And the data from this survey shows this And it's expected to spend more and plan to invest accordingly, that can snap to and the great nuggets in there. Alteryx is going to join me. that data analytics is for the few, Alan, it's great to that being data driven is very important. And really the first step the lines of business and more skills to really keep of the leading sports teams. between the domains industry to industry. to be compared to the same is that the majority of them said So faster, able to So across really all of the is to be able to outperform that is E enabled, that's the challenge. and mature to be competitive, around the globe to teach finance and the ROI, the speed, that they had to run to comply And also the business of the employees, and they of the demanding customer, to see how we can help you the power in it for organizations and Jacqui Van der Leij 1200 hours of wind tunnel testing, to make sense of it all. back to the program. going to start with you. So at the end of the day, one of the 7% of organizations to be centralized until we of the roadblocks to analytics adoption and to get the information if you want it. that the audience is watching and the confidence to be able to be a part to really be data driven. in their different phases to And the business outcome and to work hand in hand Jacqui, let's go over to you now. We have to share this Paula, let's go back to in the opportunity to unlock and eBay is a great example of that. and be able to solve problems that way. that keeps coming to mind, Thank you so much. in each of the four Es,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JacquiPERSON

0.99+

PaulaPERSON

0.99+

Jason KleinPERSON

0.99+

Paula HansenPERSON

0.99+

Lisa MartinPERSON

0.99+

Paula HansenPERSON

0.99+

Alan JacobsonPERSON

0.99+

AlteryxORGANIZATION

0.99+

eBayORGANIZATION

0.99+

JasonPERSON

0.99+

International Institute of AnalyticsORGANIZATION

0.99+

LisaPERSON

0.99+

AlanPERSON

0.99+

Alan JacobsonPERSON

0.99+

60%QUANTITY

0.99+

Kevin RubinPERSON

0.99+

Jacqui Van der Leij GreylingPERSON

0.99+

14QUANTITY

0.99+

International Institute of AnalyticsORGANIZATION

0.99+

10%QUANTITY

0.99+

50 employeesQUANTITY

0.99+

63%QUANTITY

0.99+

93%QUANTITY

0.99+

90%QUANTITY

0.99+

nineQUANTITY

0.99+

75%QUANTITY

0.99+

70 entriesQUANTITY

0.99+

16 yearQUANTITY

0.99+

1200 hoursQUANTITY

0.99+

The Great Supercloud Debate | Supercloud22


 

[Music] welcome to the great super cloud debate a power panel of three top technology industry analysts maribel lopez is here she's the founder and principal analyst at lopez research keith townsend is ceo and founder of the cto advisor and sanjeev mohan is principal at sanjmo super cloud is a term that we've used to describe the future of cloud architectures the idea is that super clouds are built on top of hyperscaler capex infrastructure and the idea is it goes beyond multi-cloud the premise being that multi-cloud is primarily a symptom of multi-vendor or m a or both and results in more stove we're going to talk about that super cloud's meant to connote a new architecture that leverages the underlying primitives of hyperscale clouds but hides and abstracts that complexity of each of their respective clouds and adds new value on top of that with services and a continuous experience a similar or identical experience across more than one cloud people may say hey that's multi-cloud we're going to talk about that as well so with that as brief background um i'd like to first welcome our painless guys thanks so much for coming on thecube it's great to see you all again great to be here thank you to be here so i'm going to start with maribel you know what i just described what's your reaction to that is it just like what like cloud is supposed to be is that really what multi-cloud is do you agree with the premise that multi-cloud has really been you know what like chuck whitten from dell calls it it's been multi-cloud by default i call it a symptom of multi-vendor what's your take on on what this is oh wow dave another term here we go right more more to define for people but okay the reality is i agree that it's time for something new something evolved right whether we call that super cloud or something else i you know i don't want to really debate the term but we need to move beyond where we are today in multi-cloud and into if we want to call it cloud 5 multi-cloud 2 whatever we want to call it i believe that we're at the next generation that we have to define what that next generation is but if you think about it we went from public to private to hybrid to multi and every time you have a discussion with somebody about cloud you spend 10 minutes defining what you're talking about so this doesn't seem any different to me so let's just go with super cloud for the moment and see where we go and you know if you're interested after everybody else makes their comments i got a few thoughts about what super cloud might mean as well yeah great so i and i agree with you when we like i said in a recent post you could call it cl cloud you know multi-cloud 2.0 but it's something different is happening and sanjeev i know you're not a you're not a big fan of buzz words either but i wonder if you could weigh in on this topic uh you mean by the way sanjeev is at the mit cdo iq conference a great conference uh in boston uh and so he's it's a public place so we're going to have i think you viewed his line when he's not speaking please go ahead yeah so you know i come from a pedigree of uh being an analyst of uh firms that love inventing new terms i am not a big fan of inventing new terms i feel that when we come up with a new term i spend all my time standing on a stage trying to define what it is it takes me away from trying to solve the problem so so i'm you know i find these terms to be uh words of convenience like for example big data you know big data to me may not mean anything but big data connotes some of this modern way of handling vast volumes of data that traditional systems could not handle so from that point of view i'm i'm completely okay with super cloud but just inventing a new term is what i have called in my previous sessions tyranny of jargons where we have just too many jargons and uh and they resonate with i.t people they do not resonate with the business people business people care about the problem they don't care about what we and i t called them yeah and i think this is a really important point that you make and by the way we're not trying to create a new industry category per se yeah we leave that to gartner that's why actually i like super cloud because nobody's going to use that no vendor's going to use the term super cloud it's just too buzzy so so but but but it brings up the point about practitioners and so keith i want to bring you in so the what we've talked about and i'll just sort of share some some thoughts on the problems that we see and and get keith get your practitioner view most clouds most companies use multiple clouds we all kind of agree on that i think and largely these clouds operate in silos and they have their own development environment their own operating environment different apis different primitives and the functionality of a particular cloud doesn't necessarily extend to other clouds so the problem is that increases friction for customers increases cost increases security risk and so there's this promise maribel multi-cloud 2.0 that's going to solve that problem so keith my question to you is is is that an accurate description of the problem that practitioners face today do what did i miss and i wonder if you could elaborate so i think we'll get into some of the detail later on why this is a problem specifically around technologies but if we think about it in the abstract most customers have their hands full dealing with one cloud like we'll you know through m a and such and you zoom in and you look at companies that have multiple clouds or multi-cloud from result of mma mna m a activity you'll see that most of that is in silos so organizationally the customer may have multiple clouds but sub orchid silos they're generally a single silo in a single cloud so as you think about being able to take advantage of of tooling across the multicloud of what dave you guys are calling the super cloud this becomes a serious problem it's just a skill problem it's too much capability uh across too many things that look completely different than another okay so dave can i pick up on that please i'd love i was gonna just go to you maribel please chime in here okay so if we think about what we're talking about with super cloud and what keith just mentioned remember when we went to see tcp ip and the whole idea was like how do we get computers to talk to each other in a more standardized way how do we get data to move in a more standardized way i think that the problem we have with multi-cloud right now is that we don't have that so i think that's sort of a ground level of getting us to your super cloud premise is that and and you know google's tried it with anthony's like everybody every hyperscaler has tried their like right one to run anywhere but that abstraction layer you talk about what whatever we want to call it is super necessary and it's sort of the foundation so if you really think about it we've spent like 15 years or so building out all the various components of cloud and now's the time to take it so that cloud is actually more of an operating model versus a place there's at least a base level of it that is vendor neutral and then to your point the value that's going to be built on top of that you know people been trying to commoditize the basic infrastructure for a while now and i think that's what you're seeing in your super cloud multi-cloud whatever you want to call it the infrastructure is the infrastructure and then what would have been traditionally that past layer and above is where we're going to start to see some real innovation but we still haven't gotten to that point where you can do visibility observability manageability across that really complex cloud stack that we have the reason i the reason i love that tcpip example hm is because it changed the industry and it had an ecosystem effect in sanjiv the the the example that i first example that i used was snowflake a company that you're very familiar with that is sort of hiding all that complexity and right and so we're not there yet but please chime in on this topic uh you gotta you gotta view it again uh after you building upon what maribel said you know to me uh this sounds like a multi-cloud operating system where uh you know you need that kind of a common uh set of primitives and layers because if you go in in the typical multi-cloud process you've got multiple identities and you can't have that you how can you govern if i'm if i have multiple identities i don't have observability i don't know what's going on across my different stacks so to me super cloud is that call it single pane of glass or or one way through which i'm unifying my experience my my technology interfaces my integration and uh and i as an end user don't even care which uh which cloud i'm in it makes no difference to me it makes a difference to the vendor the vendor may say this is coming from aws and this is coming from gcp or azure but to the end user it is a consistent experience with consistent id and and observability and governance so that to me makes it a big difference and so one of floyer's contribution conversation was in order to have a super cloud you got to have a super pass i'm like oh boy people are going to love that but the point being that that allows a consistent developer experience and to maribel's earlier point about tcp it explodes the ecosystem because the ecosystem can now write to that super pass if you will those apis so keith do you do do you buy that number one and number two do you see that industries financial services and healthcare are actually going to be on clouds or what we call super clouds so sanjeev hit on a really key aspect of this is identity let's make this real they you love talk about data collaboration i love senji's point on the business user kind of doesn't care if this is aws versus super cloud versus etc i was collaborating with the client and he wanted to send video file and the video file uh his organization's access control policy didn't allow him to upload or share the file from their preferred platform so he had to go out to another cloud provider and create yet another identity for that data on the cloud same data different identity a proper super cloud will enable me to simply say as a end user here's a set of data or data sets and i want to share a collaboration a collaborator and that requires cross identity across multiple clouds so even before we get to the past layer and the apis we have to solve the most basic problem which is data how do we stop data scientists from shipping snowballs to a location because we can't figure out the identity the we're duplicating the same data within the same cloud because we can't share identity across customer accounts or etc we we have to solve these basic thoughts before we get to supercloud otherwise we get to us a turtles all the way down thing so we'll get into snowflake and what snowflake can do but that's what happens when i want to share my snowflake data across multiple clouds to a different platform yeah you have to go inside the snowflake cloud which leads right so i would say to keith's question sanjeev snowflake i think is solving that problem but then he brings up the other problem which is what if i want to share share data outside the snowflake cloud so that gets to the point of visit open is it closed and so sanji chime in on the sort of snowflake example and in maribel i wonder if there are networking examples because that's that's keith's saying you got to fix the plumbing before you get these higher level abstractions but sanji first yeah so i so i actually want to go and talk a little bit about network but from a data and analytics point of view so i never built upon what what keith said so i i want to give an example let's say i am getting fantastic web logs i and i know who uh uh how much time they're spending on my web pages and which pages they're looking at so i have all of that now all of that is going into cloud a now it turns out that i use google analytics or maybe i use adobe's you know analytics uh suite now that is giving me the business view and i'm trying to do customer journey analytics and guess what i now have two separate identities two separate products two separate clouds if i and i as an id person no problem i can solve any problem by writing tons of code but why would i do that if i can have that super pass or a multi-cloud layout where i've got like a single way of looking at my network traffic my customer metrics and i can do my customer journey analytics it solves a huge problem and then i can share that data with my with my partners so they can see data about their products which is a combination of data from different uh clouds great thank you uh maribel please i think we're having a lord of the rings moment here with the run one room to rule them all concept and i'm not sure that anybody's actually incented to do that right so i think there's two levels of the stack i think in the basic we're talking a lot about we don't have the basic fundamentals of how do you move data authenticate data secure data do data lineage all that stuff across different clouds right we haven't even spoken right now i feel like we're really just talking about the public cloud venue and we haven't even pulled in the fact that people are doing hybrid cloud right so hybrid cloud you know then you're talking about you've got hardware vendors and you've got hyperscaler vendors and there's two or three different ways of doing things so i honestly think that something will emerge like if we think about where we are in technology today it's almost like we need back to that operating system that sanji was talking about like we need a next generation operating system like nobody wants to build the cloud mouse driver of the 21st century over and over again right we need something like that as a foundation layer but then on top of it you know there's obviously a lot of opportunity to build differentiation like when i think back on what happened with cloud amazon remained aws remained very powerful and popular because people invested in building things on amazon right they created a platform and it took a while for anybody else to catch up to that or to have that kind of presence and i still feel that way when i talk to companies but having said that i talked to retail the other day and they were like hey we spent a long time building an abstraction layer on top of the clouds so that our developers could basically write once and run anywhere but they were a massive global presence retailer that's not something that everybody can do so i think that we are still missing a gap i don't know if that exactly answers your question but i i do feel like we're kind of in this chicken and egg thing which comes first and nobody wants to necessarily invest in like oh well you know amazon has built a way to do this so we're all just going to do it the amazon way right it seems like that's not going to work either but i think you bring up a really important point which there is going to be no one ring to rule them all you're going to have you know vmware is going to solve its multi-cloud problem snowflake's going to do a very has a very specific you know purpose-built system for it itself databricks is going to do its thing and it's going to be you know more open source i would companies like aviatrix i would say cisco even is going to go out and solve this problem dell showed at uh at dell tech world a thing called uh project alpine which is basically storage across clouds they're going to be many super clouds we're going to get maybe super cloud stove pipes but but the point is however for a specific problem in a set of use cases they will be addressing those and solving incremental value so keith maybe we won't have that single cloud operating you know system but we'll have multiple ones what are your thoughts on that yeah we're definitely going to have multiple ones uh the there is no um there is no community large enough or influential enough to push a design take maribel's example of the mega retailer they've solved it but they're not going to that's that's competitive that's their competitive advantage they're not going to share that with the rest of us and open source that and force that upon the industry via just agreement from everyone else so we're not going to get uh the level of collaboration either originated by the cloud provider originated from user groups that solves this problem big for us we will get silos in which this problem is solved we'll get groups working together inside of maybe uh industry or subgroups within the industry to say that hey we're going to share or federate identity across our three or four or five or a dozen organizations we'll be able to share data we're going to solve that data problem but in the same individual organizations in another part of the super cloud problem are going to again just be silos i can't uh i can't run machine learning against my web assets for the community group that i run because that's not part of the working group that solved a different data science problem so yes we're going to have these uh bifurcations and forks within the super cloud the question is where is the focus for each individual organization where do i point my smart people and what problems they solve okay i want to throw out a premise and get you guys reaction to it because i think this again i go back to the maribel's tcpip example it changed the industry it opened up an ecosystem and to me this is what digital transformation is all about you've got now industry participants marc andreessen says every company is a software company you've now got industry participants and here's some examples it's not i wouldn't call them true super clouds yet but walmart's doing their hybrid thing with azure you got goldman sachs announced at the last reinvent and it's going to take its tools its software its data and which is on-prem and connect that to the aws cloud and actually deliver a service capital one we saw sanjiv at the snowflake summit is is taking their tooling and doing it now granted just within snowflake and aws but i fully expect them to expand that across other clouds these are industry examples capital one software is the name of the division that are now it's to the re reason why i don't get so worried that we're not solving the lord of the rings problem that maribel mentioned is because it opens up tremendous opportunities for companies we got like just under five minutes left i want to throw that out there and see what you guys think yeah i would just i want to build upon what maribel said i love what she said you're not going to build a mouse driver so if multi-cloud supercloud is a multi-cloud os the mouse driver would be identity or maybe it's data quality and to teach point that data quality is not going to come from a single vendor that is going to come from a different vendor whose job is to to harmonize data because there might be data might be for the same identity but it may be a different granularity level so you cannot just mix and match so you need to have some sort of like resolution and that is is an example of a driver for multi-cloud interesting okay so you know octa might be the identity cloud or z scaler might be the security cloud or calibre has its cloud etc any thoughts on that keith or maribel yeah so let's talk about where the practical challenges run into this we did some really great research that was sponsored by one of the large cloud providers in which we took all we looked at all the vmware cloud solutions when i say vmware cloud vmware has a lot of products across multi-cloud now in the rock broadcloud portfolio but we're talking about the og solution vmware vsphere it would seem like on paper if i put vmware vsphere in each cloud that is therefore a super cloud i think we would all agree to that in principle what we found in our research was that when we put hands on keyboard the differences of the clouds show themselves in the training gap and that skills gap between the clouds show themselves if i needed to expose less our favorite friend a friend a tc pip address to the public internet that is a different process on each one of the clouds that needs to be done on each one of the clouds and not abstracted in vmware vsphere so as we look at the nuance yes we can give the big controls but where the capital ones the uh jp morgan chase just spent two billion dollars on this type of capability where the spin effort is done is taking it from that 80 percent to that 90 95 experience and that's where the effort and money is spent on that last mile maribel we're out of time but please you know bring us home give us your closing thoughts hey i think we're still going to be working on what the multi-cloud thing is for a while and you know super cloud i think is a direction of the future of cloud computing but we got some real problems to solve around authentication uh identity data lineage data security so i think those are going to be sort of the tactical things that we're working on for the next couple years right guys always a pleasure having you on the cube i hope we see you around keith i understand you're you're bringing your airstream to vmworld or vmware explorer putting it on the on the floor i can't wait to see that and uh mrs cto advisor i'm sure we'll be uh by your side so looking forward to that hopefully sanjeev and maribel we'll see you uh on the circuit as well yes hope to see you there right looking forward to hopefully even doing some content with you guys at vmware explorer too awesome looking forward all right keep it right there for more content from super cloud 22 right back [Music] you

Published Date : Jul 20 2022

SUMMARY :

that problem so keith my question to you

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
marc andreessenPERSON

0.99+

maribel lopezPERSON

0.99+

threeQUANTITY

0.99+

amazonORGANIZATION

0.99+

10 minutesQUANTITY

0.99+

twoQUANTITY

0.99+

two billion dollarsQUANTITY

0.99+

maribelPERSON

0.99+

sanjeevPERSON

0.99+

fourQUANTITY

0.99+

ciscoORGANIZATION

0.99+

fiveQUANTITY

0.99+

keithPERSON

0.99+

80 percentQUANTITY

0.99+

sanjiPERSON

0.99+

walmartORGANIZATION

0.99+

aviatrixORGANIZATION

0.99+

bostonLOCATION

0.99+

sanjmoORGANIZATION

0.99+

cto advisorORGANIZATION

0.99+

two levelsQUANTITY

0.98+

15 yearsQUANTITY

0.98+

sanjeev mohanPERSON

0.98+

21st centuryDATE

0.98+

more than one cloudQUANTITY

0.97+

uh project alpineORGANIZATION

0.96+

each oneQUANTITY

0.96+

awsORGANIZATION

0.96+

lopezORGANIZATION

0.96+

each cloudQUANTITY

0.96+

under five minutesQUANTITY

0.96+

senjiPERSON

0.96+

todayDATE

0.95+

oneQUANTITY

0.94+

first exampleQUANTITY

0.94+

firstQUANTITY

0.94+

vmwareTITLE

0.93+

bothQUANTITY

0.93+

one roomQUANTITY

0.92+

vmworldORGANIZATION

0.92+

azureTITLE

0.92+

single cloudQUANTITY

0.92+

keith townsendPERSON

0.91+

one wayQUANTITY

0.91+

googleORGANIZATION

0.9+

three different waysQUANTITY

0.89+

two separateQUANTITY

0.89+

single wayQUANTITY

0.89+

eachQUANTITY

0.88+

adobeTITLE

0.88+

each individual organizationQUANTITY

0.86+

gartnerORGANIZATION

0.86+

dellORGANIZATION

0.86+

awsTITLE

0.86+

vmwareORGANIZATION

0.85+

uhORGANIZATION

0.85+

single paneQUANTITY

0.84+

next couple yearsDATE

0.83+

single vendorQUANTITY

0.83+

a dozen organizationsQUANTITY

0.83+

floyerPERSON

0.82+

tons of codeQUANTITY

0.81+

one cloudQUANTITY

0.81+

super cloudTITLE

0.8+

maribelLOCATION

0.79+

three top technology industry analystsQUANTITY

0.78+

dell tech worldORGANIZATION

0.78+

davePERSON

0.77+

cloudsORGANIZATION

0.77+

Gil Shneorson, Dell | Dell Technologies World 2022


 

>>The cube presents. Dell technologies world brought to you by Dell. >>Welcome to Las Vegas. Lisa Martin, with Dave Volante. The cube is live at Dell technologies world 2022. Dave, hope you say live, live <laugh>. We are live. We are in person. We are three-D. We are also here on the first day of our coverage with an eight time, right? Eight time cube alum, GA Norris joins us the senior vice president of edge portfolio solutions at Dell technologies. Welcome back our friend. >>Thank you. It's great to be here in this forum with live people, you know, and 3d, >>Isn't it. We're amazing. We're not, we're not via a screen. This is actually real. So Gill a a lot, a lot of buzz, great attendance at this first event, since 20, lot's been going on since then, we're talking a lot about edge. It's not new, but there's a lot changing what's going on there. >>Well, you know, edge has been around for a while. Um, actually since, you know, the beginning of time people were doing, you know, compute and, and applications, they in the, um, in the physical space where data it, but more and more, um, data is based on sensors in cameras and machine vision. And if you wanna make real time decisions, there's a few reasons why you can't just send everything back to a data center or a cloud. Maybe you don't have the right latency, maybe, um, you it's too costly. Maybe you don't have the right end with maybe you have security challenges, maybe have compliance challenges. So the world's moving more and more resources towards where the data is created and to make real time decisions and to generate new business values, things are changing and they're becoming much more, um, um, involved than before, much more. Um, so basically that that's, what's changing. You know, we need to deal with distributed architectures much more than we needed before. >>I think one of the things we've learned in the last very dynamic two years is that access to realtime data is no longer a nice to have it's table stakes for whether we're talking about retail, healthcare, et cetera. So that the, the realtime data access is critical for everybody to these days. >>Right? And it, it could be a real time decision, or it could even be data collection either way. You need to place some device, some comput next to the source. And then, you know, you have a lot of them and you just multiply by multiple use cases and you be, you basically, you have a very complex problem to solve. And if you ask me what's new is that complexity is big coming more and more, um, critical to solve >>Critical. >>Oh, go ahead, please. >>I was just gonna say, talk to me about some of the, from a, from a complexity resolution perspective, what are some of the things that Dell is doing to help organizations as they spread out to the edge more to meet that consumer demand, but reduce that complexity from an infrastructure standpoint. >>So we focus on simplifying. I think that's what people need right now. So there are two things we do. We, we optimize our products, um, whether they need regularization or different temperature envelopes or, uh, management capability, remote management capability, and we create solutions. And so we develop, um, solutions that look at specific, um, outcomes and we size it and we create deployment guides. Um, we do everything we can, um, to simplify the, uh, the edge uses for our customers. >>You know, you guys is talking about, it's not new. I, and I know you do a lot in retail. I think of like the NCR cash register as the, the original edge, you know, but there's other use cases. Uh there's you Gil, you and I have talked about AI inferencing in, in real time, there was a question today in the analyst forum, uh, I think it went to Jeff or nobody wanted to take it. No, maybe it was Michael, but the metaverse, but that there's edge space is the edge industrial I OT. So how do you, I mean, the Tam is enormous. How do you think about the use cases? Are there ones that, that aren't necessarily sort of horizontal for you that you don't go after, like EVs and TA the cars? Or how are you thinking about >>It? Depends. I agree that the, uh, edge business is very verticalized. Um, at the same time, there are very, uh, there is, there are themes that emerge across every industry. Um, so we're trying to solve things horizontally being Dell, we need to solve for, um, repeatability and scale, but we do package, you know, vertical solutions on top of them because that's what people need. Um, so for example, you know, you said, um, NCR being the, uh, the original edge. If I asked you today, name how many applications are, are running in a retail store to enable your experience? You'd say, well, there's self checkout. Maybe there is a, um, fraud detection, >>Let's say a handful >>It's handful. The fact is it's not, it's about 30 different applications, 30 that are running. So you have, you know, digital labels and you have, you know, a curbside delivery and you have inventory management and you have crowd management and you have safety and security. And what happens today is that every one of those solar is purchased separately and deployed separately and connected to the network separately and secured separately. Hence you see the problem, right? And so I know what we do, and we create a solution. For example, we see, okay, infrastructure, what can we consolidate onto an infrastructure that could scale over time? And then we look at it in the context of a solution. So, you know, the solution we're announcing, or we announced last week does just that on the left side, it looks at a consolidated infrastructure based on VxRail and VMware stack. So you can run multiple applications on the right side, it working with a company called deep north for Inso analytics and actually people that, um, and the show they can go and see this in action, um, in our, um, you know, fake retail store, uh, back at the edge booth. Um, but the point is those elements of siloed applications and the need to consolidate their true for every industry. And that's what we're trying to solve for. >>I was just wondering, you said they're true for every industry. Every industry is facing the same challenges there. What, what makes retail so prime for transformation right now? >>That's a great question. So, you know, using my example from before, if you are faced with this set, have a shopper that buys online and they now are coming back to the stores and they need to, they want the same experience. They want the stuff that they search for. They want it available to them. Um, and in fact, we research that 80% of people say, if they have a bad experience will not come back to a retail store. So you've got all of those use cases that you need to put to, you've got this savvy shopping that comes in, you've got heightened labor costs. You've got a supply chain problem in most of those markets, labor >>Shortages as >>Well. It's a perfect storm. And you wanna give an experience, right? So CIOs are looking at this and they go, how do I do all of that? Um, and they, they, as I said before, the key management, the key problem is management of all of those things is why they can innovate faster. And so retail is in this perfect storm where they need to innovate and they want to innovate. And now they're looking for options and we're here to help them. >>You know, a lot of times we talk about the in industrial IOT, we talk about the it and the OT schism. Is there a similar sort of dissonance between it, your peeps, Dell's traditional market, and what's happening, you know, at the near edge, the retail infrastructure sort of different requirements. How are you thinking about that and managing that >>About, um, 50% of edge projects today are, are somehow involving it. Um, usually every project will involve it for networking and security, so they have to manage it either way. And today there's a lot of what we used to call shadow it. When we talked about cloud, this has happens at the edge as well. Now this happened for a good reason because the expertise are the OT people expertise on the, the specific use case. It's true for manufacturing. It's also for true for, for retail. Um, our traditional audience is the it audience and, and we will never be able to merger two worlds unless it was better able to service the OT buyers. And even in the show, I I've had multiple conversations today. We, with people to talk about the divide, how to bring it together, it will come together when it can deliver a better service to the OT, um, constituents. And that's definitely a job for Dell, right? This is what we do. If we enable our it buyer to do a better job in servicing the OT crowd or their business crowd in retail, um, more innovation will happen, you know, across the, those different dimensions. So I'm happy you asked that because that's actually part of the mission we're taking on. >>Where is one of the things I think about when you, you talk about that consumer experience and we're very demanding as consumers. We wanna ha as you described, we wanna have the same experience we expect to have that regardless of where we are. And if that doesn't happen, you, you mentioned that number of 80% of people's survey said, if I have a bad experience with a merchant, I'm out, I'm going somewhere else. Right. So where is the rest of the Csuite in the conversation? I can think of, um, a COO the chief marketing officer from brand value, brand reputation perspective. Are you talking with those folks as well to help make the connective so reality? >>Um, I, I, I don't know that we're having those conversation with those business owners. We we're a, um, a system, an infrastructure company. So, you know, we get involved once they understand, you know, what they want to do. We just look at it in. And so if you solve it one way, it's gonna be one outcome. Maybe there is a better way to look at it. Maybe there's an architecture, maybe there's a more, you know, thoughtful way to think about, you know, the problems before they happen. And, um, but the fact that they're all looking shows you, that their business owners are very, very concerned with, with this reality, their >>Key stakeholders. Can >>We come back to your announcement? Can you, can we unpack that a little bit, uh, for those who might not be familiar with it? What, what, what is it called again? And give us a peel, the onion a little bit Gil. Yeah. >>So, so we call it a Dell technologies validated design. Um, it is essentially reference architecture. Um, we take a use case, we size it. So we, you know, we, um, we save customers, the effort of, of testing and sizing. We document the deployment step by step. We just make it simpler. And as says, before we look for consolidation, so we took a VXL, which is our leading ACI product based on VMware technology with a VMware application management stack with Tansu. Um, and then we, we, we look at that as the infrastructure, and then we test it with a company called deep north and deep north, um, are, um, store analytics. So through machine vision, they can tell you where people are queuing up. If there is somebody in the store that needs help and nobody's approaching, if there is a water spill and somebody might, you know, slip and hurt themselves, if a fridge is open and something may get spot. >>And so all of those things together through machine vision and realtime decisions can have this much better experience. So we put all of this together, we created a design and now it's out there in the market for our partners to use for our customers to use. Um, this is an extension of our manufacturing solutions, where we did the same thing. We partner with a company called PTC. I know of obviously in a company called Litmos, um, to create, um, industrial and the leading solution. So this whole word of solutioning is supposed to look at the infrastructure and a use case and bring them together and document in a way that simplifies things for >>Customers. Do you ever see that becoming a Aku at some point in time or, >>Um, personal, if you ask me? I don't think so. And the reason is there's still a lot of variability in those and skewing, but that's a very formal, you know, internal discussion. Yeah. Um, the point is we are, we want people to buy as much of it as they need to, and, and we really want to help them if Aku could help them, we will get there, but we need to see repeatability before creating skews. >>Can you give us an example of a, of a retail or a manufacturing customer that's using this Dell validated design, this DVD, and that really has reduced or eliminated that complexity that was there before. >>So this solution is new. I mean, it's brand new, we just announced it. So, no, but, um, I don't know what names I can call out, cuz referenceability is probably examples though about generic, but I will tell you that most of the large retailers in the us are based in their stores on Dell technologies. Um, a lot of the trail is in, in those stores and you're talking about thousands of locations with remote management. Um, what we're doing here is we're taking it to the next step by looking at new use cases that they have not been implementing before and saying, look, same infrastructure is valid. You know, scalable is it's scalable. And here are the new use cases with machine vision and other things that here is how you do that. But we're seeing a lot of success in retail in the last few years. >>So what should we expect looking forward, you know, any gaps that customers are asking for trying to fill? What, what two to three years out, what should we expect? >>Um, I think we're gonna stay very true to our simplification message. We want to help people simplify. So if it's simplifying, um, maintenance, if it's simplifying management, if it's simplifying through solutioning, you're gonna see us more and more and more, um, investing in simplification of edge. Um, and that's through our own IP, through our partnerships. Um, there, there is a lot more coming if, if I may say it myself, but, but it's, it's a little too early to, uh, to talk about it. >>So for those folks that are here at the show that get to see it and play with it and touch it and feel it, what would you say some of the biggest impacts are that this technology can deliver tomorrow? >>Well, first of all, it's enabling to do what they want. See, we don't have to go and, and tell people, oh, you probably really need to move things through the edge. They know they need to do it. Our job is to tell them how to do it in a secure way, in a simplified way. So that's, that's a nice thing about this, this market it's happening, whether we want it or not. Um, people in this show can go see some things in action. They can see the solution in action. They can see the manufacturing solution in action and even more so. And I forgot to say part of our announcement was a set of solution centers in Limerick island and in Singapore, that was just open. And soon enough in Austin, Texas saw that, and we will have people come in and have the full experience of IOT OT and edge device devices in action. So AR and VR, I T IEN technology and scanning technology. So they could be, um, thinking about the art of the possible, right? Thinking about this immersive experience that will help them invent with us. And so we're expecting a lot of innovation to come out of those conversations for us and for them. >>So doing a lot of testing before deployment and really gleaning that testing >>Before deployment solution architecture, just ideation, if they're not there yet. So, and I've just been to Singapore in one of those, um, they asked me to, um, pretend I was a, um, retail ski enter in a distribution center and I didn't do so well, but I was still impressed with the technology. So, >>Well, eight time Q alumni. Now you have a career to fall back on if you need to. Exactly. >><laugh> >>GA it's been great to have you. Thank you so much for coming back, talking to us about what's new on day one of Dell technologies world 22. Thank >>You for having me again, >>Our pleasure for Dave Volante. I'm Lisa Martin, coming to you live from the Venetian in Las Vegas at Dell technologies world 2022. This is day one of our coverage stick around Dave and I will be right back with our next guest.

Published Date : May 3 2022

SUMMARY :

Dell technologies world brought to you by Dell. Dave, hope you say live, live <laugh>. It's great to be here in this forum with live people, you know, and 3d, a lot of buzz, great attendance at this first event, since 20, lot's been going on since then, have the right latency, maybe, um, you it's too costly. So that the, the realtime data access is critical for everybody to these days. you know, you have a lot of them and you just multiply by multiple use cases and you be, out to the edge more to meet that consumer demand, but reduce that complexity from an infrastructure standpoint. And so we develop, um, solutions that look at specific, um, outcomes and we size it and I think of like the NCR cash register as the, the original edge, you know, you know, you said, um, NCR being the, uh, the original edge. um, in our, um, you know, fake retail store, uh, back at the edge booth. I was just wondering, you said they're true for every industry. So, you know, using my example from before, if you are faced with And you wanna give an experience, right? you know, at the near edge, the retail infrastructure sort of different requirements. more innovation will happen, you know, across the, those different dimensions. We wanna ha as you described, we wanna have the same experience we expect to have that regardless And so if you solve it one way, it's gonna be one outcome. Can We come back to your announcement? So we, you know, So we put all of this together, we created a design Do you ever see that becoming a Aku at some point in time or, a lot of variability in those and skewing, but that's a very formal, you know, Can you give us an example of a, of a retail or a manufacturing customer that's using this Dell validated but I will tell you that most of the large retailers in the us are based in their stores So if it's simplifying, um, maintenance, and tell people, oh, you probably really need to move things through the edge. and I've just been to Singapore in one of those, um, they asked me to, um, pretend I was Now you have a career to fall back on if you need to. Thank you so much for coming back, talking to us about what's new on day one of Dell technologies I'm Lisa Martin, coming to you live from the Venetian

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

SingaporeLOCATION

0.99+

Dave VolantePERSON

0.99+

Gil ShneorsonPERSON

0.99+

JeffPERSON

0.99+

DavePERSON

0.99+

50%QUANTITY

0.99+

DellORGANIZATION

0.99+

twoQUANTITY

0.99+

Las VegasLOCATION

0.99+

80%QUANTITY

0.99+

last weekDATE

0.99+

three yearsQUANTITY

0.99+

MichaelPERSON

0.99+

todayDATE

0.99+

two thingsQUANTITY

0.99+

GilPERSON

0.99+

LitmosORGANIZATION

0.99+

ACIORGANIZATION

0.99+

Austin, TexasLOCATION

0.99+

oneQUANTITY

0.99+

eight timeQUANTITY

0.99+

30QUANTITY

0.99+

PTCORGANIZATION

0.98+

VenetianLOCATION

0.98+

Limerick islandLOCATION

0.98+

first dayQUANTITY

0.98+

thousandsQUANTITY

0.97+

VxRailTITLE

0.97+

tomorrowDATE

0.97+

Dell technologiesORGANIZATION

0.96+

TansuORGANIZATION

0.96+

two worldsQUANTITY

0.95+

Dell technologies world 2022EVENT

0.94+

Inso analyticsORGANIZATION

0.93+

2022DATE

0.93+

two yearsQUANTITY

0.92+

first eventQUANTITY

0.91+

Eight time cubeQUANTITY

0.91+

AkuORGANIZATION

0.9+

about 30 different applicationsQUANTITY

0.89+

day oneQUANTITY

0.88+

VMwareTITLE

0.88+

Technologies World 2022EVENT

0.87+

technologies world 22EVENT

0.84+

one wayQUANTITY

0.83+

VMwareORGANIZATION

0.83+

VXLORGANIZATION

0.81+

yearsQUANTITY

0.8+

deep northORGANIZATION

0.79+

one outcomeQUANTITY

0.77+

GAPERSON

0.74+

20QUANTITY

0.74+

GillPERSON

0.72+

NorrisPERSON

0.58+

edgeORGANIZATION

0.56+

GALOCATION

0.49+

NCRORGANIZATION

0.34+

Gian Merlino, Imply.io | AWS Startup Showcase S2 E2


 

(upbeat music) >> Hello, and welcome to theCUBE's presentation of the AWS Startup Showcase: Data as Code. This is Season 2, Episode 2 of the ongoing SaaS covering exciting startups from the AWS ecosystem and we're going to talk about the future of enterprise data analytics. I'm your host, John Furrier and today we're joined by Gian Merlino CTO and co-founder of Imply.io. Welcome to theCUBE. >> Hey, thanks for having me. >> Building analytics apps with Apache Druid and Imply is what the focus of this talk is and your company being showcased today. So thanks for coming on. You guys have been in the streaming data large scale for many, many years of pioneer going back. This past decade has been the key focus. Druid's unique position in that market has been key, you guys been empowering it. Take a minute to explain what you guys are doing over there at Imply. >> Yeah, for sure. So I guess to talk about Imply, I'll talk about Druid first. Imply is a open source based company and Apache Druid is the open source project that the Imply product's built around. So what Druid's all about is it's a database to power analytical applications. And there's a couple things I want to talk about there. The first off is, is why do we need that? And the second is why are we good at, and I'll just a little flavor of both. So why do we need database to power analytical apps? It's the same reason we need databases to power transactional apps. I mean, the requirements of these applications are different analytical applications, apps where you have tons of data coming in, you have lots of different people wanting to interact with that data, see what's happening both real time and historical. The requirements of that kind of application have sort of given rise to a new kind of database that Druid is one example of. There's others, of course, out there in both the open source and non open source world. And what makes Druid really good at it is, people often say what is Druid's big secret? How is it so good? Why is it so fast? And I never know what to say to that. I always sort of go to, well it's just getting all the little details right. It's a lot of pieces that individually need to be engineered, you build up software in layers, you build up a database in layers, just like any other piece of software. And to have really high performance and to do really well at a specific purpose, you kind of have to get each layer right and have each layer have as little overhead as possible. And so just a lot of kind of nitty gritty engineering work. >> What's interesting about the trends over the past 10 years in particular, maybe you can go back 10, 15 years is state of the art database was, stream a bunch of data put it into a pile, index it, interrogate it, get some reports, pretty basic stuff and then all of a sudden now you have with cloud, thousands of databases out there, potentially hundreds of databases living in the wild. So now data with Kafka and Kinesis, these kinds of technologies streaming data's happening in real time so you don't have time to put it in a pile or index it. You want real time analytics. And so perhaps whether they're mobile app, Instagrams of the world, this is now what people want in the enterprise. You guys are the heart of this. Can you talk about that dynamic of getting data quickly at scale? >> So our thinking is that actually both things matter. Realtime data matters but also historical context matters. And the best way to get historical context out of data is to put it in a pile, index it, so to speak, and then the best way to get realtime context to what's happening right now is to be able to operate on these streams. And so one of the things that we do in Druid, I wish I had more time to talk about it but one of the things that we do in Druid is we kind of integrate this real time processing and this historical processing. So we actually have a system that we call the historical system that does what you're saying, take all this data, put in a pile, index it for all your historical data. And we have a system that we call the realtime system that is pulling data in from things like Kafka, Kinesis, getting data pushed into it as the case may be. And this system is responsible for all the data that's recent, maybe the last hour or two of data will be handled by this system and then the older stuff handled by historical system. And our query layer blends these two together seamlessly so a user never needs to think about whether they're querying realtime data or historical data. It's presented as a blended view. >> It's interesting and you know a lot of the people just say, Hey, I don't really have the expertise, and now they're trying to learn it so their default was throw into a data lake. So that brings back that historical. So the rise of the data lake, you're seeing Databricks and others out there doing very well with the data lakes. How do you guys fit into that 'cause that makes it a lot of sense too cause that looks like historical information? >> So data lakes are great technology. We love that kind of stuff. I would say that a really popular pattern, with Druid there's actually two very popular patterns. One is, I would say streaming forward. So stream focus where you connect up to something like Kafka and you load data to stream and then we will actually take that data, we'll store all the historical data that came from the stream and instead of blend those two together. And another other pattern that's also very common is the data lake pattern. So you have a data lake and then you're sort of mirroring that data from the data lake into Druid. This is really common when you have a data lake that you want to be able to build an application on top of, you want to say I have this data in the data lake, I have my table, I want to build an application that has hundreds of people using it, that has really fast response time, that is always online. And so when I mirror that data into Druid and then build my app on top of that. >> Gian take me through the progression of the maturity cycle here. As you look back even a few years, the pioneers and the hardcore streaming data using data analytics at scale that you guys are doing with Druid was really a few percentage of the population doing that. And then as the hyperscale became mainstream, it's now in the enterprise, how stable is it? What's the current state of the art relative to the stability and adoption of the techniques that you guys are seeing? >> I think what we're seeing right now at this stage in the game, and this is something that we kind of see at the commercial side of Imply, what we're seeing at this stage of the game is that these kinds of realization that you actually can get a lot of value out of data by building interactive apps around it and by allowing people to kind of slice and dice it and play with it and just kind of getting out there to everybody, that there is a lot of value here and that it is actually very feasible to do with current technology. So I've been working on this problem, just in my own career for the past decade, 10 years ago where we were is even the most high tech of tech companies were like, well, I could sort of see the value. It seems like it might be difficult. And we're kind of getting from there to the high tech companies realizing that it is valuable and it is very doable. And I think that was something there was a tipping point that I saw a few years ago when these Druid and database like really started to blow up. And I think now we're seeing that beyond sort of the high tech companies, which is great to see. >> And a lot of people see the value of the data and they see the application as data as code means the application developers really want to have that functionality. Can you share the roadmap for the next 12 months for you guys on what's coming next? What's coming around the corner? >> Yeah, for sure. I mentioned during the Apache open source community, different products we're one member of that community, very prominent one but one member so I'll talk a bit about what we're doing for the Druid project as part of our effort to make Druid better and take it to the next level. And then I'll talk about some of the stuff we're doing on the, I guess, the Druid sort of commercial side. So on the Druid side, stuff that we're doing to make Druid better, take it to the next level, the big thing is something that we really started writing about a few weeks ago, the multi-stage query engine that we're working on, a new multi-stage query engine. If you're interested, the full details are on blog on our website and also on GitHub on Apache Druid GitHub, but short version is Druid's. We're sort of extending Druid's Query engine to support more and varied kinds of queries with a focus on sort of reporting queries, more complex queries. Druid's core query engine has classically been extremely good at doing rapid fire queries very quickly, so think thousands of queries per second where each query is maybe something that involves a filter in a group eye like a relatively straightforward query but we're just doing thousands of them constantly. Historically folks have not reached for technologies like Druid is, really complex and a thousand line sequel queries, complex supporting needs. Although people really do need to do both interactive stuff and complex stuff on the same dataset and so that's why we're building out these capabilities in Druid. And then on the implied commercial side, the big effort for this year is Polaris which is our cloud based Druid offering. >> Talk about the relationship between Druid and Imply? Share with the folks out there how that works. >> So Druid is, like I mentioned before, it's Apache Druid so it's a community based project. It's not a project that is owned by Imply, some open source projects are sort of owned or sponsored by a particular organization. Druid is not, Druid is an independent project. Imply is the biggest contributor to Druid. So the imply engineering team is contributing tons of stuff constantly and we're really putting a lot of the work in to improve Druid although it is a community effort. >> You guys are launching a new SaaS service on AWS. Can you tell me about what that's happening there, what it's all about? >> Yeah, so we actually launched that a couple weeks ago. It's called Polaris. It's very cool. So historically there's been two ways, you can either get started with Apache Druid, it's open source, you install it yourself, or you can get started with Imply Enterprise which is our enterprise offering. And these are the two ways you can get started historically. One of the issues of getting started with Apache Druid is that it is a very complicated distributed database. It's simple enough to run on a single server but once you want to scale things out, once you get all these things set up, you may want someone to take some of that operational burden off your hands. And on the Imply Enterprise side, it says right there in the name, it's enterprise product. It's something that may take a little bit of time to get started with. It's not something you can just roll up with a credit card and sign up for. So Polaris is really about of having a cloud product that's sort of designed to be really easy to get started with, really self-service that kind of stuff. So kind of providing a really nice getting started experience that does take that maintenance burden and operational burden away from you but is also sort of as easy to get started with as something that's database would be. >> So a more developer friendly than from an onboarding standpoint, classic. >> Exactly. Much more developer friendly is what we're going for with that product. >> So take me through the state of the art data as code in your mind 'cause infrastructure is code, DevOps has been awesome, that's cloud scale, we've seen that. Data as Code is a term we coined but means data's in the developer process. How do you see data being integrated into the workflow for developers in the future? >> Great question. I mean all kinds of ways. Part of the reason that, I kind of alluded to this earlier, building analytical applications, building applications based on data and based on letting people do analysis, how valuable it is and I guess to develop in that context there's kind of two big ways that we sort of see these things getting pushed out. One is developers building apps for other people to use. So think like, I want to build something like Google analytics, I want to build something that clicks my web traffic and then lets the marketing team slice and dice through it and make decisions about how well the marketing's doing. You can build something like that with databases like Druid and products like what we're having in Imply. I guess the other way is things that are actually helping developers do their own job. So kind of like use your own product or use it for yourself. And in this world, you kind of have things like... So going beyond what I think my favorite use case, I'll just talk about one. My favorite use case is so I'm really into performance, I spend the last 10 years of my life working on high performance database so obviously I'm into this kind of stuff. I love when people use our product to help make their own products faster. So this concept of performance monitoring and performance management for applications. One thing that I've seen some of our customers do and some of our users do that I really love is when you kind of take that performance data of your own app, as far as it can possibly go take it to the next level. I think the basic level of using performance data is I collect performance data from my application deployed out there in the world and I can just use it for monitoring. I can say, okay my response times are getting high in this region, maybe there's something wrong with that region. One of the very original use cases for Druid was that Netflix doing performance analysis, performance analysis more exciting than monitoring because you're not just understanding that there's a performance, is good or bad in whatever region sort of getting very fine grain. You're saying in this region, on this server rack for these devices, I'm seeing a degradation or I'm seeing a increase. You can see things like Apple just rolled out a new version of iOS and on that new version of iOS, my app is performing worse than the older version. And even though not many devices are on that new version yet I can kind of see that because I have the ability to get really deep in the data and then I can start slicing nice that more. I can say for those new iOS people, is it all iOS devices? Is it just the iPhone? Is it just the iPad? And that kind of stuff is just one example but it's an example that I really like. >> It's kind of like the data about the data was always good to have context, you're like data analytics for data analytics to see how it's working at scale. This is interesting because now you're bringing up the classic finding the needle in the haystack of needles, so to speak where you have so much data out there like edge cases, edge computing, for instance, you have devices sending data off. There's so much data coming in, the scale is a big issue. This is kind of where you guys seem to be a nice fit for, large scale data ingestion, large scaled data management, large scale data insights kind of all rolled in to one. Is that kind of-? >> Yeah, for sure. One of the things that we knew we had to do with Druid was we were building it for the sort of internet age and so we knew it had to scale well. So the original use case for Druid, the very first one that we ended up building for, the reason we build in the first place is because that original use case had massive scale and we struggled finding something, we were literally trying to do what we see people doing now which is we're trying to build an app on a massive data set and we're struggling to do it. And so we knew it had to scale to massive data sets. And so that's a little flavor of kind know how that works is, like I was mentioning earlier this, this realtime system and historical system, the realtime system is scalable, it's scalable out if you're reading from Kafka, we scale out just like any other Kafka consumer. And then the historical system is all based on what we call segments which are these files that has a few million rows per file. And a cluster is really big, might have thousands of servers, millions of segments, but it's a design that is kind of, it's a design that does scale to these multi-trillion road tables. >> It's interesting, you go back when you probably started, you had Twitter, Netflix, Facebook, I mean a handful of companies that were at the scale. Now, the trend is you're on this wave where those hyperscalers and, or these unique huge scale app companies are now mainstream enterprise. So as you guys roll out the enterprise version of building analytics and applications, which Druid and Imply, they got to going to get religion on this. And I think it's not hard because it's distributed computing which they're used to. So how is that enterprise transition going because I can imagine people would want it and are just kicking the tires or learning and then trying to put it into action. How are you seeing the adoption of the enterprise piece of it? >> The thing that's driving the interest is for sure doing more and more stuff on the internet because anything that happens on the internet whether it's apps or web based, there's more and more happening there and anything that is connected to the internet, anything that's serving customers on the internet, it's going to generate an absolute mountain of data. And the only question is not if you're going to have that much data, you do if you're doing anything on the internet, the only question is what are you going to do with it? So that's I think what drives the interest, is people want to try to get value out of this. And then what drives the actual adoption is I think, I don't want to necessarily talk about specific folks but within every industry I would say there's people that are leaders, there's organizations that are leaders, teams that are leaders, what drives a lot of interest is seeing someone in your own industry that has adopted new technology and has gotten a lot of value out of it. So a big part of what we do at Imply is that identify those leaders, work with them and then you can talk about how it's helped them in their business. And then also I guess the classic enterprise thing, what they're looking for is a sense of stability, a sense of supportability, a sense of robustness and this is something that comes with maturity. I think that the super high tech companies are comfortable using some open source software that's rolled off the presses a few months ago; he big enterprises are looking for something that has corporate backing, they're looking for something that's been around for a while and I think that Druid technologies like it are breaching that little maturity right now. >> It's interesting that supply chain has come up in the software side. That conversation is a lot now, you're hearing about open source being great, but in the cloud scale, you can get the data in there to identify opportunities and also potentially vulnerabilities is big discussion. Question for you on the cloud native side, how do you see cloud native, cloud scale with services like serverless Lambda, edge merging, it's easier to get into the cloud scale. How do you see the enterprise being hardened out with Druid and Imply? >> I think the cloud stuff is great, we love using it to build all of our own stuff, our product is of course built on other cloud technologies and I think these technologies built on each other, you sort of have like I mentioned earlier, all software is built in layers and cloud architecture is the same thing. What we see ourselves as doing is we're building the next layer of that stack. So we're building the analytics database layer. You saw when people first started doing these in public cloud, the very first two services that came out you can get a virtual machine and you can store some data and you can retrieve that data but there's no real analytics on it, there's just kind of storage and retrieval. And then as time goes on higher and higher levels get built out delivering more and more value and then the levels mature as they go up. And so the the bottom of layers are incredibly mature, the top most layers are cutting edge and there's a kind of a maturity gradient between those two. And so what we're doing is we're building out one of those layers. >> Awesome extraction layers, faster performance, great stuff. Final question for you, Gian, what's your vision for the future? How do you Imply and Druid it going? What's it look like five years from now? >> I think that for sure it seems like that there's two big trends that are happening in the world and it's going to sound a little bit self serving for me to say it but I believe what we're doing here says, I'm here 'cause I believe it, I believe in open source and I believe in cloud stuff. That's why I'm really excited that what we're doing is we're building a great cloud product based on a great open source project. I think that's the kind of company that I would want to buy from if I wasn't at this company and I was just building something, I would want to buy a great cloud product that's backed by a great open source project. So I think the kind of the way I see the industry going, the way I see us going and I think would be a great place to end up just kind of as an engineering world, as an industry is a lot of these really great open source projects doing things like what Kubernetes doing containers, we're doing with analytics et cetera. And then really first class really well done cloud versions of each one of them and so you can kind of choose, do you want to get down and dirty with the open source or do you want to choose just kind of have the abstraction of the cloud. >> That's awesome. Cloud scale, cloud flexibility, community getting down and dirty open source, the best of both worlds. Great solution. Goin, thanks for coming on and thanks for sharing here in the Showcase. Thanks for coming on theCUBE. >> Thank you too. >> Okay, this is theCUBE Showcase Season 2, Episode 2. I'm John Furrier, your host. Data as Code is the theme of this episode. Thanks for watching. (upbeat music)

Published Date : Apr 26 2022

SUMMARY :

of the AWS Startup Showcase: Data as Code. Take a minute to explain what you guys are And the second is why are we good at, Instagrams of the world, And so one of the things know a lot of the people data that came from the of the art relative to the that beyond sort of the the next 12 months for you So on the Druid side, Talk about the relationship Imply is the biggest contributor to Druid. Can you tell me about what And on the Imply Enterprise side, So a more developer friendly than from we're going for with that product. means data's in the developer process. I have the ability to get It's kind of like the One of the things that of the enterprise piece of it? I guess the classic enterprise thing, but in the cloud scale, And so the the bottom of How do you Imply and Druid it going? and so you can kind of choose, here in the Showcase. Data as Code is the theme of this episode.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
John FurrierPERSON

0.99+

Gian MerlinoPERSON

0.99+

AWSORGANIZATION

0.99+

AppleORGANIZATION

0.99+

two waysQUANTITY

0.99+

iOSTITLE

0.99+

NetflixORGANIZATION

0.99+

10QUANTITY

0.99+

each layerQUANTITY

0.99+

iPhoneCOMMERCIAL_ITEM

0.99+

millionsQUANTITY

0.99+

DruidTITLE

0.99+

iPadCOMMERCIAL_ITEM

0.99+

firstQUANTITY

0.99+

OneQUANTITY

0.99+

secondQUANTITY

0.99+

thousandsQUANTITY

0.99+

twoQUANTITY

0.99+

ImplyORGANIZATION

0.99+

bothQUANTITY

0.99+

oneQUANTITY

0.99+

each queryQUANTITY

0.99+

theCUBEORGANIZATION

0.98+

TwitterORGANIZATION

0.98+

FacebookORGANIZATION

0.98+

GianPERSON

0.98+

KafkaTITLE

0.98+

Imply.ioORGANIZATION

0.97+

one exampleQUANTITY

0.97+

first two servicesQUANTITY

0.97+

hundreds of peopleQUANTITY

0.97+

each oneQUANTITY

0.97+

two big waysQUANTITY

0.97+

10 years agoDATE

0.96+

past decadeDATE

0.96+

first classQUANTITY

0.96+

one memberQUANTITY

0.96+

LambdaTITLE

0.96+

two big trendsQUANTITY

0.96+

ApacheORGANIZATION

0.95+

both worldsQUANTITY

0.95+

PolarisORGANIZATION

0.95+

one memberQUANTITY

0.95+

todayDATE

0.95+

Breaking Analysis: Technology & Architectural Considerations for Data Mesh


 

>> From theCUBE Studios in Palo Alto and Boston, bringing you data driven insights from theCUBE in ETR, this is Breaking Analysis with Dave Vellante. >> The introduction in socialization of data mesh has caused practitioners, business technology executives, and technologists to pause, and ask some probing questions about the organization of their data teams, their data strategies, future investments, and their current architectural approaches. Some in the technology community have embraced the concept, others have twisted the definition, while still others remain oblivious to the momentum building around data mesh. Here we are in the early days of data mesh adoption. Organizations that have taken the plunge will tell you that aligning stakeholders is a non-trivial effort, but necessary to break through the limitations that monolithic data architectures and highly specialized teams have imposed over frustrated business and domain leaders. However, practical data mesh examples often lie in the eyes of the implementer, and may not strictly adhere to the principles of data mesh. Now, part of the problem is lack of open technologies and standards that can accelerate adoption and reduce friction, and that's what we're going to talk about today. Some of the key technology and architecture questions around data mesh. Hello, and welcome to this week's Wikibon CUBE Insights powered by ETR, and in this Breaking Analysis, we welcome back the founder of data mesh and director of Emerging Technologies at Thoughtworks, Zhamak Dehghani. Hello, Zhamak. Thanks for being here today. >> Hi Dave, thank you for having me back. It's always a delight to connect and have a conversation. Thank you. >> Great, looking forward to it. Okay, so before we get into it in the technology details, I just want to quickly share some data from our friends at ETR. You know, despite the importance of data initiative since the pandemic, CIOs and IT organizations have had to juggle of course, a few other priorities, this is why in the survey data, cyber and cloud computing are rated as two most important priorities. Analytics and machine learning, and AI, which are kind of data topics, still make the top of the list, well ahead of many other categories. And look, a sound data architecture and strategy is fundamental to digital transformations, and much of the past two years, as we've often said, has been like a forced march into digital. So while organizations are moving forward, they really have to think hard about the data architecture decisions that they make, because it's going to impact them, Zhamak, for years to come, isn't it? >> Yes, absolutely. I mean, we are moving really from, slowly moving from reason based logical algorithmic to model based computation and decision making, where we exploit the patterns and signals within the data. So data becomes a very important ingredient, of not only decision making, and analytics and discovering trends, but also the features and applications that we build for the future. So we can't really ignore it, and as we see, some of the existing challenges around getting value from data is not necessarily that no longer is access to computation, is actually access to trustworthy, reliable data at scale. >> Yeah, and you see these domains coming together with the cloud and obviously it has to be secure and trusted, and that's why we're here today talking about data mesh. So let's get into it. Zhamak, first, your new book is out, 'Data Mesh: Delivering Data-Driven Value at Scale' just recently published, so congratulations on getting that done, awesome. Now in a recent presentation, you pulled excerpts from the book and we're going to talk through some of the technology and architectural considerations. Just quickly for the audience, four principles of data mesh. Domain driven ownership, data as product, self-served data platform and federated computational governance. So I want to start with self-serve platform and some of the data that you shared recently. You say that, "Data mesh serves autonomous domain oriented teams versus existing platforms, which serve a centralized team." Can you elaborate? >> Sure. I mean the role of the platform is to lower the cognitive load for domain teams, for people who are focusing on the business outcomes, the technologies that are building the applications, to really lower the cognitive load for them, to be able to work with data. Whether they are building analytics, automated decision making, intelligent modeling. They need to be able to get access to data and use it. So the role of the platform, I guess, just stepping back for a moment is to empower and enable these teams. Data mesh by definition is a scale out model. It's a decentralized model that wants to give autonomy to cross-functional teams. So it is core requires a set of tools that work really well in that decentralized model. When we look at the existing platforms, they try to achieve this similar outcome, right? Lower the cognitive load, give the tools to data practitioners, to manage data at scale because today centralized teams, really their job, the centralized data teams, their job isn't really directly aligned with a one or two or different, you know, business units and business outcomes in terms of getting value from data. Their job is manage the data and make the data available for then those cross-functional teams or business units to use the data. So the platforms they've been given are really centralized around or tuned to work with this structure as a team, structure of centralized team. Although on the surface, it seems that why not? Why can't I use my, you know, cloud storage or computation or data warehouse in a decentralized way? You should be able to, but some changes need to happen to those online platforms. As an example, some cloud providers simply have hard limits on the number of like account storage, storage accounts that you can have. Because they never envisaged you have hundreds of lakes. They envisage one or two, maybe 10 lakes, right. They envisage really centralizing data, not decentralizing data. So I think we see a shift in thinking about enabling autonomous independent teams versus a centralized team. >> So just a follow up if I may, we could be here for a while. But so this assumes that you've sorted out the organizational considerations? That you've defined all the, what a data product is and a sub product. And people will say, of course we use the term monolithic as a pejorative, let's face it. But the data warehouse crowd will say, "Well, that's what data march did. So we got that covered." But Europe... The primest of data mesh, if I understand it is whether it's a data march or a data mart or a data warehouse, or a data lake or whatever, a snowflake warehouse, it's a node on the mesh. Okay. So don't build your organization around the technology, let the technology serve the organization is that-- >> That's a perfect way of putting it, exactly. I mean, for a very long time, when we look at decomposition of complexity, we've looked at decomposition of complexity around technology, right? So we have technology and that's maybe a good segue to actually the next item on that list that we looked at. Oh, I need to decompose based on whether I want to have access to raw data and put it on the lake. Whether I want to have access to model data and put it on the warehouse. You know I need to have a team in the middle to move the data around. And then try to figure organization into that model. So data mesh really inverses that, and as you said, is look at the organizational structure first. Then scale boundaries around which your organization and operation can scale. And then the second layer look at the technology and how you decompose it. >> Okay. So let's go to that next point and talk about how you serve and manage autonomous interoperable data products. Where code, data policy you say is treated as one unit. Whereas your contention is existing platforms of course have independent management and dashboards for catalogs or storage, et cetera. Maybe we double click on that a bit. >> Yeah. So if you think about that functional, or technical decomposition, right? Of concerns, that's one way, that's a very valid way of decomposing, complexity and concerns. And then build solutions, independent solutions to address them. That's what we see in the technology landscape today. We will see technologies that are taking care of your management of data, bring your data under some sort of a control and modeling. You'll see technology that moves that data around, will perform various transformations and computations on it. And then you see technology that tries to overlay some level of meaning. Metadata, understandability, discovery was the end policy, right? So that's where your data processing kind of pipeline technologies versus data warehouse, storage, lake technologies, and then the governance come to play. And over time, we decomposed and we compose, right? Deconstruct and reconstruct back this together. But, right now that's where we stand. I think for data mesh really to become a reality, as in independent sources of data and teams can responsibly share data in a way that can be understood right then and there can impose policies, right then when the data gets accessed in that source and in a resilient manner, like in a way that data changes structure of the data or changes to the scheme of the data, doesn't have those downstream down times. We've got to think about this new nucleus or new units of data sharing. And we need to really bring back transformation and governing data and the data itself together around these decentralized nodes on the mesh. So that's another, I guess, deconstruction and reconstruction that needs to happen around the technology to formulate ourselves around the domains. And again the data and the logic of the data itself, the meaning of the data itself. >> Great. Got it. And we're going to talk more about the importance of data sharing and the implications. But the third point deals with how operational, analytical technologies are constructed. You've got an app DevStack, you've got a data stack. You've made the point many times actually that we've contextualized our operational systems, but not our data systems, they remain separate. Maybe you could elaborate on this point. >> Yes. I think this is, again, has a historical background and beginning. For a really long time, applications have dealt with features and the logic of running the business and encapsulating the data and the state that they need to run that feature or run that business function. And then we had for anything analytical driven, which required access data across these applications and across the longer dimension of time around different subjects within the organization. This analytical data, we had made a decision that, "Okay, let's leave those applications aside. Let's leave those databases aside. We'll extract the data out and we'll load it, or we'll transform it and put it under the analytical kind of a data stack and then downstream from it, we will have analytical data users, the data analysts, the data sciences and the, you know, the portfolio of users that are growing use that data stack. And that led to this really separation of dual stack with point to point integration. So applications went down the path of transactional databases or urban document store, but using APIs for communicating and then we've gone to, you know, lake storage or data warehouse on the other side. If we are moving and that again, enforces the silo of data versus app, right? So if we are moving to the world that our missions that are ambitions around making applications, more intelligent. Making them data driven. These two worlds need to come closer. As in ML Analytics gets embedded into those app applications themselves. And the data sharing, as a very essential ingredient of that, gets embedded and gets closer, becomes closer to those applications. So, if you are looking at this now cross-functional, app data, based team, right? Business team, then the technology stacks can't be so segregated, right? There has to be a continuum of experience from app delivery, to sharing of the data, to using that data, to embed models back into those applications. And that continuum of experience requires well integrated technologies. I'll give you an example, which actually in some sense, we are somewhat moving to that direction. But if we are talking about data sharing or data modeling and applications use one set of APIs, you know, HTTP compliant, GraQL or RAC APIs. And on the other hand, you have proprietary SQL, like connect to my database and run SQL. Like those are very two different models of representing and accessing data. So we kind of have to harmonize or integrate those two worlds a bit more closely to achieve that domain oriented cross-functional teams. >> Yeah. We are going to talk about some of the gaps later and actually you look at them as opportunities, more than barriers. But they are barriers, but they're opportunities for more innovation. Let's go on to the fourth one. The next point, it deals with the roles that the platform serves. Data mesh proposes that domain experts own the data and take responsibility for it end to end and are served by the technology. Kind of, we referenced that before. Whereas your contention is that today, data systems are really designed for specialists. I think you use the term hyper specialists a lot. I love that term. And the generalist are kind of passive bystanders waiting in line for the technical teams to serve them. >> Yes. I mean, if you think about the, again, the intention behind data mesh was creating a responsible data sharing model that scales out. And I challenge any organization that has a scaled ambitions around data or usage of data that relies on small pockets of very expensive specialists resources, right? So we have no choice, but upscaling cross-scaling. The majority population of our technologists, we often call them generalists, right? That's a short hand for people that can really move from one technology to another technology. Sometimes we call them pandric people sometimes we call them T-shaped people. But regardless, like we need to have ability to really mobilize our generalists. And we had to do that at Thoughtworks. We serve a lot of our clients and like many other organizations, we are also challenged with hiring specialists. So we have tested the model of having a few specialists, really conveying and translating the knowledge to generalists and bring them forward. And of course, platform is a big enabler of that. Like what is the language of using the technology? What are the APIs that delight that generalist experience? This doesn't mean no code, low code. We have to throw away in to good engineering practices. And I think good software engineering practices remain to exist. Of course, they get adopted to the world of data to build resilient you know, sustainable solutions, but specialty, especially around kind of proprietary technology is going to be a hard one to scale. >> Okay. I'm definitely going to come back and pick your brain on that one. And, you know, your point about scale out in the examples, the practical examples of companies that have implemented data mesh that I've talked to. I think in all cases, you know, there's only a handful that I've really gone deep with, but it was their hadoop instances, their clusters wouldn't scale, they couldn't scale the business and around it. So that's really a key point of a common pattern that we've seen now. I think in all cases, they went to like the data lake model and AWS. And so that maybe has some violation of the principles, but we'll come back to that. But so let me go on to the next one. Of course, data mesh leans heavily, toward this concept of decentralization, to support domain ownership over the centralized approaches. And we certainly see this, the public cloud players, database companies as key actors here with very large install bases, pushing a centralized approach. So I guess my question is, how realistic is this next point where you have decentralized technologies ruling the roost? >> I think if you look at the history of places, in our industry where decentralization has succeeded, they heavily relied on standardization of connectivity with, you know, across different components of technology. And I think right now you are right. The way we get value from data relies on collection. At the end of the day, collection of data. Whether you have a deep learning machinery model that you're training, or you have, you know, reports to generate. Regardless, the model is bring your data to a place that you can collect it, so that we can use it. And that leads to a naturally set of technologies that try to operate as a full stack integrated proprietary with no intention of, you know, opening, data for sharing. Now, conversely, if you think about internet itself, web itself, microservices, even at the enterprise level, not at the planetary level, they succeeded as decentralized technologies to a large degree because of their emphasis on open net and openness and sharing, right. API sharing. We don't talk about, in the API worlds, like we don't say, you know, "I will build a platform to manage your logical applications." Maybe to a degree but we actually moved away from that. We say, "I'll build a platform that opens around applications to manage your APIs, manage your interfaces." Right? Give you access to API. So I think the shift needs to... That definition of decentralized there means really composable, open pieces of the technology that can play nicely with each other, rather than a full stack, all have control of your data yet being somewhat decentralized within the boundary of my platform. That's just simply not going to scale if data needs to come from different platforms, different locations, different geographical locations, it needs to rethink. >> Okay, thank you. And then the final point is, is data mesh favors technologies that are domain agnostic versus those that are domain aware. And I wonder if you could help me square the circle cause it's nuanced and I'm kind of a 100 level student of your work. But you have said for example, that the data teams lack context of the domain and so help us understand what you mean here in this case. >> Sure. Absolutely. So as you said, we want to take... Data mesh tries to give autonomy and decision making power and responsibility to people that have the context of those domains, right? The people that are really familiar with different business domains and naturally the data that that domain needs, or that naturally the data that domains shares. So if the intention of the platform is really to give the power to people with most relevant and timely context, the platform itself naturally becomes as a shared component, becomes domain agnostic to a large degree. Of course those domains can still... The platform is a (chuckles) fairly overloaded world. As in, if you think about it as a set of technology that abstracts complexity and allows building the next level solutions on top, those domains may have their own set of platforms that are very much doing agnostic. But as a generalized shareable set of technologies or tools that allows us share data. So that piece of technology needs to relinquish the knowledge of the context to the domain teams and actually becomes domain agnostic. >> Got it. Okay. Makes sense. All right. Let's shift gears here. Talk about some of the gaps and some of the standards that are needed. You and I have talked about this a little bit before, but this digs deeper. What types of standards are needed? Maybe you could walk us through this graphic, please. >> Sure. So what I'm trying to depict here is that if we imagine a world that data can be shared from many different locations, for a variety of analytical use cases, naturally the boundary of what we call a node on the mesh will encapsulates internally a fair few pieces. It's not just the boundary of that, not on the mesh, is the data itself that it's controlling and updating and maintaining. It's of course a computation and the code that's responsible for that data. And then the policies that continue to govern that data as long as that data exists. So if that's the boundary, then if we shift that focus from implementation details, that we can leave that for later, what becomes really important is the scene or the APIs and interfaces that this node exposes. And I think that's where the work that needs to be done and the standards that are missing. And we want the scene and those interfaces be open because that allows, you know, different organizations with different boundaries of trust to share data. Not only to share data to kind of move that data to yes, another location, to share the data in a way that distributed workloads, distributed analytics, distributed machine learning model can happen on the data where it is. So if you follow that line of thinking around the centralization and connection of data versus collection of data, I think the very, very important piece of it that needs really deep thinking, and I don't claim that I have done that, is how do we share data responsibly and sustainably, right? That is not brittle. If you think about it today, the ways we share data, one of the very common ways is around, I'll give you a JDC endpoint, or I give you an endpoint to your, you know, database of choice. And now as technology, whereas a user actually, you can now have access to the schema of the underlying data and then run various queries or SQL queries on it. That's very simple and easy to get started with. That's why SQL is an evergreen, you know, standard or semi standard, pseudo standard that we all use. But it's also very brittle, because we are dependent on a underlying schema and formatting of the data that's been designed to tell the computer how to store and manage the data. So I think that the data sharing APIs of the future really need to think about removing this brittle dependencies, think about sharing, not only the data, but what we call metadata, I suppose. Additional set of characteristics that is always shared along with data to make the data usage, I suppose ethical and also friendly for the users and also, I think we have to... That data sharing API, the other element of it, is to allow kind of computation to run where the data exists. So if you think about SQL again, as a simple primitive example of computation, when we select and when we filter and when we join, the computation is happening on that data. So maybe there is a next level of articulating, distributed computational data that simply trains models, right? Your language primitives change in a way to allow sophisticated analytical workloads run on the data more responsibly with policies and access control and force. So I think that output port that I mentioned simply is about next generation data sharing, responsible data sharing APIs. Suitable for decentralized analytical workloads. >> So I'm not trying to bait you here, but I have a follow up as well. So you schema, for all its good creates constraints. No schema on right, that didn't work, cause it was just a free for all and it created the data swamps. But now you have technology companies trying to solve that problem. Take Snowflake for example, you know, enabling, data sharing. But it is within its proprietary environment. Certainly Databricks doing something, you know, trying to come at it from its angle, bringing some of the best to data warehouse, with the data science. Is your contention that those remain sort of proprietary and defacto standards? And then what we need is more open standards? Maybe you could comment. >> Sure. I think the two points one is, as you mentioned. Open standards that allow... Actually make the underlying platform invisible. I mean my litmus test for a technology provider to say, "I'm a data mesh," (laughs) kind of compliant is, "Is your platform invisible?" As in, can I replace it with another and yet get the similar data sharing experience that I need? So part of it is that. Part of it is open standards, they're not really proprietary. The other angle for kind of sharing data across different platforms so that you know, we don't get stuck with one technology or another is around APIs. It is around code that is protecting that internal schema. So where we are on the curve of evolution of technology, right now we are exposing the internal structure of the data. That is designed to optimize certain modes of access. We're exposing that to the end client and application APIs, right? So the APIs that use the data today are very much aware that this database was optimized for machine learning workloads. Hence you will deal with a columnar storage of the file versus this other API is optimized for a very different, report type access, relational access and is optimized around roles. I think that should become irrelevant in the API sharing of the future. Because as a user, I shouldn't care how this data is internally optimized, right? The language primitive that I'm using should be really agnostic to the machine optimization underneath that. And if we did that, perhaps this war between warehouse or lake or the other will become actually irrelevant. So we're optimizing for that human best human experience, as opposed to the best machine experience. We still have to do that but we have to make that invisible. Make that an implementation concern. So that's another angle of what should... If we daydream together, the best experience and resilient experience in terms of data usage than these APIs with diagnostics to the internal storage structure. >> Great, thank you for that. We've wrapped our ankles now on the controversy, so we might as well wade all the way in, I can't let you go without addressing some of this. Which you've catalyzed, which I, by the way, I see as a sign of progress. So this gentleman, Paul Andrew is an architect and he gave a presentation I think last night. And he teased it as quote, "The theory from Zhamak Dehghani versus the practical experience of a technical architect, AKA me," meaning him. And Zhamak, you were quick to shoot back that data mesh is not theory, it's based on practice. And some practices are experimental. Some are more baked and data mesh really avoids by design, the specificity of vendor or technology. Perhaps you intend to frame your post as a technology or vendor specific, specific implementation. So touche, that was excellent. (Zhamak laughs) Now you don't need me to defend you, but I will anyway. You spent 14 plus years as a software engineer and the better part of a decade consulting with some of the most technically advanced companies in the world. But I'm going to push you a little bit here and say, some of this tension is of your own making because you purposefully don't talk about technologies and vendors. Sometimes doing so it's instructive for us neophytes. So, why don't you ever like use specific examples of technology for frames of reference? >> Yes. My role is pushes to the next level. So, you know everybody picks their fights, pick their battles. My role in this battle is to push us to think beyond what's available today. Of course, that's my public persona. On a day to day basis, actually I work with clients and existing technology and I think at Thoughtworks we have given the talk we gave a case study talk with a colleague of mine and I intentionally got him to talk about (indistinct) I want to talk about the technology that we use to implement data mesh. And the reason I haven't really embraced, in my conversations, the specific technology. One is, I feel the technology solutions we're using today are still not ready for the vision. I mean, we have to be in this transitional step, no matter what we have to be pragmatic, of course, and practical, I suppose. And use the existing vendors that exist and I wholeheartedly embrace that, but that's just not my role, to show that. I've gone through this transformation once before in my life. When microservices happened, we were building microservices like architectures with technology that wasn't ready for it. Big application, web application servers that were designed to run these giant monolithic applications. And now we're trying to run little microservices onto them. And the tail was riding the dock, the environmental complexity of running these services was consuming so much of our effort that we couldn't really pay attention to that business logic, the business value. And that's where we are today. The complexity of integrating existing technologies is really overwhelmingly, capturing a lot of our attention and cost and effort, money and effort as opposed to really focusing on the data product themselves. So it's just that's the role I have, but it doesn't mean that, you know, we have to rebuild the world. We've got to do with what we have in this transitional phase until the new generation, I guess, technologies come around and reshape our landscape of tools. >> Well, impressive public discipline. Your point about microservice is interesting because a lot of those early microservices, weren't so micro and for the naysayers look past this, not prologue, but Thoughtworks was really early on in the whole concept of microservices. So be very excited to see how this plays out. But now there was some other good comments. There was one from a gentleman who said the most interesting aspects of data mesh are organizational. And that's how my colleague Sanji Mohan frames data mesh versus data fabric. You know, I'm not sure, I think we've sort of scratched the surface today that data today, data mesh is more. And I still think data fabric is what NetApp defined as software defined storage infrastructure that can serve on-prem and public cloud workloads back whatever, 2016. But the point you make in the thread that we're showing you here is that you're warning, and you referenced this earlier, that the segregating different modes of access will lead to fragmentation. And we don't want to repeat the mistakes of the past. >> Yes, there are comments around. Again going back to that original conversation that we have got this at a macro level. We've got this tendency to decompose complexity based on technical solutions. And, you know, the conversation could be, "Oh, I do batch or you do a stream and we are different."' They create these bifurcations in our decisions based on the technology where I do events and you do tables, right? So that sort of segregation of modes of access causes accidental complexity that we keep dealing with. Because every time in this tree, you create a new branch, you create new kind of new set of tools and then somehow need to be point to point integrated. You create new specialization around that. So the least number of branches that we have, and think about really about the continuum of experiences that we need to create and technologies that simplify, that continuum experience. So one of the things, for example, give you a past experience. I was really excited around the papers and the work that came around on Apache Beam, and generally flow based programming and stream processing. Because basically they were saying whether you are doing batch or whether you're doing streaming, it's all one stream. And sometimes the window of time, narrows and sometimes the window of time over which you're computing, widens and at the end of today, is you are just getting... Doing the stream processing. So it is those sort of notions that simplify and create continuum of experience. I think resonate with me personally, more than creating these tribal fights of this type versus that mode of access. So that's why data mesh naturally selects kind of this multimodal access to support end users, right? The persona of end users. >> Okay. So the last topic I want to hit, this whole discussion, the topic of data mesh it's highly nuanced, it's new, and people are going to shoehorn data mesh into their respective views of the world. And we talked about lake houses and there's three buckets. And of course, the gentleman from LinkedIn with Azure, Microsoft has a data mesh community. See you're going to have to enlist some serious army of enforcers to adjudicate. And I wrote some of the stuff down. I mean, it's interesting. Monte Carlo has a data mesh calculator. Starburst is leaning in, chaos. Search sees themselves as an enabler. Oracle and Snowflake both use the term data mesh. And then of course you've got big practitioners J-P-M-C, we've talked to Intuit, Orlando, HelloFresh has been on, Netflix has this event based sort of streaming implementation. So my question is, how realistic is it that the clarity of your vision can be implemented and not polluted by really rich technology companies and others? (Zhamak laughs) >> Is it even possible, right? Is it even possible? That's a yes. That's why I practice then. This is why I should practice things. Cause I think, it's going to be hard. What I'm hopeful, is that the socio-technical, Leveling Data mentioned that this is a socio-technical concern or solution, not just a technology solution. Hopefully always brings us back to, you know, the reality that vendors try to sell you safe oil that solves all of your problems. (chuckles) All of your data mesh problems. It's just going to cause more problem down the track. So we'll see, time will tell Dave and I count on you as one of those members of, (laughs) you know, folks that will continue to share their platform. To go back to the roots, as why in the first place? I mean, I dedicated a whole part of the book to 'Why?' Because we get, as you said, we get carried away with vendors and technology solution try to ride a wave. And in that story, we forget the reason for which we even making this change and we are going to spend all of this resources. So hopefully we can always come back to that. >> Yeah. And I think we can. I think you have really given this some deep thought and as we pointed out, this was based on practical knowledge and experience. And look, we've been trying to solve this data problem for a long, long time. You've not only articulated it well, but you've come up with solutions. So Zhamak, thank you so much. We're going to leave it there and I'd love to have you back. >> Thank you for the conversation. I really enjoyed it. And thank you for sharing your platform to talk about data mesh. >> Yeah, you bet. All right. And I want to thank my colleague, Stephanie Chan, who helps research topics for us. Alex Myerson is on production and Kristen Martin, Cheryl Knight and Rob Hoff on editorial. Remember all these episodes are available as podcasts, wherever you listen. And all you got to do is search Breaking Analysis Podcast. Check out ETR's website at etr.ai for all the data. And we publish a full report every week on wikibon.com, siliconangle.com. You can reach me by email david.vellante@siliconangle.com or DM me @dvellante. Hit us up on our LinkedIn post. This is Dave Vellante for theCUBE Insights powered by ETR. Have a great week, stay safe, be well. And we'll see you next time. (bright music)

Published Date : Apr 20 2022

SUMMARY :

bringing you data driven insights Organizations that have taken the plunge and have a conversation. and much of the past two years, and as we see, and some of the data and make the data available But the data warehouse crowd will say, in the middle to move the data around. and talk about how you serve and the data itself together and the implications. and the logic of running the business and are served by the technology. to build resilient you I think in all cases, you know, And that leads to a that the data teams lack and naturally the data and some of the standards that are needed. and formatting of the data and it created the data swamps. We're exposing that to the end client and the better part of a decade So it's just that's the role I have, and for the naysayers look and at the end of today, And of course, the gentleman part of the book to 'Why?' and I'd love to have you back. And thank you for sharing your platform etr.ai for all the data.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Kristen MartinPERSON

0.99+

Rob HoffPERSON

0.99+

Cheryl KnightPERSON

0.99+

Stephanie ChanPERSON

0.99+

Alex MyersonPERSON

0.99+

DavePERSON

0.99+

ZhamakPERSON

0.99+

oneQUANTITY

0.99+

Dave VellantePERSON

0.99+

AWSORGANIZATION

0.99+

10 lakesQUANTITY

0.99+

Sanji MohanPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Paul AndrewPERSON

0.99+

twoQUANTITY

0.99+

NetflixORGANIZATION

0.99+

Zhamak DehghaniPERSON

0.99+

Data Mesh: Delivering Data-Driven Value at ScaleTITLE

0.99+

BostonLOCATION

0.99+

OracleORGANIZATION

0.99+

14 plus yearsQUANTITY

0.99+

Palo AltoLOCATION

0.99+

two pointsQUANTITY

0.99+

siliconangle.comOTHER

0.99+

second layerQUANTITY

0.99+

2016DATE

0.99+

LinkedInORGANIZATION

0.99+

todayDATE

0.99+

SnowflakeORGANIZATION

0.99+

hundreds of lakesQUANTITY

0.99+

theCUBEORGANIZATION

0.99+

david.vellante@siliconangle.comOTHER

0.99+

theCUBE StudiosORGANIZATION

0.98+

SQLTITLE

0.98+

one unitQUANTITY

0.98+

firstQUANTITY

0.98+

100 levelQUANTITY

0.98+

third pointQUANTITY

0.98+

DatabricksORGANIZATION

0.98+

EuropeLOCATION

0.98+

three bucketsQUANTITY

0.98+

ETRORGANIZATION

0.98+

DevStackTITLE

0.97+

OneQUANTITY

0.97+

wikibon.comOTHER

0.97+

bothQUANTITY

0.97+

ThoughtworksORGANIZATION

0.96+

one setQUANTITY

0.96+

one streamQUANTITY

0.96+

IntuitORGANIZATION

0.95+

one wayQUANTITY

0.93+

two worldsQUANTITY

0.93+

HelloFreshORGANIZATION

0.93+

this weekDATE

0.93+

last nightDATE

0.91+

fourth oneQUANTITY

0.91+

SnowflakeTITLE

0.91+

two different modelsQUANTITY

0.91+

ML AnalyticsTITLE

0.91+

Breaking AnalysisTITLE

0.87+

two worldsQUANTITY

0.84+

Wen Phan, Ahana & Satyam Krishna, Blinkit & Akshay Agarwal, Blinkit | AWS Startup Showcase S2 E2


 

(gentle music) >> Welcome everyone to theCUBE's presentation of the AWS Startup Showcase. The theme is Data as Code; The Future of Enterprise Data and Analytics. This is the season two, episode two of the ongoing series of covering the exciting startups in the AWS ecosystem around data analytics and cloud computing. I'm your host, John Furrier. Today we're joined by great guests here. Three guests. Wen Phan, who's a Director of Product Management at Ahana, Satyam Krishna, Engineering Manager at Blinkit, and we have Akshay Agarwal, Senior Engineer at Blinkit as well. We're going to get into the relationship there. Let's get into. We're going to talk about how Blinkit's using open data lake, data house with Presto on AWS. Gentlemen, thanks for joining us. >> Thanks for having us. >> So we're going to get into the deep dive on the open data lake, but I want to just quickly get your thoughts on what it is for the folks out there. Set the table. What is the open data lakehouse? Why it is important? What's in it for the customers? Why are we seeing adoption around this because this is a big story. >> Sure. Yeah, the open data lakehouse is really being able to run a gamut of analytics, whether it be BI, SQL, machine learning, data science, on top of the data lake, which is based on inexpensive, low cost, scalable storage. And more importantly, it's also on top of open formats. And this to the end customer really offers a tremendous range of flexibility. They can run a bunch of use cases on the same storage and great price performance. >> You guys have any other thoughts on what's your reaction to the lakehouse? What is your experience with it? What's going on with Blinkit? >> No, I think for us also, it has been the primary driver of how as a company we have shifted our completely delivery model from us delivering in one day to someone who is delivering in 10 minutes, right? And a lot of this was made possible by having this kind of architecture in place, which helps us to be more open-source, more... where the tools are open-source, we have an open table format which helps us be very modular in nature, meaning we can pick solutions which works best for us, right? And that is the kind of architecture that we want to be in. >> Awesome. Wen, you know last time we chat with Ahana, we had a great conversation around Presto, data. The theme of this episode is Data as Code, which is interesting because in all the conversations in these episodes all around developers, which administrators are turning into developers, there's a developer vibe with data. And with opensource, it's software. Now you've got data taking a similar trajectory as how software development was with code, but the people running data they're not developers, they're administrators, they're operators. Now they're turning into DataOps. So it's kind of a similar vibe going on with branches and taking stuff out of and putting it back in, and testing it. Datasets becoming much more stable, iterating on machine learning algorithm. This is a movement. What's your guys reaction before we get into the relationships here with you guys. But, what's your reaction to this Data as Code movement? >> Yeah, so I think the folks at Blinkit are doing a great job there. I mean, they have a pretty compact data engineering team and they have some pretty stringent SLAs, as well as in terms of time to value and reliability. And what that ultimately translates for them is not only flexibility but reliability. So they've done some very fantastic work on a lot of automation, a lot of integration with code, and their data pipelines. And I'm sure they can give the details on that. >> Yes. Satyam and Akshay, you guys are engineers' software, but this is becoming a whole another paradigm where the frontline coding and or work or engineer data engineering is implementing the operations as well. It's kind of like DevOps for data. >> For sure. Right. And I think whenever you're working, even as a software engineer, the understanding of business is equally important. You cannot be working on something and be away from business, right? And that's where, like I mentioned earlier, when we realized that we have to completely move our stack and start giving analytics at 10 minutes, right. Because when you're delivering in 10 minutes, your leaders want to take decisions in your real-time. That means you need to move with them. You need to move with business. And when you do that, the kind of flexibility these softwares give is what enables the businesses at the end of the day. >> Awesome. This is the really kind of like, is there going to be a book called agile data warehouses? I don't think so. >> I think so. (laughing) >> The agile cloud data. This is cool. So let's get into what you guys do. What is Blinkit up to? What do you guys do? Can you take a minute to explain the company and your product? >> Sure. I'll take that. So Blinkit is India's biggest 10 minute delivery platform. It pioneered the delivery model in the country with over 10 million Indian shopping on our platform, ranging from everything: grocery staples, vegetables, emergency services, electronics, and much more, right. It currently delivers over 200,000 orders every day, and is in a hurry to bring the future of farmers to everyone in India. >> What's the relationship with Ahana and Blinkit? Wen, what's the tie in? >> Yeah, so Blinkit had a pretty well formed stack. They needed a little bit more flexibility and control. They thought a managed service was the way to go. And here at Ahana, we provide a SaaS managed service for Presto. So they engaged us and they evaluated our offering. And more importantly, we're able to partner. As a early stage startup, we really rely on very strong partners with great use cases that are willing to collaborate. And the folks at Blinkit have been really great in helping us push our product, develop our product. And we've been very happy about the value that we've been able to deliver to them as well. >> Okay. So let's unpack the open data lakehouse. What is it? What's under the covers? Let's get into it. >> Sure. So if bring up a slide. Like I said before, it's really a paradigm on being able to run a gamut of analytics on top of the open data lake. So what does that mean? How did it come about? So on the left hand side of the slide, we are coming out of this world where for the last several decades, the primary workhorse for SQL based processing and reporting and dashboarding use cases was really the data warehouse. And what we're seeing is a shift due to the trends in inexpensive scalable storage, cloud storage. The proliferation of open formats to facilitate using this storage to get certain amounts of reliability and performance, and the adoption of frameworks that can operate on top of this cloud data lake. So while here at Ahana, we're primarily focused on SQL workloads and Presto, this architecture really allows for other types of frameworks. And you see the ML and AI side. And like to Satyam's point earlier, offers a great amount of flexibility modularity for many use cases in the cloud. So really, that's really the lakehouse, and people like it for the performance, the openness, and the price performance. >> How's the open-source open side of it playing in the open-source? It's kind of open formats. What is the open-source angle on this because there's a lot of different approaches. I'm hearing open formats. You know, you have data stores which are a big part of seeing that. You got SQL, you mentioned SQL. There's got a mishmash of opportunities. Is it all coexisting? Is it one tool to rule the world or is it interchangeable? What's the open-source angle? >> There's multiple angles and I'll let definitely Satyam add to what I'm saying. This was definitely a big piece for Blinkit. So on one hand, you have the open formats. And what really the open formats enable is multiple compute engines to work on that data. And that's very huge. 'Cause it's open, you're not locked in. I think the other part of open that is important and I think it was important to Blinkit was the governance around that. So in particular Presto is governed by the Linux Foundation. And so, as a customer of open-source technology, they want some assurances for things like how's it governed? Is the license going to change? So there's that aspect of openness that I think is very important. >> Yeah. Blinkit, what's the data strategy here with lakehouse and you guys? Why are you adopting this type of architecture? >> So adding to what... Yeah, I think adding to Wen said, right. When we are thinking in terms of all these OpenStacks, you have got these open table formats, everything which is deployed over cloud, the primary reason there is modularity. It's as simple as that, right. You can plug and play so many different table formats from one thing to another based on the use case that you're trying to serve, so that you get the most value out of data. Right? I'll give you a very simple example. So for us we use... not even use one single table format. It's not that one thing solves for everything, right? We use both Hudi and Iceberg to solve for different use cases. One is good for when you're working for a certain data site. Icebergs works well when you're in the SQL kind of interface, right. Hudi's still trying to reach there. It's going to go there very soon. So having the ability to plug and play different formats based on the use case helps you to grow faster, helps you to take decisions faster because you now you're not stuck on one thing. They will have to implement it. Right. So I think that's what it is great about this data lake strategy. Keeping yourself cost effective. Yeah, please. >> So the enablement is basically use case driven. You don't have to be rearchitecturing for use cases. You can simply plug can play based on what you need for the use case. >> Yeah. You can... and again, you can focus on your business use case. You can figure out what your business users need and not worry about these things because that's where Presto comes in, helps you stitch that data together with multiple data formats, give you the performance that you need and it works out the best there. And that's something that you don't get to with traditional warehouse these days. Right? The kind of thing that we need, you don't get that. >> I do want to add. This is just to riff on what Satyam said. I think it's pretty interesting. So, it really allowed him to take the best-of-breed of what he was seeing in the community, right? So in the case of table formats, you've got Delta, you've got Hudi, you've got Iceberg, and they all have got their own roadmap and it's kind of organic of how these different communities want to evolve, and I think that's great, but you have these end consumers like Blinkit who have different maybe use cases overlapping, and they're not forced to pick one. When you have an open architecture, they can really put together best-of-breed. And as these projects evolve, they can continue to monitor it and then make decisions and continue to remain agile based on the landscape and how it's evolving. >> So the agility is a key point. Flexibility and agility, and time to valuing with your data. >> Yeah. >> All right. Wen, I got to get in to why the Presto is important here. Where does that fit in? Why is Presto important? >> Yeah. For me, it all comes down to the use cases and the needs. And reporting and dashboarding is not going to go away anytime soon. It's a very common use case. Many of our customers like Blinkit come to us for that use case. The difference now is today, people want to do that particular use case on top of the modern data lake, on top of scalable, inexpensive, low cost storage. Right? In addition to that, there's a need for this low latency interactive ability to engage with the data. This is often arises when you need to do things in a ad hoc basis or you're in the developmental phase of building things up. So if that's what your need is. And latency's important and getting your arms around the problems, very important. You have a certain SLA, I need to deliver something. That puts some requirements in the technology. And Presto is a perfect for that ideal use case. It's ideal for that use case. It's distributed, it's scalable, it's in memory. And so it's able to really provide that. I think the other benefit for Presto and why we're bidding on Presto is it works well on the data lakes, but you have to think about how are these organizations maturing with this technology. So it's not necessarily an all or nothing. You have organizations that have maybe the data lake and it's augmented with other analytical data stores like Snowflake or Redshift. So Presto also... a core aspect is its ability to federate or connect and query across different data sources. So this can be a permanent thing. This could also be a transitionary thing. We have some customers that are moving and slowly shifting their data portfolio from maybe all data warehouse into 80% data lake. But it gives that optionality, it gives that ability to transition over a timeframe. But for all those reasons, the latency, the scalability, the federation, is why Presto for this particular use case. >> And you can connect with other databases. It can be purpose built database, could be whatever. Right? >> Sure. Yes, yes. Presto has a very pluggable architecture. >> Okay. Here's the question for the Blinkit team? Why did you choose Presto and what led you to Ahana? >> So I'll take this better, over this what Presto sits well in that reach is, is how it is designed. Like basically, Presto decouples your storage with the compute. Basically like, people can use any storage and Presto just works as a query engine for them. So basically, it has a constant connectors where you can connect with a real-time databases like Pinot or a Druid, along with your warehouses like Redshift, along with your data lake that's like based on Hudi or Iceberg. So it's like a very landscape that you can use with the Presto. And consumers like the analytics doesn't need to learn the SQL or different paradigms of the querying for different sources. They just need to learn a single source. And, they get a single place to consume from. They get a single consumer on their single destination to write on also. So, it's a homologous architecture, which allows you to put a central security like which Presto integrates. So it's also based on open architecture, that's Apache engine. And it has also certain innovative features that you can see based on caching, which reduces a lot of the cost. And since you have further decoupled your storage with the compute, you can further reduce your cost, because now the biggest part of our tradition warehouse is a storage. And the cost goes massively upwards with the amount of data that you've added. Like basically, each time that you add more data, you require more storage, and warehouses ask you to write the data in their own format. Over here since we have decoupled that, the storage cost have gone down. It's literally that your cost that you are writing, and you just pay for the compute, and you can scale in scale out based on the requirements. If you have high traffic, you scale out. If you have low traffic, you scale in. So all those. >> So huge cost savings. >> Yeah. >> Yeah. Cost effectiveness, for sure. >> Cost effectiveness and you get a very good price value out of it. Like for each query, you can estimate what's the cost for you based on that tracking and all those things. >> I mean, if you think about the other classic Iceberg and what's under the water you don't know, it's the hidden cost. You think about the tooling, right, and also, time it takes to do stuff. So if you have flexibility on choice, when we were riffing on this last time we chatted with you guys and you brought it up earlier around, you can have the open formats to have different use cases in different tools or different platforms to work on it. Redshift, you can use Redshift here, or use something over there. You don't have to get locking >> Absolutely. >> Satyam & Akshay: Yeah. >> Locking is a huge problem. How do you guys see that 'cause sounds like here there's not a lot of locking. You got the open formats, and you got choice. >> Yeah. So you get best of the both worlds. Like you get with Ahana or with the Presto, you can get the best of the both worlds. Since it's cloud native, you can easily deploy your clusters very easily within like five minutes. Your cluster is up, you can start working on it. You can deploy multiple clusters for multiple teams. You get also flexibility of adding new connectors since it's open and further it's also much more secure since it's based on cloud native. So basically, you can control your security endpoints very well. So all those things comes in together with this architecture. So you can definitely go more on the lakehouse architecture than warehousing when you want to deliver data value faster. And basically, you get the much more high value out of your data in a sorted template. >> So Satyam, it sounds like the old warehousing was like the application person, not a lot of usage, old, a lot of latency. Okay. Here and there. But now you got more speed to deploy clusters, scale up scale down. Application developers are as everyone. It's not one person. It's not one group. It's whenever you want. So, you got speed. You got more diversity in the data opportunities, and your coding. >> Yeah. I think data warehouses are a way to start for every organization who is getting into data. I don't think data warehousing is still a solution and will be a solution for a lot of teams which are still getting into data. But as soon as you start scaling, as you start seeing the cost going up, as you start seeing the number of use cases adding up, having an open format definitely helps. So, I would say that's where we are also heading into and that's how our journey as well started with Presto as well, why we even thought about Ahana, right. >> (John chuckles) >> So, like you mentioned, one of the things that happened was as we were moving to the lakehouse and the open table format, I think Ahana is one of the first ones in the market to have Hudi as a first class citizen completely supported with all the things which are not even present at the time of... even with Presto, right. So we see Ahana working behind the scenes, improving even some of the things already over the open-source ecosystem. And that's where we get the most value out of Ahana as well. >> This is the convergence of open-source magic and commercialization. Wen, because you think about Data as Code, reminds me, I hear, "Data warehouse, it's not going to go away." But you got cloud scale or scale. It reminds me of the old, "Oh yeah, I have a data center." Well, here comes the cloud. So, doesn't really kill the data center, although Amazon would say that the data center's going to be eliminated. No, you just use it for whatever you need it for. You use it for specific use cases, but everyone, all the action goes to the cloud for scale. The same things happen with data, and look at the open-source community. It's kind of coming together. Data as Code is coming together. >> Yeah, absolutely. >> Absolutely. >> I do want to again to connect on another dot in terms of cost and that. You know, we've been talking a little bit about price performance, but there's an implicit cost, and I think this was also very important to Blinkit, and also why we're offering a managed service. So one piece of it. And it really revolves around the people, right? So outside of the technology, the performance. One thing that Akshay brought up and it's another important piece that I should have highlighted a little bit more is, Presto exposes the ability to interact your data in a widely adopted way, which is basically ANSI SQL. So the ability for your practitioners to use this technology is huge. That's just regular Presto. In terms of a managed service, the guys at Blinkit are a great high performing team, but they have to be very efficient with their time and what they manage. And what we're trying to do is provide leverage for them. So take a lot of the heavy lifting away, but at the same time, figuring out the right things to expose so that they have that same flexibility. And that's been the balancing point that we've been trying to balance at Ahana, but that goes back to cost. How do I total cost of ownership? And that not doesn't include just the actual querying processing time, but the ability for the organization to go ahead and absorb the solution. And what does it cost in terms of the people involved? >> Yeah. Great conversation. I mean, this brings up the question of back in the data center, the cloud days, you had the concept of an SRE, which is now popular, site reliability engineer. One person does all the clusters and manages all the scale. Is the data engineer the new SRE for data? Are we seeing a similar trajectory? Just want to get your reaction. What do you guys think? >> Yes, so I would say, definitely. It depends on the teams and the sizes of that. We are high performing team so each automation takes bits on the pieces of the architecture, like where they want to invest in. And it comes out with the value of the engineer's time and basically like how much they can invest in, how much they need to configure the architecture, and how much time it'll take to time to market. So basically like, this is what I would also highlight as an engineer. I found Ahana like the... I would say as a Presto in a cloud native environment, or I think so there's the one in the market that seamlessly scales and then scales out. And further, with a team of us, I would say our team size like three to four engineers managing cluster day in day out, conferring, tuning and all those things takes a lot of time. And Ahana came in and takes it off our plate and the hands in a solution which works out of box. So that's where this comes in. Ahana it's also based on open-source community. >> So the time of the engineer's time is so valuable. >> Yeah. >> My take on it really in terms of the data engineering being the SRE. I think that can work, it depends on the actual person, and we definitely try to make the process as easy as possible. I think in Blinkit's case, you guys are... There are data platform owners, but they definitely are aware of the pipelines. >> John: Yeah. >> So they have very intimate knowledge of what data engineers do, but I think in their case, you guys, you're managing a ton of systems. So it's not just even Presto. They have a ton of systems and surfacing that interface so they can cater to all the data engineers across their data systems, I think is the big need for them. I know you guys you want to chime in. I mean, we've seen the architecture and things like that. I think you guys did an amazing job there. >> So, and to adding to Wen's point, right. Like I generally think what DevOps is to the tech team. I think, what is data engineer or the data teams are to the data organization, right? Like they play a very similar role that you have to act as a guardrail to ensure that everyone has access to the data so the democratizing and everything is there, but that has to also come with security, right? And when you do that, there are (indistinct) a lot of points where someone can interact with data. We have... And again, there's a mixed match of open-source tools that works well, as well. And there are some paid tools as well. So for us like for visualization, we use Redash for our ad hoc analysis. And we use Tableau as well whenever we want to give a very concise reporting. We have Jupyter notebooks in place and we have EMRs as well. So we always have a mixed batch of things where people can interact with data. And most of our time is spent in acting as that guardrail to ensure that everyone should have access to data, but it shouldn't be exploited, right. And I think that's where we spend most of our time in. >> Yeah. And I think the time is valuable, but that your point about the democratization aspect of it, there seems to be a bigger step function value that you're enabling and needs to be talked out. The 10x engineer, it's more like 50x, right? If you get it done right, the enablement downstream at the scale that we're seeing with this new trend is significant. It's not just, oh yeah, visualization and get some data quicker, there's actually real advantages on a multiple with that engineering. So, and we saw that with DevOps, right? Like, you do this right and then magic happens on the edges. So, yeah, it's interesting. You guys, congratulations. Great environment. Thanks for sharing the insight Blinkit. Wen, great to see you. Ahana again with Presto, congratulations. The open-source meets data engineering. Thanks so much. >> Thanks, John. >> Appreciate it. >> Okay. >> Thanks John. >> Thanks. >> Thanks for having us. >> This season two, episode two of our ongoing series. This one is Data as Code. This is theCUBE. I'm John furrier. Thanks for watching. (gentle music)

Published Date : Apr 1 2022

SUMMARY :

This is the season two, episode What is the open data lakehouse? And this to the end customer And that is the kind of into the relationships here with you guys. give the details on that. is implementing the operations as well. You need to move with business. This is the really kind of like, I think so. So let's get into what you guys do. and is in a hurry to bring And the folks at Blinkit the open data lakehouse. So on the left hand side of the slide, What is the open-source angle on this Is the license going to change? with lakehouse and you guys? So having the ability to plug So the enablement is and again, you can focus So in the case of table formats, So the agility is a key point. Wen, I got to get in and the needs. And you can connect Presto has a very pluggable architecture. and what led you to Ahana? And consumers like the analytics and you get a very good and also, time it takes to do stuff. and you got choice. best of the both worlds. like the old warehousing as you start seeing the cost going up, and the open table format, the data center's going to be eliminated. figuring out the right things to expose and manages all the scale. and the sizes of that. So the time of the it depends on the actual person, I think you guys did an amazing job there. So, and to adding Thanks for sharing the insight Blinkit. This is theCUBE.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
John FurrierPERSON

0.99+

Wen PhanPERSON

0.99+

Akshay AgarwalPERSON

0.99+

JohnPERSON

0.99+

AmazonORGANIZATION

0.99+

AhanaPERSON

0.99+

IndiaLOCATION

0.99+

BlinkitORGANIZATION

0.99+

Satyam KrishnaPERSON

0.99+

Linux FoundationORGANIZATION

0.99+

AhanaORGANIZATION

0.99+

five minutesQUANTITY

0.99+

AkshayPERSON

0.99+

AWSORGANIZATION

0.99+

10 minutesQUANTITY

0.99+

Three guestsQUANTITY

0.99+

SatyamPERSON

0.99+

BlinkitPERSON

0.99+

one dayQUANTITY

0.99+

10 minuteQUANTITY

0.99+

RedshiftTITLE

0.99+

both worldsQUANTITY

0.99+

over 200,000 ordersQUANTITY

0.99+

PrestoPERSON

0.99+

over 10 millionQUANTITY

0.99+

SQLTITLE

0.99+

10xQUANTITY

0.99+

WenPERSON

0.98+

50xQUANTITY

0.98+

agileTITLE

0.98+

one pieceQUANTITY

0.98+

bothQUANTITY

0.98+

threeQUANTITY

0.98+

todayDATE

0.98+

oneQUANTITY

0.98+

single destinationQUANTITY

0.97+

One personQUANTITY

0.97+

each timeQUANTITY

0.96+

eachQUANTITY

0.96+

PrestoORGANIZATION

0.96+

one personQUANTITY

0.96+

single sourceQUANTITY

0.96+

TableauTITLE

0.96+

one toolQUANTITY

0.96+

IcebergsORGANIZATION

0.96+

TodayDATE

0.95+

OneQUANTITY

0.95+

one thingQUANTITY

0.95+

Venkat Venkataramani, Rockset & Doug Moore, Command Alkon | AWS Startup Showcase S2 E2


 

(upbeat music) >> Hey everyone. Welcome to theCUBE's presentation of the AWS Startup Showcase. This is Data as Code, The Future of Enterprise Data and Analytics. This is also season two, episode two of our ongoing series with exciting partners from the AWS ecosystem who are here to talk with us about data and analytics. I'm your host, Lisa Martin. Two guests join me, one, a cube alumni. Venkat Venkataramani is here CEO & Co-Founder of Rockset. Good to see you again. And Doug Moore, VP of cloud platforms at Command Alkon. You're here to talk to me about how Command Alkon implemented real time analytics in just days with Rockset. Guys, welcome to the program. >> Thanks for having us. >> Yeah, great to be here. >> Doug, give us a little bit of a overview of Command Alkon, what type of business you are? what your mission is? That good stuff. >> Yeah, great. I'll pref it by saying I've been in this industry for only three years. The 30 years prior I was in financial services. So this was really exciting and eye opening. It actually plays into the story of how we met Rockset. So that's why I wanted to preface that. But Command Alkon is in the business, is in the what's called The Heavy Building Materials Industry. And I had never heard of it until I got here. But if you think about large projects like building buildings, cities, roads anything that requires concrete asphalt or just really big trucks, full of bulky materials that's the heavy building materials industry. So for over 40 years Command Alkon has been the north American leader in providing software to quarries and production facilities to help mine and load these materials and to produce them and then get them to the job site. So that's what our supply chain is, is from the quarry through the development of these materials, then out to the to a heavy building material job site. >> Got it, and now how historically in the past has the movement of construction materials been coordinated? What was that like before you guys came on the scene? >> You'll love this answer. So 'cause, again, it's like a step back in time. When I got here the people told me that we're trying to come up with the platform that there are 27 industries studied globally. And our industry is second to last in terms of automation which meant that literally everything is still being done with paper and a lot of paper. So when one of those, let's say material is developed, concrete asphalt is produced and then needs to get to the job site. They start by creating a five part printed ticket or delivery description that then goes to multiple parties. It ends up getting touched physically over 50 times for every delivery. And to give you some idea what kind of scale it is there are over 330 million of these type deliveries in north America every year. So it's really a lot of favor and a lot of manual work. So that was the state of really where we were. And obviously there are compelling reasons certainly today but even 3, 4, 5 years ago to automate that and digitize it. >> Wow, tremendous potential to go nowhere but up with the amount of paper, the lack of, of automation. So, you guys Command Alkon built a platform, a cloud software construction software platform. Talk to me of about that. Why you built it, what was the compelling event? I mean, I think you've kind of already explained the compelling event of all the paper but give us a little bit more context. >> Yeah. That was the original. And then we'll get into what happened two years ago which has made it even more compelling but essentially with everything on premises there's really in a huge amount of inefficiency. So, people have heard the enormous numbers that it takes to build up a highway or a really large construction project. And a lot of that is tied up in these inefficiencies. So we felt like with our significant presence in this market, that if we could figure out how to automate getting this data into the cloud so that at least the partners in the supply chain could begin sharing information. That's not on paper a little bit closer to real time that we could make has an impact on everything from the timing it takes to do a project to even the amount of carbon dioxide that's admitted, for example from trucks running around and being delayed and not being coordinated well. >> So you built the connect platform you started on Amazon DynamoDB and ran into some performance challenges. Talk to us about the, some of those performance bottlenecks and how you found Venkat and Rockset. >> So from the beginning, we were fortunate, if you start building a cloud three years ago you're you have a lot of opportunity to use some of the what we call more fully managed or serverless offerings from Amazon and all the cloud vendors have them but Amazon is the one we're most familiar with throughout the past 10 years. So we went head first into saying, we're going to do everything we can to not manage infrastructure ourselves. So we can really focus on solving this problem efficiently. And it paid off great. And so we chose dynamo as our primary database and it still was a great decision. We have obviously hundreds of millions of billions of these data points in dynamo. And it's great from a transactional perspective, but at some point you need to get the data back out. And what plays into the story of the beginning when I came here with no background basically in this industry, is that, and as did most of the other people on my team, we weren't really sure what questions were going to be asked of the data. And that's super, super important with a NoSQL database like dynamo. You sort of have to know in advance what those usage patterns are going to be and what people are going to want to get back out of it. And that's what really began to strain us on both performance and just availability of information. >> Got it. Venkat, let's bring you into the conversation. Talk to me about some of the challenges that Doug articulated the, is industry with such little automation so much paper. Are you finding that still out there for in quite a few industries that really have nowhere to go but up? >> I think that's a very good point. We talk about digital transformation 2.0 as like this abstract thing. And then you meet like disruptors and innovators like Doug, and you realize how much impact, it has on the real world. But now it's not just about disrupting, and digitizing all of these records but doing it at a faster pace than ever before, right. I think this is really what digital transformation in the cloud really enable tools you do that, a small team in a, with a very very big mission and responsibility like what Doug team have been, shepherding here. They're able to move very, very, very fast, to be able to kind of accelerate this. And, they're not only on the forefront of digitizing and transforming a very big, paper-heavy kind of process, but real-time analytics and real time reporting is a requirement, right? Nobody's wondering where is my supply chain three days ago? Are my, one of the most important thing in heavy construction is to keep running on a schedule. If you fall behind, there's no way to catch up because there's so many things that falls apart. Now, how do you make sure you don't fall behind, realtime analytics and realtime reporting on how many trucks are supposed to be delivered today? Halfway through the day, are they on track? Are they getting behind? And all of those things is not just able to manage the data but also be able to get reporting and analytics on that is a extremely important aspect of this. So this is like a combination of digital transformation happening in the cloud in realtime and realtime analytics being in the forefront of it. And so we are very, very happy to partner with digital disruptors like Doug and his team to be part of this movement. >> Doug, as Venkat mentioned, access to real time data is a requirement that is just simple truth these days. I'm just curious, compelling event wise was COVID and accelerator? 'Cause we all know of the supply chain challenges that we're all facing in one way or the other, was that part of the compelling event that had you guys go and say, we want to do DynamoDB plus Rockset? >> Yeah, that is a fantastic question. In fact, more so than you can imagine. So anytime you come into an industry and you're going to try to completely change or revolutionize the way it operates it takes a long time to get the message out. Sometimes years, I remember in insurance it took almost 10 years really to get that message out and get great adoption and then COVID came along. And when COVID came along, we all of a sudden had a situation where drivers and the foreman on the job site didn't want to exchange the paperwork. I heard one story of a driver taping the ticket for signature to the foreman on a broomstick and putting it out his windows so that he didn't get too close. It really was that dramatic. And again, this is the early days and no one really has any idea what's happening and we're all working from home. So we launched, we saw that as an opportunity to really help people solve that problem and understand more what this transformation would mean in the long term. So we launched internally what we called Project Lemonade obviously from, make lemonade out of lemons, that's the situation that we were in and we immediately made some enhancements to a mobile app and then launched that to the field. So that basically there's now a digital acceptance capability where the driver can just stay in the vehicle and the foreman can be anywhere, look at the material say it's acceptable for delivery and go from there. So yeah, it made a, it actually immediately caused many of our customers hundreds to begin, to want to push their data to the cloud for that reason just to take advantage of that one capability >> Project lemonade, sounds like it's made a lot of lemonade out of a lot of lemons. Can you comment Doug on kind of the larger trend of real time analytics and logistics? >> Yeah, obviously, and this is something I didn't think about much either not knowing anything about concrete other than it was in my driveway before I got here. And that it's a perishable product and you've got that basically no more than about an hour and a half from the time you mix it, put it in the drum and get it to the job site and pour it. And then the next one has to come behind it. And I remember I, the trend is that we can't really do that on paper anymore and stay on top of what has to be done we'll get into the field. So a foreman, I recall saying that when you're in the field waiting on delivery, that you have people standing around and preparing the site ready to make a pour that two minutes is an eternity. And so, working a real time is all always a controversial word because it means something different to anyone, but that gave it real, a real clarity to mean, what it really meant to have real time analytics and how we are doing and where are my vehicles and how is this job performing today? And I think that a lot of people are still trying to figure out how to do that. And fortunately, we found a great tool set that's allowing us to do that at scale. Thankfully, for Rockset primarily. >> Venkat talk about it from your perspective the larger trend of real time analytics not just in logistics, but in other key industries. >> Yeah. I think we're seeing this across the board. I think, whether, even we see a huge trend even within an enterprise different teams from the marketing team to the support teams to more and more business operations team to the security team, really moving more and more of their use cases from real time. So we see this, the industries that are the innovators and the pioneers here are the ones for whom real times that requirement like Doug and his team here or where, if it is all news, it's no news, it's useless, right? But I think even within, across all industries, whether it is, gaming whether it is, FinTech, Bino related companies, e-learning platforms, so across, ed tech and so many different platforms, there is always this need for business operations. Some, certain aspects certain teams within large organizations to, have to tell me how to win the game and not like, play Monday morning quarterback after the game is over. >> Right, Doug, let's go back at you, I'm curious with connects, have you been able to scale the platform since you integrated with Rockset? Talk to us about some of the outcomes that you've achieved so far? >> Yeah, we have, and of course we knew and we made our database selection with dynamo that it really doesn't have a top end in terms of how much information that we can throw at it. But that's very, very challenging when it comes to using that information from reporting. But we've found the same thing as we've scaled the analytics side with Rockset indexing and searching of that database. So the scale in terms of the number of customers and the amount of data we've been able to take on has been, not been a problem. And honestly, for the first time in my career, I can say that we've always had to add people every time we add a certain number of customers. And that has absolutely not been the case with this platform. >> Well, and I imagine the team that you do have is far more, sorry Venkat, far more strategic and able to focus on bigger projects. >> It, is, and, you've amazed at, I mean Venkat hit on a couple of points that it's in terms of the adoption of analytics. What we found is that we are as big a customer of this analytic engine as our customers are because our marketing team and our sales team are always coming to us. Well how many customers are doing this? How many partners are connected in this way? Which feature flags are turned on the platform? And the way this works is all data that we push into the platform is automatically just indexed and ready for reporting analytics. So we really it's no additional ad of work, to answer these questions, which is really been phenomenal. >> I think the thing I want to add here is the speed at which they were able to build a scalable solution and also how little, operational and administrative overhead that it has cost of their teams, right. I think, this is again, realtime analytics. If you go and ask hundred people, do you want fast analytics on realtime data or slow analytics on scale data, people, no one would say give me slow and scale. So, I think it goes back to again our fundamental pieces that you have to remove all the cost and complexity barriers for realtime analytics to be the new default, right? Today companies try to get away with batch and the pioneers and the innovators are forced to solve, I know, kind of like address some of these realtime analytics challenges. I think with the platforms like the realtime analytics platform, like Rockset, we want to completely flip it on its head. You can do everything in real time. And there may be some extreme situations where you're dealing with like, hundreds of petabytes of data and you just need an analyst to generate like, quarterly reports out of that, go ahead and use some really, really good batch base system but you should be able to get anything, and everything you want without additional cost or complexity, in real time. That is really the vision. That is what we are really enabling here. >> Venkat, I want to also get your perspective and Doug I'd like your perspective on this as well but that is the role of cloud native and serverless technologies in digital disruption. And what do you see there? >> Yeah, I think it's huge. I think, again and again, every customer, and we meet, Command Alkon and Doug and his team is a great example of this where they really want to spend as much time and energies and calories that they have to, help their business, right? Like what, are we accomplishing trying to accomplish as a business? How do we enable, how do we build better products? How do we grow revenue? How do we eliminate risk that is inherent in the business? And that is really where they want to spend all of their energy not trying to like, install some backend software, administer build IDL pipelines and so on and so forth. And so, doing serverless on the compute side of that things like AWS lambda does and what have you. And, it's a very important innovation but that isn't, complete the story or your data stack also have to become serverless. And, that is really the vision with Rockset that your entire realtime analytics stack can be operating and managing. It could be as simple as managing a serverless stack for your compute environments like your APS servers and what have you. And so I think that is going to be a that is for here to stay. This is a path towards simplicity and simplicity scales really, really well, right? Complexity will always be the killer that'll limit, how far you can use this solution and how many problems can you solve with that solution? So, simplicity is a very, very important aspect here. And serverless helps you, deliver that. >> And Doug your thoughts on cloud native and serverless in terms of digital disruption >> Great point, and there are two parts to the scalability part. The second one is the one that's more subtle unless you're in charge of the budget. And that is, with enough effort and enough money that you can make almost any technology scale whether it's multiple copies of it, it may take a long time to get there but you can get there with most technologies but what is least scalable, at least that I as I see that this industry is the people, everybody knows we have a talent shortage and these other ways of getting the real time analytics and scaling infrastructure for compute and database storage, it really takes a highly skilled set of resources. And the more your company grows, the more of those you need. And that is what we really can't find. And that's actually what drove our team in our last industry to even go this way we reached a point where our growth was limited by the people we could find. And so we really wanted to break out of that. So now we had the best of both scalable people because we don't have to scale them and scalable technology. >> Excellent. The best of both worlds. Isn't it great when those two things come together? Gentlemen, thank you so much for joining me on "theCUBE" today. Talking about what Rockset and Command Alkon are doing together better together what you're enabling from a supply chain digitization perspective. We appreciate your insights. >> Great. Thank you. >> Thanks, Lisa. Thanks for having us. >> My pleasure. For Doug Moore and Venkat Venkatramani, I'm Lisa Martin. Keep it right here for more coverage of "theCUBE", your leader in high tech event coverage. (upbeat music)

Published Date : Mar 30 2022

SUMMARY :

Good to see you again. what type of business you are? and to produce them and then And to give you some idea Talk to me of about that. And a lot of that is tied and how you found Venkat and Rockset. and as did most of the that really have nowhere to go but up? and his team to be part of this movement. and say, we want to do and then launched that to the field. kind of the larger trend and get it to the job site and pour it. the larger trend of real time analytics team to the support teams And that has absolutely not been the case and able to focus on bigger projects. that it's in terms of the and the pioneers and the but that is the role of cloud native And so I think that is going to be a And that is what we really can't find. and Command Alkon are doing Thank you. Moore and Venkat Venkatramani,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

Doug MoorePERSON

0.99+

DougPERSON

0.99+

Venkat VenkataramaniPERSON

0.99+

Command AlkonORGANIZATION

0.99+

RocksetORGANIZATION

0.99+

LisaPERSON

0.99+

Doug MoorePERSON

0.99+

AmazonORGANIZATION

0.99+

Two guestsQUANTITY

0.99+

AWSORGANIZATION

0.99+

27 industriesQUANTITY

0.99+

two minutesQUANTITY

0.99+

bothQUANTITY

0.99+

VenkatORGANIZATION

0.99+

north AmericaLOCATION

0.99+

Monday morningDATE

0.99+

two partsQUANTITY

0.99+

over 50 timesQUANTITY

0.99+

oneQUANTITY

0.99+

over 330 millionQUANTITY

0.99+

Venkat VenkatramaniPERSON

0.99+

hundred peopleQUANTITY

0.99+

three days agoDATE

0.99+

two thingsQUANTITY

0.99+

over 40 yearsQUANTITY

0.99+

two years agoDATE

0.98+

three years agoDATE

0.98+

secondQUANTITY

0.98+

five partQUANTITY

0.98+

first timeQUANTITY

0.98+

todayDATE

0.98+

VenkatPERSON

0.97+

hundredsQUANTITY

0.97+

30 years priorDATE

0.97+

both worldsQUANTITY

0.97+

TodayDATE

0.97+

three yearsQUANTITY

0.96+

one storyQUANTITY

0.95+

DynamoDBTITLE

0.94+

almost 10 yearsQUANTITY

0.94+

hundreds of millions of billionsQUANTITY

0.93+

dynamoORGANIZATION

0.92+

second oneQUANTITY

0.91+

about an hour and a halfQUANTITY

0.9+

theCUBEORGANIZATION

0.9+

NoSQLTITLE

0.89+

3DATE

0.87+

BinoORGANIZATION

0.85+

past 10 yearsDATE

0.84+

every yearQUANTITY

0.84+

DougORGANIZATION

0.83+

AnalyticsTITLE

0.83+

5 years agoDATE

0.82+

north AmericanOTHER

0.81+

Startup ShowcaseEVENT

0.81+

Accelerating Automated Analytics in the Cloud with Alteryx


 

>>Alteryx is a company with a long history that goes all the way back to the late 1990s. Now the one consistent theme over 20 plus years has been that Ultrix has always been a data company early in the big data and Hadoop cycle. It saw the need to combine and prep different data types so that organizations could analyze data and take action Altrix and similar companies played a critical role in helping companies become data-driven. The problem was the decade of big data, brought a lot of complexities and required immense skills just to get the technology to work as advertised this in turn limited, the pace of adoption and the number of companies that could really lean in and take advantage of the cloud began to change all that and set the foundation for today's theme to Zuora of digital transformation. We hear that phrase a ton digital transformation. >>People used to think it was a buzzword, but of course we learned from the pandemic that if you're not a digital business, you're out of business and a key tenant of digital transformation is democratizing data, meaning enabling, not just hypo hyper specialized experts, but anyone business users to put data to work. Now back to Ultrix, the company has embarked on a major transformation of its own. Over the past couple of years, brought in new management, they've changed the way in which it engaged with customers with the new subscription model and it's topgraded its talent pool. 2021 was even more significant because of two acquisitions that Altrix made hyper Ana and trifecta. Why are these acquisitions important? Well, traditionally Altryx sold to business analysts that were part of the data pipeline. These were fairly technical people who had certain skills and were trained in things like writing Python code with hyper Ana Altryx has added a new persona, the business user, anyone in the business who wanted to gain insights from data and, or let's say use AI without having to be a deep technical expert. >>And then Trifacta a company started in the early days of big data by cube alum, Joe Hellerstein and his colleagues at Berkeley. They knocked down the data engineering persona, and this gives Altryx a complimentary extension into it where things like governance and security are paramount. So as we enter 2022, the post isolation economy is here and we do so with a digital foundation built on the confluence of cloud native technologies, data democratization and machine intelligence or AI, if you prefer. And Altryx is entering that new era with an expanded portfolio, new go-to market vectors, a recurring revenue business model, and a brand new outlook on how to solve customer problems and scale a company. My name is Dave Vellante with the cube and I'll be your host today. And the next hour, we're going to explore the opportunities in this new data market. And we have three segments where we dig into these trends and themes. First we'll talk to Jay Henderson, vice president of product management at Ultrix about cloud acceleration and simplifying complex data operations. Then we'll bring in Suresh Vetol who's the chief product officer at Altrix and Adam Wilson, the CEO of Trifacta, which of course is now part of Altrix. And finally, we'll hear about how Altryx is partnering with snowflake and the ecosystem and how they're integrating with data platforms like snowflake and what this means for customers. And we may have a few surprises sprinkled in as well into the conversation let's get started. >>We're kicking off the program with our first segment. Jay Henderson is the vice president of product management Altryx and we're going to talk about the trends and data, where we came from, how we got here, where we're going. We get some launch news. Well, Jay, welcome to the cube. >>Great to be here, really excited to share some of the things we're working on. >>Yeah. Thank you. So look, you have a deep product background, product management, product marketing, you've done strategy work. You've been around software and data, your entire career, and we're seeing the collision of software data cloud machine intelligence. Let's start with the customer and maybe we can work back from there. So if you're an analytics or data executive in an organization, w J what's your north star, where are you trying to take your company from a data and analytics point of view? >>Yeah, I mean, you know, look, I think all organizations are really struggling to get insights out of their data. I think one of the things that we see is you've got digital exhaust, creating large volumes of data storage is really cheap, so it doesn't cost them much to keep it. And that results in a situation where the organization's, you know, drowning in data, but somehow still starving for insights. And so I think, uh, you know, when I talk to customers, they're really excited to figure out how they can put analytics in the hands of every single person in their organization, and really start to democratize the analytics, um, and, you know, let the, the business users and the whole organization get value out of all that data they have. >>And we're going to dig into that throughout this program data, I like to say is plentiful insights, not always so much. Tell us about your launch today, Jay, and thinking about the trends that you just highlighted, the direction that your customers want to go and the problems that you're solving, what role does the cloud play in? What is what you're launching? How does that fit in? >>Yeah, we're, we're really excited today. We're launching the Altryx analytics cloud. That's really a portfolio of cloud-based solutions that have all been built from the ground up to be cloud native, um, and to take advantage of things like based access. So that it's really easy to give anyone access, including folks on a Mac. Um, it, you know, it also lets you take advantage of elastic compute so that you can do, you know, in database processing and cloud native, um, solutions that are gonna scale to solve the most complex problems. So we've got a portfolio of solutions, things like designer cloud, which is our flagship designer product in a browser and on the cloud, but we've got ultra to machine learning, which helps up-skill regular old analysts with advanced machine learning capabilities. We've got auto insights, which brings a business users into the fold and automatically unearths insights using AI and machine learning. And we've got our latest edition, which is Trifacta that helps data engineers do data pipelining and really, um, you know, create a lot of the underlying data sets that are used in some of this, uh, downstream analytics. >>Let's dig into some of those roles if we could a little bit, I mean, you've traditionally Altryx has served the business analysts and that's what designer cloud is fit for, I believe. And you've explained, you know, kind of the scope, sorry, you've expanded that scope into the, to the business user with hyper Anna. And we're in a moment we're going to talk to Adam Wilson and Suresh, uh, about Trifacta and that recent acquisition takes you, as you said, into the data engineering space in it. But in thinking about the business analyst role, what's unique about designer cloud cloud, and how does it help these individuals? >>Yeah, I mean, you know, really, I go back to some of the feedback we've had from our customers, which is, um, you know, they oftentimes have dozens or hundreds of seats of our designer desktop product, you know, really, as they look to take the next step, they're trying to figure out how do I give access to that? Those types of analytics to thousands of people within the organization and designer cloud is, is really great for that. You've got the browser-based interface. So if folks are on a Mac, they can really easily just pop, open the browser and get access to all of those, uh, prep and blend capabilities to a lot of the analysis we're doing. Um, it's a great way to scale up access to the analytics and then start to put it in the hands of really anyone in the organization, not just those highly skilled power users. >>Okay, great. So now then you add in the hyper Anna acquisition. So now you're targeting the business user Trifacta comes into the mix that deeper it angle that we talked about, how does this all fit together? How should we be thinking about the new Altryx portfolio? >>Yeah, I mean, I think it's pretty exciting. Um, you know, when you think about democratizing analytics and providing access to all these different groups of people, um, you've not been able to do it through one platform before. Um, you know, it's not going to be one interface that meets the, of all these different groups within the organization. You really do need purpose built specialized capabilities for each group. And finally, today with the announcement of the alternates analytics cloud, we brought together all of those different capabilities, all of those different interfaces into a single in the end application. So really finally delivering on the promise of providing analytics to all, >>How much of this you've been able to share with your customers and maybe your partners. I mean, I know OD is fairly new, but if you've been able to get any feedback from them, what are they saying about it? >>Uh, I mean, it's, it's pretty amazing. Um, we ran a early access, limited availability program that led us put a lot of this technology in the hands of over 600 customers, um, over the last few months. So we have gotten a lot of feedback. I tell you, um, it's been overwhelmingly positive. I think organizations are really excited to unlock the insights that have been hidden in all this data. They've got, they're excited to be able to use analytics in every decision that they're making so that the decisions they have or more informed and produce better business outcomes. Um, and, and this idea that they're going to move from, you know, dozens to hundreds or thousands of people who have access to these kinds of capabilities, I think has been a really exciting thing that is going to accelerate the transformation that these customers are on. >>Yeah, those are good. Good, good numbers for, for preview mode. Let's, let's talk a little bit about vision. So it's democratizing data is the ultimate goal, which frankly has been elusive for most organizations over time. How's your cloud going to address the challenges of putting data to work across the entire enterprise? >>Yeah, I mean, I tend to think about the future and some of the investments we're making in our products and our roadmap across four big themes, you know, in the, and these are really kind of enduring themes that you're going to see us making investments in over the next few years, the first is having cloud centricity. You know, the data gravity has been moving to the cloud. We need to be able to provide access, to be able to ingest and manipulate that data, to be able to write back to it, to provide cloud solution. So the first one is really around cloud centricity. The second is around big data fluency. Once you have all of the data, you need to be able to manipulate it in a performant manner. So having the elastic cloud infrastructure and in database processing is so important, the third is around making AI a strategic advantage. >>So, uh, you know, getting everyone involved and accessing AI and machine learning to unlock those insights, getting it out of the hands of the small group of data scientists, putting it in the hands of analysts and business users. Um, and then the fourth thing is really providing access across the entire organization. You know, it and data engineers, uh, as well as business owners and analysts. So, um, cloud centricity, big data fluency, um, AI is a strategic advantage and, uh, personas across the organization are really the four big themes you're going to see us, uh, working on over the next few months and, uh, coming coming year. >>That's good. Thank you for that. So, so on a related question, how do you see the data organizations evolving? I mean, traditionally you've had, you know, monolithic organizations, uh, very specialized or I might even say hyper specialized roles and, and your, your mission of course is the customer. You, you, you, you and your customers, they want to democratize the data. And so it seems logical that domain leaders are going to take more responsibility for data, life cycles, data ownerships, low code becomes more important. And perhaps this kind of challenges, the historically highly centralized and really specialized roles that I just talked about. How do you see that evolving and, and, and what role will Altryx play? >>Yeah. Um, you know, I think we'll see sort of a more federated systems start to emerge. Those centralized groups are going to continue to exist. Um, but they're going to start to empower, you know, in a much more de-centralized way, the people who are closer to the business problems and have better business understanding. I think that's going to let the centralized highly skilled teams work on, uh, problems that are of higher value to the organization. The kinds of problems where one or 2% lift in the model results in millions of dollars a day for the business. And then by pushing some of the analytics out to, uh, closer to the edge and closer to the business, you'll be able to apply those analytics in every single decision. So I think you're going to see, you know, both the decentralized and centralized models start to work in harmony and a little bit more about almost a federated sort of a way. And I think, you know, the exciting thing for us at Altryx is, you know, we want to facilitate that. We want to give analytic capabilities and solutions to both groups and types of people. We want to help them collaborate better, um, and drive business outcomes with the analytics they're using. >>Yeah. I mean, I think my take on another one, if you could comment is to me, the technology should be an operational detail and it has been the, the, the dog that wags the tail, or maybe the other way around, you mentioned digital exhaust before. I mean, essentially it's digital exhaust coming out of operationals systems that then somehow, eventually end up in the hand of the domain users. And I wonder if increasingly we're going to see those domain users, users, those, those line of business experts get more access. That's your goal. And then even go beyond analytics, start to build data products that could be monetized, and that maybe it's going to take a decade to play out, but that is sort of a new era of data. Do you see it that way? >>Absolutely. We're actually making big investments in our products and capabilities to be able to create analytic applications and to enable somebody who's an analyst or business user to create an application on top of the data and analytics layers that they have, um, really to help democratize the analytics, to help prepackage some of the analytics that can drive more insights. So I think that's definitely a trend we're going to see more. >>Yeah. And to your point, if you can federate the governance and automate that, then that can happen. I mean, that's a key part of it, obviously. So, all right, Jay, we have to leave it there up next. We take a deep dive into the Altryx recent acquisition of Trifacta with Adam Wilson who led Trifacta for more than seven years. It's the recipe. Tyler is the chief product officer at Altryx to explain the rationale behind the acquisition and how it's going to impact customers. Keep it right there. You're watching the cube. You're a leader in enterprise tech coverage. >>It's go time, get ready to accelerate your data analytics journey with a unified cloud native platform. That's accessible for everyone on the go from home to office and everywhere in between effortless analytics to help you go from ideas to outcomes and no time. It's your time to shine. It's Altryx analytics cloud time. >>Okay. We're here with. Who's the chief product officer at Altryx and Adam Wilson, the CEO of Trifacta. Now of course, part of Altryx just closed this quarter. Gentlemen. Welcome. >>Great to be here. >>Okay. So let me start with you. In my opening remarks, I talked about Altrix is traditional position serving business analysts and how the hyper Anna acquisition brought you deeper into the business user space. What does Trifacta bring to your portfolio? Why'd you buy the company? >>Yeah. Thank you. Thank you for the question. Um, you know, we see, uh, we see a massive opportunity of helping, um, brands, um, democratize the use of analytics across their business. Um, every knowledge worker, every individual in the company should have access to analytics. It's no longer optional, um, as they navigate their businesses with that in mind, you know, we know designer and are the products that Altrix has been selling the past decade or so do a really great job, um, addressing the business analysts, uh, with, um, hyper Rana now kind of renamed, um, Altrix auto. We even speak with the business owner and the line of business owner. Who's looking for insights that aren't real in traditional dashboards and so on. Um, but we see this opportunity of really helping the data engineering teams and it organizations, um, to also make better use of analytics. Um, and that's where the drive factor comes in for us. Um, drive factor has the best data engineering cloud in the planet. Um, they have an established track record of working across multiple cloud platforms and helping data engineers, um, do better data pipelining and work better with, uh, this massive kind of cloud transformation that's happening in every business. Um, and so fact made so much sense for us. >>Yeah. Thank you for that. I mean, you, look, you could have built it yourself would have taken, you know, who knows how long, you know, but, uh, so definitely a great time to market move, Adam. I wonder if we could dig into Trifacta some more, I mean, I remember interviewing Joe Hellerstein in the early days. You've talked about this as well, uh, on the cube coming at the problem of taking data from raw refined to an experience point of view. And Joe in the early days, talked about flipping the model and starting with data visualization, something Jeff, her was expert at. So maybe explain how we got here. We used to have this cumbersome process of ETL and you may be in some others changed that model with ELL and then T explain how Trifacta really changed the data engineering game. >>Yeah, that's exactly right. Uh, David, it's been a really interesting journey for us because I think the original hypothesis coming out of the campus research, uh, at Berkeley and Stanford that really birth Trifacta was, you know, why is it that the people who know the data best can't do the work? You know, why is this become the exclusive purview of the highly technical? And, you know, can we rethink this and make this a user experience, problem powered by machine learning that will take some of the more complicated things that people want to do with data and really help to automate those. So, so a broader set of, of users can, um, can really see for themselves and help themselves. And, and I think that, um, there was a lot of pent up frustration out there because people have been told for, you know, for a decade now to be more data-driven and then the whole time they're saying, well, then give me the data, you know, in the shape that I could use it with the right level of quality and I'm happy to be, but don't tell me to be more data-driven and then, and, and not empower me, um, to, to get in there and to actually start to work with the data in meaningful ways. >>And so, um, that was really, you know, what, you know, the origin story of the company and I think is, as we, um, saw over the course of the last 5, 6, 7 years that, um, you know, uh, real, uh, excitement to embrace this idea of, of trying to think about data engineering differently, trying to democratize the, the ETL process and to also leverage all these exciting new, uh, engines and platforms that are out there that allow for processing, you know, ever more diverse data sets, ever larger data sets and new and interesting ways. And that's where a lot of the push-down or the ELT approaches that, you know, I think it could really won the day. Um, and that, and that for us was a hallmark of the solution from the very beginning. >>Yeah, this is a huge point that you're making is, is first of all, there's a large business, it's probably about a hundred billion dollar Tam. Uh, and the, the point you're making, because we've looked, we've contextualized most of our operational systems, but the big data pipeline is hasn't gotten there. But, and maybe we could talk about that a little bit because democratizing data is Nirvana, but it's been historically very difficult. You've got a number of companies it's very fragmented and they're all trying to attack their little piece of the problem to achieve an outcome, but it's been hard. And so what's going to be different about Altryx as you bring these puzzle pieces together, how is this going to impact your customers who would like to take that one? >>Yeah, maybe, maybe I'll take a crack at it. And Adam will, um, add on, um, you know, there hasn't been a single platform for analytics, automation in the enterprise, right? People have relied on, uh, different products, um, to solve kind of, uh, smaller problems, um, across this analytics, automation, data transformation domain. Um, and, um, I think uniquely Alcon's has that opportunity. Uh, we've got 7,000 plus customers who rely on analytics for, um, data management, for analytics, for AI and ML, uh, for transformations, uh, for reporting and visualization for automated insights and so on. Um, and so by bringing drive factor, we have the opportunity to scale this even further and solve for more use cases, expand the scenarios where it's applied and so multiple personas. Um, and we just talked about the data engineers. They are really a growing stakeholder in this transformation of data and analytics. >>Yeah, good. Maybe we can stay on this for a minute cause you, you you're right. You bring it together. Now at least three personas the business analyst, the end user slash business user. And now the data engineer, which is really out of an it role in a lot of companies, and you've used this term, the data engineering cloud, what is that? How is it going to integrate in with, or support these other personas? And, and how's it going to integrate into the broader ecosystem of clouds and cloud data warehouses or any other data stores? >>Yeah, no, that's great. Uh, yeah, I think for us, we really looked at this and said, you know, we want to build an open and interactive cloud platform for data engineers, you know, to collaboratively profile pipeline, um, and prepare data for analysis. And that really meant collaborating with the analysts that were in the line of business. And so this is why a big reason why this combination is so magic because ultimately if we can get the data engineers that are creating the data products together with the analysts that are in the line of business that are driving a lot of the decision making and allow for that, what I would describe as collaborative curation of the data together, so that you're starting to see, um, uh, you know, increasing returns to scale as this, uh, as this rolls out. I just think that is an incredibly powerful combination and, and frankly, something that the market is not crack the code on yet. And so, um, I think when we, when I sat down with Suresh and with mark and the team at Ultrix, that was really part of the, the, the big idea, the big vision that was painted and got us really energized about the acquisition and about the potential of the combination. >>And you're really, you're obviously writing the cloud and the cloud native wave. Um, and, but specifically we're seeing, you know, I almost don't even want to call it a data warehouse anyway, because when you look at what's, for instance, Snowflake's doing, of course their marketing is around the data cloud, but I actually think there's real justification for that because it's not like the traditional data warehouse, right. It's, it's simplified get there fast, don't necessarily have to go through the central organization to share data. Uh, and, and, and, but it's really all about simplification, right? Isn't that really what the democratization comes down to. >>Yeah. It's simplification and collaboration. Right. I don't want to, I want to kind of just what Adam said resonates with me deeply. Um, analytics is one of those, um, massive disciplines inside an enterprise that's really had the weakest of tools. Um, and we just have interfaces to collaborate with, and I think truly this was all drinks and a superpower was helping the analysts get more out of their data, get more out of the analytics, like imagine a world where these people are collaborating and sharing insights in real time and sharing workflows and getting access to new data sources, um, understanding data models better, I think, um, uh, curating those insights. I boring Adam's phrase again. Um, I think that creates a real value inside the organization because frankly in scaling analytics and democratizing analytics and data, we're still in such early phases of this journey. >>So how should we think about designer cloud, which is from Altrix it's really been the on-prem and the server desktop offering. And of course Trifacta is with cloud cloud data warehouses. Right. Uh, how, how should we think about those two products? Yeah, >>I think, I think you should think about them. And, uh, um, as, as very complimentary right designer cloud really shares a lot of DNA and heritage with, uh, designer desktop, um, the low code tooling and that interface, uh, the really appeals to the business analysts, um, and gets a lot of the things that they do well, we've also built it with interoperability in mind, right. So if you started building your workflows in designer desktop, you want to share that with design and cloud, we want to make it super easy for you to do that. Um, and I think over time now we're only a week into, um, this Alliance with, um, with, um, Trifacta, um, I think we have to get deeper inside to think about what does the data engineer really need? What's the business analysts really need and how to design a cloud, and Trifacta really support both of those requirements, uh, while kind of continue to build on the trifecta on the amazing Trifacta cloud platform. >>You know, >>I think we're just going to say, I think that's one of the things that, um, you know, creates a lot of, uh, opportunity as we go forward, because ultimately, you know, Trifacta took a platform, uh, first mentality to everything that we built. So thinking about openness and extensibility and, um, and how over time people could build things on top of factor that are a variety of analytic tool chain, or analytic applications. And so, uh, when you think about, um, Ultrix now starting to, uh, to move some of its capabilities or to provide additional capabilities, uh, in the cloud, um, you know, Trifacta becomes a platform that can accelerate, you know, all of that work and create, uh, uh, a cohesive set of, of cloud-based services that, um, share a common platform. And that maintains independence because both companies, um, have been, uh, you know, fiercely independent, uh, and, and really giving people choice. >>Um, so making sure that whether you're, uh, you know, picking one cloud platform and other, whether you're running things on the desktop, uh, whether you're running in hybrid environments, that, um, no matter what your decision, um, you're always in a position to be able to get out your data. You're always in a position to be able to cleanse transform shape structure, that data, and ultimately to deliver, uh, the analytics that you need. And so I think in that sense, um, uh, you know, this, this again is another reason why the combination, you know, fits so well together, giving people, um, the choice. Um, and as they, as they think about their analytics strategy and their platform strategy going forward, >>Yeah. I make a chuckle, but one of the reasons I always liked Altrix is cause you kinda did the little end run on it. It can be a blocker sometimes, but that created problems, right? Because the organization said, wow, this big data stuff has taken off, but we need security. We need governance. And it's interesting because you've got, you know, ETL has been complex, whereas the visualization tools, they really, you know, really weren't great at governance and security. It took some time there. So that's not, not their heritage. You're bringing those worlds together. And I'm interested, you guys just had your sales kickoff, you know, what was their reaction like? Uh, maybe Suresh, you could start off and maybe Adam, you could bring us home. >>Um, thanks for asking about our sales kickoff. So we met for the first time and you've got a two years, right. For, as, as it is for many of us, um, in person, uh, um, which I think was a, was a real breakthrough as Qualtrics has been on its transformation journey. Uh, we added a Trifacta to, um, the, the potty such as the tour, um, and getting all of our sales teams and product organizations, um, to meet in person in one location. I thought that was very powerful for other the company. Uh, but then I tell you, um, um, the reception for Trifacta was beyond anything I could have imagined. Uh, we were working out him and I will, when he's so hot on, on the deal and the core hypotheses and so on. And then you step back and you're going to share the vision with the field organization, and it blows you away, the energy that it creates among our sellers out of partners. >>And I'm sure Madam will and his team were mocked, um, every single day, uh, with questions and opportunities to bring them in. But Adam, maybe you should share. Yeah, no, it was, uh, it was through the roof. I mean, uh, uh, the, uh, the amount of energy, the, uh, certainly how welcoming everybody was, uh, uh, you know, just, I think the story makes so much sense together. I think culturally, the company is, are very aligned. Um, and, uh, it was a real, uh, real capstone moment, uh, to be able to complete the acquisition and to, and to close and announced, you know, at the kickoff event. And, um, I think, you know, for us, when we really thought about it, you know, when we ended, the story that we told was just, you have this opportunity to really cater to what the end users care about, which is a lot about interactivity and self-service, and at the same time. >>And that's, and that's a lot of the goodness that, um, that Altryx is, has brought, you know, through, you know, you know, years and years of, of building a very vibrant community of, you know, thousands, hundreds of thousands of users. And on the other side, you know, Trifacta bringing in this data engineering focus, that's really about, uh, the governance things that you mentioned and the openness, um, that, that it cares deeply about. And all of a sudden, now you have a chance to put that together into a complete story where the data engineering cloud and analytics, automation, you know, coming together. And, um, and I just think, you know, the lights went on, um, you know, for people instantaneously and, you know, this is a story that, um, that I think the market is really hungry for. And certainly the reception we got from, uh, from the broader team at kickoff was, uh, was a great indication. >>Well, I think the story hangs together really well, you know, one of the better ones I've seen in, in this space, um, and, and you guys coming off a really, really strong quarter. So congratulations on that jets. We have to leave it there. I really appreciate your time today. Yeah. Take a look at this short video. And when we come back, we're going to dig into the ecosystem and the integration into cloud data warehouses and how leading organizations are creating modern data teams and accelerating their digital businesses. You're watching the cube you're leader in enterprise tech coverage. >>This is your data housed neatly insecurely in the snowflake data cloud. And all of it has potential the potential to solve complex business problems, deliver personalized financial offerings, protect supply chains from disruption, cut costs, forecast, grow and innovate. All you need to do is put your data in the hands of the right people and give it an opportunity. Luckily for you. That's the easy part because snowflake works with Alteryx and Alteryx turns data into breakthroughs with just a click. Your organization can automate analytics with drag and drop building blocks, easily access snowflake data with both sequel and no SQL options, share insights, powered by Alteryx data science and push processing to snowflake for lightning, fast performance, you get answers you can put to work in your teams, get repeatable processes they can share in that's exciting because not only is your data no longer sitting around in silos, it's also mobilized for the next opportunity. Turn your data into a breakthrough Alteryx and snowflake >>Okay. We're back here in the queue, focusing on the business promise of the cloud democratizing data, making it accessible and enabling everyone to get value from analytics, insights, and data. We're now moving into the eco systems segment the power of many versus the resources of one. And we're pleased to welcome. Barb Hills camp was the senior vice president partners and alliances at Ultrix and a special guest Terek do week head of technology alliances at snowflake folks. Welcome. Good to see you. >>Thank you. Thanks for having me. Good to see >>Dave. Great to see you guys. So cloud migration, it's one of the hottest topics. It's the top one of the top initiatives of senior technology leaders. We have survey data with our partner ETR it's number two behind security, and just ahead of analytics. So we're hovering around all the hot topics here. Barb, what are you seeing with respect to customer, you know, cloud migration momentum, and how does the Ultrix partner strategy fit? >>Yeah, sure. Partners are central company's strategy. They always have been. We recognize that our partners have deep customer relationships. And when you connect that with their domain expertise, they're really helping customers on their cloud and business transformation journey. We've been helping customers achieve their desired outcomes with our partner community for quite some time. And our partner base has been growing an average of 30% year over year, that partner community and strategy now addresses several kinds of partners, spanning solution providers to global SIS and technology partners, such as snowflake and together, we help our customers realize the business promise of their journey to the cloud. Snowflake provides a scalable storage system altereds provides the business user friendly front end. So for example, it departments depend on snowflake to consolidate data across systems into one data cloud with Altryx business users can easily unlock that data in snowflake solving real business outcomes. Our GSI and solution provider partners are instrumental in providing that end to end benefit of a modern analytic stack in the cloud providing platform, guidance, deployment, support, and other professional services. >>Great. Let's get a little bit more into the relationship between Altrix and S in snowflake, the partnership, maybe a little bit about the history, you know, what are the critical aspects that we should really focus on? Barb? Maybe you could start an Interra kindly way in as well. >>Yeah, so the relationship started in 2020 and all shirts made a big bag deep with snowflake co-innovating and optimizing cloud use cases together. We are supporting customers who are looking for that modern analytic stack to replace an old one or to implement their first analytic strategy. And our joint customers want to self-serve with data-driven analytics, leveraging all the benefits of the cloud, scalability, accessibility, governance, and optimizing their costs. Um, Altrix proudly achieved. Snowflake's highest elite tier in their partner program last year. And to do that, we completed a rigorous third party testing process, which also helped us make some recommended improvements to our joint stack. We wanted customers to have confidence. They would benefit from high quality and performance in their investment with us then to help customers get the most value out of the destroyed solution. We developed two great assets. One is the officer starter kit for snowflake, and we coauthored a joint best practices guide. >>The starter kit contains documentation, business workflows, and videos, helping customers to get going more easily with an altered since snowflake solution. And the best practices guide is more of a technical document, bringing together experiences and guidance on how Altryx and snowflake can be deployed together. Internally. We also built a full enablement catalog resources, right? We wanted to provide our account executives more about the value of the snowflake relationship. How do we engage and some best practices. And now we have hundreds of joint customers such as Juniper and Sainsbury who are actively using our joint solution, solving big business problems much faster. >>Cool. Kara, can you give us your perspective on the partnership? >>Yeah, definitely. Dave, so as Barb mentioned, we've got this standing very successful partnership going back years with hundreds of happy joint customers. And when I look at the beginning, Altrix has helped pioneer the concept of self-service analytics, especially with use cases that we worked on with for, for data prep for BI users like Tableau and as Altryx has evolved to now becoming from data prep to now becoming a full end to end data science platform. It's really opened up a lot more opportunities for our partnership. Altryx has invested heavily over the last two years in areas of deep integration for customers to fully be able to expand their investment, both technologies. And those investments include things like in database pushed down, right? So customers can, can leverage that elastic platform, that being the snowflake data cloud, uh, with Alteryx orchestrating the end to end machine learning workflows Alteryx also invested heavily in snow park, a feature we released last year around this concept of data programmability. So all users were regardless of their business analysts, regardless of their data, scientists can use their tools of choice in order to consume and get at data. And now with Altryx cloud, we think it's going to open up even more opportunities. It's going to be a big year for the partnership. >>Yeah. So, you know, Terike, we we've covered snowflake pretty extensively and you initially solve what I used to call the, I still call the snake swallowing the basketball problem and cloud data warehouse changed all that because you had virtually infinite resources, but so that's obviously one of the problems that you guys solved early on, but what are some of the common challenges or patterns or trends that you see with snowflake customers and where does Altryx come in? >>Sure. Dave there's there's handful, um, that I can come up with today, the big challenges or trends for us, and Altrix really helps us across all of them. Um, there are three particular ones I'm going to talk about the first one being self-service analytics. If we think about it, every organization is trying to democratize data. Every organization wants to empower all their users, business users, um, you know, the, the technology users, but the business users, right? I think every organization has realized that if everyone has access to data and everyone can do something with data, it's going to make them competitively, give them a competitive advantage with Altrix is something we share that vision of putting that power in the hands of everyday users, regardless of the skillsets. So, um, with self-service analytics, with Ultrix designer they've they started out with self-service analytics as the forefront, and we're just scratching the surface. >>I think there was an analyst, um, report that shows that less than 20% of organizations are truly getting self-service analytics to their end users. Now, with Altryx going to Ultrix cloud, we think that's going to be a huge opportunity for us. Um, and then that opens up the second challenge, which is machine learning and AI, every organization is trying to get predictive analytics into every application that they have in order to be competitive in order to be competitive. Um, and with Altryx creating this platform so they can cater to both the everyday business user, the quote unquote, citizen data scientists, and making a code friendly for data scientists to be able to get at their notebooks and all the different tools that they want to use. Um, they fully integrated in our snow park platform, which I talked about before, so that now we get an end to end solution caring to all, all lines of business. >>And then finally this concept of data marketplaces, right? We, we created snowflake from the ground up to be able to solve the data sharing problem, the big data problem, the data sharing problem. And Altryx um, if we look at mobilizing your data, getting access to third-party datasets, to enrich with your own data sets, to enrich with, um, with your suppliers and with your partners, data sets, that's what all customers are trying to do in order to get a more comprehensive 360 view, um, within their, their data applications. And so with Altryx alterations, we're working on third-party data sets and marketplaces for quite some time. Now we're working on how do we integrate what Altrix is providing with the snowflake data marketplace so that we can enrich these workflows, these great, great workflows that Altrix writing provides. Now we can add third party data into that workflow. So that opens up a ton of opportunities, Dave. So those are three I see, uh, easily that we're going to be able to solve a lot of customer challenges with. >>So thank you for that. Terrick so let's stay on cloud a little bit. I mean, Altrix is undergoing a major transformation, big focus on the cloud. How does this cloud launch impact the partnership Terike from snowflakes perspective and then Barb, maybe, please add some color. >>Yeah, sure. Dave snowflake started as a cloud data platform. We saw our founders really saw the challenges that customers are having with becoming data-driven. And the biggest challenge was the complexity of having imagine infrastructure to even be able to do it, to get applications off the ground. And so we created something to be cloud-native. We created to be a SAS managed service. So now that that Altrix is moving to the same model, right? A cloud platform, a SAS managed service, we're just, we're just removing more of the friction. So we're going to be able to start to package these end to end solutions that are SAS based that are fully managed. So customers can, can go faster and they don't have to worry about all of the underlying complexities of, of, of stitching things together. Right? So, um, so that's, what's exciting from my viewpoint >>And I'll follow up. So as you said, we're investing heavily in the cloud a year ago, we had two pre desktop products, and today we have four cloud products with cloud. We can provide our users with more flexibility. We want to make it easier for the users to leverage their snowflake data in the Alteryx platform, whether they're using our beloved on-premise solution or the new cloud products were committed to that continued investment in the cloud, enabling our joint partner solutions to meet customer requirements, wherever they store their data. And we're working with snowflake, we're doing just that. So as customers look for a modern analytic stack, they expect that data to be easily accessible, right within a fast, secure and scalable platform. And the launch of our cloud strategy is a huge leap forward in making Altrix more widely accessible to all users in all types of roles, our GSI and our solution provider partners have asked for these cloud capabilities at scale, and they're excited to better support our customers, cloud and analytic >>Are. How about you go to market strategy? How would you describe your joint go to market strategy with snowflake? >>Sure. It's simple. We've got to work backwards from our customer's challenges, right? Driving transformation to solve problems, gain efficiencies, or help them save money. So whether it's with snowflake or other GSI, other partner types, we've outlined a joint journey together from recruit solution development, activation enablement, and then strengthening our go to market strategies to optimize our results together. We launched an updated partner program and within that framework, we've created new benefits for our partners around opportunity registration, new role based enablement and training, basically extending everything we do internally for our own go-to-market teams to our partners. We're offering partner, marketing resources and funding to reach new customers together. And as a matter of fact, we recently launched a fantastic video with snowflake. I love this video that very simply describes the path to insights starting with your snowflake data. Right? We do joint customer webinars. We're working on joint hands-on labs and have a wonderful landing page with a lot of assets for our customers. Once we have an interested customer, we engage our respective account managers, collaborating through discovery questions, proof of concepts really showcasing the desired outcome. And when you combine that with our partners technology or domain expertise, it's quite powerful, >>Dark. How do you see it? You'll go to market strategy. >>Yeah. Dave we've. Um, so we initially started selling, we initially sold snowflake as technology, right? Uh, looking at positioning the diff the architectural differentiators and the scale and concurrency. And we noticed as we got up into the larger enterprise customers, we're starting to see how do they solve their business problems using the technology, as well as them coming to us and saying, look, we want to also know how do you, how do you continue to map back to the specific prescriptive business problems we're having? And so we shifted to an industry focus last year, and this is an area where Altrix has been mature for probably since their inception selling to the line of business, right? Having prescriptive use cases that are particular to an industry like financial services, like retail, like healthcare and life sciences. And so, um, Barb talked about these, these starter kits where it's prescriptive, you've got a demo and, um, a way that customers can get off the ground and running, right? >>Cause we want to be able to shrink that time to market, the time to value that customers can watch these applications. And we want to be able to, to tell them specifically how we can map back to their business initiatives. So I see a huge opportunity to align on these industry solutions. As BARR mentioned, we're already doing that where we've released a few around financial services working in healthcare and retail as well. So that is going to be a way for us to allow customers to go even faster and start to map two lines of business with Alteryx. >>Great. Thanks Derek. Bob, what can we expect if we're observing this relationship? What should we look for in the coming year? >>A lot specifically with snowflake, we'll continue to invest in the partnership. Uh, we're co innovators in this journey, including snow park extensibility efforts, which Derek will tell you more about shortly. We're also launching these great news strategic solution blueprints, and extending that at no charge to our partners with snowflake, we're already collaborating with their retail and CPG team for industry blueprints. We're working with their data marketplace team to highlight solutions, working with that data in their marketplace. More broadly, as I mentioned, we're relaunching the ultra partner program designed to really better support the unique partner types in our global ecosystem, introducing new benefits so that with every partner, achievement or investment with ultra score, providing our partners with earlier access to benefits, um, I could talk about our program for 30 minutes. I know we don't have time. The key message here Alteryx is investing in our partner community across the business, recognizing the incredible value that they bring to our customers every day. >>Tarik will give you the last word. What should we be looking for from, >>Yeah, thanks. Thanks, Dave. As BARR mentioned, Altrix has been the forefront of innovating with us. They've been integrating into, uh, making sure again, that customers get the full investment out of snowflake things like in database push down that I talked about before that extensibility is really what we're excited about. Um, the ability for Ultrix to plug into this extensibility framework that we call snow park and to be able to extend out, um, ways that the end users can consume snowflake through, through sequel, which has traditionally been the way that you consume snowflake as well as Java and Scala, not Python. So we're excited about those, those capabilities. And then we're also excited about the ability to plug into the data marketplace to provide third party data sets, right there probably day sets in, in financial services, third party, data sets and retail. So now customers can build their data applications from end to end using ultrasound snowflake when the comprehensive 360 view of their customers, of their partners, of even their employees. Right? I think it's exciting to see what we're going to be able to do together with these upcoming innovations. Great >>Barb Tara, thanks so much for coming on the program, got to leave it right there in a moment, I'll be back with some closing thoughts in a summary, don't go away. >>1200 hours of wind tunnel testing, 30 million race simulations, 2.4 second pit stops make that 2.3. The sector times out the wazoo, whites are much of this velocity's pressures, temperatures, 80,000 components generating 11.8 billion data points and one analytics platform to make sense of it all. When McLaren needs to turn complex data into insights, they turn to Altryx Qualtrics analytics, automation, >>Okay, let's summarize and wrap up the session. We can pretty much agree the data is plentiful, but organizations continue to struggle to get maximum value out of their data investments. The ROI has been elusive. There are many reasons for that complexity data, trust silos, lack of talent and the like, but the opportunity to transform data operations and drive tangible value is immense collaboration across various roles. And disciplines is part of the answer as is democratizing data. This means putting data in the hands of those domain experts that are closest to the customer and really understand where the opportunity exists and how to best address them. We heard from Jay Henderson that we have all this data exhaust and cheap storage. It allows us to keep it for a long time. It's true, but as he pointed out that doesn't solve the fundamental problem. Data is spewing out from our operational systems, but much of it lacks business context for the data teams chartered with analyzing that data. >>So we heard about the trend toward low code development and federating data access. The reason this is important is because the business lines have the context and the more responsibility they take for data, the more quickly and effectively organizations are going to be able to put data to work. We also talked about the harmonization between centralized teams and enabling decentralized data flows. I mean, after all data by its very nature is distributed. And importantly, as we heard from Adam Wilson and Suresh Vittol to support this model, you have to have strong governance and service the needs of it and engineering teams. And that's where the trifecta acquisition fits into the equation. Finally, we heard about a key partnership between Altrix and snowflake and how the migration to cloud data warehouses is evolving into a global data cloud. This enables data sharing across teams and ecosystems and vertical markets at massive scale all while maintaining the governance required to protect the organizations and individuals alike. >>This is a new and emerging business model that is very exciting and points the way to the next generation of data innovation in the coming decade. We're decentralized domain teams get more facile access to data. Self-service take more responsibility for quality value and data innovation. While at the same time, the governance security and privacy edicts of an organization are centralized in programmatically enforced throughout an enterprise and an external ecosystem. This is Dave Volante. All these videos are available on demand@theqm.net altrix.com. Thanks for watching accelerating automated analytics in the cloud made possible by Altryx. And thanks for watching the queue, your leader in enterprise tech coverage. We'll see you next time.

Published Date : Mar 1 2022

SUMMARY :

It saw the need to combine and prep different data types so that organizations anyone in the business who wanted to gain insights from data and, or let's say use AI without the post isolation economy is here and we do so with a digital We're kicking off the program with our first segment. So look, you have a deep product background, product management, product marketing, And that results in a situation where the organization's, you know, the direction that your customers want to go and the problems that you're solving, what role does the cloud and really, um, you know, create a lot of the underlying data sets that are used in some of this, into the, to the business user with hyper Anna. of our designer desktop product, you know, really, as they look to take the next step, comes into the mix that deeper it angle that we talked about, how does this all fit together? analytics and providing access to all these different groups of people, um, How much of this you've been able to share with your customers and maybe your partners. Um, and, and this idea that they're going to move from, you know, So it's democratizing data is the ultimate goal, which frankly has been elusive for most You know, the data gravity has been moving to the cloud. So, uh, you know, getting everyone involved and accessing AI and machine learning to unlock seems logical that domain leaders are going to take more responsibility for data, And I think, you know, the exciting thing for us at Altryx is, you know, we want to facilitate that. the tail, or maybe the other way around, you mentioned digital exhaust before. the data and analytics layers that they have, um, really to help democratize the We take a deep dive into the Altryx recent acquisition of Trifacta with Adam Wilson It's go time, get ready to accelerate your data analytics journey the CEO of Trifacta. serving business analysts and how the hyper Anna acquisition brought you deeper into the with that in mind, you know, we know designer and are the products And Joe in the early days, talked about flipping the model that really birth Trifacta was, you know, why is it that the people who know the data best can't And so, um, that was really, you know, what, you know, the origin story of the company but the big data pipeline is hasn't gotten there. um, you know, there hasn't been a single platform for And now the data engineer, which is really And so, um, I think when we, when I sat down with Suresh and with mark and the team and, but specifically we're seeing, you know, I almost don't even want to call it a data warehouse anyway, Um, and we just have interfaces to collaborate And of course Trifacta is with cloud cloud data warehouses. What's the business analysts really need and how to design a cloud, and Trifacta really support both in the cloud, um, you know, Trifacta becomes a platform that can You're always in a position to be able to cleanse transform shape structure, that data, and ultimately to deliver, And I'm interested, you guys just had your sales kickoff, you know, what was their reaction like? And then you step back and you're going to share the vision with the field organization, and to close and announced, you know, at the kickoff event. And certainly the reception we got from, Well, I think the story hangs together really well, you know, one of the better ones I've seen in, in this space, And all of it has potential the potential to solve complex business problems, We're now moving into the eco systems segment the power of many Good to see So cloud migration, it's one of the hottest topics. on snowflake to consolidate data across systems into one data cloud with Altryx business the partnership, maybe a little bit about the history, you know, what are the critical aspects that we should really focus Yeah, so the relationship started in 2020 and all shirts made a big bag deep with snowflake And the best practices guide is more of a technical document, bringing together experiences and guidance So customers can, can leverage that elastic platform, that being the snowflake data cloud, one of the problems that you guys solved early on, but what are some of the common challenges or patterns or trends everyone has access to data and everyone can do something with data, it's going to make them competitively, application that they have in order to be competitive in order to be competitive. to enrich with your own data sets, to enrich with, um, with your suppliers and with your partners, So thank you for that. So now that that Altrix is moving to the same model, And the launch of our cloud strategy How would you describe your joint go to market strategy the path to insights starting with your snowflake data. You'll go to market strategy. And so we shifted to an industry focus So that is going to be a way for us to allow What should we look for in the coming year? blueprints, and extending that at no charge to our partners with snowflake, we're already collaborating with Tarik will give you the last word. Um, the ability for Ultrix to plug into this extensibility framework that we call Barb Tara, thanks so much for coming on the program, got to leave it right there in a moment, I'll be back with 11.8 billion data points and one analytics platform to make sense of it all. This means putting data in the hands of those domain experts that are closest to the customer are going to be able to put data to work. While at the same time, the governance security and privacy edicts

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DerekPERSON

0.99+

Dave VellantePERSON

0.99+

Suresh VetolPERSON

0.99+

AltryxORGANIZATION

0.99+

JayPERSON

0.99+

Joe HellersteinPERSON

0.99+

DavePERSON

0.99+

Dave VolantePERSON

0.99+

AltrixORGANIZATION

0.99+

Jay HendersonPERSON

0.99+

DavidPERSON

0.99+

AdamPERSON

0.99+

BarbPERSON

0.99+

JeffPERSON

0.99+

2020DATE

0.99+

BobPERSON

0.99+

TrifactaORGANIZATION

0.99+

Suresh VittolPERSON

0.99+

TylerPERSON

0.99+

JuniperORGANIZATION

0.99+

AlteryxORGANIZATION

0.99+

UltrixORGANIZATION

0.99+

30 minutesQUANTITY

0.99+

TerikePERSON

0.99+

Adam WilsonPERSON

0.99+

JoePERSON

0.99+

SureshPERSON

0.99+

TerrickPERSON

0.99+

demand@theqm.netOTHER

0.99+

thousandsQUANTITY

0.99+

AlconORGANIZATION

0.99+

KaraPERSON

0.99+

last yearDATE

0.99+

threeQUANTITY

0.99+

QualtricsORGANIZATION

0.99+

less than 20%QUANTITY

0.99+

hundredsQUANTITY

0.99+

oneQUANTITY

0.99+

OneQUANTITY

0.99+

JavaTITLE

0.99+

more than seven yearsQUANTITY

0.99+

two acquisitionsQUANTITY

0.99+

Rahul Pathak, AWS | AWS re:Invent 2021


 

>>Hey, welcome back everyone. We're live here in the cube in Las Vegas Raiders reinvent 2021. I'm Jeffrey hosted the key we're in person this year. It's a hybrid event online. Great action. Going on. I'm rolling. Vice-president of ADF analytics. David is great to see you. Thanks for coming on. >>It's great to be here, John. Thanks for having me again. >>Um, so you've got a really awesome job. You've got serverless, you've got analytics. You're in the middle of all the action for AWS. What's the big news. What are you guys announcing? What's going on? >>Yeah, well, it's been an awesome reinvent for us. Uh, we've had a number of several us analytics launches. So red shift, our petabyte scale data warehouse, EMR for open source analytics. Uh, and then we've also had, uh, managed streaming for Kafka go serverless and then on demand for Kinesis. And then a couple of other big ones. We've got RO and cell based security for AWS lake formation. So you can get really fine grain controls over your data lakes and then asset transactions. You can actually have a inserts, updates and deletes on data lakes, which is a big step forward. >>Uh, so Swami on stage and the keynote he's actually finishing up now. But even last night I saw him in the hallway. We were talking about as much as about AI. Of course, he's got the AI title, but AI is the outcome. It's the application of all the data and this and a new architecture. He said on stage just now like, Hey, it's not about the old databases from the nineties, right? There's multiple data stores now available. And there's the unification is the big trend. And he said something interesting. Governance can be an advantage, not an inhibitor. This is kind of this new horizontally scalable, um, kind of idea that enables the vertical specialization around machine learning to be effective. It's not a new architecture, but it's now becoming more popular. People are realizing it. It's sort of share your thoughts on this whole not shift, but the acceleration of horizontally scalable and vertically integrated. Yeah, >>No, I think the way Swami put it is exactly right. What you want is the right tool for the right job. And you want to be able to deliver that to customers. So you're not compromising on performance or functionality of scale, but then you wanted all of these to be interconnected. So they're, well-integrated, you can stay in your favorite interface and take advantage of other technologies. So you can have things like Redshift integrated with Sage makers, you get analytics and machine learning. And then in Swami's absolutely right. Governance is actually an enabler of velocity. Once you've got the right guardrails in place, you can actually set people free because they can innovate. You don't have to be in the way, but you know that your data is protected. It's being used in the way that you expect by the people that you are allowing to use that data. And so it becomes a very powerful way for customers to set data free. And then, because things are elastic and serverless, uh, you can really just match capacity with demand. And so as you see spikes in usage, the system can scale out as those dwindle, they can scale back down, and it just becomes a very efficient way for customers to operate with data at scale >>Every year it reinvented. So it was kind of like a pinch me moment. It's like, well, more that's really good technology. Oh my God, it's getting easier and easier. As the infrastructure as code becomes more programmable, it's becoming easier, more Lambda, more serverless action. Uh, you got new offerings. How are customers benefiting for instance, from the three new offerings that you guys announced here? What specifically is the value proposition that you guys are putting out there? Yeah, so the, >>Um, you know, as we've tried to do with AWS over the years, customers get to focus on the things that really differentiate them and differentiate their businesses. So we take away in Redshift serverless, for example, all of the work that's needed to manage clusters, provision them, scale them, optimize them. Uh, and that's all been automated and made invisible to customers, the customers to think about data, what they want to do with it, what insights they can derive from it. And they know they're getting the most efficient infrastructure possible to make that a reality for them with high performance and low costs. So, uh, better results, more ability to focus on what differentiates their business and lower cost structure over time. >>Yeah. I had the essential guys on it's interesting. They had part of the soul cloud. Continuous is their word for what Adam was saying is clouds everywhere. And they're saying it's faster to match what you want to do with the outcomes, but the capabilities and outcomes kind of merging together where it's easy to say, this is what we want to do. And here's the outcome it supports that's right with that. What are some of the key trends on those outcomes that you see with the data analytics that's most popular right now? And kind of where's that, where's that going? >>Yeah. I mean, I think what we've seen is that data's just becoming more and more critical and top of mind for customers and, uh, you know, the pandemic has also accelerated that we found that customers are really looking to data and analytics and machine learning to find new opportunities. How can they, uh, really expand their business, take advantage of what's happening? And then the other part is how can they find efficiencies? And so, um, really everything that we're trying to do is we're trying to connect it to business outcomes for customers. How can you deepen your relationship with your customers? How can you create new customer experiences and how can you do that more efficiently, uh, with more agility and take advantage of, uh, the ability to be flexible. And you know, what is a very unpredictable world, as we've seen, >>I noticed a lot of purpose-built discussion going on in the keynote with Swami as well. How are you creating this next layer of what I call purpose-built platform like features? I mean, tools are great. You see a lot of tools in the data market tools are tools of your hammer. You want to look for a nail. We see people over by too many tools and you have ultimately a platform, but this seems to be a new trend where there's this connect phenomenon was showing me that you've got these platform capabilities that people can build on top of it, because there's a huge ecosystem of data tools out there that you guys have as partners that want to snap together. So the trend is things are starting to snap together, less primitive, roll your own, which you can do, but there's now more easier ways. Take me through that. Explain that, unpack that that phenomenon role rolling your own firm is, which has been the way now to here. Here's, here's some prefabricated software go. >>Yeah. Um, so it's a great observation and you're absolutely right. I mean, I think there's some customers that want to roll their own and they'll start with instances, they'll install software, they'll write their own code, build their own bespoke systems. And, uh, and we provide what the customers need to do that. But I think increasingly you're starting to see these higher level abstractions that take away all of that detail. And mark has Adam put it and allow customers to compose these. And we think it's important when you do that, uh, to be modular. So customers don't have to have these big bang all or nothing approaches you can pick what's appropriate, uh, but you're never on a dead end. You can always evolve and scale as you need to. And then you want to bring these ideas of unified governance and cohesive interfaces across so that customers find it easy to adopt the next thing. And so you can start off say with batch analytics, you can expand into real time. You can bring in machine learning and predictive capabilities. You can add natural language, and it's a big ecosystem of managed services as well as third parties and partners. >>And what's interesting. I want to get your thoughts while I got you here, because I think this is such an important trend and historic moment in time, Jerry chin, who one of the smartest VCs that we know from Greylock and coin castles in the cloud, which kind of came out of a cube conversation here in the queue years ago, where we saw the movement of that someone's going to build real value on AWS, not just an app. And you see the rise of the snowflakes and Databricks and other companies. And he was pointing out that you can get a very narrow wedge and get a position with these platforms, build on top of them and then build value. And I think that's, uh, the number one question people ask me, it's like, okay, how do I build value on top of these analytic packages? So if I'm a startup or I'm a big company, I also want to leverage these high level abstractions and build on top of it. How do you talk about that? How do you explain that? Because that's what people kind of want to know is like, okay, is it enabling me or do I have to fend for myself later? This is kind of, it comes up a lot. >>That's a great question. And, um, you know, if you saw, uh, Goldman's announcement this week, which is about bringing, building their cloud on top of AWS, it's a great example of using our capabilities in terms of infrastructure and analytics and machine learning to really allow them to take what's value added about Goldman and their position to financial markets, to build something value, add, and create a ton of value for Goldman, uh, by leveraging the things that we offer. And to us, that's an ideal outcome because it's a win-win for us in Goldman, but it's also a win for Goldman and their customers. >>That's what we call the Supercloud that's the opportunity. So is there a lot of Goldmans opportunities out there? Is that just a, these unicorns, are these sites? I mean, how do you, I mean, that's Goldman Sachs, they're huge. Is there, is this open to everybody? >>Absolutely. I mean, that's been one of the, uh, you know, one of the core ideas behind AWS was we wanted to give anybody any developer access to the same technology that the world's largest corporations had. And, uh, that's what you have today. The things that Goldman uses to build that cloud are available to anybody. And you can start for a few pennies scale up, uh, you know, into the petabytes and beyond >>When I was talking to Adams, Lipski when I met with him prior to re-invent, I noticed that he was definitely had an affinity towards the data, obviously he's Amazonia, but he spent time at Tableau. So, so as he's running that company, so you see that kind of mindset of the data advantage. So I have to ask you, because it's something that I've been talking about for a while and I'm waiting for it to emerge, but I'm not sure it's going to happen yet. But what infrastructure is code was for dev ops and then dev sec ops, there's almost like a data ops developing where data as code or programmable data. If I can connect the dots of what Swami's saying, what you're doing is this is like a new horizontal layer of data of freely available data with some government governance built in that's right. So it's, data's being baked into everything. So data is any ingredient, not a query to some database, it's gotta be baked into the apps, that's data as code that's. Right. So it's almost a data DevOps kind of vibe. >>Yeah, no, you're absolutely right. And you know, you've seen it with things like ML ops and so on. It's all the special case of dev ops. But what you're really trying to do is to get programmatic and systematic about how you deal with data. And it's not just data that you have. It's also publicly available data sets and it's customers sharing with each other. So building the ecosystem, our data, and we've got things like our open data program where we've got publicly hosted data sets or things like the AWS data exchange where customers can actually monetize data. So it's not just data as code, but now data as a monetizeable asset. So it's a really exciting time to be in the data business. >>Yeah. And I think it's so many too. So I've got to ask you while I got you here since you're an expert. Um, okay. Here's my problem. I have a lot of data. I'm nervous about it. I want to secure it. So if I try to secure it, I'm not making it available. So I want to feed the machine learning. How do I create an architecture where I can make it freely available, but yet maintain the control and the comfort that this is going to be secure. So what products do I buy? >>Yeah. So, uh, you know, a great place to start at as three. Um, you know, it's one of the best places for data lakes, uh, for all the reasons. That's why we talked about your ability scale costs. You can then use lake formation to really protect and govern that data so you can decide who's allowed to see it and what they're allowed to see, and you don't have to create multiple copies. So you can define that, you know, this group of partners can see a, B and C. This group can see D E and F and the system enforces that. And you have a central point of control where you can monitor what's happening. And if you want to change your mind, you can do that instantly. And all access can be locked down that you've got a variety of encryption capabilities with things like KMS. And so you can really lock down your data, but yet keep it open to the parties that you want and give them specifically the access that you want to give them. And then once you've done that, they're free to use that data, according to the rules that you defined with the analytics tools that we offer to go drive value, create insight, and do something >>That's lake formation. And then you got a Thena querying. Yes, we got all kinds of tooling on top of it. >>It's all right. You can have, uh, Athena query and your data in S3 lake formation, protecting it. And then SageMaker is integrated with Athena. So you can pull that data into SageMaker for machine learning, interrogate that data, using natural language with things like QuickSight Q a like we demoed. So just a ton of power without having to really think too deeply about, uh, developing expert skill sets in this. >>So the next question I want to ask you is because that first part of the great, great, great description, thank you very much. Now, 5g in the edges here, outpost, how was the analytics going on that as edge becomes more pervasive in the architecture? >>Yeah, it's going to be a key part of this ecosystem and it's really a continuum. So, uh, you know, we find customers are collecting data at the edge. They might be making local ML or inference type decisions on edge devices, or, you know, automobiles, for example. Uh, but typically that data with some point will come back into the cloud, into S3 will be used to do heavy duty training, and then those models get pushed back out to the edge. And then some of the things that we've done in Athena, for example, with federated query, as long as you have a network path, and you can understand what the data format or the database is, you can actually run a query on that data. So you can run real-time queries on data, wherever it lives, whether it's on an edge device, on an outpost, in a local zone or in your cloud region and combine all of that together in one place. >>Yeah. And I think having that data copies everywhere is a big thing deal. I've got to ask you now that we're here at reinvent, what's your take we're back in person last year was all virtual. Finally, not 60,000 people, like a couple of years ago, it's still 27,000 people here, all lining up for the sessions, all having a great time. Um, all good. What's the most important story from your, your area that people should pay attention to? What's the headline, what's the top news? What should people pay attention to? >>Yeah, so I think first off it is awesome to be back in person. It's just so fun to see customers and to see, I mean, you, like, we've been meeting here over the years and it's, it's great to so much energy in person. It's been really nice. Uh, you know, I think from an analytics perspective, there's just been a ton of innovation. I think the core idea for us is we want to make it easy for customers to use the right tool for the right job to get insight from all of their data as cost effectively as possible. And I think, uh, you know, I think if customers walk away and think about it as being, it's now easier than ever for me to take advantage of everything that AWS has to offer, uh, to make sense of all the data that I'm generating and use it to drive business value, but I think we'll have done our jobs. Right. >>What's the coolest thing that you're seeing here is that the serverless innovation, is it, um, the new abstraction layer with data high level services in your mind? What's the coolest thing. Got it. >>It's hard to pick the coolest that sticks like kicking the candies. I mean, I think the, uh, you know, the continued innovation in terms of, uh, performance and functionality in each of our services is a big deal. I think serverless is a game changer for customers. Uh, and then I think really the infusion of machine learning throughout all of these systems. So things like Redshift ML, Athena ML, Pixar, Q a just really enabling new experiences for customers, uh, in a way that's easier than it ever has been. And I think that's a, that's a big deal and I'm really excited to see what customers do with it. >>Yeah. And I think the performance thing to me, the coolest thing that I'm seeing is the graviton three and the gravitron progression with the custom stacks with all this ease of use, it's just going to be just a real performance advantage and the costs are getting lowered. So I think the ECE two instances around the compute is phenomenal. No, >>Absolutely. I mean, I think the hardware and Silicon innovation is huge and it's not just performance. It's also the energy efficiency. It's a big deal for the future reality. >>We're at an inflection point where this modern applications are being built. And in my history, I'm old, my birthday is today. I'm in my fifties. So I remember back in the eighties, every major inflection point when there was a shift in how things were developed from mainframe client server, PC inter network, you name it every time the apps change, the app owners, app developers all went to the best platform processing. And so I think, you know, that idea of system software applications being bundled together, um, is a losing formula. I think you got to have that decoupling large-scale was seeing that with cloud. And I think now if I'm an app developer, whether whether I'm in a large ISV in your ecosystem or in the APN partner or a startup, I'm going to go with my software runs the best period and where I can create value. That's right. I get distribution, I create value and it runs fast. I mean, that's, I mean, it's pretty simple. So I think the ecosystem is going to be a big action for the next couple of years. >>Absolutely. Right. And I mean, the ecosystem's huge and I think, um, and we're also grateful to have all these partners here. It's a huge deal for us. And I think it really matters for customers >>What's on your roadmap this year, what you got going on. What can you share a little bit of a trajectory without kind of, uh, breaking the rules of the Amazonian, uh, confidentiality. Um, what's, what's the focus for the year? What do you what's next? >>Well, you know, as you know, we're always talking to customers and, uh, I think we're going to make things better, faster, cheaper, easier to use. And, um, I think you've seen some of the things that we're doing with integration now, you'll see more of that. And, uh, really the goal is how can customers get value as quickly as possible for as low cost as possible? That's how we went to >>Yeah. They're in the longterm. Yeah. We've always say every time we see each other data is at the center of the value proposition. I've been saying that for 10 years now, it's actually the value proposition, powering AI. And you're seeing because of it, the rise of superclouds and then the superclouds are emerging. I think you guys are the under innings of these emerging superclouds. And so it's a huge treading, the Goldman Sachs things of validation. So again, more data, the better, sorry, cool things happening. >>It is just it's everywhere. And the, uh, the diversity of use cases is amazing. I mean, I think from, you know, the Australia swimming team to, uh, to formula one to NASDAQ, it's just incredible to see what our >>Customers do. We see the great route. Good to see you. Thanks for coming on the cube. >>Pleasure to be here as always John. Great to see you. Thank you. Yeah. >>Thanks for, thanks for sharing. All of the data is the key to the success. Data is the value proposition. You've seen the rise of superclouds because of the data advantage. If you can expose it, protect it and govern it, unleashes creativity and opportunities for entrepreneurs and businesses. Of course, you got to have the scale and the price performance. That's what doing this is the cube coverage. You're watching the leader in worldwide tech coverage here in person for any of us reinvent 2021 I'm John ferry. Thanks for watching.

Published Date : Dec 1 2021

SUMMARY :

David is great to see you. It's great to be here, John. What are you guys announcing? So you can get really fine grain controls over your data lakes and then asset transactions. It's the application of all the data and this and a new architecture. And so as you see spikes in usage, the system can scale out How are customers benefiting for instance, from the three new offerings that you guys announced the customers to think about data, what they want to do with it, what insights they can derive from it. And they're saying it's faster to match what you want to do with the outcomes, And you know, what is a very unpredictable world, as we've seen, tools out there that you guys have as partners that want to snap together. So customers don't have to have these big bang all or nothing approaches you can pick And he was pointing out that you can get a very narrow wedge and get a position And, um, you know, if you saw, uh, Goldman's announcement this week, Is there, is this open to everybody? I mean, that's been one of the, uh, you know, one of the core ideas behind AWS was we wanted to give so you see that kind of mindset of the data advantage. And it's not just data that you have. So I've got to ask you while I got you here since you're an expert. And so you can really lock down your data, but yet And then you got a Thena querying. So you can pull that data into SageMaker for machine learning, So the next question I want to ask you is because that first part of the great, great, great description, thank you very much. data format or the database is, you can actually run a query on that data. I've got to ask you now that we're here at reinvent, And I think, uh, you know, I think if customers walk away and think about it as being, What's the coolest thing that you're seeing here is that the serverless innovation, I think the, uh, you know, the continued innovation in terms of, uh, So I think the ECE two instances around the compute is phenomenal. It's a big deal for the future reality. And so I think, you know, And I think it really matters for customers What can you share a little bit of a trajectory without kind of, Well, you know, as you know, we're always talking to customers and, uh, I think we're going to make things better, I think you guys are the under innings of these emerging superclouds. I mean, I think from, you know, the Australia swimming team to, uh, to formula one to NASDAQ, Thanks for coming on the cube. Great to see you. All of the data is the key to the success.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

GoldmanORGANIZATION

0.99+

Rahul PathakPERSON

0.99+

AWSORGANIZATION

0.99+

JohnPERSON

0.99+

Goldman SachsORGANIZATION

0.99+

AdamPERSON

0.99+

Jerry chinPERSON

0.99+

NASDAQORGANIZATION

0.99+

AthenaLOCATION

0.99+

JeffreyPERSON

0.99+

2021DATE

0.99+

60,000 peopleQUANTITY

0.99+

27,000 peopleQUANTITY

0.99+

10 yearsQUANTITY

0.99+

last yearDATE

0.99+

John ferryPERSON

0.99+

todayDATE

0.99+

threeQUANTITY

0.98+

KafkaTITLE

0.98+

SwamiPERSON

0.98+

oneQUANTITY

0.98+

firstQUANTITY

0.98+

ADF analyticsORGANIZATION

0.97+

AmazoniaORGANIZATION

0.97+

eightiesDATE

0.97+

PixarORGANIZATION

0.97+

fiftiesQUANTITY

0.97+

eachQUANTITY

0.97+

three new offeringsQUANTITY

0.96+

RedshiftTITLE

0.96+

first partQUANTITY

0.96+

this yearDATE

0.96+

last nightDATE

0.95+

LipskiPERSON

0.94+

couple of years agoDATE

0.94+

next couple of yearsDATE

0.94+

SageORGANIZATION

0.93+

GoldmansORGANIZATION

0.92+

pandemicEVENT

0.92+

this weekDATE

0.92+

DatabricksORGANIZATION

0.9+

one placeQUANTITY

0.9+

markPERSON

0.88+

SupercloudORGANIZATION

0.87+

TableauORGANIZATION

0.84+

S3TITLE

0.84+

5gQUANTITY

0.8+

Athena MLORGANIZATION

0.78+

AthenaORGANIZATION

0.78+

RaidersORGANIZATION

0.77+

years agoDATE

0.75+

ninetiesDATE

0.74+

MLORGANIZATION

0.73+

AdamsPERSON

0.73+

two instancesQUANTITY

0.72+

LambdaTITLE

0.71+

ThenaORGANIZATION

0.71+

SageMakerTITLE

0.66+

VegasLOCATION

0.66+

InventEVENT

0.64+

VicePERSON

0.63+

graviton threeOTHER

0.62+

AustraliaLOCATION

0.59+

LasORGANIZATION

0.59+

KinesisTITLE

0.57+

AmazonianORGANIZATION

0.56+

Pat Conte, Opsani | AWS Startup Showcase


 

(upbeat music) >> Hello and welcome to this CUBE conversation here presenting the "AWS Startup Showcase: "New Breakthroughs in DevOps, Data Analytics "and Cloud Management Tools" featuring Opsani for the cloud management and migration track here today, I'm your host John Furrier. Today, we're joined by Patrick Conte, Chief Commercial Officer, Opsani. Thanks for coming on. Appreciate you coming on. Future of AI operations. >> Thanks, John. Great to be here. Appreciate being with you. >> So congratulations on all your success being showcased here as part of the Startups Showcase, future of AI operations. You've got the cloud scale happening. A lot of new transitions in this quote digital transformation as cloud scales goes next generation. DevOps revolution as Emily Freeman pointed out in her keynote. What's the problem statement that you guys are focused on? Obviously, AI involves a lot of automation. I can imagine there's a data problem in there somewhere. What's the core problem that you guys are focused on? >> Yeah, it's interesting because there are a lot of companies that focus on trying to help other companies optimize what they're doing in the cloud, whether it's cost or whether it's performance or something else. We felt very strongly that AI was the way to do that. I've got a slide prepared, and maybe we can take a quick look at that, and that'll talk about the three elements or dimensions of the problem. So we think about cloud services and the challenge of delivering cloud services. You've really got three things that customers are trying to solve for. They're trying to solve for performance, they're trying to solve for the best performance, and, ultimately, scalability. I mean, applications are growing really quickly especially in this current timeframe with cloud services and whatnot. They're trying to keep costs under control because certainly, it can get way out of control in the cloud since you don't own the infrastructure, and more importantly than anything else which is why it's at the bottom sort of at the foundation of all this, is they want their applications to be a really a good experience for their customers. So our customer's customer is actually who we're trying to solve this problem for. So what we've done is we've built a platform that uses AI and machine learning to optimize, meaning tune, all of the key parameters of a cloud application. So those are things like the CPU usage, the memory usage, the number of replicas in a Kubernetes or container environment, those kinds of things. It seems like it would be simple just to grab some values and plug 'em in, but it's not. It's actually the combination of them has to be right. Otherwise, you get delays or faults or other problems with the application. >> Andrew, if you can bring that slide back up for a second. I want to just ask one quick question on the problem statement. You got expenditures, performance, customer experience kind of on the sides there. Do you see this tip a certain way depending upon use cases? I mean, is there one thing that jumps out at you, Patrick, from your customer's customer's standpoint? Obviously, customer experience is the outcome. That's the app, whatever. That's whatever we got going on there. >> Sure. >> But is there patterns 'cause you can have good performance, but then budget overruns. Or all of them could be failing. Talk about this dynamic with this triangle. >> Well, without AI, without machine learning, you can solve for one of these, only one, right? So if you want to solve for performance like you said, your costs may overrun, and you're probably not going to have control of the customer experience. If you want to solve for one of the others, you're going to have to sacrifice the other two. With machine learning though, we can actually balance that, and it isn't a perfect balance, and the question you asked is really a great one. Sometimes, you want to over-correct on something. Sometimes, scalability is more important than cost, but what we're going to do because of our machine learning capability, we're going to always make sure that you're never spending more than you should spend, so we're always going to make sure that you have the best cost for whatever the performance and reliability factors that you you want to have are. >> Yeah, I can imagine. Some people leave services on. Happened to us one time. An intern left one of the services on, and like where did that bill come from? So kind of looked back, we had to kind of fix that. There's a ton of action, but I got to ask you, what are customers looking for with you guys? I mean, as they look at Opsani, what you guys are offering, what's different than what other people might be proposing with optimization solutions? >> Sure. Well, why don't we bring up the second slide, and this'll illustrate some of the differences, and we can talk through some of this stuff as well. So really, the area that we play in is called AIOps, and that's sort of a new area, if you will, over the last few years, and really what it means is applying intelligence to your cloud operations, and those cloud operations could be development operations, or they could be production operations. And what this slide is really representing is in the upper slide, that's sort of the way customers experience their DevOps model today. Somebody says we need an application or we need a feature, the developers pull down something from get. They hack an early version of it. They run through some tests. They size it whatever way they know that it won't fail, and then they throw it over to the SREs to try to tune it before they shove it out into production, but nobody really sizes it properly. It's not optimized, and so it's not tuned either. When it goes into production, it's just the first combination of settings that work. So what happens is undoubtedly, there's some type of a problem, a fault or a delay, or you push new code, or there's a change in traffic. Something happens, and then, you've got to figure out what the heck. So what happens then is you use your tools. First thing you do is you over-provision everything. That's what everybody does, they over-provision and try to soak up the problem. But that doesn't solve it because now, your costs are going crazy. You've got to go back and find out and try as best you can to get root cause. You go back to the tests, and you're trying to find something in the test phase that might be an indicator. Eventually your developers have to hack a hot fix, and the conveyor belt sort of keeps on going. We've tested this model on every single customer that we've spoken to, and they've all said this is what they experience on a day-to-day basis. Now, if we can go back to the side, let's talk about the second part which is what we do and what makes us different. So on the bottom of this slide, you'll see it's really a shift-left model. What we do is we plug in in the production phase, and as I mentioned earlier, what we're doing is we're tuning all those cloud parameters. We're tuning the CPU, the memory, the Replicas, all those kinds of things. We're tuning them all in concert, and we're doing it at machine speed, so that's how the customer gets the best performance, the best reliability at the best cost. That's the way we're able to achieve that is because we're iterating this thing in machine speed, but there's one other place where we plug in and we help the whole concept of AIOps and DevOps, and that is we can plug in in the test phase as well. And so if you think about it, the DevOps guy can actually not have to over-provision before he throws it over to the SREs. He can actually optimize and find the right size of the application before he sends it through to the SREs, and what this does is collapses the timeframe because it means the SREs don't have to hunt for a working set of parameters. They get one from the DevOps guys when they send it over, and this is how the future of AIOps is being really affected by optimization and what we call autonomous optimization which means that it's happening without humans having to press a button on it. >> John: Andrew, bring that slide back up. I want to just ask another question. Tuning in concert thing is very interesting to me. So how does that work? Are you telegraphing information to the developer from the autonomous workload tuning engine piece? I mean, how does the developer know the right knobs or where does it get that provisioning information? I see the performance lag. I see where you're solving that problem. >> Sure. >> How does that work? >> Yeah, so actually, if we go to the next slide, I'll show you exactly how it works. Okay, so this slide represents the architecture of a typical application environment that we would find ourselves in, and inside the dotted line is the customer's application namespace. That's where the app is. And so, it's got a bunch of pods. It's got a horizontal pod. It's got something for replication, probably an HPA. And so, what we do is we install inside that namespace two small instances. One is a tuning pod which some people call a canary, and that tuning pod joins the rest of the pods, but it's not part of the application. It's actually separate, but it gets the same traffic. We also install somebody we call Servo which is basically an action engine. What Servo does is Servo takes the metrics from whatever the metric system is is collecting all those different settings and whatnot from the working application. It could be something like Prometheus. It could be an Envoy Sidecar, or more likely, it's something like AppDynamics, or we can even collect metrics off of Nginx which is at the front of the service. We can plug into anywhere where those metrics are. We can pull the metrics forward. Once we see the metrics, we send them to our backend. The Opsani SaaS service is our machine learning backend. That's where all the magic happens, and what happens then is that service sees the settings, sends a recommendation to Servo, Servo sends it to the tuning pod, and we tune until we find optimal. And so, that iteration typically takes about 20 steps. It depends on how big the application is and whatnot, how fast those steps take. It could be anywhere from seconds to minutes to 10 to 20 minutes per step, but typically within about 20 steps, we can find optimal, and then we'll come back and we'll say, "Here's optimal, and do you want to "promote this to production," and the customer says, "Yes, I want to promote it to production "because I'm saving a lot of money or because I've gotten "better performance or better reliability." Then, all he has to do is press a button, and all that stuff gets sent right to the production pods, and all of those settings get put into production, and now he's now he's actually saving the money. So that's basically how it works. >> It's kind of like when I want to go to the beach, I look at the weather.com, I check the forecast, and I decide whether I want to go or not. You're getting the data, so you're getting a good look at the information, and then putting that into a policy standpoint. I get that, makes total sense. Can I ask you, if you don't mind, expanding on the performance and reliability and the cost advantage? You mentioned cost. How is that impacting? Give us an example of some performance impact, reliability, and cost impacts. >> Well, let's talk about what those things mean because like a lot of people might have different ideas about what they think those mean. So from a cost standpoint, we're talking about cloud spend ultimately, but it's represented by the settings themselves, so I'm not talking about what deal you cut with AWS or Azure or Google. I'm talking about whatever deal you cut, we're going to save you 30, 50, 70% off of that. So it doesn't really matter what cost you negotiated. What we're talking about is right-sizing the settings for CPU and memory, Replica. Could be Java. It could be garbage collection, time ratios, or heap sizes or things like that. Those are all the kinds of things that we can tune. The thing is most of those settings have an unlimited number of values, and this is why machine learning is important because, if you think about it, even if they only had eight settings or eight values per setting, now you're talking about literally billions of combinations. So to find optimal, you've got to have machine speed to be able to do it, and you have to iterate very, very quickly to make it happen. So that's basically the thing, and that's really one of the things that makes us different from anybody else, and if you put that last slide back up, the architecture slide, for just a second, there's a couple of key words at the bottom of it that I want to want to focus on, continuous. So continuous really means that we're on all the time. We're not plug us in one time, make a change, and then walk away. We're actually always measuring and adjusting, and the reason why this is important is in the modern DevOps world, your traffic level is going to change. You're going to push new code. Things are going to happen that are going to change the basic nature of the software, and you have to be able to tune for those changes. So continuous is very important. Second thing is autonomous. This is designed to take pressure off of the SREs. It's not designed to replace them, but to take the pressure off of them having to check pager all the time and run in and make adjustments, or try to divine or find an adjustment that might be very, very difficult for them to do so. So we're doing it for them, and that scale means that we can solve this for, let's say, one big monolithic application, or we can solve it for literally hundreds of applications and thousands of microservices that make up those applications and tune them all at the same time. So the same platform can be used for all of those. You originally asked about the parameters and the settings. Did I answer the question there? >> You totally did. I mean, the tuning in concert. You mentioned early as a key point. I mean, you're basically tuning the engine. It's not so much negotiating a purchase SaaS discount. It's essentially cost overruns by the engine, either over burning or heating or whatever you want to call it. I mean, basically inefficiency. You're tuning the core engine. >> Exactly so. So the cost thing is I mentioned is due to right-sizing the settings and the number of Replicas. The performance is typically measured via latency, and the reliability is typically measured via error rates. And there's some other measures as well. We have a whole list of them that are in the application itself, but those are the kinds of things that we look for as results. When we do our tuning, we look for reducing error rates, or we look for holding error rates at zero, for example, even if we improve the performance or we improve the cost. So we're looking for the best result, the best combination result, and then a customer can decide if they want to do so to actually over-correct on something. We have the whole concept of guard rail, so if performance is the most important thing, or maybe some customers, cost is the most important thing, they can actually say, "Well, give us the best cost, "and give us the best performance and the best reliability, "but at this cost," and we can then use that as a service-level objective and tune around it. >> Yeah, it reminds me back in the old days when you had filtering white lists of black lists of addresses that can go through, say, a firewall or a device. You have billions of combinations now with machine learning. It's essentially scaling the same concept to unbelievable. These guardrails are now in place, and that's super cool and I think really relevant call-out point, Patrick, to kind of highlight that. At this kind of scale, you need machine learning, you need the AI to essentially identify quickly the patterns or combinations that are actually happening so a human doesn't have to waste their time that can be filled by basically a bot at that point. >> So John, there's just one other thing I want to mention around this, and that is one of the things that makes us different from other companies that do optimization. Basically, every other company in the optimization space creates a static recommendation, basically their recommendation engines, and what you get out of that is, let's say it's a manifest of changes, and you hand that to the SREs, and they put it into effect. Well, the fact of the matter is is that the traffic could have changed then. It could have spiked up, or it could have dropped below normal. You could have introduced a new feature or some other code change, and at that point in time, you've already instituted these changes. They may be completely out of date. That's why the continuous nature of what we do is important and different. >> It's funny, even the language that we're using here: network, garbage collection. I mean, you're talking about tuning an engine, am operating system. You're talking about stuff that's moving up the stack to the application layer, hence this new kind of eliminating of these kind of siloed waterfall, as you pointed out in your second slide, is kind of one integrated kind of operating environment. So when you have that or think about the data coming in, and you have to think about the automation just like self-correcting, error-correcting, tuning, garbage collection. These are words that we've kind of kicking around, but at the end of the day, it's an operating system. >> Well in the old days of automobiles, which I remember cause I'm I'm an old guy, if you wanted to tune your engine, you would probably rebuild your carburetor and turn some dials to get the air-oxygen-gas mix right. You'd re-gap your spark plugs. You'd probably make sure your points were right. There'd be four or five key things that you would do. You couldn't do them at the same time unless you had a magic wand. So we're the magic wand that basically, or in modern world, we're sort of that thing you plug in that tunes everything at once within that engine which is all now electronically controlled. So that's the big differences as you think about what we used to do manually, and now, can be done with automation. It can be done much, much faster without humans having to get their fingernails greasy, let's say. >> And I think the dynamic versus static is an interesting point. I want to bring up the SRE which has become a role that's becoming very prominent in the DevOps kind of plus world that's happening. You're seeing this new revolution. The role of the SRE is not just to be there to hold down and do the manual configuration. They had a scale. They're a developer, too. So I think this notion of offloading the SRE from doing manual tasks is another big, important point. Can you just react to that and share more about why the SRE role is so important and why automating that away through when you guys have is important? >> The SRE role is becoming more and more important, just as you said, and the reason is because somebody has to get that application ready for production. The DevOps guys don't do it. That's not their job. Their job is to get the code finished and send it through, and the SREs then have to make sure that that code will work, so they have to find a set of settings that will actually work in production. Once they find that set of settings, the first one they find that works, they'll push it through. It's not optimized at that point in time because they don't have time to try to find optimal, and if you think about it, the difference between a machine learning backend and an army of SREs that work 24-by-seven, we're talking about being able to do the work of many, many SREs that never get tired, that never need to go play video games, to unstress or whatever. We're working all the time. We're always measuring, adjusting. A lot of the companies we talked to do a once-a-month adjustment on their software. So they put an application out, and then they send in their SREs once a month to try to tune the application, and maybe they're using some of these other tools, or maybe they're using just their smarts, but they'll do that once a month. Well, gosh, they've pushed code probably four times during the month, and they probably had a bunch of different spikes and drops in traffic and other things that have happened. So we just want to help them spend their time on making sure that the application is ready for production. Want to make sure that all the other parts of the application are where they should be, and let us worry about tuning CPU, memory, Replica, job instances, and things like that so that they can work on making sure that application gets out and that it can scale, which is really important for them, for their companies to make money is for the apps to scale. >> Well, that's a great insight, Patrick. You mentioned you have a lot of great customers, and certainly if you have your customer base are early adopters, pioneers, and grow big companies because they have DevOps. They know that they're seeing a DevOps engineer and an SRE. Some of the other enterprises that are transforming think the DevOps engineer is the SRE person 'cause they're having to get transformed. So you guys are at the high end and getting now the new enterprises as they come on board to cloud scale. You have a huge uptake in Kubernetes, starting to see the standardization of microservices. People are getting it, so I got to ask you can you give us some examples of your customers, how they're organized, some case studies, who uses you guys, and why they love you? >> Sure. Well, let's bring up the next slide. We've got some customer examples here, and your viewers, our viewers, can probably figure out who these guys are. I can't tell them, but if they go on our website, they can sort of put two and two together, but the first one there is a major financial application SaaS provider, and in this particular case, they were having problems that they couldn't diagnose within the stack. Ultimately, they had to apply automation to it, and what we were able to do for them was give them a huge jump in reliability which was actually the biggest problem that they were having. We gave them 5,000 hours back a month in terms of the application. They were they're having pager duty alerts going off all the time. We actually gave them better performance. We gave them a 10% performance boost, and we dropped their cloud spend for that application by 72%. So in fact, it was an 80-plus % price performance or cost performance improvement that we gave them, and essentially, we helped them tune the entire stack. This was a hybrid environment, so this included VMs as well as more modern architecture. Today, I would say the overwhelming majority of our customers have moved off of the VMs and are in a containerized environment, and even more to the point, Kubernetes which we find just a very, very high percentage of our customers have moved to. So most of the work we're doing today with new customers is around that, and if we look at the second and third examples here, those are examples of that. In the second example, that's a company that develops websites. It's one of the big ones out in the marketplace that, let's say, if you were starting a new business and you wanted a website, they would develop that website for you. So their internal infrastructure is all brand new stuff. It's all Kubernetes, and what we were able to do for them is they were actually getting decent performance. We held their performance at their SLO. We achieved a 100% error-free scenario for them at runtime, and we dropped their cost by 80%. So for them, they needed us to hold-serve, if you will, on performance and reliability and get their costs under control because everything in that, that's a cloud native company. Everything there is cloud cost. So the interesting thing is it took us nine steps because nine of our iterations to actually get to optimal. So it was very, very quick, and there was no integration required. In the first case, we actually had to do a custom integration for an underlying platform that was used for CICD, but with the- >> John: Because of the hybrid, right? >> Patrick: Sorry? >> John: Because it was hybrid, right? >> Patrick: Yes, because it was hybrid, exactly. But within the second one, we just plugged right in, and we were able to tune the Kubernetes environment just as I showed in that architecture slide, and then the third one is one of the leading application performance monitoring companies on the market. They have a bunch of their own internal applications and those use a lot of cloud spend. They're actually running Kubernetes on top of VMs, but we don't have to worry about the VM layer. We just worry about the Kubernetes layer for them, and what we did for them was we gave them a 48% performance improvement in terms of latency and throughput. We dropped their error rates by 90% which is pretty substantial to say the least, and we gave them a 50% cost delta from where they had been. So this is the perfect example of actually being able to deliver on all three things which you can't always do. It has to be, sort of all applications are not created equal. This was one where we were able to actually deliver on all three of the key objectives. We were able to set them up in about 25 minutes from the time we got started, no extra integration, and needless to say, it was a big, happy moment for the developers to be able to go back to their bosses and say, "Hey, we have better performance, "better reliability. "Oh, by the way, we saved you half." >> So depending on the stack situation, you got VMs and Kubernetes on the one side, cloud-native, all Kubernetes, that's dream scenario obviously. Not many people like that. All the new stuff's going cloud-native, so that's ideal, and then the mixed ones, Kubernetes, but no VMs, right? >> Yeah, exactly. So Kubernetes with no VMs, no problem. Kubernetes on top of VMs, no problem, but we don't manage the VMs. We don't manage the underlay at all, in fact. And the other thing is we don't have to go back to the slide, but I think everybody will remember the slide that had the architecture, and on one side was our cloud instance. The only data that's going between the application and our cloud instance are the settings, so there's never any data. There's never any customer data, nothing for PCI, nothing for HIPPA, nothing for GDPR or any of those things. So no personal data, no health data. Nothing is passing back and forth. Just the settings of the containers. >> Patrick, while I got you here 'cause you're such a great, insightful guest, thank you for coming on and showcasing your company. Kubernetes real quick. How prevalent is this mainstream trend is because you're seeing such great examples of performance improvements. SLAs being met, SLOs being met. How real is Kubernetes for the mainstream enterprise as they're starting to use containers to tip their legacy and get into the cloud-native and certainly hybrid and soon to be multi-cloud environment? >> Yeah, I would not say it's dominant yet. Of container environments, I would say it's dominant now, but for all environments, it's not. I think the larger legacy companies are still going through that digital transformation, and so what we do is we catch them at that transformation point, and we can help them develop because as we remember from the AIOps slide, we can plug in at that test level and help them sort of pre-optimize as they're coming through. So we can actually help them be more efficient as they're transforming. The other side of it is the cloud-native companies. So you've got the legacy companies, brick and mortar, who are desperately trying to move to digitization. Then, you've got the ones that are born in the cloud. Most of them aren't on VMs at all. Most of them are on containers right from the get-go, but you do have some in the middle who have started to make a transition, and what they've done is they've taken their native VM environment and they've put Kubernetes on top of it so that way, they don't have to scuttle everything underneath it. >> Great. >> So I would say it's mixed at this point. >> Great business model, helping customers today, and being a bridge to the future. Real quick, what licensing models, how to buy, promotions you have for Amazon Web Services customers? How do people get involved? How do you guys charge? >> The product is licensed as a service, and the typical service is an annual. We license it by application, so let's just say you have an application, and it has 10 microservices. That would be a standard application. We'd have an annual cost for optimizing that application over the course of the year. We have a large application pack, if you will, for let's say applications of 20 services, something like that, and then we also have a platform, what we call Opsani platform, and that is for environments where the customer might have hundreds of applications and-or thousands of services, and we can plug into their deployment platform, something like a harness or Spinnaker or Jenkins or something like that, or we can plug into their their cloud Kubernetes orchestrator, and then we can actually discover the apps and optimize them. So we've got environments for both single apps and for many, many apps, and with the same platform. And yes, thanks for reminding me. We do have a promotion for for our AWS viewers. If you reference this presentation, and you look at the URL there which is opsani.com/awsstartupshowcase, can't forget that, you will, number one, get a free trial of our software. If you optimize one of your own applications, we're going to give you an Oculus set of goggles, the augmented reality goggles. And we have one other promotion for your viewers and for our joint customers here, and that is if you buy an annual license, you're going to get actually 15 months. So that's what we're putting on the table. It's actually a pretty good deal. The Oculus isn't contingent. That's a promotion. It's contingent on you actually optimizing one of your own services. So it's not a synthetic app. It's got to be one of your own apps, but that's what we've got on the table here, and I think it's a pretty good deal, and I hope your guys take us up on it. >> All right, great. Get Oculus Rift for optimizing one of your apps and 15 months for the price of 12. Patrick, thank you for coming on and sharing the future of AIOps with you guys. Great product, bridge to the future, solving a lot of problems. A lot of use cases there. Congratulations on your success. Thanks for coming on. >> Thank you so much. This has been excellent, and I really appreciate it. >> Hey, thanks for sharing. I'm John Furrier, your host with theCUBE. Thanks for watching. (upbeat music)

Published Date : Sep 22 2021

SUMMARY :

for the cloud management and Appreciate being with you. of the Startups Showcase, and that'll talk about the three elements kind of on the sides there. 'cause you can have good performance, and the question you asked An intern left one of the services on, and find the right size I mean, how does the and the customer says, and the cost advantage? and that's really one of the things I mean, the tuning in concert. So the cost thing is I mentioned is due to in the old days when you had and that is one of the things and you have to think about the automation So that's the big differences of offloading the SRE and the SREs then have to make sure and certainly if you So most of the work we're doing today "Oh, by the way, we saved you half." So depending on the stack situation, and our cloud instance are the settings, and get into the cloud-native that are born in the cloud. So I would say it's and being a bridge to the future. and the typical service is an annual. and 15 months for the price of 12. and I really appreciate it. I'm John Furrier, your host with theCUBE.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Emily FreemanPERSON

0.99+

PatrickPERSON

0.99+

JohnPERSON

0.99+

AndrewPERSON

0.99+

John FurrierPERSON

0.99+

Pat ContePERSON

0.99+

10%QUANTITY

0.99+

50%QUANTITY

0.99+

Patrick ContePERSON

0.99+

15 monthsQUANTITY

0.99+

secondQUANTITY

0.99+

90%QUANTITY

0.99+

AWSORGANIZATION

0.99+

thousandsQUANTITY

0.99+

fourQUANTITY

0.99+

nine stepsQUANTITY

0.99+

30QUANTITY

0.99+

OculusORGANIZATION

0.99+

100%QUANTITY

0.99+

72%QUANTITY

0.99+

48%QUANTITY

0.99+

10 microservicesQUANTITY

0.99+

second partQUANTITY

0.99+

FirstQUANTITY

0.99+

second slideQUANTITY

0.99+

first caseQUANTITY

0.99+

TodayDATE

0.99+

Amazon Web ServicesORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

oneQUANTITY

0.99+

20 servicesQUANTITY

0.99+

PrometheusTITLE

0.99+

second exampleQUANTITY

0.99+

second oneQUANTITY

0.99+

five keyQUANTITY

0.99+

OneQUANTITY

0.99+

firstQUANTITY

0.99+

third oneQUANTITY

0.99+

80-plus %QUANTITY

0.99+

eight settingsQUANTITY

0.99+

OpsaniPERSON

0.99+

third examplesQUANTITY

0.99+

twoQUANTITY

0.99+

todayDATE

0.99+

servicesQUANTITY

0.99+

50QUANTITY

0.99+

eight valuesQUANTITY

0.99+

bothQUANTITY

0.99+

nineQUANTITY

0.98+

three elementsQUANTITY

0.98+

ServoORGANIZATION

0.98+

80%QUANTITY

0.98+

opsani.com/awsstartupshowcaseOTHER

0.98+

first oneQUANTITY

0.98+

two small instancesQUANTITY

0.98+

10QUANTITY

0.97+

three thingsQUANTITY

0.97+

once a monthQUANTITY

0.97+

one timeQUANTITY

0.97+

70%QUANTITY

0.97+

GDPRTITLE

0.97+

zeroQUANTITY

0.97+

ServoTITLE

0.97+

about 20 stepsQUANTITY

0.97+

12QUANTITY

0.96+

KubernetesTITLE

0.96+

four timesQUANTITY

0.96+

Manufacturing Reduce Costs and Improve Quality with IoT Analytics


 

>>Okay. We're here in the second manufacturing drill down session with Michael Gerber. He was the managing director for automotive and manufacturing solutions at Cloudera. And we're going to continue the discussion with a look at how to lower costs and drive quality in IOT analytics with better uptime and hook. When you do the math, that's really quite obvious when the system is down, productivity is lost and it hits revenue and the bottom line improve quality drives, better service levels and reduces lost opportunities. Michael. Great to see you, >>Dave. All right, guys. Thank you so much. So I'll tell you, we're going to talk a little bit about connected manufacturing, right? And how those IOT IOT around connected manufacturing can do as Dave talked about improved quality outcomes for manufacturing improve and improve your plant uptime. So just a little bit quick, quick, little indulgent, quick history lesson. I promise to be quick. We've all heard about industry 4.0, right? That is the fourth industrial revolution. And that's really what we're here to talk about today. First industrial revolution, real simple, right? You had steam power, right? You would reduce backbreaking work. Second industrial revolution, mass assembly line. Right. So think about Henry Ford and motorized conveyor belts, mass automation, third industrial revolution. Things got interesting, right? You started to see automation, but that automation was done essentially programmed a robot to do something. It did the same thing over and over and over irrespective about of how your outside operations, your outside conditions change fourth industrial revolution, very different breakfasts. >>Now we're connecting, um, equipment and processes and getting feedback from it. And through machine learning, we can make those, um, those processes adapted right through machine learning. That's really what we're talking about in the fourth industrial revolution. And it is intrinsically connected to data and a data life cycle. And by the way, it's important, not just for a little bit of a slight issue. There we'll issue that, but it's important, not for technology sake, right? It's important because it actually drives very important business outcomes. First of all, falling, right? If you look at the cost of quality, even despite decades of, of, uh, companies and manufacturers moving to improve while its quality prompts still account to 20% of sales, right? So every fifth of what you meant or manufactured from a revenue perspective, you've got quality issues that are costing you a lot. Plant downtime, cost companies, $50 billion a year. >>So when we're talking about using data and these industry 4.0 types of use cases, connected data types of use cases, we're not doing it just narrowly to implement technology. We're doing it to move these from adverse, improving quality, reducing downtime. So let's talk about how a connected manufacturing data life cycle with what like, right. But so this is actually the business that cloud areas is in. Let's talk a little bit about that. So we call this manufacturing edge to AI. This is analytics, life something, and it starts with having your plants, right? Those plants are increasingly connected. As I said, sensor prices have come down two thirds over the last decade, right? And those sensors are connected over the internet. So suddenly we can collect all this data from your, um, manufacturing plants, and what do we want to be able to do? You know, we want to be able to collect it. >>We want to be able to analyze that data as it's coming across. Right? So, uh, in scream, right, we want to be able to analyze it and take intelligent real-time actions. Right? We might do some simple processing and filtering at the edge, but we really want to take real-time actions on that data. But, and this is the inference part of things, right? Taking that time. But this, the ability to take these real-time actions, um, is actually the result of a machine learning life cycle. I want to walk you through this, right? And it starts with, um, ingesting this data for the first time, putting it into our enterprise data lake, right? And that data lake enterprise data lake can be either within your data center or it could be in the cloud. You're going to, you're going to ingest that data. You're going to store it. >>You're going to enrich it with enterprise data sources. So now you'll have say sensor data and you'll have maintenance repair orders from your maintenance management systems. Right now you can start to think about do you're getting really nice data sets. You can start to say, Hey, which sensor values correlate to the need for machine maintenance, right? You start to see the data sets. They're becoming very compatible with machine learning, but so you bring these datasets together. You process that you align your time series data from your sensors to your timestamp data from your, um, you know, from your enterprise systems that your maintenance management system, as I mentioned, you know, once you've done that, we could put a query layer on top. So now we can start to do advanced analytics query across all these different types of data sets. But as I mentioned to you, and what's really important here is the fact that once you've stored one histories that say that you can build out those machine learning models I talked to you about earlier. >>So like I said, you can start to say, which sensor values drove the need of correlated to the need for equipment maintenance for my maintenance management systems, right? And then you can build out those models and say, Hey, here are the sensor values of the conditions that predict the need for maintenance. And once you understand that you can actually then build out those models, you deploy the models out to the edge where they will then work in that inference mode, that photographer, I will continuously sniff that data as it's coming and say, Hey, which are the, are we experiencing those conditions that, that predicted the need for maintenance? If so, let's take real-time action, right? Let's schedule a work order and equipment maintenance work order in the past, let's in the future, let's order the parts ahead of time before that a piece of equipment fails and allows us to be very, very proactive. >>So, you know, we have, this is a, one of the Mo the most popular use cases we're seeing in terms of connected, connected manufacturing. And we're working with many different, um, manufacturers around the world. I want to just highlight one of them. Cause I thought it's really interesting. This company is bought for Russia. And for SIA for ACA is the, um, is the, is the, um, the, uh, a supplier associated with out of France. They are huge, right? This is a multi-national automotive, um, parts and systems supplier. And as you can see, they operate in 300 sites in 35 countries. So very global, they connected 2000 machines, right. Um, I mean at once be able to take data from that. They started off with learning how to ingest the data. They started off very well with, um, you know, with, uh, manufacturing control towers, right? >>To be able to just monitor the data from coming in, you know, monitor the process. That was the first step, right. Uh, and you know, 2000 machines, 300 different variables, things like, um, vibration pressure temperature, right? So first let's do performance monitoring. Then they said, okay, let's start doing machine learning on some of these things, just start to build out things like equipment, um, predictive maintenance models, or compute. What they really focused on is computer vision while the inspection. So let's take pictures of, um, parts as they go through a process and then classify what that was this picture associated with the good or bad quality outcome. Then you teach the machine to make that decision on its own. So now, now the machine, the camera is doing the inspections for you. And so they both had those machine learning models. They took that data, all this data was on-prem, but they pushed that data up to the cloud to do the machine learning models, develop those machine learning models. >>Then they push the machine learning models back into the plants where they, where they could take real-time actions through these computer vision, quality inspections. So great use case. Um, great example of how you start with monitoring, move to machine learning, but at the end of the day, or improving quality and improving, um, uh, equipment uptime. And that is the goal of most manufacturers. So with that being said, um, I would like to say, if you want to learn some more, um, we've got a wealth of information on our website. You see the URL in front of you, please go, then you'll learn. There's a lot of information there in terms of the use cases that we're seeing in manufacturing and a lot more detail and a lot more talk about a lot more customers we'll work with. If you need that information, please do find it. Um, with that, I'm going to turn it over to Dave, to Steve. I think you had some questions you want to run by. >>I do, Michael, thank you very much for that. And before I get into the questions, I just wanted to sort of make some observations that was, you know, struck by what you're saying about the phases of industry. We talk about industry 4.0, and my observation is that, you know, traditionally, you know, machines have always replaced humans, but it's been around labor and, and the difference with 4.0, and what you talked about with connecting equipment is you're injecting machine intelligence. Now the camera inspection example, and then the machines are taking action, right? That's, that's different and, and is a really new kind of paradigm here. I think the, the second thing that struck me is, you know, the costs, you know, 20% of, of sales and plant downtime costing, you know, many tens of billions of dollars a year. Um, so that was huge. I mean, the business case for this is I'm going to reduce my expected loss quite dramatically. >>And then I think the third point, which we turned in the morning sessions, and the main stage is really this, the world is hybrid. Everybody's trying to figure out hybrid, get hybrid, right. And it certainly applies here. Uh, this is, this is a hybrid world you've got to accommodate, you know, regardless of where the data is, you've got to be able to get to it, blend it, enrich it, and then act on it. So anyway, those are my big, big takeaways. Um, so first question. So in thinking about implementing connected manufacturing initiatives, what are people going to run into? What are the big challenges that they're going to, they're going to hit? >>No, there's, there's there, there's a few of the, but I think, you know, one of the, uh, one of the key ones is bridging what we'll call the it and OT data divide, right. And what we mean by the it, you know, your, it systems are the ones, your ERP systems, your MES system, Freightos your transactional systems that run on relational databases and your it departments are brilliant at running on that, right? The difficulty becomes an implementing these use cases that you also have to deal with operational technology, right? And those are all of the, that's all the equipment in your manufacturing plant that runs on its proprietary network with proprietary pro protocols. That information can be very, very difficult to get to. Right? So, and it's uncertain, it's a much more unstructured than from your OT. So the key challenge is being able to bring these data sets together in a single place where you can start to do advanced analytics and leverage that diverse data to do machine learning. Right? So that is one of the, if I had to boil it down to the single hardest thing in this, uh, in this, in this type of environment, nectar manufacturing is that that operational technology has kind of run on its own in its own. And for a long time, the silos, the silos, a bound, but at the end of the day, this is incredibly valuable data that now can be tapped, um, um, to, to, to, to move those, those metrics we talked about right around quality and uptime. So a huge opportunity. >>Well, and again, this is a hybrid team and you, you've kind of got this world, that's going toward an equilibrium. You've got the OT side and, you know, pretty hardcore engineers. And we know, we know it. A lot of that data historically has been analog data. This is Chris now is getting, you know, instrumented and captured. Uh, and so you've got that, that cultural challenge and, you know, you got to blend those two worlds. That's critical. Okay. So Michael, let's talk about some of the use cases you touched on, on some, but let's peel the onion a bit when you're thinking about this world of connected manufacturing and analytics in that space, when you talk to customers, you know, what are the most common use cases that you see? >>Yeah, that's a great, that's a great question. And you're right. I did allude to a little bit earlier, but there really is. I want people to think about this, a spectrum of use cases ranging from simple to complex, but you can get value even in the simple phases. And when I talk about the simple use cases, the simplest use cases really is really around monitoring, right? So in this, you monitor your equipment or monitor your processes, right? And you just make sure that you're staying within the bounds of your control plan, right? And this is much easier to do now. Right? Cause some of these sensors are a more sensors and those sensors are moving more and more towards the internet types of technology. So, Hey, you've got the opportunity now to be able to do some monitoring. Okay. No machine learning, we're just talking about simple monitoring next level down. >>And we're seeing is something we would call quality event forensic announces. And now on this one, you say, imagine I'm got warranty plans in the, in the field, right? So I'm starting to see warranty claims kicked off on them. And what you simply want to be able to do is do the forensic analysis back to what was the root cause of within the manufacturing process that caused it. So this is about connecting the dots I've got, I've got warranty issues. What were the manufacturing conditions of the day that caused it? Then you could also say which other, which other products were impacted by those same conditions. And we call those proactively rather than, and, and selectively rather than say, um, recalling an entire year's fleet of a car. So, and that, again, also not machine learning is simply connecting the dots from a warranty claims in the field to the manufacturing conditions of the day so that you could take corrective actions, but then you get into a whole slew of machine learning use case, you know, and, and that ranges from things like quality or say yield optimization, where you start to collect sensor values and, um, manufacturing yield, uh, values from your ERP system. >>And you're certain start to say, which, um, you know, which map a sensor values or factors drove good or bad yield outcomes. And you can identify those factors that are the most important. So you, um, you, you measure those, you monitor those and you optimize those, right. That's how you optimize your, and then you go down to the more traditional machine learning use cases around predictive maintenance. So the key point here, Dave is, look, there's a huge, you know, depending on a customer's maturity around big data, you could start simply with monitoring, get a lot of value, start, then bring together more diverse datasets to do things like connect the.analytics then all and all the way then to, to, to the more advanced machine learning use cases this value to be had throughout. >>I remember when the, you know, the it industry really started to think about, or in the early days, you know, IOT and IOT. Um, it reminds me of when, you know, there was, uh, the, the old days of football field, we were grass and, and a new player would come in and he'd be perfectly white uniform and you had it. We had to get dirty as an industry, you know, it'll learn. And so, so my question relates to other technology partners that you might be working with that are maybe new in this space that, that to accelerate some of these solutions that we've been talking about. >>Yeah. That's a great question. I kind of, um, goes back to one of the things I alluded a little bit about earlier. We've got some great partners, a partner, for example, litmus automation, whose whole world is the OT world. And what they've done is for example, they built some adapters to be able to get to practically every industrial protocol. And they've said, Hey, we can do that. And then give a single interface of that data to the Idera data platform. So now, you know, we're really good at ingesting it data and things like that. We can leverage say a company like litmus that can open the flood gates of that OT data, making it much easier to get that data into our platform. And suddenly you've got all the data you need to, to implement those types of, um, industry 4.0, uh, analytics use cases. And it really boils down to, can I get to that? Can I break down that it OT, um, you know, uh, uh, barrier that we've always had and, and bring together those data sets that really move the needle in terms of improving manufacturing performance. >>Okay. Thank you for that last question. Speaking to moving the needle, I want to Lee lead this discussion on the technology advances. I'd love to talk tech here. Uh, what are the key technology enablers and advancers, if you will, that are going to move connected manufacturing and machine learning forward in this transportation space. Sorry. Manufacturing in >>Factor space. Yeah, I know in the manufacturing space, there's a few things, first of all, I think the fact that obviously I know we touched upon this, the fact that sensor prices have come down and it had become ubiquitous that number one, we can w we're finally been able to get to the OT data, right? That's that's number one, number, number two, I think, you know, um, we, we have the ability that now to be able to store that data a whole lot more efficiently, you know, we've got, we've got great capabilities to be able to do that, to put it over into the cloud, to do the machine learning types of workloads. You've got things like if you're doing computer vision, while in analyst respect GPU's to make those machine learning models much more, um, much more effective, if that 5g technology that starts to blur at least from a latency perspective where you do your computer, whether it be on the edge or in the cloud, you've, you've got more, you know, super business critical stuff. >>You probably don't want to rely on, uh, any type of network connection, but from a latency perspective, you're starting to see, uh, you know, the ability to do compute where it's the most effective now. And that's really important. And again, the machine learning capabilities, and they believed the book, bullet, uh, GP, you know, GPU level, machine learning, all that, those models, and then deployed by over the air updates to your equipment. All of those things are making this, um, there's, you know, there's the advanced analytics and machine learning, uh, data life cycle just faster and better. And at the end of the day, to your point, Dave, that equipment and processes are getting much smarter, uh, very much more quickly. >>Yep. We've got a lot of data and we have way lower costs, uh, processing platforms I'll throw in NP use as well. Watch that space neural processing units. Okay. Michael, we're going to leave it there. Thank you so much. Really appreciate your time, >>Dave. I really appreciate it. And thanks. Thanks for, uh, for everybody who joined. Uh, thanks. Thanks for joining today. Yes. Thank you for watching. Keep it right there.

Published Date : Aug 3 2021

SUMMARY :

When you do the math, that's really quite obvious when the system is down, productivity is lost and it hits revenue and the bottom Thank you so much. So every fifth of what you meant or manufactured from a revenue perspective, And those sensors are connected over the internet. I want to walk you through those machine learning models I talked to you about earlier. And then you can build out those models and say, Hey, here are the sensor values of the conditions And as you can see, they operate in 300 sites To be able to just monitor the data from coming in, you know, monitor the process. And that is the goal of most manufacturers. I think the, the second thing that struck me is, you know, the costs, you know, 20% of, And then I think the third point, which we turned in the morning sessions, and the main stage is really this, And what we mean by the it, you know, your, it systems are the ones, So Michael, let's talk about some of the use cases you touched on, on some, And you just make sure that you're staying within the bounds of your control plan, And now on this one, you say, imagine I'm got warranty plans in the, in the field, And you can identify those factors that I remember when the, you know, the it industry really started to think about, or in the early days, litmus that can open the flood gates of that OT data, making it much easier to if you will, that are going to move connected manufacturing and machine learning forward that data a whole lot more efficiently, you know, we've got, we've got great capabilities to be able to do that, And at the end of the day, to your point, Dave, that equipment and processes are getting much smarter, Thank you so much. Thank you for watching.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
StevePERSON

0.99+

DavePERSON

0.99+

MichaelPERSON

0.99+

FranceLOCATION

0.99+

Michael GerberPERSON

0.99+

300 sitesQUANTITY

0.99+

2000 machinesQUANTITY

0.99+

20%QUANTITY

0.99+

ClouderaORGANIZATION

0.99+

ChrisPERSON

0.99+

2000 machinesQUANTITY

0.99+

RussiaLOCATION

0.99+

third pointQUANTITY

0.99+

todayDATE

0.99+

35 countriesQUANTITY

0.99+

secondQUANTITY

0.99+

300 different variablesQUANTITY

0.99+

first timeQUANTITY

0.99+

first questionQUANTITY

0.99+

first stepQUANTITY

0.99+

Henry FordPERSON

0.98+

oneQUANTITY

0.98+

bothQUANTITY

0.98+

firstQUANTITY

0.98+

SecondQUANTITY

0.97+

two thirdsQUANTITY

0.96+

litmusORGANIZATION

0.96+

FirstQUANTITY

0.96+

ACAORGANIZATION

0.95+

decadesQUANTITY

0.95+

two worldsQUANTITY

0.94+

singleQUANTITY

0.94+

fourth industrial revolutionEVENT

0.93+

second thingQUANTITY

0.93+

LeePERSON

0.92+

single interfaceQUANTITY

0.92+

last decadeDATE

0.92+

single placeQUANTITY

0.92+

fourth industrial revolutionEVENT

0.92+

IderaORGANIZATION

0.91+

tens of billions of dollars a yearQUANTITY

0.88+

$50 billion a yearQUANTITY

0.84+

industrial revolutionEVENT

0.82+

20% of salesQUANTITY

0.81+

fifthQUANTITY

0.75+

4.0 typesQUANTITY

0.72+

thirdQUANTITY

0.61+

lakeORGANIZATION

0.61+

AnalyticsTITLE

0.47+

twoQUANTITY

0.38+

oneOTHER

0.34+

DockerCon2021 Keynote


 

>>Individuals create developers, translate ideas to code, to create great applications and great applications. Touch everyone. A Docker. We know that collaboration is key to your innovation sharing ideas, working together. Launching the most secure applications. Docker is with you wherever your team innovates, whether it be robots or autonomous cars, we're doing research to save lives during a pandemic, revolutionizing, how to buy and sell goods online, or even going into the unknown frontiers of space. Docker is launching innovation everywhere. Join us on the journey to build, share, run the future. >>Hello and welcome to Docker con 2021. We're incredibly excited to have more than 80,000 of you join us today from all over the world. As it was last year, this year at DockerCon is 100% virtual and 100% free. So as to enable as many community members as possible to join us now, 100%. Virtual is also an acknowledgement of the continuing global pandemic in particular, the ongoing tragedies in India and Brazil, the Docker community is a global one. And on behalf of all Dr. Khan attendees, we are donating $10,000 to UNICEF support efforts to fight the virus in those countries. Now, even in those regions of the world where the pandemic is being brought under control, virtual first is the new normal. It's been a challenging transition. This includes our team here at Docker. And we know from talking with many of you that you and your developer teams are challenged by this as well. So to help application development teams better collaborate and ship faster, we've been working on some powerful new features and we thought it would be fun to start off with a demo of those. How about it? Want to have a look? All right. Then no further delay. I'd like to introduce Youi Cal and Ben, gosh, over to you and Ben >>Morning, Ben, thanks for jumping on real quick. >>Have you seen the email from Scott? The one about updates and the docs landing page Smith, the doc combat and more prominence. >>Yeah. I've got something working on my local machine. I haven't committed anything yet. I was thinking we could try, um, that new Docker dev environments feature. >>Yeah, that's cool. So if you hit the share button, what I should do is it will take all of your code and the dependencies and the image you're basing it on and wrap that up as one image for me. And I can then just monitor all my machines that have been one click, like, and then have it side by side, along with the changes I've been looking at as well, because I was also having a bit of a look and then I can really see how it differs to what I'm doing. Maybe I can combine it to do the best of both worlds. >>Sounds good. Uh, let me get that over to you, >>Wilson. Yeah. If you pay with the image name, I'll get that started up. >>All right. Sen send it over >>Cheesy. Okay, great. Let's have a quick look at what you he was doing then. So I've been messing around similar to do with the batter. I've got movie at the top here and I think it looks pretty cool. Let's just grab that image from you. Pick out that started on a dev environment. What this is doing. It's just going to grab the image down, which you can take all of the code, the dependencies only get brunches working on and I'll get that opened up in my idea. Ready to use. It's a here close. We can see our environment as my Molly image, just coming down there and I've got my new idea. >>We'll load this up and it'll just connect to my dev environment. There we go. It's connected to the container. So we're working all in the container here and now give it a moment. What we'll do is we'll see what changes you've been making as well on the code. So it's like she's been working on a landing page as well, and it looks like she's been changing the banner as well. So let's get this running. Let's see what she's actually doing and how it looks. We'll set up our checklist and then we'll see how that works. >>Great. So that's now rolling. So let's just have a look at what you use doing what changes she had made. Compare those to mine just jumped back into my dev container UI, see that I've got both of those running side by side with my changes and news changes. Okay. So she's put Molly up there rather than mobi or somebody had the same idea. So I think in a way I can make us both happy. So if we just jumped back into what we'll do, just add Molly and Moby and here I'll save that. And what we can see is, cause I'm just working within the container rather than having to do sort of rebuild of everything or serve, or just reload my content. No, that's straight the page. So what I can then do is I can come up with my browser here. Once that's all refreshed, refresh the page once hopefully, maybe twice, we should then be able to see your refresh it or should be able to see that we get Malia mobi come up. So there we go, got Molly mobi. So what we'll do now is we'll describe that state. It sends us our image and then we'll just create one of those to share with URI or share. And we'll get a link for that. I guess we'll send that back over to you. >>So I've had a look at what you were doing and I'm actually going to change. I think that might work for both of us. I wondered if you could take a look at it. If I send it over. >>Sounds good. Let me grab the link. >>Yeah, it's a dev environment link again. So if you just open that back in the doc dashboard, it should be able to open up the code that I've changed and then just run it in the same way you normally do. And that shouldn't interrupt what you're already working on because there'll be able to run side by side with your other brunch. You already got, >>Got it. Got it. Loading here. Well, that's great. It's Molly and movie together. I love it. I think we should ship it. >>Awesome. I guess it's chip it and get on with the rest of.com. Wasn't that cool. Thank you Joey. Thanks Ben. Everyone we'll have more of this later in the keynote. So stay tuned. Let's say earlier, we've all been challenged by this past year, whether the COVID pandemic, the complete evaporation of customer demand in many industries, unemployment or business bankruptcies, we all been touched in some way. And yet, even to miss these tragedies last year, we saw multiple sources of hope and inspiration. For example, in response to COVID we saw global communities, including the tech community rapidly innovate solutions for analyzing the spread of the virus, sequencing its genes and visualizing infection rates. In fact, if all in teams collaborating on solutions for COVID have created more than 1,400 publicly shareable images on Docker hub. As another example, we all witnessed the historic landing and exploration of Mars by the perseverance Rover and its ingenuity drone. >>Now what's common in these examples, these innovative and ambitious accomplishments were made possible not by any single individual, but by teams of individuals collaborating together. The power of teams is why we've made development teams central to Docker's mission to build tools and content development teams love to help them get their ideas from code to cloud as quickly as possible. One of the frictions we've seen that can slow down to them in teams is that the path from code to cloud can be a confusing one, riddle with multiple point products, tools, and images that need to be integrated and maintained an automated pipeline in order for teams to be productive. That's why a year and a half ago we refocused Docker on helping development teams make sense of all this specifically, our goal is to provide development teams with the trusted content, the sharing capabilities and the pipeline integrations with best of breed third-party tools to help teams ship faster in short, to provide a collaborative application development platform. >>Everything a team needs to build. Sharon run create applications. Now, as I noted earlier, it's been a challenging year for everyone on our planet and has been similar for us here at Docker. Our team had to adapt to working from home local lockdowns caused by the pandemic and other challenges. And despite all this together with our community and ecosystem partners, we accomplished many exciting milestones. For example, in open source together with the community and our partners, we open sourced or made major contributions to many projects, including OCI distribution and the composed plugins building on these open source projects. We had powerful new capabilities to the Docker product, both free and subscription. For example, support for WSL two and apple, Silicon and Docker, desktop and vulnerability scanning audit logs and image management and Docker hub. >>And finally delivering an easy to use well-integrated development experience with best of breed tools and content is only possible through close collaboration with our ecosystem partners. For example, this last year we had over 100 commercialized fees, join our Docker verified publisher program and over 200 open source projects, join our Docker sponsored open source program. As a result of these efforts, we've seen some exciting growth in the Docker community in the 12 months since last year's Docker con for example, the number of registered developers grew 80% to over 8 million. These developers created many new images increasing the total by 56% to almost 11 million. And the images in all these repositories were pulled by more than 13 million monthly active IP addresses totaling 13 billion pulls a month. Now while the growth is exciting by Docker, we're even more excited about the stories we hear from you and your development teams about how you're using Docker and its impact on your businesses. For example, cancer researchers and their bioinformatics development team at the Washington university school of medicine needed a way to quickly analyze their clinical trial results and then share the models, the data and the analysis with other researchers they use Docker because it gives them the ease of use choice of pipeline tools and speed of sharing so critical to their research. And most importantly to the lives of their patients stay tuned for another powerful customer story later in the keynote from Matt fall, VP of engineering at Oracle insights. >>So with this last year behind us, what's next for Docker, but challenge you this last year of force changes in how development teams work, but we felt for years to come. And what we've learned in our discussions with you will have long lasting impact on our product roadmap. One of the biggest takeaways from those discussions that you and your development team want to be quicker to adapt, to changes in your environment so you can ship faster. So what is DACA doing to help with this first trusted content to own the teams that can focus their energies on what is unique to their businesses and spend as little time as possible on undifferentiated work are able to adapt more quickly and ship faster in order to do so. They need to be able to trust other components that make up their app together with our partners. >>Docker is doubling down and providing development teams with trusted content and the tools they need to use it in their applications. Second, remote collaboration on a development team, asking a coworker to take a look at your code used to be as easy as swiveling their chair around, but given what's happened in the last year, that's no longer the case. So as you even been hinted in the demo at the beginning, you'll see us deliver more capabilities for remote collaboration within a development team. And we're enabling development team to quickly adapt to any team configuration all on prem hybrid, all work from home, helping them remain productive and focused on shipping third ecosystem integrations, those development teams that can quickly take advantage of innovations throughout the ecosystem. Instead of getting locked into a single monolithic pipeline, there'll be the ones able to deliver amps, which impact their businesses faster. >>So together with our ecosystem partners, we are investing in more integrations with best of breed tools, right? Integrated automated app pipelines. Furthermore, we'll be writing more public API APIs and SDKs to enable ecosystem partners and development teams to roll their own integrations. We'll be sharing more details about remote collaboration and ecosystem integrations. Later in the keynote, I'd like to take a moment to share with Docker and our partners are doing for trusted content, providing development teams, access to content. They can trust, allows them to focus their coding efforts on what's unique and differentiated to that end Docker and our partners are bringing more and more trusted content to Docker hub Docker official images are 160 images of popular upstream open source projects that serve as foundational building blocks for any application. These include operating systems, programming, languages, databases, and more. Furthermore, these are updated patch scan and certified frequently. So I said, no image is older than 30 days. >>Docker verified publisher images are published by more than 100 commercialized feeds. The image Rebos are explicitly designated verify. So the developers searching for components for their app know that the ISV is actively maintaining the image. Docker sponsored open source projects announced late last year features images for more than 200 open source communities. Docker sponsors these communities through providing free storage and networking resources and offering their community members unrestricted access repos for businesses allow businesses to update and share their apps privately within their organizations using role-based access control and user authentication. No, and finally, public repos for communities enable community projects to be freely shared with anonymous and authenticated users alike. >>And for all these different types of content, we provide services for both development teams and ISP, for example, vulnerability scanning and digital signing for enhanced security search and filtering for discoverability packaging and updating services and analytics about how these products are being used. All this trusted content, we make available to develop teams for them directly to discover poll and integrate into their applications. Our goal is to meet development teams where they live. So for those organizations that prefer to manage their internal distribution of trusted content, we've collaborated with leading container registry partners. We announced our partnership with J frog late last year. And today we're very pleased to announce our partnerships with Amazon and Miranda's for providing an integrated seamless experience for joint for our joint customers. Lastly, the container images themselves and this end to end flow are built on open industry standards, which provided all the teams with flexibility and choice trusted content enables development teams to rapidly build. >>As I let them focus on their unique differentiated features and use trusted building blocks for the rest. We'll be talking more about trusted content as well as remote collaboration and ecosystem integrations later in the keynote. Now ecosystem partners are not only integral to the Docker experience for development teams. They're also integral to a great DockerCon experience, but please join me in thanking our Dr. Kent on sponsors and checking out their talks throughout the day. I also want to thank some others first up Docker team. Like all of you this last year has been extremely challenging for us, but the Docker team rose to the challenge and worked together to continue shipping great product, the Docker community of captains, community leaders, and contributors with your welcoming newcomers, enthusiasm for Docker and open exchanges of best practices and ideas talker, wouldn't be Docker without you. And finally, our development team customers. >>You trust us to help you build apps. Your businesses rely on. We don't take that trust for granted. Thank you. In closing, we often hear about the tenant's developer capable of great individual feeds that can transform project. But I wonder if we, as an industry have perhaps gotten this wrong by putting so much emphasis on weight, on the individual as discussed at the beginning, great accomplishments like innovative responses to COVID-19 like landing on Mars are more often the results of individuals collaborating together as a team, which is why our mission here at Docker is delivered tools and content developers love to help their team succeed and become 10 X teams. Thanks again for joining us, we look forward to having a great DockerCon with you today, as well as a great year ahead of us. Thanks and be well. >>Hi, I'm Dana Lawson, VP of engineering here at get hub. And my job is to enable this rich interconnected community of builders and makers to build even more and hopefully have a great time doing it in order to enable the best platform for developers, which I know is something we are all passionate about. We need to partner across the ecosystem to ensure that developers can have a great experience across get hub and all the tools that they want to use. No matter what they are. My team works to build the tools and relationships to make that possible. I am so excited to join Scott on this virtual stage to talk about increasing developer velocity. So let's dive in now, I know this may be hard for some of you to believe, but as a former CIS admin, some 21 years ago, working on sense spark workstations, we've come such a long way for random scripts and desperate systems that we've stitched together to this whole inclusive developer workflow experience being a CIS admin. >>Then you were just one piece of the siloed experience, but I didn't want to just push code to production. So I created scripts that did it for me. I taught myself how to code. I was the model lazy CIS admin that got dangerous and having pushed a little too far. I realized that working in production and building features is really a team sport that we had the opportunity, all of us to be customer obsessed today. As developers, we can go beyond the traditional dev ops mindset. We can really focus on adding value to the customer experience by ensuring that we have work that contributes to increasing uptime via and SLS all while being agile and productive. We get there. When we move from a pass the Baton system to now having an interconnected developer workflow that increases velocity in every part of the cycle, we get to work better and smarter. >>And honestly, in a way that is so much more enjoyable because we automate away all the mundane and manual and boring tasks. So we get to focus on what really matters shipping, the things that humans get to use and love. Docker has been a big part of enabling this transformation. 10, 20 years ago, we had Tomcat containers, which are not Docker containers. And for y'all hearing this the first time go Google it. But that was the way we built our applications. We had to segment them on the server and give them resources. Today. We have Docker containers, these little mini Oasys and Docker images. You can do it multiple times in an orchestrated manner with the power of actions enabled and Docker. It's just so incredible what you can do. And by the way, I'm showing you actions in Docker, which I hope you use because both are great and free for open source. >>But the key takeaway is really the workflow and the automation, which you certainly can do with other tools. Okay, I'm going to show you just how easy this is, because believe me, if this is something I can learn and do anybody out there can, and in this demo, I'll show you about the basic components needed to create and use a package, Docker container actions. And like I said, you won't believe how awesome the combination of Docker and actions is because you can enable your workflow to do no matter what you're trying to do in this super baby example. We're so small. You could take like 10 seconds. Like I am here creating an action due to a simple task, like pushing a message to your logs. And the cool thing is you can use it on any the bit on this one. Like I said, we're going to use push. >>You can do, uh, even to order a pizza every time you roll into production, if you wanted, but at get hub, that'd be a lot of pizzas. And the funny thing is somebody out there is actually tried this and written that action. If you haven't used Docker and actions together, check out the docs on either get hub or Docker to get you started. And a huge shout out to all those doc writers out there. I built this demo today using those instructions. And if I can do it, I know you can too, but enough yapping let's get started to save some time. And since a lot of us are Docker and get hub nerds, I've already created a repo with a Docker file. So we're going to skip that step. Next. I'm going to create an action's Yammel file. And if you don't Yammer, you know, actions, the metadata defines my important log stuff to capture and the input and my time out per parameter to pass and puts to the Docker container, get up a build image from your Docker file and run the commands in a new container. >>Using the Sigma image. The cool thing is, is you can use any Docker image in any language for your actions. It doesn't matter if it's go or whatever in today's I'm going to use a shell script and an input variable to print my important log stuff to file. And like I said, you know me, I love me some. So let's see this action in a workflow. When an action is in a private repo, like the one I demonstrating today, the action can only be used in workflows in the same repository, but public actions can be used by workflows in any repository. So unfortunately you won't get access to the super awesome action, but don't worry in the Guild marketplace, there are over 8,000 actions available, especially the most important one, that pizza action. So go try it out. Now you can do this in a couple of ways, whether you're doing it in your preferred ID or for today's demo, I'm just going to use the gooey. I'm going to navigate to my actions tab as I've done here. And I'm going to in my workflow, select new work, hello, probably load some workflows to Claire to get you started, but I'm using the one I've copied. Like I said, the lazy developer I am in. I'm going to replace it with my action. >>That's it. So now we're going to go and we're going to start our commitment new file. Now, if we go over to our actions tab, we can see the workflow in progress in my repository. I just click the actions tab. And because they wrote the actions on push, we can watch the visualization under jobs and click the job to see the important stuff we're logging in the input stamp in the printed log. And we'll just wait for this to run. Hello, Mona and boom. Just like that. It runs automatically within our action. We told it to go run as soon as the files updated because we're doing it on push merge. That's right. Folks in just a few minutes, I built an action that writes an entry to a log file every time I push. So I don't have to do it manually. In essence, with automation, you can be kind to your future self and save time and effort to focus on what really matters. >>Imagine what I could do with even a little more time, probably order all y'all pieces. That is the power of the interconnected workflow. And it's amazing. And I hope you all go try it out, but why do we care about all of that? Just like in the demo, I took a manual task with both tape, which both takes time and it's easy to forget and automated it. So I don't have to think about it. And it's executed every time consistently. That means less time for me to worry about my human errors and mistakes, and more time to focus on actually building the cool stuff that people want. Obviously, automation, developer productivity, but what is even more important to me is the developer happiness tools like BS, code actions, Docker, Heroku, and many others reduce manual work, which allows us to focus on building things that are awesome. >>And to get into that wonderful state that we call flow. According to research by UC Irvine in Humboldt university in Germany, it takes an average of 23 minutes to enter optimal creative state. What we call the flow or to reenter it after distraction like your dog on your office store. So staying in flow is so critical to developer productivity and as a developer, it just feels good to be cranking away at something with deep focus. I certainly know that I love that feeling intuitive collaboration and automation features we built in to get hub help developer, Sam flow, allowing you and your team to do so much more, to bring the benefits of automation into perspective in our annual October's report by Dr. Nicole, Forsgren. One of my buddies here at get hub, took a look at the developer productivity in the stork year. You know what we found? >>We found that public GitHub repositories that use the Automational pull requests, merge those pull requests. 1.2 times faster. And the number of pooled merged pull requests increased by 1.3 times, that is 34% more poor requests merged. And other words, automation can con can dramatically increase, but the speed and quantity of work completed in any role, just like an open source development, you'll work more efficiently with greater impact when you invest the bulk of your time in the work that adds the most value and eliminate or outsource the rest because you don't need to do it, make the machines by elaborate by leveraging automation in their workflows teams, minimize manual work and reclaim that time for innovation and maintain that state of flow with development and collaboration. More importantly, their work is more enjoyable because they're not wasting the time doing the things that the machines or robots can do for them. >>And I remember what I said at the beginning. Many of us want to be efficient, heck even lazy. So why would I spend my time doing something I can automate? Now you can read more about this research behind the art behind this at October set, get hub.com, which also includes a lot of other cool info about the open source ecosystem and how it's evolving. Speaking of the open source ecosystem we at get hub are so honored to be the home of more than 65 million developers who build software together for everywhere across the globe. Today, we're seeing software development taking shape as the world's largest team sport, where development teams collaborate, build and ship products. It's no longer a solo effort like it was for me. You don't have to take my word for it. Check out this globe. This globe shows real data. Every speck of light you see here represents a contribution to an open source project, somewhere on earth. >>These arts reach across continents, cultures, and other divides. It's distributed collaboration at its finest. 20 years ago, we had no concept of dev ops, SecOps and lots, or the new ops that are going to be happening. But today's development and ops teams are connected like ever before. This is only going to continue to evolve at a rapid pace, especially as we continue to empower the next hundred million developers, automation helps us focus on what's important and to greatly accelerate innovation. Just this past year, we saw some of the most groundbreaking technological advancements and achievements I'll say ever, including critical COVID-19 vaccine trials, as well as the first power flight on Mars. This past month, these breakthroughs were only possible because of the interconnected collaborative open source communities on get hub and the amazing tools and workflows that empower us all to create and innovate. Let's continue building, integrating, and automating. So we collectively can give developers the experience. They deserve all of the automation and beautiful eye UIs that we can muster so they can continue to build the things that truly do change the world. Thank you again for having me today, Dr. Khan, it has been a pleasure to be here with all you nerds. >>Hello. I'm Justin. Komack lovely to see you here. Talking to developers, their world is getting much more complex. Developers are being asked to do everything security ops on goal data analysis, all being put on the rockers. Software's eating the world. Of course, and this all make sense in that view, but they need help. One team. I told you it's shifted all our.net apps to run on Linux from windows, but their developers found the complexity of Docker files based on the Linux shell scripts really difficult has helped make these things easier for your teams. Your ones collaborate more in a virtual world, but you've asked us to make this simpler and more lightweight. You, the developers have asked for a paved road experience. You want things to just work with a simple options to be there, but it's not just the paved road. You also want to be able to go off-road and do interesting and different things. >>Use different components, experiments, innovate as well. We'll always offer you both those choices at different times. Different developers want different things. It may shift for ones the other paved road or off road. Sometimes you want reliability, dependability in the zone for day to day work, but sometimes you have to do something new, incorporate new things in your pipeline, build applications for new places. Then you knew those off-road abilities too. So you can really get under the hood and go and build something weird and wonderful and amazing. That gives you new options. Talk as an independent choice. We don't own the roads. We're not pushing you into any technology choices because we own them. We're really supporting and driving open standards, such as ISEI working opensource with the CNCF. We want to help you get your applications from your laptops, the clouds, and beyond, even into space. >>Let's talk about the key focus areas, that frame, what DACA is doing going forward. These are simplicity, sharing, flexibility, trusted content and care supply chain compared to building where the underlying kernel primitives like namespaces and Seagraves the original Docker CLI was just amazing Docker engine. It's a magical experience for everyone. It really brought those innovations and put them in a world where anyone would use that, but that's not enough. We need to continue to innovate. And it was trying to get more done faster all the time. And there's a lot more we can do. We're here to take complexity away from deeply complicated underlying things and give developers tools that are just amazing and magical. One of the area we haven't done enough and make things magical enough that we're really planning around now is that, you know, Docker images, uh, they're the key parts of your application, but you know, how do I do something with an image? How do I, where do I attach volumes with this image? What's the API. Whereas the SDK for this image, how do I find an example or docs in an API driven world? Every bit of software should have an API and an API description. And our vision is that every container should have this API description and the ability for you to understand how to use it. And it's all a seamless thing from, you know, from your code to the cloud local and remote, you can, you can use containers in this amazing and exciting way. >>One thing I really noticed in the last year is that companies that started off remote fast have constant collaboration. They have zoom calls, apron all day terminals, shattering that always working together. Other teams are really trying to learn how to do this style because they didn't start like that. We used to walk around to other people's desks or share services on the local office network. And it's very difficult to do that anymore. You want sharing to be really simple, lightweight, and informal. Let me try your container or just maybe let's collaborate on this together. Um, you know, fast collaboration on the analysts, fast iteration, fast working together, and he wants to share more. You want to share how to develop environments, not just an image. And we all work by seeing something someone else in our team is doing saying, how can I do that too? I can, I want to make that sharing really, really easy. Ben's going to talk about this more in the interest of one minute. >>We know how you're excited by apple. Silicon and gravis are not excited because there's a new architecture, but excited because it's faster, cooler, cheaper, better, and offers new possibilities. The M one support was the most asked for thing on our public roadmap, EFA, and we listened and share that we see really exciting possibilities, usership arm applications, all the way from desktop to production. We know that you all use different clouds and different bases have deployed to, um, you know, we work with AWS and Azure and Google and more, um, and we want to help you ship on prime as well. And we know that you use huge number of languages and the containers help build applications that use different languages for different parts of the application or for different applications, right? You can choose the best tool. You have JavaScript hat or everywhere go. And re-ask Python for data and ML, perhaps getting excited about WebAssembly after hearing about a cube con, you know, there's all sorts of things. >>So we need to make that as easier. We've been running the whole month of Python on the blog, and we're doing a month of JavaScript because we had one specific support about how do I best put this language into production of that language into production. That detail is important for you. GPS have been difficult to use. We've added GPS suppose in desktop for windows, but we know there's a lot more to do to make the, how multi architecture, multi hardware, multi accelerator world work better and also securely. Um, so there's a lot more work to do to support you in all these things you want to do. >>How do we start building a tenor has applications, but it turns out we're using existing images as components. I couldn't assist survey earlier this year, almost half of container image usage was public images rather than private images. And this is growing rapidly. Almost all software has open source components and maybe 85% of the average application is open source code. And what you're doing is taking whole container images as modules in your application. And this was always the model with Docker compose. And it's a model that you're already et cetera, writing you trust Docker, official images. We know that they might go to 25% of poles on Docker hub and Docker hub provides you the widest choice and the best support that trusted content. We're talking to people about how to make this more helpful. We know, for example, that winter 69 four is just showing us as support, but the image doesn't yet tell you that we're working with canonical to improve messaging from specific images about left lifecycle and support. >>We know that you need more images, regularly updated free of vulnerabilities, easy to use and discover, and Donnie and Marie neuro, going to talk about that more this last year, the solar winds attack has been in the, in the news. A lot, the software you're using and trusting could be compromised and might be all over your organization. We need to reduce the risk of using vital open-source components. We're seeing more software supply chain attacks being targeted as the supply chain, because it's often an easier place to attack and production software. We need to be able to use this external code safely. We need to, everyone needs to start from trusted sources like photography images. They need to scan for known vulnerabilities using Docker scan that we built in partnership with sneak and lost DockerCon last year, we need just keep updating base images and dependencies, and we'll, we're going to help you have the control and understanding about your images that you need to do this. >>And there's more, we're also working on the nursery V2 project in the CNCF to revamp container signings, or you can tell way or software comes from we're working on tooling to make updates easier, and to help you understand and manage all the principals carrier you're using security is a growing concern for all of us. It's really important. And we're going to help you work with security. We can't achieve all our dreams, whether that's space travel or amazing developer products ever see without deep partnerships with our community to cloud is RA and the cloud providers aware most of you ship your occasion production and simple routes that take your work and deploy it easily. Reliably and securely are really important. Just get into production simply and easily and securely. And we've done a bunch of work on that. And, um, but we know there's more to do. >>The CNCF on the open source cloud native community are an amazing ecosystem of creators and lovely people creating an amazing strong community and supporting a huge amount of innovation has its roots in the container ecosystem and his dreams beyond that much of the innovation is focused around operate experience so far, but developer experience is really a growing concern in that community as well. And we're really excited to work on that. We also uses appraiser tool. Then we know you do, and we know that you want it to be easier to use in your environment. We just shifted Docker hub to work on, um, Kubernetes fully. And, um, we're also using many of the other projects are Argo from atheists. We're spending a lot of time working with Microsoft, Amazon right now on getting natural UV to ready to ship in the next few. That's a really detailed piece of collaboration we've been working on for a long term. Long time is really important for our community as the scarcity of the container containers and, um, getting content for you, working together makes us stronger. Our community is made up of all of you have. Um, it's always amazing to be reminded of that as a huge open source community that we already proud to work with. It's an amazing amount of innovation that you're all creating and where perhaps it, what with you and share with you as well. Thank you very much. And thank you for being here. >>Really excited to talk to you today and share more about what Docker is doing to help make you faster, make your team faster and turn your application delivery into something that makes you a 10 X team. What we're hearing from you, the developers using Docker everyday fits across three common themes that we hear consistently over and over. We hear that your time is super important. It's critical, and you want to move faster. You want your tools to get out of your way, and instead to enable you to accelerate and focus on the things you want to be doing. And part of that is that finding great content, great application components that you can incorporate into your apps to move faster is really hard. It's hard to discover. It's hard to find high quality content that you can trust that, you know, passes your test and your configuration needs. >>And it's hard to create good content as well. And you're looking for more safety, more guardrails to help guide you along that way so that you can focus on creating value for your company. Secondly, you're telling us that it's a really far to collaborate effectively with your team and you want to do more, to work more effectively together to help your tools become more and more seamless to help you stay in sync, both with yourself across all of your development environments, as well as with your teammates so that you can more effectively collaborate together. Review each other's work, maintain things and keep them in sync. And finally, you want your applications to run consistently in every single environment, whether that's your local development environment, a cloud-based development environment, your CGI pipeline, or the cloud for production, and you want that micro service to provide that consistent experience everywhere you go so that you have similar tools, similar environments, and you don't need to worry about things getting in your way, but instead things make it easy for you to focus on what you wanna do and what Docker is doing to help solve all of these problems for you and your colleagues is creating a collaborative app dev platform. >>And this collaborative application development platform consists of multiple different pieces. I'm not going to walk through all of them today, but the overall view is that we're providing all the tooling you need from the development environment, to the container images, to the collaboration services, to the pipelines and integrations that enable you to focus on making your applications amazing and changing the world. If we start zooming on a one of those aspects, collaboration we hear from developers regularly is that they're challenged in synchronizing their own setups across environments. They want to be able to duplicate the setup of their teammates. Look, then they can easily get up and running with the same applications, the same tooling, the same version of the same libraries, the same frameworks. And they want to know if their applications are good before they're ready to share them in an official space. >>They want to collaborate on things before they're done, rather than feeling like they have to officially published something before they can effectively share it with others to work on it, to solve this. We're thrilled today to announce Docker, dev environments, Docker, dev environments, transform how your team collaborates. They make creating, sharing standardized development environments. As simple as a Docker poll, they make it easy to review your colleagues work without affecting your own work. And they increase the reproducibility of your own work and decreased production issues in doing so because you've got consistent environments all the way through. Now, I'm going to pass it off to our principal product manager, Ben Gotch to walk you through more detail on Docker dev environments. >>Hi, I'm Ben. I work as a principal program manager at DACA. One of the areas that doc has been looking at to see what's hard today for developers is sharing changes that you make from the inner loop where the inner loop is a better development, where you write code, test it, build it, run it, and ultimately get feedback on those changes before you merge them and try and actually ship them out to production. Most amount of us build this flow and get there still leaves a lot of challenges. People need to jump between branches to look at each other's work. Independence. Dependencies can be different when you're doing that and doing this in this new hybrid wall of work. Isn't any easier either the ability to just save someone, Hey, come and check this out. It's become much harder. People can't come and sit down at your desk or take your laptop away for 10 minutes to just grab and look at what you're doing. >>A lot of the reason that development is hard when you're remote, is that looking at changes and what's going on requires more than just code requires all the dependencies and everything you've got set up and that complete context of your development environment, to understand what you're doing and solving this in a remote first world is hard. We wanted to look at how we could make this better. Let's do that in a way that let you keep working the way you do today. Didn't want you to have to use a browser. We didn't want you to have to use a new idea. And we wanted to do this in a way that was application centric. We wanted to let you work with all the rest of the application already using C for all the services and all those dependencies you need as part of that. And with that, we're excited to talk more about docket developer environments, dev environments are new part of the Docker experience that makes it easier you to get started with your whole inner leap, working inside a container, then able to share and collaborate more than just the code. >>We want it to enable you to share your whole modern development environment, your whole setup from DACA, with your team on any operating system, we'll be launching a limited beta of dev environments in the coming month. And a GA dev environments will be ID agnostic and supporting composts. This means you'll be able to use an extend your existing composed files to create your own development environment in whatever idea, working in dev environments designed to be local. First, they work with Docker desktop and say your existing ID, and let you share that whole inner loop, that whole development context, all of your teammates in just one collect. This means if you want to get feedback on the working progress change or the PR it's as simple as opening another idea instance, and looking at what your team is working on because we're using compose. You can just extend your existing oppose file when you're already working with, to actually create this whole application and have it all working in the context of the rest of the services. >>So it's actually the whole environment you're working with module one service that doesn't really understand what it's doing alone. And with that, let's jump into a quick demo. So you can see here, two dev environments up and running. First one here is the same container dev environment. So if I want to go into that, let's see what's going on in the various code button here. If that one open, I can get straight into my application to start making changes inside that dev container. And I've got all my dependencies in here, so I can just run that straight in that second application I have here is one that's opened up in compose, and I can see that I've also got my backend, my front end and my database. So I've got all my services running here. So if I want, I can open one or more of these in a dev environment, meaning that that container has the context that dev environment has the context of the whole application. >>So I can get back into and connect to all the other services that I need to test this application properly, all of them, one unit. And then when I've made my changes and I'm ready to share, I can hit my share button type in the refund them on to share that too. And then give that image to someone to get going, pick that up and just start working with that code and all my dependencies, simple as putting an image, looking ahead, we're going to be expanding development environments, more of your dependencies for the whole developer worst space. We want to look at backing up and letting you share your volumes to make data science and database setups more repeatable and going. I'm still all of this under a single workspace for your team containing images, your dev environments, your volumes, and more we've really want to allow you to create a fully portable Linux development environment. >>So everyone you're working with on any operating system, as I said, our MVP we're coming next month. And that was for vs code using their dev container primitive and more support for other ideas. We'll follow to find out more about what's happening and what's coming up next in the future of this. And to actually get a bit of a deeper dive in the experience. Can we check out the talk I'm doing with Georgie and girl later on today? Thank you, Ben, amazing story about how Docker is helping to make developer teams more collaborative. Now I'd like to talk more about applications while the dev environment is like the workbench around what you're building. The application itself has all the different components, libraries, and frameworks, and other code that make up the application itself. And we hear developers saying all the time things like, how do they know if their images are good? >>How do they know if they're secure? How do they know if they're minimal? How do they make great images and great Docker files and how do they keep their images secure? And up-to-date on every one of those ties into how do I create more trust? How do I know that I'm building high quality applications to enable you to do this even more effectively than today? We are pleased to announce the DACA verified polisher program. This broadens trusted content by extending beyond Docker official images, to give you more and more trusted building blocks that you can incorporate into your applications. It gives you confidence that you're getting what you expect because Docker verifies every single one of these publishers to make sure they are who they say they are. This improves our secure supply chain story. And finally it simplifies your discovery of the best building blocks by making it easy for you to find things that you know, you can trust so that you can incorporate them into your applications and move on and on the right. You can see some examples of the publishers that are involved in Docker, official images and our Docker verified publisher program. Now I'm pleased to introduce you to marina. Kubicki our senior product manager who will walk you through more about what we're doing to create a better experience for you around trust. >>Thank you, Dani, >>Mario Andretti, who is a famous Italian sports car driver. One said that if everything feels under control, you're just not driving. You're not driving fast enough. Maya Andretti is not a software developer and a software developers. We know that no matter how fast we need to go in order to drive the innovation that we're working on, we can never allow our applications to spin out of control and a Docker. As we continue talking to our, to the developers, what we're realizing is that in order to reach that speed, the developers are the, the, the development community is looking for the building blocks and the tools that will, they will enable them to drive at the speed that they need to go and have the trust in those building blocks. And in those tools that they will be able to maintain control over their applications. So as we think about some of the things that we can do to, to address those concerns, uh, we're realizing that we can pursue them in a number of different venues, including creating reliable content, including creating partnerships that expands the options for the reliable content. >>Um, in order to, in a we're looking at creating integrations, no link security tools, talk about the reliable content. The first thing that comes to mind are the Docker official images, which is a program that we launched several years ago. And this is a set of curated, actively maintained, open source images that, uh, include, uh, operating systems and databases and programming languages. And it would become immensely popular for, for, for creating the base layers of, of the images of, of the different images, images, and applications. And would we realizing that, uh, many developers are, instead of creating something from scratch, basically start with one of the official images for their basis, and then build on top of that. And this program has become so popular that it now makes up a quarter of all of the, uh, Docker poles, which essentially ends up being several billion pulse every single month. >>As we look beyond what we can do for the open source. Uh, we're very ability on the open source, uh, spectrum. We are very excited to announce that we're launching the Docker verified publishers program, which is continuing providing the trust around the content, but now working with, uh, some of the industry leaders, uh, in multiple, in multiple verticals across the entire technology technical spec, it costs entire, uh, high tech in order to provide you with more options of the images that you can use for building your applications. And it still comes back to trust that when you are searching for content in Docker hub, and you see the verified publisher badge, you know, that this is, this is the content that, that is part of the, that comes from one of our partners. And you're not running the risk of pulling the malicious image from an employee master source. >>As we look beyond what we can do for, for providing the reliable content, we're also looking at some of the tools and the infrastructure that we can do, uh, to create a security around the content that you're creating. So last year at the last ad, the last year's DockerCon, we announced partnership with sneak. And later on last year, we launched our DACA, desktop and Docker hub vulnerability scans that allow you the options of writing scans in them along multiple points in your dev cycle. And in addition to providing you with information on the vulnerability on, on the vulnerabilities, in, in your code, uh, it also provides you with a guidance on how to re remediate those vulnerabilities. But as we look beyond the vulnerability scans, we're also looking at some of the other things that we can do, you know, to, to, to, uh, further ensure that the integrity and the security around your images, your images, and with that, uh, later on this year, we're looking to, uh, launch the scope, personal access tokens, and instead of talking about them, I will simply show you what they look like. >>So if you can see here, this is my page in Docker hub, where I've created a four, uh, tokens, uh, read-write delete, read, write, read only in public read in public creeper read only. So, uh, earlier today I went in and I, I logged in, uh, with my read only token. And when you see, when I'm going to pull an image, it's going to allow me to pull an image, not a problem success. And then when I do the next step, I'm going to ask to push an image into the same repo. Uh, would you see is that it's going to give me an error message saying that they access is denied, uh, because there is an additional authentication required. So these are the things that we're looking to add to our roadmap. As we continue thinking about the things that we can do to provide, um, to provide additional building blocks, content, building blocks, uh, and, and, and tools to build the trust so that our DACA developer and skinned code faster than Mario Andretti could ever imagine. Uh, thank you to >>Thank you, marina. It's amazing what you can do to improve the trusted content so that you can accelerate your development more and move more quickly, move more collaboratively and build upon the great work of others. Finally, we hear over and over as that developers are working on their applications that they're looking for, environments that are consistent, that are the same as production, and that they want their applications to really run anywhere, any environment, any architecture, any cloud one great example is the recent announcement of apple Silicon. We heard from developers on uproar that they needed Docker to be available for that architecture before they could add those to it and be successful. And we listened. And based on that, we are pleased to share with you Docker, desktop on apple Silicon. This enables you to run your apps consistently anywhere, whether that's developing on your team's latest dev hardware, deploying an ARM-based cloud environments and having a consistent architecture across your development and production or using multi-year architecture support, which enables your whole team to collaborate on its application, using private repositories on Docker hub, and thrilled to introduce you to Hughie cower, senior director for product management, who will walk you through more of what we're doing to create a great developer experience. >>Senior director of product management at Docker. And I'd like to jump straight into a demo. This is the Mac mini with the apple Silicon processor. And I want to show you how you can now do an end-to-end arm workflow from my M one Mac mini to raspberry PI. As you can see, we have vs code and Docker desktop installed on a, my, the Mac mini. I have a small example here, and I have a raspberry PI three with an led strip, and I want to turn those LEDs into a moving rainbow. This Dockerfile here, builds the application. We build the image with the Docker, build X command to make the image compatible for all raspberry pies with the arm. 64. Part of this build is built with the native power of the M one chip. I also add the push option to easily share the image with my team so they can give it a try to now Dr. >>Creates the local image with the application and uploads it to Docker hub after we've built and pushed the image. We can go to Docker hub and see the new image on Docker hub. You can also explore a variety of images that are compatible with arm processors. Now let's go to the raspberry PI. I have Docker already installed and it's running Ubuntu 64 bit with the Docker run command. I can run the application and let's see what will happen from there. You can see Docker is downloading the image automatically from Docker hub and when it's running, if it's works right, there are some nice colors. And with that, if we have an end-to-end workflow for arm, where continuing to invest into providing you a great developer experience, that's easy to install. Easy to get started with. As you saw in the demo, if you're interested in the new Mac, mini are interested in developing for our platforms in general, we've got you covered with the same experience you've come to expect from Docker with over 95,000 arm images on hub, including many Docker official images. >>We think you'll find what you're looking for. Thank you again to the community that helped us to test the tech previews. We're so delighted to hear when folks say that the new Docker desktop for apple Silicon, it just works for them, but that's not all we've been working on. As Dani mentioned, consistency of developer experience across environments is so important. We're introducing composed V2 that makes compose a first-class citizen in the Docker CLI you no longer need to install a separate composed biter in order to use composed, deploying to production is simpler than ever with the new compose integration that enables you to deploy directly to Amazon ECS or Azure ACI with the same methods you use to run your application locally. If you're interested in running slightly different services, when you're debugging versus testing or, um, just general development, you can manage that all in one place with the new composed service to hear more about what's new and Docker desktop, please join me in the three 15 breakout session this afternoon. >>And now I'd love to tell you a bit more about bill decks and convince you to try it. If you haven't already it's our next gen build command, and it's no longer experimental as shown in the demo with built X, you'll be able to do multi architecture builds, share those builds with your team and the community on Docker hub. With build X, you can speed up your build processes with remote caches or build all the targets in your composed file in parallel with build X bake. And there's so much more if you're using Docker, desktop or Docker, CE you can use build X checkout tonus is talk this afternoon at three 45 to learn more about build X. And with that, I hope everyone has a great Dr. Khan and back over to you, Donnie. >>Thank you UA. It's amazing to hear about what we're doing to create a better developer experience and make sure that Docker works everywhere you need to work. Finally, I'd like to wrap up by showing you everything that we've announced today and everything that we've done recently to make your lives better and give you more and more for the single price of your Docker subscription. We've announced the Docker verified publisher program we've announced scoped personal access tokens to make it easier for you to have a secure CCI pipeline. We've announced Docker dev environments to improve your collaboration with your team. Uh, we shared with you Docker, desktop and apple Silicon, to make sure that, you know, Docker runs everywhere. You need it to run. And we've announced Docker compose version two, finally making it a first-class citizen amongst all the other great Docker tools. And we've done so much more recently as well from audit logs to advanced image management, to compose service profiles, to improve where you can run Docker more easily. >>Finally, as we look forward, where we're headed in the upcoming year is continuing to invest in these themes of helping you build, share, and run modern apps more effectively. We're going to be doing more to help you create a secure supply chain with which only grows more and more important as time goes on. We're going to be optimizing your update experience to make sure that you can easily understand the current state of your application, all its components and keep them all current without worrying about breaking everything as you're doing. So we're going to make it easier for you to synchronize your work. Using cloud sync features. We're going to improve collaboration through dev environments and beyond, and we're going to do make it easy for you to run your microservice in your environments without worrying about things like architecture or differences between those environments. Thank you so much. I'm thrilled about what we're able to do to help make your lives better. And now you're going to be hearing from one of our customers about what they're doing to launch their business with Docker >>I'm Matt Falk, I'm the head of engineering and orbital insight. And today I want to talk to you a little bit about data from space. So who am I like many of you, I'm a software developer and a software developer about seven companies so far, and now I'm a head of engineering. So I spend most of my time doing meetings, but occasionally I'll still spend time doing design discussions, doing code reviews. And in my free time, I still like to dabble on things like project oiler. So who's Oberlin site. What do we do? Portal insight is a large data supplier and analytics provider where we take data geospatial data anywhere on the planet, any overhead sensor, and translate that into insights for the end customer. So specifically we have a suite of high performance, artificial intelligence and machine learning analytics that run on this geospatial data. >>And we build them to specifically determine natural and human service level activity anywhere on the planet. What that really means is we take any type of data associated with a latitude and longitude and we identify patterns so that we can, so we can detect anomalies. And that's everything that we do is all about identifying those patterns to detect anomalies. So more specifically, what type of problems do we solve? So supply chain intelligence, this is one of the use cases that we we'd like to talk about a lot. It's one of our main primary verticals that we go after right now. And as Scott mentioned earlier, this had a huge impact last year when COVID hit. So specifically supply chain intelligence is all about identifying movement patterns to and from operating facilities to identify changes in those supply chains. How do we do this? So for us, we can do things where we track the movement of trucks. >>So identifying trucks, moving from one location to another in aggregate, same thing we can do with foot traffic. We can do the same thing for looking at aggregate groups of people moving from one location to another and analyzing their patterns of life. We can look at two different locations to determine how people are moving from one location to another, or going back and forth. All of this is extremely valuable for detecting how a supply chain operates and then identifying the changes to that supply chain. As I said last year with COVID, everything changed in particular supply chains changed incredibly, and it was hugely important for customers to know where their goods or their products are coming from and where they were going, where there were disruptions in their supply chain and how that's affecting their overall supply and demand. So to use our platform, our suite of tools, you can start to gain a much better picture of where your suppliers or your distributors are going from coming from or going to. >>So what's our team look like? So my team is currently about 50 engineers. Um, we're spread into four different teams and the teams are structured like this. So the first team that we have is infrastructure engineering and this team largely deals with deploying our Dockers using Kubernetes. So this team is all about taking Dockers, built by other teams, sometimes building the Dockers themselves and putting them into our production system, our platform engineering team, they produce these microservices. So they produce microservice, Docker images. They develop and test with them locally. Their entire environments are dockerized. They produce these doctors, hand them over to him for infrastructure engineering to be deployed. Similarly, our product engineering team does the same thing. They develop and test with Dr. Locally. They also produce a suite of Docker images that the infrastructure team can then deploy. And lastly, we have our R and D team, and this team specifically produces machine learning algorithms using Nvidia Docker collectively, we've actually built 381 Docker repositories and 14 million. >>We've had 14 million Docker pools over the lifetime of the company, just a few stats about us. Um, but what I'm really getting to here is you can see actually doctors becoming almost a form of communication between these teams. So one of the paradigms in software engineering that you're probably familiar with encapsulation, it's really helpful for a lot of software engineering problems to break the problem down, isolate the different pieces of it and start building interfaces between the code. This allows you to scale different pieces of the platform or different pieces of your code in different ways that allows you to scale up certain pieces and keep others at a smaller level so that you can meet customer demands. And for us, one of the things that we can largely do now is use Dockers as that interface. So instead of having an entire platform where all teams are talking to each other, and everything's kind of, mishmashed in a monolithic application, we can now say this team is only able to talk to this team by passing over a particular Docker image that defines the interface of what needs to be built before it passes to the team and really allows us to scalp our development and be much more efficient. >>Also, I'd like to say we are hiring. Um, so we have a number of open roles. We have about 30 open roles in our engineering team that we're looking to fill by the end of this year. So if any of this sounds really interesting to you, please reach out after the presentation. >>So what does our platform do? Really? Our platform allows you to answer any geospatial question, and we do this at three different inputs. So first off, where do you want to look? So we did this as what we call an AOI or an area of interest larger. You can think of this as a polygon drawn on the map. So we have a curated data set of almost 4 million AOIs, which you can go and you can search and use for your analysis, but you're also free to build your own. Second question is what you want to look for. We do this with the more interesting part of our platform of our machine learning and AI capabilities. So we have a suite of algorithms that automatically allow you to identify trucks, buildings, hundreds of different types of aircraft, different types of land use, how many people are moving from one location to another different locations that people in a particular area are moving to or coming from all of these different analyses or all these different analytics are available at the click of a button, and then determine what you want to look for. >>Lastly, you determine when you want to find what you're looking for. So that's just, uh, you know, do you want to look for the next three hours? Do you want to look for the last week? Do you want to look every month for the past two, whatever the time cadence is, you decide that you hit go and out pops a time series, and that time series tells you specifically where you want it to look what you want it to look for and how many, or what percentage of the thing you're looking for appears in that area. Again, we do all of this to work towards patterns. So we use all this data to produce a time series from there. We can look at it, determine the patterns, and then specifically identify the anomalies. As I mentioned with supply chain, this is extremely valuable to identify where things change. So we can answer these questions, looking at a particular operating facility, looking at particular, what is happening with the level of activity is at that operating facility where people are coming from, where they're going to, after visiting that particular facility and identify when and where that changes here, you can just see it's a picture of our platform. It's actually showing all the devices in Manhattan, um, over a period of time. And it's more of a heat map view. So you can actually see the hotspots in the area. >>So really the, and this is the heart of the talk, but what happened in 2020? So for men, you know, like many of you, 2020 was a difficult year COVID hit. And that changed a lot of what we're doing, not from an engineering perspective, but also from an entire company perspective for us, the motivation really became to make sure that we were lowering our costs and increasing innovation simultaneously. Now those two things often compete with each other. A lot of times you want to increase innovation, that's going to increase your costs, but the challenge last year was how to do both simultaneously. So here's a few stats for you from our team. In Q1 of last year, we were spending almost $600,000 per month on compute costs prior to COVID happening. That wasn't hugely a concern for us. It was a lot of money, but it wasn't as critical as it was last year when we really needed to be much more efficient. >>Second one is flexibility for us. We were deployed on a single cloud environment while we were cloud thought ready, and that was great. We want it to be more flexible. We want it to be on more cloud environments so that we could reach more customers. And also eventually get onto class side networks, extending the base of our customers as well from a custom analytics perspective. This is where we get into our traction. So last year, over the entire year, we computed 54,000 custom analytics for different users. We wanted to make sure that this number was steadily increasing despite us trying to lower our costs. So we didn't want the lowering cost to come as the sacrifice of our user base. Lastly, of particular percentage here that I'll say definitely needs to be improved is 75% of our projects never fail. So this is where we start to get into a bit of stability of our platform. >>Now I'm not saying that 25% of our projects fail the way we measure this is if you have a particular project or computation that runs every day and any one of those runs sale account, that is a failure because from an end-user perspective, that's an issue. So this is something that we know we needed to improve on and we needed to grow and make our platform more stable. I'm going to something that we really focused on last year. So where are we now? So now coming out of the COVID valley, we are starting to soar again. Um, we had, uh, back in April of last year, we had the entire engineering team. We actually paused all development for about four weeks. You had everyone focused on reducing our compute costs in the cloud. We got it down to 200 K over the period of a few months. >>And for the next 12 months, we hit that number every month. This is huge for us. This is extremely important. Like I said, in the COVID time period where costs and operating efficiency was everything. So for us to do that, that was a huge accomplishment last year and something we'll keep going forward. One thing I would actually like to really highlight here, two is what allowed us to do that. So first off, being in the cloud, being able to migrate things like that, that was one thing. And we were able to use there's different cloud services in a more particular, in a more efficient way. We had a very detailed tracking of how we were spending things. We increased our data retention policies. We optimized our processing. However, one additional piece was switching to new technologies on, in particular, we migrated to get lab CICB. >>Um, and this is something that the costs we use Docker was extremely, extremely easy. We didn't have to go build new new code containers or repositories or change our code in order to do this. We were simply able to migrate the containers over and start using a new CIC so much. In fact, that we were able to do that migration with three engineers in just two weeks from a cloud environment and flexibility standpoint, we're now operating in two different clouds. We were able to last night, I've over the last nine months to operate in the second cloud environment. And again, this is something that Docker helped with incredibly. Um, we didn't have to go and build all new interfaces to all new, different services or all different tools in the next cloud provider. All we had to do was build a base cloud infrastructure that ups agnostic the way, all the different details of the cloud provider. >>And then our doctors just worked. We can move them to another environment up and running, and our platform was ready to go from a traction perspective. We're about a third of the way through the year. At this point, we've already exceeded the amount of customer analytics we produce last year. And this is thanks to a ton more albums, that whole suite of new analytics that we've been able to build over the past 12 months and we'll continue to build going forward. So this is really, really great outcome for us because we were able to show that our costs are staying down, but our analytics and our customer traction, honestly, from a stability perspective, we improved from 75% to 86%, not quite yet 99 or three nines or four nines, but we are getting there. Um, and this is actually thanks to really containerizing and modularizing different pieces of our platform so that we could scale up in different areas. This allowed us to increase that stability. This piece of the code works over here, toxin an interface to the rest of the system. We can scale this piece up separately from the rest of the system, and that allows us much more easily identify issues in the system, fix those and then correct the system overall. So basically this is a summary of where we were last year, where we are now and how much more successful we are now because of the issues that we went through last year and largely brought on by COVID. >>But that this is just a screenshot of the, our, our solution actually working on supply chain. So this is in particular, it is showing traceability of a distribution warehouse in salt lake city. It's right in the center of the screen here. You can see the nice kind of orange red center. That's a distribution warehouse and all the lines outside of that, all the dots outside of that are showing where people are, where trucks are moving from that location. So this is really helpful for supply chain companies because they can start to identify where their suppliers are, are coming from or where their distributors are going to. So with that, I want to say, thanks again for following along and enjoy the rest of DockerCon.

Published Date : May 27 2021

SUMMARY :

We know that collaboration is key to your innovation sharing And we know from talking with many of you that you and your developer Have you seen the email from Scott? I was thinking we could try, um, that new Docker dev environments feature. So if you hit the share button, what I should do is it will take all of your code and the dependencies and Uh, let me get that over to you, All right. It's just going to grab the image down, which you can take all of the code, the dependencies only get brunches working It's connected to the container. So let's just have a look at what you use So I've had a look at what you were doing and I'm actually going to change. Let me grab the link. it should be able to open up the code that I've changed and then just run it in the same way you normally do. I think we should ship it. For example, in response to COVID we saw global communities, including the tech community rapidly teams make sense of all this specifically, our goal is to provide development teams with the trusted We had powerful new capabilities to the Docker product, both free and subscription. And finally delivering an easy to use well-integrated development experience with best of breed tools and content And what we've learned in our discussions with you will have long asking a coworker to take a look at your code used to be as easy as swiveling their chair around, I'd like to take a moment to share with Docker and our partners are doing for trusted content, providing development teams, and finally, public repos for communities enable community projects to be freely shared with anonymous Lastly, the container images themselves and this end to end flow are built on open industry standards, but the Docker team rose to the challenge and worked together to continue shipping great product, the again for joining us, we look forward to having a great DockerCon with you today, as well as a great year So let's dive in now, I know this may be hard for some of you to believe, I taught myself how to code. And by the way, I'm showing you actions in Docker, And the cool thing is you can use it on any And if I can do it, I know you can too, but enough yapping let's get started to save Now you can do this in a couple of ways, whether you're doing it in your preferred ID or for today's In essence, with automation, you can be kind to your future self And I hope you all go try it out, but why do we care about all of that? And to get into that wonderful state that we call flow. and eliminate or outsource the rest because you don't need to do it, make the machines Speaking of the open source ecosystem we at get hub are so to be here with all you nerds. Komack lovely to see you here. We want to help you get your applications from your laptops, And it's all a seamless thing from, you know, from your code to the cloud local And we all And we know that you use So we need to make that as easier. We know that they might go to 25% of poles we need just keep updating base images and dependencies, and we'll, we're going to help you have the control to cloud is RA and the cloud providers aware most of you ship your occasion production Then we know you do, and we know that you want it to be easier to use in your It's hard to find high quality content that you can trust that, you know, passes your test and your configuration more guardrails to help guide you along that way so that you can focus on creating value for your company. that enable you to focus on making your applications amazing and changing the world. Now, I'm going to pass it off to our principal product manager, Ben Gotch to walk you through more doc has been looking at to see what's hard today for developers is sharing changes that you make from the inner dev environments are new part of the Docker experience that makes it easier you to get started with your whole inner leap, We want it to enable you to share your whole modern development environment, your whole setup from DACA, So you can see here, So I can get back into and connect to all the other services that I need to test this application properly, And to actually get a bit of a deeper dive in the experience. Docker official images, to give you more and more trusted building blocks that you can incorporate into your applications. We know that no matter how fast we need to go in order to drive The first thing that comes to mind are the Docker official images, And it still comes back to trust that when you are searching for content in And in addition to providing you with information on the vulnerability on, So if you can see here, this is my page in Docker hub, where I've created a four, And based on that, we are pleased to share with you Docker, I also add the push option to easily share the image with my team so they can give it a try to now continuing to invest into providing you a great developer experience, a first-class citizen in the Docker CLI you no longer need to install a separate composed And now I'd love to tell you a bit more about bill decks and convince you to try it. image management, to compose service profiles, to improve where you can run Docker more easily. So we're going to make it easier for you to synchronize your work. And today I want to talk to you a little bit about data from space. What that really means is we take any type of data associated with a latitude So to use our platform, our suite of tools, you can start to gain a much better picture of where your So the first team that we have is infrastructure This allows you to scale different pieces of the platform or different pieces of your code in different ways that allows So if any of this sounds really interesting to you, So we have a suite of algorithms that automatically allow you to identify So you can actually see the hotspots in the area. the motivation really became to make sure that we were lowering our costs and increasing innovation simultaneously. of particular percentage here that I'll say definitely needs to be improved is 75% Now I'm not saying that 25% of our projects fail the way we measure this is if you have a particular And for the next 12 months, we hit that number every month. night, I've over the last nine months to operate in the second cloud environment. And this is thanks to a ton more albums, they can start to identify where their suppliers are, are coming from or where their distributors are going

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Mario AndrettiPERSON

0.99+

DaniPERSON

0.99+

Matt FalkPERSON

0.99+

Dana LawsonPERSON

0.99+

AmazonORGANIZATION

0.99+

Maya AndrettiPERSON

0.99+

DonniePERSON

0.99+

MicrosoftORGANIZATION

0.99+

MonaPERSON

0.99+

NicolePERSON

0.99+

UNICEFORGANIZATION

0.99+

25%QUANTITY

0.99+

GermanyLOCATION

0.99+

14 millionQUANTITY

0.99+

75%QUANTITY

0.99+

ManhattanLOCATION

0.99+

KhanPERSON

0.99+

10 minutesQUANTITY

0.99+

last yearDATE

0.99+

99QUANTITY

0.99+

1.3 timesQUANTITY

0.99+

1.2 timesQUANTITY

0.99+

ClairePERSON

0.99+

DockerORGANIZATION

0.99+

ScottPERSON

0.99+

BenPERSON

0.99+

UC IrvineORGANIZATION

0.99+

85%QUANTITY

0.99+

OracleORGANIZATION

0.99+

34%QUANTITY

0.99+

JustinPERSON

0.99+

JoeyPERSON

0.99+

80%QUANTITY

0.99+

160 imagesQUANTITY

0.99+

2020DATE

0.99+

$10,000QUANTITY

0.99+

10 secondsQUANTITY

0.99+

23 minutesQUANTITY

0.99+

JavaScriptTITLE

0.99+

AprilDATE

0.99+

twoQUANTITY

0.99+

56%QUANTITY

0.99+

PythonTITLE

0.99+

MollyPERSON

0.99+

Mac miniCOMMERCIAL_ITEM

0.99+

Hughie cowerPERSON

0.99+

two weeksQUANTITY

0.99+

100%QUANTITY

0.99+

GeorgiePERSON

0.99+

Matt fallPERSON

0.99+

MarsLOCATION

0.99+

Second questionQUANTITY

0.99+

KubickiPERSON

0.99+

MobyPERSON

0.99+

IndiaLOCATION

0.99+

DockerConEVENT

0.99+

Youi CalPERSON

0.99+

three ninesQUANTITY

0.99+

J frogORGANIZATION

0.99+

200 KQUANTITY

0.99+

appleORGANIZATION

0.99+

SharonPERSON

0.99+

AWSORGANIZATION

0.99+

10 XQUANTITY

0.99+

COVID-19OTHER

0.99+

windowsTITLE

0.99+

381QUANTITY

0.99+

NvidiaORGANIZATION

0.99+

Nipun Agarwal, Oracle | CUBEconversation


 

(bright upbeat music) >> Hello everyone, and welcome to the special exclusive CUBE Conversation, where we continue our coverage of the trends of the database market. With me is Nipun Agarwal, who's the vice president, MySQL HeatWave in advanced development at Oracle. Nipun, welcome. >> Thank you Dave. >> I love to have technical people on the Cube to educate, debate, inform, and we've extensively covered this market. We were all over the Snowflake IPO and at that time I remember, I challenged organizations bring your best people. Because I want to better understand what's happening at Database. After Oracle kind of won the Database wars 20 years ago, Database kind of got boring. And then it got really exciting with the big data movement, and all the, not only SQL stuff coming out, and Hadoop and blah, blah, blah. And now it's just exploding. You're seeing huge investments from many of your competitors, VCs are trying to get into the action. Meanwhile, as I've said many, many times, your chairman and head of technology, CTO, Larry Ellison, continues to invest to keep Oracle relevant. So it's really been fun to watch and I really appreciate you coming on. >> Sure thing. >> We have written extensively, we talked to a lot of Oracle customers. You get the leading mission critical database in the world. Everybody from Fortune 100, we evaluated what Gardner said about the operational databases. I think there's not a lot of question there. And we've written about that on WikiBound about you're converged databases, and the strategy there, and we're going to get into that. We've covered Autonomous Data Warehouse Exadata Cloud at Customer, and then we just want to really try to get into your area, which has been, kind of caught our attention recently. And I'm talking about the MySQL Database Service with HeatWave. I love the name, I laugh. It was an unveiled, I don't know, a few months ago. So Nipun, let's start the discussion today. Maybe you can update our viewers on what is HeatWave? What's the overall focus with Oracle? And how does it fit into the Cloud Database Service? >> Sure Dave. So HeatWave is a in-memory query accelerator for the MySQL Database Service for speeding up analytic queries as well as long running complex OLTP queries. And this is all done in the context of a single database which is the MySQL Database Service. Also, all existing MySQL applications or MySQL compatible tools and applications continue to work as is. So there is no change. And with this HeatWave, Oracle is delivering the only MySQL service which provides customers with a single unified platform for both analytic as well as transaction processing workloads. >> Okay, so, we've seen open source databases in the cloud growing very rapidly. I mentioned Snowflake, I think Google's BigQuery, get some mention, we'll talk, we'll maybe talk more about Redshift later on, but what I'm wondering, well let's talk about now, how does MySQL HeatWave service, how does that compare to MySQL-based services from other cloud vendors? I can get MySQL from others. In fact, I think we do. I think we run WikiBound on the LAMP stack. I think it's running on Amazon, but so how does your service compare? >> No other vendor, like, no other vendor offers this differentiated solution with an open source database namely, having a single database, which is optimized both for transactional processing and analytics, right? So the example is like MySQL. A lot of other cloud vendors provide MySQL service but MySQL has been optimized for transaction processing so when customs need to run analytics they need to move the data out of MySQL into some other database for any analytics, right? So we are the only vendor which is now offering this unified solution for both transactional processing analytics. That's the first point. Second thing is, most of the vendors out there have taken open source databases and they're basically hosting it in the cloud. Whereas HeatWave, has been designed from the ground up for the cloud, and it is a 100% compatible with MySQL applications. And the fact that we have designed it from the ground up for the cloud, maybe I'll spend 100s of person years of research and engineering means that we have a solution, which is very, very scalable, it's very optimized in terms of performance, and it is very inexpensive in terms of the cost. >> Are you saying, well, wait, are you saying that you essentially rewrote MySQL to create HeatWave but at the same time maintained compatibility with existing applications? >> Right. So we enhanced MySQL significantly and we wrote a whole bunch of new code which is brand new code optimized for the cloud in such a manner that yes, it is 100% compatible with all existing MySQL applications. >> What does it mean? And if I'm to optimize for the cloud, I mean, I hear that and I say, okay, it's taking advantage of cloud-native. I hear kind of the buzzwords, cloud-first, cloud-native. What does it specifically mean from a technical standpoint? >> Right. So first, let's talk about performance. What we have done is that we have looked at two aspects. We have worked with shapes like for instance, like, the compute shapes which provide the best performance for dollar, per dollar. So I'll give you a couple of examples. We have optimized for certain shifts. So, HeatWave is in-memory query accelerator. So the cost of the system is dominated by the cost. So we are working with chips which provide the cheapest cost per terabyte of memory. Secondly, we are using commodity cloud services in such a manner that it's in-optimized for both performance as well as performance per dollar. So, example is, we are not using any locally-attached SSDs. We use ObjectStore because it's very inexpensive. And then I guess at some point I will get into the details of the architecture. The system has been really, really designed for massive scalability. So as you add more compute, as you add more service, the system continues to scale almost perfectly linearly. So this is what I mean in terms of being optimized for the cloud. >> All right, great. >> And furthermore, (indistinct). >> Thank you. No, carry on. >> Over the next few months, you will see a bunch of other announcements where we're adding a whole bunch of machine learning and data driven-based automation which we believe is critical for the cloud. So optimized for performance, optimized for the cloud, and machine learning-based automation which we believe is critical for any good cloud-based service. >> All right, I want to come back and ask you more about the architecture, but you mentioned some of the others taking open source databases and shoving them into the cloud. Let's take the example of AWS. They have a series of specialized data stores and, for different workloads, Aurora is for OLTP I actually think it's based on MySQL Redshift which is based on ParAccel. And so, and I've asked Amazon about this, and their response is, actually kind of made sense to me. Look, we want the right tool for the right job, we want access to the primitives because when the market changes we can change faster as opposed to, if we put, if we start building bigger and bigger databases with more functionality, it's, we're not as agile. So that kind of made sense to me. I know we, again, we use a lot, we use, I think I said MySQL in Amazon we're using DynamoDB, works, that's cool. We're not huge. And I, we fully admit and we've researched this, when you start to get big that starts to get maybe expensive. But what do you think about that approach and why is your approach better? >> Right, we believe that there are multiple drawbacks of having different databases or different services, one, optimized for transactional processing and one for analytics and having to ETL between these different services. First of all, it's expensive because you have to manage different databases. Secondly, it's complex. From an application standpoint, applications need, now need to understand the semantics of two different databases. It's inefficient because you have to transfer data at some PRPC from one database to the other one. It's not secure because there is security aspects involved when your transferring data and also the identity of users in the two different databases is different. So it's, the approach which has been taken by Amazons and such, we believe, is more costly, complex, inefficient and not secure. Whereas with HeatWave, all the data resides in one database which is MySQL and it can run both transaction processing and analytics. So in addition to all the benefits I talked about, customers can also make their decisions in real time because there is no need to move the data. All the data resides in a single database. So as soon as you make any changes, those changes are visible to customers for queries right away, which is not the case when you have different siloed specialized databases. >> Okay, that, a lot of ways to skin a cat and that what you just said makes sense. By the way, we were saying before, companies have taken off the shelf or open source database has shoved them in the cloud. I have to give Amazon some props. They actually have done engineering to Aurora and Redshift. And they've got the engineering capabilities to do that. But you can see, for example, in Redshift the way they handle separating compute from storage it's maybe not as elegant as some of the other players like a Snowflake, for example, but they get there and they, maybe it's a little bit more brute force but so I don't want to just make it sound like they're just hosting off the shelf in the cloud. But is it fair to say that there's like a crossover point? So in other words, if I'm smaller and I'm not, like doing a bunch of big, like us, I mean, it's fine. It's easy, I spin it up. It's cheaper than having to host my own servers. So there's, presumably there's a sweet spot for that approach and a sweet spot for your approach. Is that fair or do you feel like you can cover a wider spectrum? >> We feel we can cover the entire spectrum, not wider, the entire spectrum. And we have benchmarks published which are actually available on GitHub for anyone to try. You will see that this approach you have taken with the MySQL Database Service in HeatWave, we are faster, we are cheaper without having to move the data. And the mileage or the amount of improvement you will get, surely vary. So if you have less data the amount of improvement you will get, maybe like say 100 times, right, or 500 times, but smaller data sizes. If you get to lots of data sizes this improvement amplifies to 1000 times or 10,000 times. And similarly for the cost, if the data size is smaller, the cost advantage you will have is less, maybe MySQL HeatWave is one third the cost. If the data size is larger, the cost advantage amplifies. So to your point, MySQL Database Service in HeatWave is going to be better for all sizes but the amount of mileage or the amount of benefit you will get increases as the size of the data increases. >> Okay, so you're saying you got better performance, better cost, better price performance. Let me just push back a little bit on this because I, having been around for awhile, I often see these performance and price comparisons. And what often happens is a vendor will take the latest and greatest, the one they just announced and they'll compare it to an N-1 or an N-2, running on old hardware. So, is, you're normalizing for that, is that the game you're playing here? I mean, how can you, give us confidence that this is easier kind of legitimate benchmarks in your GitHub repo. >> Absolutely. I'll give you a bunch of like, information. But let me preface this by saying that all of our scripts are available in the open source in the GitHub repo for anyone to try and we would welcome feedback otherwise. So we have taken, yes, the latest version of MySQL Database Service in HeatWave, we have optimized it, and we have run multiple benchmarks. For instance, TBC-H, TPC-DS, right? Because the amount of improvement a query will get depends upon the specific query, depends upon the predicates, it depends on the selectivity so we just wanted to use standard benchmarks. So it's not the case that if you're using certain classes of query, excuse me, benefit, get them more. So, standard benchmarks. Similarly, for the other vendors or other services like Redshift, we have run benchmarks on the latest shapes of Redshift the most optimized configuration which they recommend, running their scripts. So this is not something that, hey, we're just running out of the box. We have optimized Aurora, we have optimized (indistinct) to the best and possible extent we can based on their guidelines, based on their latest release, and that's what you're talking about in terms of the numbers. >> All right. Please continue. >> Now, for some other vendors, if we get to the benchmark section, we'll talk about, we are comparing with other services, let's say Snowflake. Well there, there are issues in terms of you can't legally run Snowflake numbers, right? So there, we have looked at some reports published by Gigaom report. and we are taking the numbers published by the Gigaom report for Snowflake, Google BigQuery and as you'll see maps numbers, right? So those, we have not won ourselves. But for AWS Redshift, as well as AWS Aurora, we have run the numbers and I believe these are the best numbers anyone can get. >> I saw that Gigaom report and I got to say, Gigaom, sometimes I'm like, eh, but I got to say that, I forget the guy's name, he knew what he was talking about. He did a good job, I thought. I was curious as to the workload. I always say, well, what's the workload. And, but I thought that report was pretty detailed. And Snowflake did not look great in that report. Oftentimes, and they've been marketing the heck out of it. I forget who sponsored it. It is, it was sponsored content. But, I did, I remember seeing that and thinking, hmm. So, I think maybe for Snowflake that sweet spot is not, maybe not that performance, maybe it's the simplicity and I think that's where they're making their mark. And most of their databases are small and a lot of read-only stuff. And so they've found a market there. But I want to come back to the architecture and really sort of understand how you've able, you've been able to get this range of both performance and cost you talked about. I thought I heard that you're optimizing the chips, you're using ObjectStore. You're, you've got an architecture that's not using SSD, it's using ObjectStore. So this, is their cashing there? I wonder if you could just give us some details of the architecture and tell us how you got to where you are. >> Right, so let me start off saying like, what are the kind of numbers we are talking about just to kind of be clear, like what the improvements are. So if you take the MySQL Database Service in HeatWave in Oracle Cloud and compare it with MySQL service in any other cloud, and if you look at smaller data sizes, say data sizes which are about half a terabyte or so, HeatWave is 400 times faster, 400 times faster. And as you get to... >> Sorry. Sorry to interrupt. What are you measuring there? Faster in terms of what? >> Latency. So we take TCP-H 22 queries, we run them on HeatWave, and we run the same queries on MySQL service on any other cloud, half a terabyte and the performance in terms of latency is 400 times faster in HeatWave. >> Thank you. Okay. >> If you go to a lot of other data sites, then the other data point of view, we're looking at say something like, 4 TB, there, we did two comparisons. One is with AWS Aurora, which is, as you said, they have taken MySQL. They have done a bunch of innovations over there and we are offering it as a premier service. So on 4 TB TPC-H, MySQL Database Service with HeatWave is 1100 times faster than Aurora. It is three times faster than the fastest shape of Redshift. So Redshift comes in different flavors some talking about dense compute too, right? And again, looking at the most recommended configuration from Redshift. So 1100 times faster that Aurora, three times faster than Redshift and at one third, the cost. So this where I just really want to point out that it is much faster and much cheaper. One third the cost. And then going back to the Gigaom report, there was a comparison done with Snowflake, Google BigQuery, Redshift, Azure Synapse. I wouldn't go into the numbers here but HeatWave was faster on both TPC-H as well as TPC-DS across all these products and cheaper compared to any of these products. So faster, cheaper on both the benchmarks across all these products. Now let's come to, like, what is the technology underneath? >> Great. >> So, basically there are three parts which you're going to see. One is, improve performance, very good scale, and improve a lower cost. So the first thing is that HeatWave has been optimized and, for the cloud. And when I say that, we talked about this a bit earlier. One is we are using the cheapest shapes which are available. We're using the cheapest services which are available without having to compromise the performance and then there is this machine learning-based automation. Now, underneath, in terms of the architecture of HeatWave there are basically, I would say, four key things. First is, HeatWave is an in-memory engine that a presentation which we have in memory is a hybrid columnar representation which is optimized for vector process. That's the first thing. And that's pretty table stakes these days for anyone who wants to do in-memory analytics except that it's hybrid columnar which is optimized for vector processing. So that's the first thing. The second thing which starts getting to be novel is that HeatWave has a massively parallel architecture which is enabled by a massively partitioned architecture. So we take the data, we read the data from MySQL into the memory of the HeatWave and we massively partition this data. So as we're reading the data, we're partitioning the data based on the workload, the sizes of these partitions is such that it fits in the cache of the underlying processor and then we're able to consume these partitions really, really fast. So that's the second bit which is like, massively parallel architecture enabled by massively partitioned architecture. Then the third thing is, that we have developed new state-of-art algorithms for distributed query processing. So for many of the workloads, we find that joints are the long pole in terms of the amount of time it takes. So we at Oracle have developed new algorithms for distributed joint processing and similarly for many other operators. And this is how we're being able to consume this data or process this data, which is in-memory really, really fast. And finally, and what we have, is that we have an eye for scalability and we have designed algorithms such that there's a lot of overlap between compute and communication, which means that as you're sending data across various nodes and there could be like, dozens of of nodes or 100s of nodes that they're able to overlap the computation time with the communication time and this is what gives us massive scalability in the cloud. >> Yeah, so, some hard core database techniques that you've brought to HeatWave, that's impressive. Thank you for that description. Let me ask you, just to go to quicker side. So, MySQL is open source, HeatWave is what? Is it like, open core? Is it open source? >> No, so, HeatWave is something which has been designed and optimized for the cloud. So it can't be open source. So any, it's not open service. >> It is a service. >> It is a service. That's correct. >> So it's a managed service that I pay Oracle to host for me. Okay. Got it. >> That's right. >> Okay, I wonder if you could talk about some of the use cases that you're seeing for HeatWave, any patterns that you're seeing with customers? >> Sure, so we've had the service, we had the HeatWave service in limited availability for almost 15 months and it's been about five months since we have gone G. And there's a very interesting trend of our customers we're seeing. The first one is, we are seeing many migrations from AWS specifically from Aurora. Similarly, we are seeing many migrations from Azure MySQL we're migrations from Google. And the number one reason customers are coming is because of ease of use. Because they have their databases currently siloed. As you were talking about some for optimized for transactional processing, some for analytics. Here, what customers find is that in a single database, they're able to get very good performance, they don't need to move the data around, they don't need to manage multiple databaes. So we are seeing many migrations from these services. And the number one reason is reduce complexity of ease of use. And the second one is, much better performance and reduced costs, right? So that's the first thing. We are very excited and delighted to see the number of migrations we're getting. The second thing which we're seeing is, initially, when we had the service announced, we were like, targeting really towards analytics. But now what are finding is, many of these customers, for instance, who have be running on Aurora, when they are moving from MySQL in HeatWave, they are finding that many of the OLTP queries as well, are seeing significant acceleration with the HeatWave. So now customers are moving their entire applications or, to HeatWave. So that's the second trend we're seeing. The third thing, and I think I kind of missed mentioning this earlier, one of the very key and unique value propositions we provide with the MySQL Database Service in HeatWave, is that we provide a mechanism where if customers have their data stored on premise they can still leverage the HeatWave service by enabling MySQL replication. So they can have their data on premise, they can replicate this data in the Oracle Cloud and then they can run analytics. So this deployment which we are calling the hybrid deployment is turning out to be very, very popular because there are customers, there are some customers who for various reasons, compliance or regulatory reasons cannot move the entire data to the cloud or migrate the data to the cloud completely. So this provides them a very good setup where they can continue to run their existing database and when it comes to getting benefits of HeatWave for query acceleration, they can set up this replication. >> And I can run that on anyone, any available server capacity or is there an appliance to facilitate that? >> No, this is just standard MySQL replication. So if a customer is running MySQL on premise they can just turn off this application. We have obviously enhanced it to support this inbound replication between on-premise and Oracle Cloud with something which can be enabled as long as the source and destination are both MySQL. >> Okay, so I want to come back to this sort of idea of the architecture a little bit. I mean, it's hard for me to go toe to toe with the, I'm not an engineer, but I'm going to try anyway. So you've talked about OLTP queries. I thought, I always thought HeatWave was optimized for analytics. But so, I want to push on this notion because people think of this the converged database, and what you're talking about here with HeatWave is sort of the Swiss army knife which is great 'cause you got a screwdriver and you got Phillips and a flathead and some scissors, maybe they're not as good. They're not as good necessarily as the purpose-built tool. But you're arguing that this is best of breed for OLTP and best of breed for analytics, both in terms of performance and cost. Am I getting that right or is this really a Swiss army knife where that flathead is really not as good as the big, long screwdriver that I have in my bag? >> Yes, so, you're getting it right but I did want to make a clarification. That HeatWave is definitely the accelerator for all your queries, all analytic queries and also for the long running complex transaction processing inquiries. So yes, HeatWave the uber query accelerator engine. However, when it comes to transaction processing in terms of your insert statements, delete statements, those are still all done and served by the MySQL database. So all, the transactions are still sent to the MySQL database and they're persistent there, it's the queries for which HeatWave is the accelerator. So what you said is correct. For all query acceleration, HeatWave is the engine. >> Makes sense. Okay, so if I'm a MySQL customer and I want to use HeatWave, what do I have to do? Do I have to make changes to my existing applications? You applied earlier that, no, it's just sort of plugs right in. But can you clarify that. >> Yes, there are absolutely no changes, which any MySQL or MySQL compatible application needs to make to take advantage of HeatWave. HeatWave is an in-memory accelerator and it's completely transparent to the application. So we have like, dozens and dozens of like, applications which have migrated to HeatWave, and they are seeing the same thing, similarly tools. So if you look at various tools which work for analytics like, Tableau, Looker, Oracle Analytics Cloud, all of them will work just seamlessly. And this is one of the reasons we had to do a lot of heavy lifting in the MySQL database itself. So the MySQL database engineering team was, has been very actively working on this. And one of the reasons is because we did the heavy lifting and we meet enhancements to the MySQL optimizer in the MySQL storage layer to do the integration of HeatWave in such a seamless manner. So there is absolutely no change which an application needs to make in order to leverage or benefit from HeatWave. >> You said earlier, Nipun, that you're seeing migrations from, I think you said Aurora and Google BigQuery, you might've said Redshift as well. Do you, what kind of tooling do you have to facilitate migrations? >> Right, now, there are multiple ways in which customers may want to do this, right? So the first tooling which we have is that customers, as I was talking about the replication or the inbound replication mechanism, customers can set up heat HeatWave in the Oracle Cloud and they can send the data, they can set up replication within their instances in their cloud and HeatWave. Second thing is we have various kinds of tools to like, facilitate the data migration in terms of like, fast ingestion sites. So there are a lot of such customers we are seeing who are kind of migrating and we have a plethora of like, tools and applications, in addition to like, setting up this inbound application, which is the most seamless way of getting customers started with HeatWave. >> So, I think you mentioned before, I have my notes, machine intelligence and machine learning. We've seen that with autonomous database it's a big, big deal obviously. How does HeatWave take advantage of machine intelligence and machine learning? >> Yeah, and I'm probably going to be talking more about this in the future, but what we have already is that HeatWave uses machine learning to intelligently automate many operations. So we know that when there's a service being offered in the cloud, our customers expect automation. And there're a lot of vendors and a lot of services which do a good job in automation. One of the places where we're going to be very unique is that HeatWave uses machine learning to automate many of these operations. And I'll give you one such example which is provisioning. Right now with HeatWave, when a customer wants to determine how many nodes are needed for running their workload, they don't need to make a guess. They invoke a provisioning advisor and this advisor uses machine learning to sample a very small percentage of the data. We're talking about, like, 0.1% sampling and it's able to predict the amount of memory with 95% accuracy, which this data is going to take. And based on that, it's able to make a prediction of how many servers are needed. So just a simple operation, the first step of provisioning, this is something which is done manually across, on any of the service, whereas at HeatWave, we have machine learning-based advisor. So this is an example of what we're doing. And in the future, we'll be offering many such innovations as a part of the MySQL Database and the HeatWave service. >> Well, I've got to say I was skeptic but I really appreciate it, you're, answering my questions. And, a lot of people when you made the acquisition and inherited MySQL, thought you were going to kill it because they thought it would be competitive to Oracle Database. I'm happy to see that you've invested and figured out a way to, hey, we can serve our community and continue to be the steward of MySQL. So Nipun, thanks very much for coming to the CUBE. Appreciate your time. >> Sure. Thank you so much for the time, Dave. I appreciate it. >> And thank you for watching everybody. This is Dave Vellante with another CUBE Conversation. We'll see you next time. (bright upbeat music)

Published Date : Apr 28 2021

SUMMARY :

of the trends of the database market. So it's really been fun to watch and the strategy there, for the MySQL Database Service on the LAMP stack. And the fact that we have designed it optimized for the cloud I hear kind of the buzzwords, So the cost of the system Thank you. critical for the cloud. So that kind of made sense to me. So it's, the approach which has been taken By the way, we were saying before, the amount of improvement you will get, is that the game you're playing here? So it's not the case All right. and we are taking the numbers published of the architecture and if you look at smaller data sizes, Sorry to interrupt. and the performance in terms of latency Thank you. So faster, cheaper on both the benchmarks So for many of the workloads, to go to quicker side. and optimized for the cloud. It is a service. So it's a managed cannot move the entire data to the cloud as long as the source and of the architecture a little bit. and also for the long running complex Do I have to make changes So the MySQL database engineering team to facilitate migrations? So the first tooling which and machine learning? and the HeatWave service. and continue to be the steward of MySQL. much for the time, Dave. And thank you for watching everybody.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

AmazonORGANIZATION

0.99+

Larry EllisonPERSON

0.99+

Nipun AgarwalPERSON

0.99+

NipunPERSON

0.99+

AWSORGANIZATION

0.99+

400 timesQUANTITY

0.99+

DavePERSON

0.99+

1000 timesQUANTITY

0.99+

OracleORGANIZATION

0.99+

10,000 timesQUANTITY

0.99+

100%QUANTITY

0.99+

HeatWaveORGANIZATION

0.99+

second bitQUANTITY

0.99+

MySQLTITLE

0.99+

95%QUANTITY

0.99+

100 timesQUANTITY

0.99+

two aspectsQUANTITY

0.99+

500 timesQUANTITY

0.99+

0.1%QUANTITY

0.99+

half a terabyteQUANTITY

0.99+

dozensQUANTITY

0.99+

1100 timesQUANTITY

0.99+

4 TBQUANTITY

0.99+

first pointQUANTITY

0.99+

FirstQUANTITY

0.99+

PhillipsORGANIZATION

0.99+

AmazonsORGANIZATION

0.99+

three timesQUANTITY

0.99+

GoogleORGANIZATION

0.99+

One thirdQUANTITY

0.99+

one databaseQUANTITY

0.99+

second thingQUANTITY

0.99+

oneQUANTITY

0.99+

OneQUANTITY

0.99+

bothQUANTITY

0.99+

SnowflakeTITLE

0.99+

Wim Coekaerts, Oracle | CUBEconversations


 

(bright upbeat music) >> Hello everyone, and welcome to this exclusive Cube Conversation. We have the pleasure today to welcome, Wim Coekaerts, senior vice president of software development at Oracle. Wim, it's good to see you. How you been, sir? >> Good, it's been a while since we last talked but I'm excited to be here, as always. >> It was during COVID though and so I hope to see you face to face soon. But so Wim, since the Barron's Article declared Oracle a Cloud giant, we've really been sort of paying attention and amping up our coverage of Oracle and asking a lot of questions like, is Oracle really a Cloud giant? And I'll say this, we've always stressed that Oracle invests in R&D and of course there's a lot of D in that equation. And over the past year, we've seen, of course the autonomous database is ramping up, especially notable on Exadata Cloud@Customer, we've covered that extensively. We covered the autonomous data warehouse announcement, the blockchain piece, which of course got me excited 'cause I get to talk about crypto with Juan. Roving Edge, which for everybody who might not be familiar with that, it's an edge cloud service, dedicated regions that you guys announced, which is a managed cloud region. And so it's clear, you guys are serious about cloud. These are all cloud first services using second gen OCI. So, Oracle's making some moves but the question is, what are customers doing? Are they buying this stuff? Are they leaning into these new deployment models for the databases? What can you tell us? >> You know, definitely. And I think, you know, the reason that we have so many different services is that not every customer is the same, right? One of the things that people don't necessarily realize, I guess, is in the early days of cloud lots of startups went there because they had no local infrastructure. It was easy for them to get started in something completely new. Our customers are mostly enterprise customers that have huge data centers in many cases, they have lots of real estate local. And when they think about cloud they're wondering how can we create an environment that doesn't cause us to have two ops teams and two ways of managing things. And so, they're trying to figure out exactly what it means to take their real estate and either move it wholesale to the cloud over a period of years, or they say, "Hey, some of these things need to be local maybe even for regulatory purposes." Or just because they want to keep some data locally within their own data centers but then they have to move other things remotely. And so, there's many different ways of solving the problem. And you can't just say, "Here's one cloud, this is where you go and that's it." So, we basically say, if you're on prem, we provide you with cloud services on-premises, like dedicated regions or Oracle Exadata Cloud@Customer and so forth so that you get the benefits of what we built for cloud and spend a lot of time on, but you can run them in your own data center or people say, "No, no, no. I want to get rid of my data centers, I do it remotely." Okay, then you do it in Oracle cloud directly. Or you have a hybrid model where you say, "Some stays local, some is remote." The nice thing is you get the exact same API, the exact same way of managing things, no matter how you deploy it. And that's a big differentiator. >> So, is it fair to say that you guys have, I think of it as a purpose built club, 'cause I talk to a lot of customers. I mean, take an insurance app like Claims, and customers tell me, "I'm not putting that into the public cloud." But you're making a case that it actually might make sense in your cloud because you can support those mission critical applications with the exact same experience, same API, same... I can get, you know, take Rack for instance, I can't get, you know, real application clusters in an Amazon cloud but presumably I can get them in your cloud. So, is it fair to say you have a purpose built cloud specifically for the most demanding applications? Is that a right way to look at it or not necessarily? >> Well, it's interesting. I think the thing to be careful of is, I guess, purpose built cloud might for some people mean, "Oh, you can only do things if it's Oracle centric." Right, and so I think that fundamentally, Oracle cloud provides a generic cloud. You can run anything you want, any application, any deployment model that you have. Whether you're an Oracle customer or not, we provide you with a full cloud service, right? However, given that we know and have known, obviously for a long time, how our products run best, when we designed OCI gen two, when we designed the networking stack, the storage layer and all that stuff, we made sure that it would be capable of running our more complex environments because our advantage is, Oracle customers have a place where they can run Oracle the best. Right, and so obviously the context of purpose-built fits that model, where yes, we've made some design choices that allow us to run Rack inside OCI and allow us to deploy Exadatas inside OCI which you cannot do in other clouds. So yes, it's purpose built in that sense but I would caution on the side of that it sometimes might imply that it's unique to Oracle products and I guess one way to look at it is if you can run Oracle, you can run everything else, right? Because it's such a complex suite of products that if you can run that then it it'll support any other (mumbling). >> Right. Right, it's like New York city. You make it there, you can make it anywhere. If I can run the most demanding mission critical applications, well, then I can run a web app for instance, okay. I got a question on tooling 'cause there's a lot of tooling, like sometimes it makes my eyes bleed when I look at all this stuff and doesn't... Square the circle for me, doesn't autonomous, an autonomous database like Autonomous Linux, for instance, doesn't it eliminate the need for all these management tools? >> You know, it does. It eliminates the need for the management at the lower level, right. So, with the autonomous Linux, what we offer and what we do is, we automatically patch the operating system for you and make sure it's secure from a security patching point of view. We eliminate the downtime, so when we do it then you don't have to restart applications. However, we don't know necessarily what the app is that is installed on top of it. You know, people can deploy their own applications, they can run third party applications, they can use it for development environments and so forth. So, there's sort of the core operating system layer and on the database side, you know, we take care of database patching and upgrades and storage management and all that stuff. So the same thing, if you run your own application inside the database, we can manage the database portion but we don't manage the application portion just like on the operating system. And so, there's still a management level that's required, no matter what, a level above that. And the other thing and I think this is what a lot of the stuff we're doing is based on is, you still have tons of stuff on-premises that needs full management. You have applications that you migrate that are not running Autonomous Linux, could be a Windows application that's running or it could be something on a different Linux distribution or you could still have some databases installed that you manage yourself, you don't want to use the autonomous or you're on a third-party. And so we want to make sure that we can address all of them with a single set of tools, right. >> Okay, so I wonder, can you give us just an overview, just briefly of the products that comprise into the cloud services, your management solution, what's in that portfolio? How should we think about it? >> Yeah, so it basically starts with Enterprise Manager on-premises, right? Which has been the tool that our Oracle database customers in particular have been using for many years and is widely used by our customer base. And so you have those customers, most of their real estate is on-premises and they can use enterprise management with local. They have it running and they don't want to change. They can keep doing that and we keep enhancing as you know, with newer versions of Enterprise Manager getting better. So, then there's the transition to cloud and so what we've been doing over the last several years is basically, looking at the things, well, one aspect is looking at things people, likes of Enterprise Manager and make sure that we provide similar functionality in Oracle cloud. So, we have Performance Hub for looking at how the database performance is working. We have APM for Application Performance Monitoring, we have Logging Analytics that looks at all the different log files and helps make sense of it for you. We have Database Management. So, a lot of the functionality that people like in Enterprise Manager mentioned the database that we've built into Oracle cloud, and, you know, a number of other things that are coming Operations Insights, to look at how databases are performing and how we can potentially do consolidation and stuff. So we've basically looked at what people have been using on-premises, how we can replicate that in Oracle cloud and then also, when you're in a cloud, how you can make make use of all the base services that a cloud vendor provides, telemetry, logging and so forth. And so, it's a broad portfolio and what it allows us to do with our customers is say, "Look, if you're predominantly on-prem, you want to stay there, keep using Enterprise Manager. If you're starting to move to Oracle cloud, you can first use EM, look at what's happening in the cloud and then switch over, start using all the management products we have in the cloud and let go of the Enterprise Manager instance on-premise. So you can gradually shift, you can start using more and more. Maybe you start with analytics first and then you start with insights and then you switch to database management. So there's a whole suite of possibilities. >> (indistinct) you mentioned APM, I've been watching that space, it's really evolved. I mean, you saw, you know, years ago, Splunk came out with sort of log analytics, maybe simplified that a little bit, now you're seeing some open source stuff come out. You're seeing a lot of startups come out, you saw Cisco made an acquisition with AppD and that whole space is transforming it seems that the future is all about that end to end visibility, simplifying the ability to remediate problems. And I'm thinking, okay, you just mentioned, you guys have a lot of these capabilities, you got Autonomous, is that sort of where you're headed with your capabilities? >> It definitely is and in fact, one of the... So, you know, APM allows you to say, "Hey, here's my web browser and it's making a connection to the database, to a middle tier" and it's hard for operations people in companies to say, hey, the end user calls and says, "You know, my order entry system is slow. Is it the browser? Is it the middle tier that they connect to? Is it the database that's overloaded in the backend?" And so, APM helps you with tracing, you know, what happens from where to where, where the delays are. Now, once you know where the delay is, you need to drill down on it. And then you need to go look at log files. And that's where the logging piece comes in. And what happens very often is that these log files are very difficult to read. You have networking log files and you have database log files and you have reslog files and you almost have to be an expert in all of these things. And so, then with Logging Analytics, we basically provide sort of an expert dashboard system on top of that, that allows us to say, "Hey! When you look at logging for the network stack, here are the most important errors that we could find." So you don't have to go and learn all the details of these things. And so, the real advantages of saying, "Hey, we have APM, we have Logging Analytics, we can tie the two together." Right, and so we can provide a solution that actually helps solve the problem, rather than, you need to use APM for one vendor, you need to use Logging Analytics from another vendor and you know, that doesn't necessarily work very well. >> Yeah and that's why you're seeing with like the ELK Stack it's cool, you're an open source guy, it's cool as an open source, but it's complicated to set up all that that brings. So, that's kind of a cool approach that you guys are taking. You mentioned Enterprise Manager, you just made a recent announcement, a new release. What's new in that new release? >> So Enterprise Manager 13.5 just got released. And so EM keeps improving, right? We've made a lot of changes over over the years and one of the things we've done in recent years is do more frequent updates sort of the cloud model frequent updates that are not just bug fixes but also introduce new functionality so people get more stuff more frequently rather than you know, once a year. And that's certainly been very attractive because it shows that it's a lively evolving product. And one of the main focus areas of course is cloud. And so a lot of work that happens in Enterprise Manager is hybrid cloud, which basically means I run Enterprise Manager and I have some stuff in Oracle cloud, I might have some other stuff in another cloud vendors environment and so we can actually see which databases are where and provide you with one consolidated view and one tool, right? And of course it supports Autonomous Database and Exadata in cloud servers and so forth. So you can from EM see both your databases on-premises and also how it's doing in in Oracle cloud as you potentially migrate things over. So that's one aspect. And then the other one is in terms of operations and automation. One of the things that we started doing again with Enterprise Manager in the last few years is making sure that everything has a REST API. So we try to make the experience with Enterprise Manager be very similar to how people work with a cloud service. Most folks now writing automation tools are used to calling REST APIs. EM in the early days didn't have REST APIs, now we're making sure everything works that way. And one of the advantages is that we can do extensibility without having to rewrite the product, that we just add the API clause in the agent and it makes it a lot easier to become part of the modern system. Another thing that we introduced last year but that we're evolving with more dashboards and so forth is the Grafana plugin. So even though Enterprise Manager provides lots of cool tools, a lot of cloud operations folks use a tool called Grafana. And so we provide a plugin that allows customers to have Grafana dashboards but the data actually comes out of Enterprise Manager. So that allows us to integrate EM into a more cloudy world in a cloud environment. I think the other important part is making sure that again, Enterprise Manager has sort of a cloud feel to it. So when you do patching and upgrades, it's near zero downtime which basically means that we do all the upgrades for you without having to bring EM down. Because even though it's a management tool, it's used for operations. So if there were downtime for patching Enterprise Manager for an hour, then for that hour, it's a blackout window for all the monitoring we do. And so we want to avoid that from happening, so now EM is upgrading, even though all the events are still happening and being processed, and then we do a very short switch. So that help our operations people to be more available. >> Yes. I mean, I've been talking about Automated Operations since, you know, lights out data centers since the eighties back in (laughs). I remember (indistinct) data center one-time lights out there were storage tech libraries in there and so... But there were a lot of unintended consequences around, you know, automated ops, and so people were sort of scared to go there, at least lean in too much but now with all this machine intelligence... So you're talking about ops automation, you mentioned the REST APIs, the Grafana plugins, the Cloud feel, is that what you're bringing to the table that's unique, is that unique to Oracle? >> Well, the integration with Oracle in that sense is unique. So one example is you mentioned the word migration, right? And so database migration tends to be something, you know, customers obviously take very serious. We go from one place, you have to move all your data to another place that runs in a slightly different environment. And so how do you know whether that migration is going to work? And you can't migrate a thousand databases manually, right? So automation, again, it's not just... Automation is not just to say, "Hey, I can do an upgrade of a system or I can make sure that nothing is done by hand when you patch something." It's more about having a huge fleet of servers and a huge fleet of databases. How can you move something from one place to another and automate that? And so with EM, you know, we start with sort of the prerequisite phase. So we're looking at the existing environment, how much memory does it need? How much storage does it use? Which version of the database does it have? How much data is there to move? Then on the target side, we see whether the target can actually run in that environment. Then we go and look at, you know, how do you want to migrate? Do you want to migrate everything from a sort of a physical model or do you want to migrate it from a logical model? Do you want to do it while your environment is still running so that you start backing up the data to the target database while your existing production system is still running? Then we do a short switch afterwards, or you say, "No, I want to bring my database down. I want to do the migrate and then bring it back up." So there's different deployment models that we can let our customers pick. And then when the migration is done, we have a ton of health checks that can validate whether the target database will run through basically the exact same way. And then you can say, "I want to migrate 10 databases or 50 databases" and it'll work, It's all automated out of the box. >> So you're saying, I mean, you've looked at the prevailing way you've done migrations, historically you'd have to freeze the code and then migrate, and it would take forever, it was a function of the number of lines of code you had. And then a lot of times, you know, people would say, "We're not going to freeze the code" and then they would almost go out of business trying to merge the two. You're saying in 2021, you can give customers the choice, you can migrate, you could change the, you know, refuel the plane while you're in midair? Is that essentially what you're saying? >> That's a good way of describing it, yeah. So your existing database is running and we can do a logical backup and restore. So while transactions are happening we're still migrating it over and then you can do a cutoff. It makes the transition a lot easier. But the other thing is that in the past, migrations would typically be two things. One is one database version to the next, more upgrades than migration. Then the second one is that old hardware or a different CPU architecture are moving to newer hardware in a new CPU architecture. Those were sort of the typical migrations that you had prior to Cloud. And from a CIS admin point of view or a DBA it was all something you could touch, that you could physically touch the boxes. When you move to cloud, it's this nebulous thing somewhere in a data center that you have no access to. And that by itself creates a barrier to a lot of admins and DBA's from saying, "Oh, it'll be okay." There's a lot of concern. And so by baking in all these tests and the prerequisites and all the dashboards to say, you know, "This is what you use. These are the features you use. We know that they're available on the other side so you can do the migration." It helps solve some of these problems and remove the barriers. >> Well that was just kind of same same vision when you guys came up with it. I don't know, quite a while ago now. And it took a while to get there with, you know, you had gen one and then gen two but that is, I think, unique to Oracle. I know maybe some others that are trying to do that as well, but you were really the first to do that and so... I want to switch topics to talk about security. It's hot topic. You guys, you know, like many companies really focused on security. Does Enterprise Manager bring any of that over? I mean, the prevailing way to do security often times is to do scripts and write, you know, custom security policy scripts are fragile, they break, what can you tell us about security? >> Yeah. So there's really two things, you know. One is, we obviously have our own best security practices. How we run a database inside Oracle for our own world, we've learned about that over the years. And so we sort of baked that knowledge into Enterprise Manager. So we can say, "Hey, if you install this way, we do the install and the configuration based on our best practice." That's one thing. The other one is there's STIG, there's PCI and they're ShipBob, those are the main ones. And so customers can do their own way. They can download the documentation and do it manually. But what we've done is, and we've done this for a long time, is basically bake those policies into Enterprise Manager. So you can say, "Here's my database this needs to be PCI compliant or it needs to be HIPAA compliant and you push a button and then we validate the policies in those documents or in those prescript described files. And we make sure that the database is combined to that. And so we take that manual work and all that stuff basically out of the picture, we say, "Push this button and we'll take care of it." >> Now, Wim, but just quick sidebar here, last time we talked, it was under a year ago. It was definitely during COVID and it's still during COVID. We talked about the state of the penguin. So I'm wondering, you know, what's the latest update for Linux, any Linux developments that we should be aware of? >> Linux, we're still working very hard on Autonomous Linux and that's something where we can really differentiate and solve a problem. Of course, one of the things to mention is that Enterprise Manager can can do HIPAA compliance on Oracle Linux as well. So the security practices are not just for the database it can also go down to the operating system. Anyway, so on the Autonomous Linux side, you know, management in an Oracle Cloud's OS management is evolving. We're spending a lot of time on integrating log capturing, and if something were to go wrong that we can analyze a log file on the fly and send you a notification saying, "Hey, you know there was this bug and here's the cause." And it was potentially a fix for it to Autonomous Linux and we're putting a lot of effort into that. And then also sort of IT/operation management where we can look at the different applications that are running. So you're running a web server on a Linux environment or you're running some Java processes, we can see what's running. We can say, "Hey, here's the CPU utilization over the past week or the past year." And then how is this evolving? Say, if something suddenly spikes we can say, "Well, that's normal, because every Monday morning at 10 o'clock there's a spike or this is abnormal." And then you can start drilling this down. And this comes back to overtime integration with whether it's APM or Logging Analytics, we can tie the dots, right? We can connect them, we can say, "Push this thing, then click on that link." We give you the information. So it's that integration with the entire cloud platform that's really happening now >> Integration, there's that theme again. I want to come back to migration and I think you did a good job of explaining how you sort of make that non-disruptive and you know, your customers, I think, you know, generally you're pushing you know, that experience which makes people more comfortable. But my question is, why do people want to migrate if it works and it's on prem, are they doing it just because they want to get out of the data center business? Or is it a better experience in the cloud? What can you tell us there? >> You know, it's a little bit of everything. You know, one is, of course the idea that data center maintenance costs are very high. The other one is that when you run your own data center, you know, we obviously have this problem but when you're a cloud vendor, you have these problems but we're in this business. But if you buy a server, then in three years that server basically is depreciated by new versions and they have to do migration stuff. And so one of the advantages with cloud is you push a button, you have a new version of the hardware, basically, right? So the refreshes happen on a regular basis. You don't have to go and recycle that yourself. Then the other part is the subscription model. It's a lot easier to pay for what you use rather than you have a data center whether it's used or not, you pay for it. So there's the cost advantages and predictability of what you need, you pay for, you can say, "Oh next year we need to get x more of EMs." And it's easier to scale that, right? We take care of dealing with capacity planning. You don't have to deal with capacity planning of hardware, we do that as the cloud vendor. So there's all these practical advantages you get from doing it remotely and that's really what the appeal is. >> Right. So, as it relates to Enterprise Manager, did you guys have to like tear down the code and rebuild it? Was it entire like redo? How did you achieve that? >> No, no, no. So, Enterprise Manager keeps evolving and you know, we changed the underlying technologies here and there, piecemeal, not sort of a wholesale replacement. And so in talking about five, there's a lot of new stuff but it's built on the existing EM core. And so we're just, you know, improving certain areas. One of the things is, stability is important for our customers, obviously. And so by picking things piecemeal, we replace one engine rather than the whole thing. It allows us to introduce change more slowly, right. And then it's well-tested as a unit and then when we go on to the next thing. And then the other one is I mentioned earlier, a lot of the automation and extensibility comes from REST APIs. And so instead of basically re-writing everything we just provide a REST endpoint and we make all the new features that we built automatically be REST enabled. So that makes it a lot easier for us to introduce new stuff. >> Got it. So if I want to poke around with this new version of Enterprise Manager, can I do that? Is there a place I can go, do I have to call a rep? How does that work? >> Yeah, so for information you can just go to oracle.com/enterprise manager. That's the website that has all the data. The other thing is if you're already playing with Oracle Cloud or you use Oracle Cloud, we have Enterprise Manager images in the marketplace. So if you have never used EM, you can go to Oracle Cloud, push a button in the marketplace and you get a full Enterprise Manager installation in a matter of minutes. And then you can just start using that as well. >> Awesome. Hey, I wanted to ask you about, you know, people forget that you guys are the stewards of MySQL and we've been looking at MySQL Database Cloud service with HeatWave Did you name that? And so I wonder if you could talk about what you're doing with regard to managing HeatWave environments? >> So, HeatWave is the MySQL option that helps with analytics, right? And it really accelerates MySQL usage by 100 x and in some cases more and it's transparent to the customer. So as a MySQL user, you connect with standard MySQL applications and APIs and SQL and everything. And the HeatWave part is all done within the MySQL server. The engine itself says, "Oh, this SQL query, we can offload to the backend HeatWave cluster," which then goes in memory operations and blazingly fast returns it to you. And so the nice thing is that it turns every single MySQL database into also a data warehouse without any change whatsoever in your application. So it's been widely popular and it's quite exciting. I didn't personally name it, HeatWave, that was not my decision, but it sounds very cool. >> That's very cool. >> Yeah, It's a very cool name. >> We love MySQL, we started our company on the lamp stack, so like many >> Oh? >> Yeah, yeah. >> Yeah, yeah. That's great. So, yeah. And so with HeatWave or MySQL in general we're basically doing the same thing as we have done for the Oracle Database. So we're going to add more functionality in our database management tools to also look at HeatWave. So whether it's doing things like performance hub or generic database management and monitoring tools, we'll expand that in, you know, in the near future, in the future. >> That's great. Well, Wim, it's always a pleasure. Thank you so much for coming back in "The Cube" and letting me ask all my Colombo questions. It was really a pleasure having you. (mumbling) >> It's good be here. Thank you so much. >> You're welcome. And thank you for watching, everybody, this is Dave Vellante. We'll see you next time. (bright music)

Published Date : Apr 27 2021

SUMMARY :

How you been, sir? but I'm excited to be here, as always. And so it's clear, you guys and so forth so that you get So, is it fair to say you that if you can run that You make it there, you and on the database side, you know, and then you switch to it seems that the future is all about and you know, that doesn't approach that you guys are taking. all the upgrades for you since, you know, lights out And so with EM, you know, of lines of code you had. and then you can do a cutoff. is to do scripts and write, you know, and you push a button and So I'm wondering, you know, And then you can start drilling this down. and you know, your customers, And so one of the advantages with cloud is did you guys have to like tear And so we're just, you know, How does that work? And then you can just And so I wonder if you could And so the nice thing is that it turns we'll expand that in, you know, Thank you so much for Thank you so much. And thank you for watching, everybody,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

OracleORGANIZATION

0.99+

Wim CoekaertsPERSON

0.99+

50 databasesQUANTITY

0.99+

CiscoORGANIZATION

0.99+

10 databasesQUANTITY

0.99+

2021DATE

0.99+

Enterprise ManagerTITLE

0.99+

New YorkLOCATION

0.99+

EnterpriseTITLE

0.99+

MySQLTITLE

0.99+

JavaTITLE

0.99+

last yearDATE

0.99+

two thingsQUANTITY

0.99+

three yearsQUANTITY

0.99+

twoQUANTITY

0.99+

LinuxTITLE

0.99+

an hourQUANTITY

0.99+

Enterprise ManagerTITLE

0.99+

WindowsTITLE

0.99+

SQLTITLE

0.99+

100 xQUANTITY

0.99+

OneQUANTITY

0.99+

next yearDATE

0.99+

one toolQUANTITY

0.99+

oneQUANTITY

0.99+

AmazonORGANIZATION

0.98+

todayDATE

0.98+

second oneQUANTITY

0.98+

firstQUANTITY

0.98+

one exampleQUANTITY

0.98+

Enterprise Manager 13.5TITLE

0.98+

one aspectQUANTITY

0.98+

one engineQUANTITY

0.97+

WimPERSON

0.97+

gen oneQUANTITY

0.97+

bothQUANTITY

0.97+

once a yearQUANTITY

0.97+

Oracle CloudTITLE

0.97+

one wayQUANTITY

0.97+

GrafanaTITLE

0.97+

BarronPERSON

0.97+

first servicesQUANTITY

0.96+

HeatWaveORGANIZATION

0.96+

past yearDATE

0.96+

one-timeQUANTITY

0.96+

gen twoQUANTITY

0.96+

one placeQUANTITY

0.96+

past weekDATE

0.96+

two waysQUANTITY

0.95+