Breaking Analysis: Databricks faces critical strategic decisions…here’s why
>> From theCUBE Studios in Palo Alto and Boston, bringing you data-driven insights from theCUBE and ETR. This is Breaking Analysis with Dave Vellante. >> Spark became a top level Apache project in 2014, and then shortly thereafter, burst onto the big data scene. Spark, along with the cloud, transformed and in many ways, disrupted the big data market. Databricks optimized its tech stack for Spark and took advantage of the cloud to really cleverly deliver a managed service that has become a leading AI and data platform among data scientists and data engineers. However, emerging customer data requirements are shifting into a direction that will cause modern data platform players generally and Databricks, specifically, we think, to make some key directional decisions and perhaps even reinvent themselves. Hello and welcome to this week's wikibon theCUBE Insights, powered by ETR. In this Breaking Analysis, we're going to do a deep dive into Databricks. We'll explore its current impressive market momentum. We're going to use some ETR survey data to show that, and then we'll lay out how customer data requirements are changing and what the ideal data platform will look like in the midterm future. We'll then evaluate core elements of the Databricks portfolio against that vision, and then we'll close with some strategic decisions that we think the company faces. And to do so, we welcome in our good friend, George Gilbert, former equities analyst, market analyst, and current Principal at TechAlpha Partners. George, good to see you. Thanks for coming on. >> Good to see you, Dave. >> All right, let me set this up. We're going to start by taking a look at where Databricks sits in the market in terms of how customers perceive the company and what it's momentum looks like. And this chart that we're showing here is data from ETS, the emerging technology survey of private companies. The N is 1,421. What we did is we cut the data on three sectors, analytics, database-data warehouse, and AI/ML. The vertical axis is a measure of customer sentiment, which evaluates an IT decision maker's awareness of the firm and the likelihood of engaging and/or purchase intent. The horizontal axis shows mindshare in the dataset, and we've highlighted Databricks, which has been a consistent high performer in this survey over the last several quarters. And as we, by the way, just as aside as we previously reported, OpenAI, which burst onto the scene this past quarter, leads all names, but Databricks is still prominent. You can see that the ETR shows some open source tools for reference, but as far as firms go, Databricks is very impressively positioned. Now, let's see how they stack up to some mainstream cohorts in the data space, against some bigger companies and sometimes public companies. This chart shows net score on the vertical axis, which is a measure of spending momentum and pervasiveness in the data set is on the horizontal axis. You can see that chart insert in the upper right, that informs how the dots are plotted, and net score against shared N. And that red dotted line at 40% indicates a highly elevated net score, anything above that we think is really, really impressive. And here we're just comparing Databricks with Snowflake, Cloudera, and Oracle. And that squiggly line leading to Databricks shows their path since 2021 by quarter. And you can see it's performing extremely well, maintaining an elevated net score and net range. Now it's comparable in the vertical axis to Snowflake, and it consistently is moving to the right and gaining share. Now, why did we choose to show Cloudera and Oracle? The reason is that Cloudera got the whole big data era started and was disrupted by Spark. And of course the cloud, Spark and Databricks and Oracle in many ways, was the target of early big data players like Cloudera. Take a listen to Cloudera CEO at the time, Mike Olson. This is back in 2010, first year of theCUBE, play the clip. >> Look, back in the day, if you had a data problem, if you needed to run business analytics, you wrote the biggest check you could to Sun Microsystems, and you bought a great big, single box, central server, and any money that was left over, you handed to Oracle for a database licenses and you installed that database on that box, and that was where you went for data. That was your temple of information. >> Okay? So Mike Olson implied that monolithic model was too expensive and inflexible, and Cloudera set out to fix that. But the best laid plans, as they say, George, what do you make of the data that we just shared? >> So where Databricks has really come up out of sort of Cloudera's tailpipe was they took big data processing, made it coherent, made it a managed service so it could run in the cloud. So it relieved customers of the operational burden. Where they're really strong and where their traditional meat and potatoes or bread and butter is the predictive and prescriptive analytics that building and training and serving machine learning models. They've tried to move into traditional business intelligence, the more traditional descriptive and diagnostic analytics, but they're less mature there. So what that means is, the reason you see Databricks and Snowflake kind of side by side is there are many, many accounts that have both Snowflake for business intelligence, Databricks for AI machine learning, where Snowflake, I'm sorry, where Databricks also did really well was in core data engineering, refining the data, the old ETL process, which kind of turned into ELT, where you loaded into the analytic repository in raw form and refine it. And so people have really used both, and each is trying to get into the other. >> Yeah, absolutely. We've reported on this quite a bit. Snowflake, kind of moving into the domain of Databricks and vice versa. And the last bit of ETR evidence that we want to share in terms of the company's momentum comes from ETR's Round Tables. They're run by Erik Bradley, and now former Gartner analyst and George, your colleague back at Gartner, Daren Brabham. And what we're going to show here is some direct quotes of IT pros in those Round Tables. There's a data science head and a CIO as well. Just make a few call outs here, we won't spend too much time on it, but starting at the top, like all of us, we can't talk about Databricks without mentioning Snowflake. Those two get us excited. Second comment zeros in on the flexibility and the robustness of Databricks from a data warehouse perspective. And then the last point is, despite competition from cloud players, Databricks has reinvented itself a couple of times over the year. And George, we're going to lay out today a scenario that perhaps calls for Databricks to do that once again. >> Their big opportunity and their big challenge for every tech company, it's managing a technology transition. The transition that we're talking about is something that's been bubbling up, but it's really epical. First time in 60 years, we're moving from an application-centric view of the world to a data-centric view, because decisions are becoming more important than automating processes. So let me let you sort of develop. >> Yeah, so let's talk about that here. We going to put up some bullets on precisely that point and the changing sort of customer environment. So you got IT stacks are shifting is George just said, from application centric silos to data centric stacks where the priority is shifting from automating processes to automating decision. You know how look at RPA and there's still a lot of automation going on, but from the focus of that application centricity and the data locked into those apps, that's changing. Data has historically been on the outskirts in silos, but organizations, you think of Amazon, think Uber, Airbnb, they're putting data at the core, and logic is increasingly being embedded in the data instead of the reverse. In other words, today, the data's locked inside the app, which is why you need to extract that data is sticking it to a data warehouse. The point, George, is we're putting forth this new vision for how data is going to be used. And you've used this Uber example to underscore the future state. Please explain? >> Okay, so this is hopefully an example everyone can relate to. The idea is first, you're automating things that are happening in the real world and decisions that make those things happen autonomously without humans in the loop all the time. So to use the Uber example on your phone, you call a car, you call a driver. Automatically, the Uber app then looks at what drivers are in the vicinity, what drivers are free, matches one, calculates an ETA to you, calculates a price, calculates an ETA to your destination, and then directs the driver once they're there. The point of this is that that cannot happen in an application-centric world very easily because all these little apps, the drivers, the riders, the routes, the fares, those call on data locked up in many different apps, but they have to sit on a layer that makes it all coherent. >> But George, so if Uber's doing this, doesn't this tech already exist? Isn't there a tech platform that does this already? >> Yes, and the mission of the entire tech industry is to build services that make it possible to compose and operate similar platforms and tools, but with the skills of mainstream developers in mainstream corporations, not the rocket scientists at Uber and Amazon. >> Okay, so we're talking about horizontally scaling across the industry, and actually giving a lot more organizations access to this technology. So by way of review, let's summarize the trend that's going on today in terms of the modern data stack that is propelling the likes of Databricks and Snowflake, which we just showed you in the ETR data and is really is a tailwind form. So the trend is toward this common repository for analytic data, that could be multiple virtual data warehouses inside of Snowflake, but you're in that Snowflake environment or Lakehouses from Databricks or multiple data lakes. And we've talked about what JP Morgan Chase is doing with the data mesh and gluing data lakes together, you've got various public clouds playing in this game, and then the data is annotated to have a common meaning. In other words, there's a semantic layer that enables applications to talk to the data elements and know that they have common and coherent meaning. So George, the good news is this approach is more effective than the legacy monolithic models that Mike Olson was talking about, so what's the problem with this in your view? >> So today's data platforms added immense value 'cause they connected the data that was previously locked up in these monolithic apps or on all these different microservices, and that supported traditional BI and AI/ML use cases. But now if we want to build apps like Uber or Amazon.com, where they've got essentially an autonomously running supply chain and e-commerce app where humans only care and feed it. But the thing is figuring out what to buy, when to buy, where to deploy it, when to ship it. We needed a semantic layer on top of the data. So that, as you were saying, the data that's coming from all those apps, the different apps that's integrated, not just connected, but it means the same. And the issue is whenever you add a new layer to a stack to support new applications, there are implications for the already existing layers, like can they support the new layer and its use cases? So for instance, if you add a semantic layer that embeds app logic with the data rather than vice versa, which we been talking about and that's been the case for 60 years, then the new data layer faces challenges that the way you manage that data, the way you analyze that data, is not supported by today's tools. >> Okay, so actually Alex, bring me up that last slide if you would, I mean, you're basically saying at the bottom here, today's repositories don't really do joins at scale. The future is you're talking about hundreds or thousands or millions of data connections, and today's systems, we're talking about, I don't know, 6, 8, 10 joins and that is the fundamental problem you're saying, is a new data error coming and existing systems won't be able to handle it? >> Yeah, one way of thinking about it is that even though we call them relational databases, when we actually want to do lots of joins or when we want to analyze data from lots of different tables, we created a whole new industry for analytic databases where you sort of mung the data together into fewer tables. So you didn't have to do as many joins because the joins are difficult and slow. And when you're going to arbitrarily join thousands, hundreds of thousands or across millions of elements, you need a new type of database. We have them, they're called graph databases, but to query them, you go back to the prerelational era in terms of their usability. >> Okay, so we're going to come back to that and talk about how you get around that problem. But let's first lay out what the ideal data platform of the future we think looks like. And again, we're going to come back to use this Uber example. In this graphic that George put together, awesome. We got three layers. The application layer is where the data products reside. The example here is drivers, rides, maps, routes, ETA, et cetera. The digital version of what we were talking about in the previous slide, people, places and things. The next layer is the data layer, that breaks down the silos and connects the data elements through semantics and everything is coherent. And then the bottom layers, the legacy operational systems feed that data layer. George, explain what's different here, the graph database element, you talk about the relational query capabilities, and why can't I just throw memory at solving this problem? >> Some of the graph databases do throw memory at the problem and maybe without naming names, some of them live entirely in memory. And what you're dealing with is a prerelational in-memory database system where you navigate between elements, and the issue with that is we've had SQL for 50 years, so we don't have to navigate, we can say what we want without how to get it. That's the core of the problem. >> Okay. So if I may, I just want to drill into this a little bit. So you're talking about the expressiveness of a graph. Alex, if you'd bring that back out, the fourth bullet, expressiveness of a graph database with the relational ease of query. Can you explain what you mean by that? >> Yeah, so graphs are great because when you can describe anything with a graph, that's why they're becoming so popular. Expressive means you can represent anything easily. They're conducive to, you might say, in a world where we now want like the metaverse, like with a 3D world, and I don't mean the Facebook metaverse, I mean like the business metaverse when we want to capture data about everything, but we want it in context, we want to build a set of digital twins that represent everything going on in the world. And Uber is a tiny example of that. Uber built a graph to represent all the drivers and riders and maps and routes. But what you need out of a database isn't just a way to store stuff and update stuff. You need to be able to ask questions of it, you need to be able to query it. And if you go back to prerelational days, you had to know how to find your way to the data. It's sort of like when you give directions to someone and they didn't have a GPS system and a mapping system, you had to give them turn by turn directions. Whereas when you have a GPS and a mapping system, which is like the relational thing, you just say where you want to go, and it spits out the turn by turn directions, which let's say, the car might follow or whoever you're directing would follow. But the point is, it's much easier in a relational database to say, "I just want to get these results. You figure out how to get it." The graph database, they have not taken over the world because in some ways, it's taking a 50 year leap backwards. >> Alright, got it. Okay. Let's take a look at how the current Databricks offerings map to that ideal state that we just laid out. So to do that, we put together this chart that looks at the key elements of the Databricks portfolio, the core capability, the weakness, and the threat that may loom. Start with the Delta Lake, that's the storage layer, which is great for files and tables. It's got true separation of compute and storage, I want you to double click on that George, as independent elements, but it's weaker for the type of low latency ingest that we see coming in the future. And some of the threats highlighted here. AWS could add transactional tables to S3, Iceberg adoption is picking up and could accelerate, that could disrupt Databricks. George, add some color here please? >> Okay, so this is the sort of a classic competitive forces where you want to look at, so what are customers demanding? What's competitive pressure? What are substitutes? Even what your suppliers might be pushing. Here, Delta Lake is at its core, a set of transactional tables that sit on an object store. So think of it in a database system, this is the storage engine. So since S3 has been getting stronger for 15 years, you could see a scenario where they add transactional tables. We have an open source alternative in Iceberg, which Snowflake and others support. But at the same time, Databricks has built an ecosystem out of tools, their own and others, that read and write to Delta tables, that's what makes the Delta Lake and ecosystem. So they have a catalog, the whole machine learning tool chain talks directly to the data here. That was their great advantage because in the past with Snowflake, you had to pull all the data out of the database before the machine learning tools could work with it, that was a major shortcoming. They fixed that. But the point here is that even before we get to the semantic layer, the core foundation is under threat. >> Yep. Got it. Okay. We got a lot of ground to cover. So we're going to take a look at the Spark Execution Engine next. Think of that as the refinery that runs really efficient batch processing. That's kind of what disrupted the DOOp in a large way, but it's not Python friendly and that's an issue because the data science and the data engineering crowd are moving in that direction, and/or they're using DBT. George, we had Tristan Handy on at Supercloud, really interesting discussion that you and I did. Explain why this is an issue for Databricks? >> So once the data lake was in place, what people did was they refined their data batch, and Spark has always had streaming support and it's gotten better. The underlying storage as we've talked about is an issue. But basically they took raw data, then they refined it into tables that were like customers and products and partners. And then they refined that again into what was like gold artifacts, which might be business intelligence metrics or dashboards, which were collections of metrics. But they were running it on the Spark Execution Engine, which it's a Java-based engine or it's running on a Java-based virtual machine, which means all the data scientists and the data engineers who want to work with Python are really working in sort of oil and water. Like if you get an error in Python, you can't tell whether the problems in Python or where it's in Spark. There's just an impedance mismatch between the two. And then at the same time, the whole world is now gravitating towards DBT because it's a very nice and simple way to compose these data processing pipelines, and people are using either SQL in DBT or Python in DBT, and that kind of is a substitute for doing it all in Spark. So it's under threat even before we get to that semantic layer, it so happens that DBT itself is becoming the authoring environment for the semantic layer with business intelligent metrics. But that's again, this is the second element that's under direct substitution and competitive threat. >> Okay, let's now move down to the third element, which is the Photon. Photon is Databricks' BI Lakehouse, which has integration with the Databricks tooling, which is very rich, it's newer. And it's also not well suited for high concurrency and low latency use cases, which we think are going to increasingly become the norm over time. George, the call out threat here is customers want to connect everything to a semantic layer. Explain your thinking here and why this is a potential threat to Databricks? >> Okay, so two issues here. What you were touching on, which is the high concurrency, low latency, when people are running like thousands of dashboards and data is streaming in, that's a problem because SQL data warehouse, the query engine, something like that matures over five to 10 years. It's one of these things, the joke that Andy Jassy makes just in general, he's really talking about Azure, but there's no compression algorithm for experience. The Snowflake guy started more than five years earlier, and for a bunch of reasons, that lead is not something that Databricks can shrink. They'll always be behind. So that's why Snowflake has transactional tables now and we can get into that in another show. But the key point is, so near term, it's struggling to keep up with the use cases that are core to business intelligence, which is highly concurrent, lots of users doing interactive query. But then when you get to a semantic layer, that's when you need to be able to query data that might have thousands or tens of thousands or hundreds of thousands of joins. And that's a SQL query engine, traditional SQL query engine is just not built for that. That's the core problem of traditional relational databases. >> Now this is a quick aside. We always talk about Snowflake and Databricks in sort of the same context. We're not necessarily saying that Snowflake is in a position to tackle all these problems. We'll deal with that separately. So we don't mean to imply that, but we're just sort of laying out some of the things that Snowflake or rather Databricks customers we think, need to be thinking about and having conversations with Databricks about and we hope to have them as well. We'll come back to that in terms of sort of strategic options. But finally, when come back to the table, we have Databricks' AI/ML Tool Chain, which has been an awesome capability for the data science crowd. It's comprehensive, it's a one-stop shop solution, but the kicker here is that it's optimized for supervised model building. And the concern is that foundational models like GPT could cannibalize the current Databricks tooling, but George, can't Databricks, like other software companies, integrate foundation model capabilities into its platform? >> Okay, so the sound bite answer to that is sure, IBM 3270 terminals could call out to a graphical user interface when they're running on the XT terminal, but they're not exactly good citizens in that world. The core issue is Databricks has this wonderful end-to-end tool chain for training, deploying, monitoring, running inference on supervised models. But the paradigm there is the customer builds and trains and deploys each model for each feature or application. In a world of foundation models which are pre-trained and unsupervised, the entire tool chain is different. So it's not like Databricks can junk everything they've done and start over with all their engineers. They have to keep maintaining what they've done in the old world, but they have to build something new that's optimized for the new world. It's a classic technology transition and their mentality appears to be, "Oh, we'll support the new stuff from our old stuff." Which is suboptimal, and as we'll talk about, their biggest patron and the company that put them on the map, Microsoft, really stopped working on their old stuff three years ago so that they could build a new tool chain optimized for this new world. >> Yeah, and so let's sort of close with what we think the options are and decisions that Databricks has for its future architecture. They're smart people. I mean we've had Ali Ghodsi on many times, super impressive. I think they've got to be keenly aware of the limitations, what's going on with foundation models. But at any rate, here in this chart, we lay out sort of three scenarios. One is re-architect the platform by incrementally adopting new technologies. And example might be to layer a graph query engine on top of its stack. They could license key technologies like graph database, they could get aggressive on M&A and buy-in, relational knowledge graphs, semantic technologies, vector database technologies. George, as David Floyer always says, "A lot of ways to skin a cat." We've seen companies like, even think about EMC maintained its relevance through M&A for many, many years. George, give us your thought on each of these strategic options? >> Okay, I find this question the most challenging 'cause remember, I used to be an equity research analyst. I worked for Frank Quattrone, we were one of the top tech shops in the banking industry, although this is 20 years ago. But the M&A team was the top team in the industry and everyone wanted them on their side. And I remember going to meetings with these CEOs, where Frank and the bankers would say, "You want us for your M&A work because we can do better." And they really could do better. But in software, it's not like with EMC in hardware because with hardware, it's easier to connect different boxes. With software, the whole point of a software company is to integrate and architect the components so they fit together and reinforce each other, and that makes M&A harder. You can do it, but it takes a long time to fit the pieces together. Let me give you examples. If they put a graph query engine, let's say something like TinkerPop, on top of, I don't even know if it's possible, but let's say they put it on top of Delta Lake, then you have this graph query engine talking to their storage layer, Delta Lake. But if you want to do analysis, you got to put the data in Photon, which is not really ideal for highly connected data. If you license a graph database, then most of your data is in the Delta Lake and how do you sync it with the graph database? If you do sync it, you've got data in two places, which kind of defeats the purpose of having a unified repository. I find this semantic layer option in number three actually more promising, because that's something that you can layer on top of the storage layer that you have already. You just have to figure out then how to have your query engines talk to that. What I'm trying to highlight is, it's easy as an analyst to say, "You can buy this company or license that technology." But the really hard work is making it all work together and that is where the challenge is. >> Yeah, and well look, I thank you for laying that out. We've seen it, certainly Microsoft and Oracle. I guess you might argue that well, Microsoft had a monopoly in its desktop software and was able to throw off cash for a decade plus while it's stock was going sideways. Oracle had won the database wars and had amazing margins and cash flow to be able to do that. Databricks isn't even gone public yet, but I want to close with some of the players to watch. Alex, if you'd bring that back up, number four here. AWS, we talked about some of their options with S3 and it's not just AWS, it's blob storage, object storage. Microsoft, as you sort of alluded to, was an early go-to market channel for Databricks. We didn't address that really. So maybe in the closing comments we can. Google obviously, Snowflake of course, we're going to dissect their options in future Breaking Analysis. Dbt labs, where do they fit? Bob Muglia's company, Relational.ai, why are these players to watch George, in your opinion? >> So everyone is trying to assemble and integrate the pieces that would make building data applications, data products easy. And the critical part isn't just assembling a bunch of pieces, which is traditionally what AWS did. It's a Unix ethos, which is we give you the tools, you put 'em together, 'cause you then have the maximum choice and maximum power. So what the hyperscalers are doing is they're taking their key value stores, in the case of ASW it's DynamoDB, in the case of Azure it's Cosmos DB, and each are putting a graph query engine on top of those. So they have a unified storage and graph database engine, like all the data would be collected in the key value store. Then you have a graph database, that's how they're going to be presenting a foundation for building these data apps. Dbt labs is putting a semantic layer on top of data lakes and data warehouses and as we'll talk about, I'm sure in the future, that makes it easier to swap out the underlying data platform or swap in new ones for specialized use cases. Snowflake, what they're doing, they're so strong in data management and with their transactional tables, what they're trying to do is take in the operational data that used to be in the province of many state stores like MongoDB and say, "If you manage that data with us, it'll be connected to your analytic data without having to send it through a pipeline." And that's hugely valuable. Relational.ai is the wildcard, 'cause what they're trying to do, it's almost like a holy grail where you're trying to take the expressiveness of connecting all your data in a graph but making it as easy to query as you've always had it in a SQL database or I should say, in a relational database. And if they do that, it's sort of like, it'll be as easy to program these data apps as a spreadsheet was compared to procedural languages, like BASIC or Pascal. That's the implications of Relational.ai. >> Yeah, and again, we talked before, why can't you just throw this all in memory? We're talking in that example of really getting down to differences in how you lay the data out on disk in really, new database architecture, correct? >> Yes. And that's why it's not clear that you could take a data lake or even a Snowflake and why you can't put a relational knowledge graph on those. You could potentially put a graph database, but it'll be compromised because to really do what Relational.ai has done, which is the ease of Relational on top of the power of graph, you actually need to change how you're storing your data on disk or even in memory. So you can't, in other words, it's not like, oh we can add graph support to Snowflake, 'cause if you did that, you'd have to change, or in your data lake, you'd have to change how the data is physically laid out. And then that would break all the tools that talk to that currently. >> What in your estimation, is the timeframe where this becomes critical for a Databricks and potentially Snowflake and others? I mentioned earlier midterm, are we talking three to five years here? Are we talking end of decade? What's your radar say? >> I think something surprising is going on that's going to sort of come up the tailpipe and take everyone by storm. All the hype around business intelligence metrics, which is what we used to put in our dashboards where bookings, billings, revenue, customer, those things, those were the key artifacts that used to live in definitions in your BI tools, and DBT has basically created a standard for defining those so they live in your data pipeline or they're defined in their data pipeline and executed in the data warehouse or data lake in a shared way, so that all tools can use them. This sounds like a digression, it's not. All this stuff about data mesh, data fabric, all that's going on is we need a semantic layer and the business intelligence metrics are defining common semantics for your data. And I think we're going to find by the end of this year, that metrics are how we annotate all our analytic data to start adding common semantics to it. And we're going to find this semantic layer, it's not three to five years off, it's going to be staring us in the face by the end of this year. >> Interesting. And of course SVB today was shut down. We're seeing serious tech headwinds, and oftentimes in these sort of downturns or flat turns, which feels like this could be going on for a while, we emerge with a lot of new players and a lot of new technology. George, we got to leave it there. Thank you to George Gilbert for excellent insights and input for today's episode. I want to thank Alex Myerson who's on production and manages the podcast, of course Ken Schiffman as well. Kristin Martin and Cheryl Knight help get the word out on social media and in our newsletters. And Rob Hof is our EIC over at Siliconangle.com, he does some great editing. Remember all these episodes, they're available as podcasts. Wherever you listen, all you got to do is search Breaking Analysis Podcast, we publish each week on wikibon.com and siliconangle.com, or you can email me at David.Vellante@siliconangle.com, or DM me @DVellante. Comment on our LinkedIn post, and please do check out ETR.ai, great survey data, enterprise tech focus, phenomenal. This is Dave Vellante for theCUBE Insights powered by ETR. Thanks for watching, and we'll see you next time on Breaking Analysis.
SUMMARY :
bringing you data-driven core elements of the Databricks portfolio and pervasiveness in the data and that was where you went for data. and Cloudera set out to fix that. the reason you see and the robustness of Databricks and their big challenge and the data locked into in the real world and decisions Yes, and the mission of that is propelling the likes that the way you manage that data, is the fundamental problem because the joins are difficult and slow. and connects the data and the issue with that is the fourth bullet, expressiveness and it spits out the and the threat that may loom. because in the past with Snowflake, Think of that as the refinery So once the data lake was in place, George, the call out threat here But the key point is, in sort of the same context. and the company that put One is re-architect the platform and architect the components some of the players to watch. in the case of ASW it's DynamoDB, and why you can't put a relational and executed in the data and manages the podcast, of
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Alex Myerson | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
Mike Olson | PERSON | 0.99+ |
2014 | DATE | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Cheryl Knight | PERSON | 0.99+ |
Ken Schiffman | PERSON | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Erik Bradley | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
thousands | QUANTITY | 0.99+ |
Sun Microsystems | ORGANIZATION | 0.99+ |
50 years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Bob Muglia | PERSON | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
Airbnb | ORGANIZATION | 0.99+ |
60 years | QUANTITY | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Ali Ghodsi | PERSON | 0.99+ |
2010 | DATE | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
Kristin Martin | PERSON | 0.99+ |
Rob Hof | PERSON | 0.99+ |
three | QUANTITY | 0.99+ |
15 years | QUANTITY | 0.99+ |
Databricks' | ORGANIZATION | 0.99+ |
two places | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
Tristan Handy | PERSON | 0.99+ |
M&A | ORGANIZATION | 0.99+ |
Frank Quattrone | PERSON | 0.99+ |
second element | QUANTITY | 0.99+ |
Daren Brabham | PERSON | 0.99+ |
TechAlpha Partners | ORGANIZATION | 0.99+ |
third element | QUANTITY | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
50 year | QUANTITY | 0.99+ |
40% | QUANTITY | 0.99+ |
Cloudera | ORGANIZATION | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
five years | QUANTITY | 0.99+ |
Jack Greenfield, Walmart | A Dive into Walmart's Retail Supercloud
>> Welcome back to SuperCloud2. This is Dave Vellante, and we're here with Jack Greenfield. He's the Vice President of Enterprise Architecture and the Chief Architect for the global technology platform at Walmart. Jack, I want to thank you for coming on the program. Really appreciate your time. >> Glad to be here, Dave. Thanks for inviting me and appreciate the opportunity to chat with you. >> Yeah, it's our pleasure. Now we call what you've built a SuperCloud. That's our term, not yours, but how would you describe the Walmart Cloud Native Platform? >> So WCNP, as the acronym goes, is essentially an implementation of Kubernetes for the Walmart ecosystem. And what that means is that we've taken Kubernetes off the shelf as open source, and we have integrated it with a number of foundational services that provide other aspects of our computational environment. So Kubernetes off the shelf doesn't do everything. It does a lot. In particular the orchestration of containers, but it delegates through API a lot of key functions. So for example, secret management, traffic management, there's a need for telemetry and observability at a scale beyond what you get from raw Kubernetes. That is to say, harvesting the metrics that are coming out of Kubernetes and processing them, storing them in time series databases, dashboarding them, and so on. There's also an angle to Kubernetes that gets a lot of attention in the daily DevOps routine, that's not really part of the open source deliverable itself, and that is the DevOps sort of CICD pipeline-oriented lifecycle. And that is something else that we've added and integrated nicely. And then one more piece of this picture is that within a Kubernetes cluster, there's a function that is critical to allowing services to discover each other and integrate with each other securely and with proper configuration provided by the concept of a service mesh. So Istio, Linkerd, these are examples of service mesh technologies. And we have gone ahead and integrated actually those two. There's more than those two, but we've integrated those two with Kubernetes. So the net effect is that when a developer within Walmart is going to build an application, they don't have to think about all those other capabilities where they come from or how they're provided. Those are already present, and the way the CICD pipelines are set up, it's already sort of in the picture, and there are configuration points that they can take advantage of in the primary YAML and a couple of other pieces of config that we supply where they can tune it. But at the end of the day, it offloads an awful lot of work for them, having to stand up and operate those services, fail them over properly, and make them robust. All of that's provided for. >> Yeah, you know, developers often complain they spend too much time wrangling and doing things that aren't productive. So I wonder if you could talk about the high level business goals of the initiative in terms of the hardcore benefits. Was the real impetus to tap into best of breed cloud services? Were you trying to cut costs? Maybe gain negotiating leverage with the cloud guys? Resiliency, you know, I know was a major theme. Maybe you could give us a sense of kind of the anatomy of the decision making process that went in. >> Sure, and in the course of answering your question, I think I'm going to introduce the concept of our triplet architecture which we haven't yet touched on in the interview here. First off, just to sort of wrap up the motivation for WCNP itself which is kind of orthogonal to the triplet architecture. It can exist with or without it. Currently does exist with it, which is key, and I'll get to that in a moment. The key drivers, business drivers for WCNP were developer productivity by offloading the kinds of concerns that we've just discussed. Number two, improving resiliency, that is to say reducing opportunity for human error. One of the challenges you tend to run into in a large enterprise is what we call snowflakes, lots of gratuitously different workloads, projects, configurations to the extent that by developing and using WCNP and continuing to evolve it as we have, we end up with cookie cutter like consistency across our workloads which is super valuable when it comes to building tools or building services to automate operations that would otherwise be manual. When everything is pretty much done the same way, that becomes much simpler. Another key motivation for WCNP was the ability to abstract from the underlying cloud provider. And this is going to lead to a discussion of our triplet architecture. At the end of the day, when one works directly with an underlying cloud provider, one ends up taking a lot of dependencies on that particular cloud provider. Those dependencies can be valuable. For example, there are best of breed services like say Cloud Spanner offered by Google or say Cosmos DB offered by Microsoft that one wants to use and one is willing to take the dependency on the cloud provider to get that functionality because it's unique and valuable. On the other hand, one doesn't want to take dependencies on a cloud provider that don't add a lot of value. And with Kubernetes, we have the opportunity, and this is a large part of how Kubernetes was designed and why it is the way it is, we have the opportunity to sort of abstract from the underlying cloud provider for stateless workloads on compute. And so what this lets us do is build container-based applications that can run without change on different cloud provider infrastructure. So the same applications can run on WCNP over Azure, WCNP over GCP, or WCNP over the Walmart private cloud. And we have a private cloud. Our private cloud is OpenStack based and it gives us some significant cost advantages as well as control advantages. So to your point, in terms of business motivation, there's a key cost driver here, which is that we can use our own private cloud when it's advantageous and then use the public cloud provider capabilities when we need to. A key place with this comes into play is with elasticity. So while the private cloud is much more cost effective for us to run and use, it isn't as elastic as what the cloud providers offer, right? We don't have essentially unlimited scale. We have large scale, but the public cloud providers are elastic in the extreme which is a very powerful capability. So what we're able to do is burst, and we use this term bursting workloads into the public cloud from the private cloud to take advantage of the elasticity they offer and then fall back into the private cloud when the traffic load diminishes to the point where we don't need that elastic capability, elastic capacity at low cost. And this is a very important paradigm that I think is going to be very commonplace ultimately as the industry evolves. Private cloud is easier to operate and less expensive, and yet the public cloud provider capabilities are difficult to match. >> And the triplet, the tri is your on-prem private cloud and the two public clouds that you mentioned, is that right? >> That is correct. And we actually have an architecture in which we operate all three of those cloud platforms in close proximity with one another in three different major regions in the US. So we have east, west, and central. And in each of those regions, we have all three cloud providers. And the way it's configured, those data centers are within 10 milliseconds of each other, meaning that it's of negligible cost to interact between them. And this allows us to be fairly agnostic to where a particular workload is running. >> Does a human make that decision, Jack or is there some intelligence in the system that determines that? >> That's a really great question, Dave. And it's a great question because we're at the cusp of that transition. So currently humans make that decision. Humans choose to deploy workloads into a particular region and a particular provider within that region. That said, we're actively developing patterns and practices that will allow us to automate the placement of the workloads for a variety of criteria. For example, if in a particular region, a particular provider is heavily overloaded and is unable to provide the level of service that's expected through our SLAs, we could choose to fail workloads over from that cloud provider to a different one within the same region. But that's manual today. We do that, but people do it. Okay, we'd like to get to where that happens automatically. In the same way, we'd like to be able to automate the failovers, both for high availability and sort of the heavier disaster recovery model between, within a region between providers and even within a provider between the availability zones that are there, but also between regions for the sort of heavier disaster recovery or maintenance driven realignment of workload placement. Today, that's all manual. So we have people moving workloads from region A to region B or data center A to data center B. It's clean because of the abstraction. The workloads don't have to know or care, but there are latency considerations that come into play, and the humans have to be cognizant of those. And automating that can help ensure that we get the best performance and the best reliability. >> But you're developing the dataset to actually, I would imagine, be able to make those decisions in an automated fashion over time anyway. Is that a fair assumption? >> It is, and that's what we're actively developing right now. So if you were to look at us today, we have these nice abstractions and APIs in place, but people run that machine, if you will, moving toward a world where that machine is fully automated. >> What exactly are you abstracting? Is it sort of the deployment model or, you know, are you able to abstract, I'm just making this up like Azure functions and GCP functions so that you can sort of run them, you know, with a consistent experience. What exactly are you abstracting and how difficult was it to achieve that objective technically? >> that's a good question. What we're abstracting is the Kubernetes node construct. That is to say a cluster of Kubernetes nodes which are typically VMs, although they can run bare metal in certain contexts, is something that typically to stand up requires knowledge of the underlying cloud provider. So for example, with GCP, you would use GKE to set up a Kubernetes cluster, and in Azure, you'd use AKS. We are actually abstracting that aspect of things so that the developers standing up applications don't have to know what the underlying cluster management provider is. They don't have to know if it's GCP, AKS or our own Walmart private cloud. Now, in terms of functions like Azure functions that you've mentioned there, we haven't done that yet. That's another piece that we have sort of on our radar screen that, we'd like to get to is serverless approach, and the Knative work from Google and the Azure functions, those are things that we see good opportunity to use for a whole variety of use cases. But right now we're not doing much with that. We're strictly container based right now, and we do have some VMs that are running in sort of more of a traditional model. So our stateful workloads are primarily VM based, but for serverless, that's an opportunity for us to take some of these stateless workloads and turn them into cloud functions. >> Well, and that's another cost lever that you can pull down the road that's going to drop right to the bottom line. Do you see a day or maybe you're doing it today, but I'd be surprised, but where you build applications that actually span multiple clouds or is there, in your view, always going to be a direct one-to-one mapping between where an application runs and the specific cloud platform? >> That's a really great question. Well, yes and no. So today, application development teams choose a cloud provider to deploy to and a location to deploy to, and they have to get involved in moving an application like we talked about today. That said, the bursting capability that I mentioned previously is something that is a step in the direction of automatic migration. That is to say we're migrating workload to different locations automatically. Currently, the prototypes we've been developing and that we think are going to eventually make their way into production are leveraging Istio to assess the load incoming on a particular cluster and start shedding that load into a different location. Right now, the configuration of that is still manual, but there's another opportunity for automation there. And I think a key piece of this is that down the road, well, that's a, sort of a small step in the direction of an application being multi provider. We expect to see really an abstraction of the fact that there is a triplet even. So the workloads are moving around according to whatever the control plane decides is necessary based on a whole variety of inputs. And at that point, you will have true multi-cloud applications, applications that are distributed across the different providers and in a way that application developers don't have to think about. >> So Walmart's been a leader, Jack, in using data for competitive advantages for decades. It's kind of been a poster child for that. You've got a mountain of IP in the form of data, tools, applications best practices that until the cloud came out was all On Prem. But I'm really interested in this idea of building a Walmart ecosystem, which obviously you have. Do you see a day or maybe you're even doing it today where you take what we call the Walmart SuperCloud, WCNP in your words, and point or turn that toward an external world or your ecosystem, you know, supporting those partners or customers that could drive new revenue streams, you know directly from the platform? >> Great questions, Dave. So there's really two things to say here. The first is that with respect to data, our data workloads are primarily VM basis. I've mentioned before some VMware, some straight open stack. But the key here is that WCNP and Kubernetes are very powerful for stateless workloads, but for stateful workloads tend to be still climbing a bit of a growth curve in the industry. So our data workloads are not primarily based on WCNP. They're VM based. Now that said, there is opportunity to make some progress there, and we are looking at ways to move things into containers that are currently running in VMs which are stateful. The other question you asked is related to how we expose data to third parties and also functionality. Right now we do have in-house, for our own use, a very robust data architecture, and we have followed the sort of domain-oriented data architecture guidance from Martin Fowler. And we have data lakes in which we collect data from all the transactional systems and which we can then use and do use to build models which are then used in our applications. But right now we're not exposing the data directly to customers as a product. That's an interesting direction that's been talked about and may happen at some point, but right now that's internal. What we are exposing to customers is applications. So we're offering our global integrated fulfillment capabilities, our order picking and curbside pickup capabilities, and our cloud powered checkout capabilities to third parties. And this means we're standing up our own internal applications as externally facing SaaS applications which can serve our partners' customers. >> Yeah, of course, Martin Fowler really first introduced to the world Zhamak Dehghani's data mesh concept and this whole idea of data products and domain oriented thinking. Zhamak Dehghani, by the way, is a speaker at our event as well. Last question I had is edge, and how you think about the edge? You know, the stores are an edge. Are you putting resources there that sort of mirror this this triplet model? Or is it better to consolidate things in the cloud? I know there are trade-offs in terms of latency. How are you thinking about that? >> All really good questions. It's a challenging area as you can imagine because edges are subject to disconnection, right? Or reduced connection. So we do place the same architecture at the edge. So WCNP runs at the edge, and an application that's designed to run at WCNP can run at the edge. That said, there are a number of very specific considerations that come up when running at the edge, such as the possibility of disconnection or degraded connectivity. And so one of the challenges we have faced and have grappled with and done a good job of I think is dealing with the fact that applications go offline and come back online and have to reconnect and resynchronize, the sort of online offline capability is something that can be quite challenging. And we have a couple of application architectures that sort of form the two core sets of patterns that we use. One is an offline/online synchronization architecture where we discover that we've come back online, and we understand the differences between the online dataset and the offline dataset and how they have to be reconciled. The other is a message-based architecture. And here in our health and wellness domain, we've developed applications that are queue based. So they're essentially business processes that consist of multiple steps where each step has its own queue. And what that allows us to do is devote whatever bandwidth we do have to those pieces of the process that are most latency sensitive and allow the queue lengths to increase in parts of the process that are not latency sensitive, knowing that they will eventually catch up when the bandwidth is restored. And to put that in a little bit of context, we have fiber lengths to all of our locations, and we have I'll just use a round number, 10-ish thousand locations. It's larger than that, but that's the ballpark, and we have fiber to all of them, but when the fiber is disconnected, When the disconnection happens, we're able to fall back to 5G and to Starlink. Starlink is preferred. It's a higher bandwidth. 5G if that fails. But in each of those cases, the bandwidth drops significantly. And so the applications have to be intelligent about throttling back the traffic that isn't essential, so that it can push the essential traffic in those lower bandwidth scenarios. >> So much technology to support this amazing business which started in the early 1960s. Jack, unfortunately, we're out of time. I would love to have you back or some members of your team and drill into how you're using open source, but really thank you so much for explaining the approach that you've taken and participating in SuperCloud2. >> You're very welcome, Dave, and we're happy to come back and talk about other aspects of what we do. For example, we could talk more about the data lakes and the data mesh that we have in place. We could talk more about the directions we might go with serverless. So please look us up again. Happy to chat. >> I'm going to take you up on that, Jack. All right. This is Dave Vellante for John Furrier and the Cube community. Keep it right there for more action from SuperCloud2. (upbeat music)
SUMMARY :
and the Chief Architect for and appreciate the the Walmart Cloud Native Platform? and that is the DevOps Was the real impetus to tap into Sure, and in the course And the way it's configured, and the humans have to the dataset to actually, but people run that machine, if you will, Is it sort of the deployment so that the developers and the specific cloud platform? and that we think are going in the form of data, tools, applications a bit of a growth curve in the industry. and how you think about the edge? and allow the queue lengths to increase for explaining the and the data mesh that we have in place. and the Cube community.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Jack Greenfield | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Jack | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Martin Fowler | PERSON | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
US | LOCATION | 0.99+ |
Zhamak Dehghani | PERSON | 0.99+ |
Today | DATE | 0.99+ |
each | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
today | DATE | 0.99+ |
two things | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
each step | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
early 1960s | DATE | 0.99+ |
Starlink | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.98+ |
a day | QUANTITY | 0.97+ |
GCP | TITLE | 0.97+ |
Azure | TITLE | 0.96+ |
WCNP | TITLE | 0.96+ |
10 milliseconds | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.94+ |
Cloud Spanner | TITLE | 0.94+ |
Linkerd | ORGANIZATION | 0.93+ |
triplet | QUANTITY | 0.92+ |
three cloud providers | QUANTITY | 0.91+ |
Cube | ORGANIZATION | 0.9+ |
SuperCloud2 | ORGANIZATION | 0.89+ |
two core sets | QUANTITY | 0.88+ |
John Furrier | PERSON | 0.88+ |
one more piece | QUANTITY | 0.86+ |
two public clouds | QUANTITY | 0.86+ |
thousand locations | QUANTITY | 0.83+ |
Vice President | PERSON | 0.8+ |
10-ish | QUANTITY | 0.79+ |
WCNP | ORGANIZATION | 0.75+ |
decades | QUANTITY | 0.75+ |
three different major regions | QUANTITY | 0.74+ |
Breaking Analysis: Enterprise Technology Predictions 2023
(upbeat music beginning) >> From the Cube Studios in Palo Alto and Boston, bringing you data-driven insights from the Cube and ETR, this is "Breaking Analysis" with Dave Vellante. >> Making predictions about the future of enterprise tech is more challenging if you strive to lay down forecasts that are measurable. In other words, if you make a prediction, you should be able to look back a year later and say, with some degree of certainty, whether the prediction came true or not, with evidence to back that up. Hello and welcome to this week's Wikibon Cube Insights, powered by ETR. In this breaking analysis, we aim to do just that, with predictions about the macro IT spending environment, cost optimization, security, lots to talk about there, generative AI, cloud, and of course supercloud, blockchain adoption, data platforms, including commentary on Databricks, snowflake, and other key players, automation, events, and we may even have some bonus predictions around quantum computing, and perhaps some other areas. To make all this happen, we welcome back, for the third year in a row, my colleague and friend Eric Bradley from ETR. Eric, thanks for all you do for the community, and thanks for being part of this program. Again. >> I wouldn't miss it for the world. I always enjoy this one. Dave, good to see you. >> Yeah, so let me bring up this next slide and show you, actually come back to me if you would. I got to show the audience this. These are the inbounds that we got from PR firms starting in October around predictions. They know we do prediction posts. And so they'll send literally thousands and thousands of predictions from hundreds of experts in the industry, technologists, consultants, et cetera. And if you bring up the slide I can show you sort of the pattern that developed here. 40% of these thousands of predictions were from cyber. You had AI and data. If you combine those, it's still not close to cyber. Cost optimization was a big thing. Of course, cloud, some on DevOps, and software. Digital... Digital transformation got, you know, some lip service and SaaS. And then there was other, it's kind of around 2%. So quite remarkable, when you think about the focus on cyber, Eric. >> Yeah, there's two reasons why I think it makes sense, though. One, the cybersecurity companies have a lot of cash, so therefore the PR firms might be working a little bit harder for them than some of their other clients. (laughs) And then secondly, as you know, for multiple years now, when we do our macro survey, we ask, "What's your number one spending priority?" And again, it's security. It just isn't going anywhere. It just stays at the top. So I'm actually not that surprised by that little pie chart there, but I was shocked that SaaS was only 5%. You know, going back 10 years ago, that would've been the only thing anyone was talking about. >> Yeah. So true. All right, let's get into it. First prediction, we always start with kind of tech spending. Number one is tech spending increases between four and 5%. ETR has currently got it at 4.6% coming into 2023. This has been a consistently downward trend all year. We started, you know, much, much higher as we've been reporting. Bottom line is the fed is still in control. They're going to ease up on tightening, is the expectation, they're going to shoot for a soft landing. But you know, my feeling is this slingshot economy is going to continue, and it's going to continue to confound, whether it's supply chains or spending. The, the interesting thing about the ETR data, Eric, and I want you to comment on this, the largest companies are the most aggressive to cut. They're laying off, smaller firms are spending faster. They're actually growing at a much larger, faster rate as are companies in EMEA. And that's a surprise. That's outpacing the US and APAC. Chime in on this, Eric. >> Yeah, I was surprised on all of that. First on the higher level spending, we are definitely seeing it coming down, but the interesting thing here is headlines are making it worse. The huge research shop recently said 0% growth. We're coming in at 4.6%. And just so everyone knows, this is not us guessing, we asked 1,525 IT decision-makers what their budget growth will be, and they came in at 4.6%. Now there's a huge disparity, as you mentioned. The Fortune 500, global 2000, barely at 2% growth, but small, it's at 7%. So we're at a situation right now where the smaller companies are still playing a little bit of catch up on digital transformation, and they're spending money. The largest companies that have the most to lose from a recession are being more trepidatious, obviously. So they're playing a "Wait and see." And I hope we don't talk ourselves into a recession. Certainly the headlines and some of their research shops are helping it along. But another interesting comment here is, you know, energy and utilities used to be called an orphan and widow stock group, right? They are spending more than anyone, more than financials insurance, more than retail consumer. So right now it's being driven by mid, small, and energy and utilities. They're all spending like gangbusters, like nothing's happening. And it's the rest of everyone else that's being very cautious. >> Yeah, so very unpredictable right now. All right, let's go to number two. Cost optimization remains a major theme in 2023. We've been reporting on this. You've, we've shown a chart here. What's the primary method that your organization plans to use? You asked this question of those individuals that cited that they were going to reduce their spend and- >> Mhm. >> consolidating redundant vendors, you know, still leads the way, you know, far behind, cloud optimization is second, but it, but cloud continues to outpace legacy on-prem spending, no doubt. Somebody, it was, the guy's name was Alexander Feiglstorfer from Storyblok, sent in a prediction, said "All in one becomes extinct." Now, generally I would say I disagree with that because, you know, as we know over the years, suites tend to win out over, you know, individual, you know, point products. But I think what's going to happen is all in one is going to remain the norm for these larger companies that are cutting back. They want to consolidate redundant vendors, and the smaller companies are going to stick with that best of breed and be more aggressive and try to compete more effectively. What's your take on that? >> Yeah, I'm seeing much more consolidation in vendors, but also consolidation in functionality. We're seeing people building out new functionality, whether it's, we're going to talk about this later, so I don't want to steal too much of our thunder right now, but data and security also, we're seeing a functionality creep. So I think there's further consolidation happening here. I think niche solutions are going to be less likely, and platform solutions are going to be more likely in a spending environment where you want to reduce your vendors. You want to have one bill to pay, not 10. Another thing on this slide, real quick if I can before I move on, is we had a bunch of people write in and some of the answer options that aren't on this graph but did get cited a lot, unfortunately, is the obvious reduction in staff, hiring freezes, and delaying hardware, were three of the top write-ins. And another one was offshore outsourcing. So in addition to what we're seeing here, there were a lot of write-in options, and I just thought it would be important to state that, but essentially the cost optimization is by and far the highest one, and it's growing. So it's actually increased in our citations over the last year. >> And yeah, specifically consolidating redundant vendors. And so I actually thank you for bringing that other up, 'cause I had asked you, Eric, is there any evidence that repatriation is going on and we don't see it in the numbers, we don't see it even in the other, there was, I think very little or no mention of cloud repatriation, even though it might be happening in this in a smattering. >> Not a single mention, not one single mention. I went through it for you. Yep. Not one write-in. >> All right, let's move on. Number three, security leads M&A in 2023. Now you might say, "Oh, well that's a layup," but let me set this up Eric, because I didn't really do a great job with the slide. I hid the, what you've done, because you basically took, this is from the emerging technology survey with 1,181 responses from November. And what we did is we took Palo Alto and looked at the overlap in Palo Alto Networks accounts with these vendors that were showing on this chart. And Eric, I'm going to ask you to explain why we put a circle around OneTrust, but let me just set it up, and then have you comment on the slide and take, give us more detail. We're seeing private company valuations are off, you know, 10 to 40%. We saw a sneak, do a down round, but pretty good actually only down 12%. We've seen much higher down rounds. Palo Alto Networks we think is going to get busy. Again, they're an inquisitive company, they've been sort of quiet lately, and we think CrowdStrike, Cisco, Microsoft, Zscaler, we're predicting all of those will make some acquisitions and we're thinking that the targets are somewhere in this mess of security taxonomy. Other thing we're predicting AI meets cyber big time in 2023, we're going to probably going to see some acquisitions of those companies that are leaning into AI. We've seen some of that with Palo Alto. And then, you know, your comment to me, Eric, was "The RSA conference is going to be insane, hopping mad, "crazy this April," (Eric laughing) but give us your take on this data, and why the red circle around OneTrust? Take us back to that slide if you would, Alex. >> Sure. There's a few things here. First, let me explain what we're looking at. So because we separate the public companies and the private companies into two separate surveys, this allows us the ability to cross-reference that data. So what we're doing here is in our public survey, the tesis, everyone who cited some spending with Palo Alto, meaning they're a Palo Alto customer, we then cross-reference that with the private tech companies. Who also are they spending with? So what you're seeing here is an overlap. These companies that we have circled are doing the best in Palo Alto's accounts. Now, Palo Alto went and bought Twistlock a few years ago, which this data slide predicted, to be quite honest. And so I don't know if they necessarily are going to go after Snyk. Snyk, sorry. They already have something in that space. What they do need, however, is more on the authentication space. So I'm looking at OneTrust, with a 45% overlap in their overall net sentiment. That is a company that's already existing in their accounts and could be very synergistic to them. BeyondTrust as well, authentication identity. This is something that Palo needs to do to move more down that zero trust path. Now why did I pick Palo first? Because usually they're very inquisitive. They've been a little quiet lately. Secondly, if you look at the backdrop in the markets, the IPO freeze isn't going to last forever. Sooner or later, the IPO markets are going to open up, and some of these private companies are going to tap into public equity. In the meantime, however, cash funding on the private side is drying up. If they need another round, they're not going to get it, and they're certainly not going to get it at the valuations they were getting. So we're seeing valuations maybe come down where they're a touch more attractive, and Palo knows this isn't going to last forever. Cisco knows that, CrowdStrike, Zscaler, all these companies that are trying to make a push to become that vendor that you're consolidating in, around, they have a chance now, they have a window where they need to go make some acquisitions. And that's why I believe leading up to RSA, we're going to see some movement. I think it's going to pretty, a really exciting time in security right now. >> Awesome. Thank you. Great explanation. All right, let's go on the next one. Number four is, it relates to security. Let's stay there. Zero trust moves from hype to reality in 2023. Now again, you might say, "Oh yeah, that's a layup." A lot of these inbounds that we got are very, you know, kind of self-serving, but we always try to put some meat in the bone. So first thing we do is we pull out some commentary from, Eric, your roundtable, your insights roundtable. And we have a CISO from a global hospitality firm says, "For me that's the highest priority." He's talking about zero trust because it's the best ROI, it's the most forward-looking, and it enables a lot of the business transformation activities that we want to do. CISOs tell me that they actually can drive forward transformation projects that have zero trust, and because they can accelerate them, because they don't have to go through the hurdle of, you know, getting, making sure that it's secure. Second comment, zero trust closes that last mile where once you're authenticated, they open up the resource to you in a zero trust way. That's a CISO of a, and a managing director of a cyber risk services enterprise. Your thoughts on this? >> I can be here all day, so I'm going to try to be quick on this one. This is not a fluff piece on this one. There's a couple of other reasons this is happening. One, the board finally gets it. Zero trust at first was just a marketing hype term. Now the board understands it, and that's why CISOs are able to push through it. And what they finally did was redefine what it means. Zero trust simply means moving away from hardware security, moving towards software-defined security, with authentication as its base. The board finally gets that, and now they understand that this is necessary and it's being moved forward. The other reason it's happening now is hybrid work is here to stay. We weren't really sure at first, large companies were still trying to push people back to the office, and it's going to happen. The pendulum will swing back, but hybrid work's not going anywhere. By basically on our own data, we're seeing that 69% of companies expect remote and hybrid to be permanent, with only 30% permanent in office. Zero trust works for a hybrid environment. So all of that is the reason why this is happening right now. And going back to our previous prediction, this is why we're picking Palo, this is why we're picking Zscaler to make these acquisitions. Palo Alto needs to be better on the authentication side, and so does Zscaler. They're both fantastic on zero trust network access, but they need the authentication software defined aspect, and that's why we think this is going to happen. One last thing, in that CISO round table, I also had somebody say, "Listen, Zscaler is incredible. "They're doing incredibly well pervading the enterprise, "but their pricing's getting a little high," and they actually think Palo Alto is well-suited to start taking some of that share, if Palo can make one move. >> Yeah, Palo Alto's consolidation story is very strong. Here's my question and challenge. Do you and me, so I'm always hardcore about, okay, you've got to have evidence. I want to look back at these things a year from now and say, "Did we get it right? Yes or no?" If we got it wrong, we'll tell you we got it wrong. So how are we going to measure this? I'd say a couple things, and you can chime in. One is just the number of vendors talking about it. That's, but the marketing always leads the reality. So the second part of that is we got to get evidence from the buying community. Can you help us with that? >> (laughs) Luckily, that's what I do. I have a data company that asks thousands of IT decision-makers what they're adopting and what they're increasing spend on, as well as what they're decreasing spend on and what they're replacing. So I have snapshots in time over the last 11 years where I can go ahead and compare and contrast whether this adoption is happening or not. So come back to me in 12 months and I'll let you know. >> Now, you know, I will. Okay, let's bring up the next one. Number five, generative AI hits where the Metaverse missed. Of course everybody's talking about ChatGPT, we just wrote last week in a breaking analysis with John Furrier and Sarjeet Joha our take on that. We think 2023 does mark a pivot point as natural language processing really infiltrates enterprise tech just as Amazon turned the data center into an API. We think going forward, you're going to be interacting with technology through natural language, through English commands or other, you know, foreign language commands, and investors are lining up, all the VCs are getting excited about creating something competitive to ChatGPT, according to (indistinct) a hundred million dollars gets you a seat at the table, gets you into the game. (laughing) That's before you have to start doing promotion. But he thinks that's what it takes to actually create a clone or something equivalent. We've seen stuff from, you know, the head of Facebook's, you know, AI saying, "Oh, it's really not that sophisticated, ChatGPT, "it's kind of like IBM Watson, it's great engineering, "but you know, we've got more advanced technology." We know Google's working on some really interesting stuff. But here's the thing. ETR just launched this survey for the February survey. It's in the field now. We circle open AI in this category. They weren't even in the survey, Eric, last quarter. So 52% of the ETR survey respondents indicated a positive sentiment toward open AI. I added up all the sort of different bars, we could double click on that. And then I got this inbound from Scott Stevenson of Deep Graham. He said "AI is recession-proof." I don't know if that's the case, but it's a good quote. So bring this back up and take us through this. Explain this chart for us, if you would. >> First of all, I like Scott's quote better than the Facebook one. I think that's some sour grapes. Meta just spent an insane amount of money on the Metaverse and that's a dud. Microsoft just spent money on open AI and it is hot, undoubtedly hot. We've only been in the field with our current ETS survey for a week. So my caveat is it's preliminary data, but I don't care if it's preliminary data. (laughing) We're getting a sneak peek here at what is the number one net sentiment and mindshare leader in the entire machine-learning AI sector within a week. It's beating Data- >> 600. 600 in. >> It's beating Databricks. And we all know Databricks is a huge established enterprise company, not only in machine-learning AI, but it's in the top 10 in the entire survey. We have over 400 vendors in this survey. It's number eight overall, already. In a week. This is not hype. This is real. And I could go on the NLP stuff for a while. Not only here are we seeing it in open AI and machine-learning and AI, but we're seeing NLP in security. It's huge in email security. It's completely transforming that area. It's one of the reasons I thought Palo might take Abnormal out. They're doing such a great job with NLP in this email side, and also in the data prep tools. NLP is going to take out data prep tools. If we have time, I'll discuss that later. But yeah, this is, to me this is a no-brainer, and we're already seeing it in the data. >> Yeah, John Furrier called, you know, the ChatGPT introduction. He said it reminded him of the Netscape moment, when we all first saw Netscape Navigator and went, "Wow, it really could be transformative." All right, number six, the cloud expands to supercloud as edge computing accelerates and CloudFlare is a big winner in 2023. We've reported obviously on cloud, multi-cloud, supercloud and CloudFlare, basically saying what multi-cloud should have been. We pulled this quote from Atif Kahn, who is the founder and CTO of Alkira, thanks, one of the inbounds, thank you. "In 2023, highly distributed IT environments "will become more the norm "as organizations increasingly deploy hybrid cloud, "multi-cloud and edge settings..." Eric, from one of your round tables, "If my sources from edge computing are coming "from the cloud, that means I have my workloads "running in the cloud. "There is no one better than CloudFlare," That's a senior director of IT architecture at a huge financial firm. And then your analysis shows CloudFlare really growing in pervasion, that sort of market presence in the dataset, dramatically, to near 20%, leading, I think you had told me that they're even ahead of Google Cloud in terms of momentum right now. >> That was probably the biggest shock to me in our January 2023 tesis, which covers the public companies in the cloud computing sector. CloudFlare has now overtaken GCP in overall spending, and I was shocked by that. It's already extremely pervasive in networking, of course, for the edge networking side, and also in security. This is the number one leader in SaaSi, web access firewall, DDoS, bot protection, by your definition of supercloud, which we just did a couple of weeks ago, and I really enjoyed that by the way Dave, I think CloudFlare is the one that fits your definition best, because it's bringing all of these aspects together, and most importantly, it's cloud agnostic. It does not need to rely on Azure or AWS to do this. It has its own cloud. So I just think it's, when we look at your definition of supercloud, CloudFlare is the poster child. >> You know, what's interesting about that too, is a lot of people are poo-pooing CloudFlare, "Ah, it's, you know, really kind of not that sophisticated." "You don't have as many tools," but to your point, you're can have those tools in the cloud, Cloudflare's doing serverless on steroids, trying to keep things really simple, doing a phenomenal job at, you know, various locations around the world. And they're definitely one to watch. Somebody put them on my radar (laughing) a while ago and said, "Dave, you got to do a breaking analysis on CloudFlare." And so I want to thank that person. I can't really name them, 'cause they work inside of a giant hyperscaler. But- (Eric laughing) (Dave chuckling) >> Real quickly, if I can from a competitive perspective too, who else is there? They've already taken share from Akamai, and Fastly is their really only other direct comp, and they're not there. And these guys are in poll position and they're the only game in town right now. I just, I don't see it slowing down. >> I thought one of your comments from your roundtable I was reading, one of the folks said, you know, CloudFlare, if my workloads are in the cloud, they are, you know, dominant, they said not as strong with on-prem. And so Akamai is doing better there. I'm like, "Okay, where would you want to be?" (laughing) >> Yeah, which one of those two would you rather be? >> Right? Anyway, all right, let's move on. Number seven, blockchain continues to look for a home in the enterprise, but devs will slowly begin to adopt in 2023. You know, blockchains have got a lot of buzz, obviously crypto is, you know, the killer app for blockchain. Senior IT architect in financial services from your, one of your insight roundtables said quote, "For enterprises to adopt a new technology, "there have to be proven turnkey solutions. "My experience in talking with my peers are, "blockchain is still an open-source component "where you have to build around it." Now I want to thank Ravi Mayuram, who's the CTO of Couchbase sent in, you know, one of the predictions, he said, "DevOps will adopt blockchain, specifically Ethereum." And he referenced actually in his email to me, Solidity, which is the programming language for Ethereum, "will be in every DevOps pro's playbook, "mirroring the boom in machine-learning. "Newer programming languages like Solidity "will enter the toolkits of devs." His point there, you know, Solidity for those of you don't know, you know, Bitcoin is not programmable. Solidity, you know, came out and that was their whole shtick, and they've been improving that, and so forth. But it, Eric, it's true, it really hasn't found its home despite, you know, the potential for smart contracts. IBM's pushing it, VMware has had announcements, and others, really hasn't found its way in the enterprise yet. >> Yeah, and I got to be honest, I don't think it's going to, either. So when we did our top trends series, this was basically chosen as an anti-prediction, I would guess, that it just continues to not gain hold. And the reason why was that first comment, right? It's very much a niche solution that requires a ton of custom work around it. You can't just plug and play it. And at the end of the day, let's be very real what this technology is, it's a database ledger, and we already have database ledgers in the enterprise. So why is this a priority to move to a different database ledger? It's going to be very niche cases. I like the CTO comment from Couchbase about it being adopted by DevOps. I agree with that, but it has to be a DevOps in a very specific use case, and a very sophisticated use case in financial services, most likely. And that's not across the entire enterprise. So I just think it's still going to struggle to get its foothold for a little bit longer, if ever. >> Great, thanks. Okay, let's move on. Number eight, AWS Databricks, Google Snowflake lead the data charge with Microsoft. Keeping it simple. So let's unpack this a little bit. This is the shared accounts peer position for, I pulled data platforms in for analytics, machine-learning and AI and database. So I could grab all these accounts or these vendors and see how they compare in those three sectors. Analytics, machine-learning and database. Snowflake and Databricks, you know, they're on a crash course, as you and I have talked about. They're battling to be the single source of truth in analytics. They're, there's going to be a big focus. They're already started. It's going to be accelerated in 2023 on open formats. Iceberg, Python, you know, they're all the rage. We heard about Iceberg at Snowflake Summit, last summer or last June. Not a lot of people had heard of it, but of course the Databricks crowd, who knows it well. A lot of other open source tooling. There's a company called DBT Labs, which you're going to talk about in a minute. George Gilbert put them on our radar. We just had Tristan Handy, the CEO of DBT labs, on at supercloud last week. They are a new disruptor in data that's, they're essentially making, they're API-ifying, if you will, KPIs inside the data warehouse and dramatically simplifying that whole data pipeline. So really, you know, the ETL guys should be shaking in their boots with them. Coming back to the slide. Google really remains focused on BigQuery adoption. Customers have complained to me that they would like to use Snowflake with Google's AI tools, but they're being forced to go to BigQuery. I got to ask Google about that. AWS continues to stitch together its bespoke data stores, that's gone down that "Right tool for the right job" path. David Foyer two years ago said, "AWS absolutely is going to have to solve that problem." We saw them start to do it in, at Reinvent, bringing together NoETL between Aurora and Redshift, and really trying to simplify those worlds. There's going to be more of that. And then Microsoft, they're just making it cheap and easy to use their stuff, you know, despite some of the complaints that we hear in the community, you know, about things like Cosmos, but Eric, your take? >> Yeah, my concern here is that Snowflake and Databricks are fighting each other, and it's allowing AWS and Microsoft to kind of catch up against them, and I don't know if that's the right move for either of those two companies individually, Azure and AWS are building out functionality. Are they as good? No they're not. The other thing to remember too is that AWS and Azure get paid anyway, because both Databricks and Snowflake run on top of 'em. So (laughing) they're basically collecting their toll, while these two fight it out with each other, and they build out functionality. I think they need to stop focusing on each other, a little bit, and think about the overall strategy. Now for Databricks, we know they came out first as a machine-learning AI tool. They were known better for that spot, and now they're really trying to play catch-up on that data storage compute spot, and inversely for Snowflake, they were killing it with the compute separation from storage, and now they're trying to get into the MLAI spot. I actually wouldn't be surprised to see them make some sort of acquisition. Frank Slootman has been a little bit quiet, in my opinion there. The other thing to mention is your comment about DBT Labs. If we look at our emerging technology survey, last survey when this came out, DBT labs, number one leader in that data integration space, I'm going to just pull it up real quickly. It looks like they had a 33% overall net sentiment to lead data analytics integration. So they are clearly growing, it's fourth straight survey consecutively that they've grown. The other name we're seeing there a little bit is Cribl, but DBT labs is by far the number one player in this space. >> All right. Okay, cool. Moving on, let's go to number nine. With Automation mixer resurgence in 2023, we're showing again data. The x axis is overlap or presence in the dataset, and the vertical axis is shared net score. Net score is a measure of spending momentum. As always, you've seen UI path and Microsoft Power Automate up until the right, that red line, that 40% line is generally considered elevated. UI path is really separating, creating some distance from Automation Anywhere, they, you know, previous quarters they were much closer. Microsoft Power Automate came on the scene in a big way, they loom large with this "Good enough" approach. I will say this, I, somebody sent me a results of a (indistinct) survey, which showed UiPath actually had more mentions than Power Automate, which was surprising, but I think that's not been the case in the ETR data set. We're definitely seeing a shift from back office to front soft office kind of workloads. Having said that, software testing is emerging as a mainstream use case, we're seeing ML and AI become embedded in end-to-end automations, and low-code is serving the line of business. And so this, we think, is going to increasingly have appeal to organizations in the coming year, who want to automate as much as possible and not necessarily, we've seen a lot of layoffs in tech, and people... You're going to have to fill the gaps with automation. That's a trend that's going to continue. >> Yep, agreed. At first that comment about Microsoft Power Automate having less citations than UiPath, that's shocking to me. I'm looking at my chart right here where Microsoft Power Automate was cited by over 60% of our entire survey takers, and UiPath at around 38%. Now don't get me wrong, 38% pervasion's fantastic, but you know you're not going to beat an entrenched Microsoft. So I don't really know where that comment came from. So UiPath, looking at it alone, it's doing incredibly well. It had a huge rebound in its net score this last survey. It had dropped going through the back half of 2022, but we saw a big spike in the last one. So it's got a net score of over 55%. A lot of people citing adoption and increasing. So that's really what you want to see for a name like this. The problem is that just Microsoft is doing its playbook. At the end of the day, I'm going to do a POC, why am I going to pay more for UiPath, or even take on another separate bill, when we know everyone's consolidating vendors, if my license already includes Microsoft Power Automate? It might not be perfect, it might not be as good, but what I'm hearing all the time is it's good enough, and I really don't want another invoice. >> Right. So how does UiPath, you know, and Automation Anywhere, how do they compete with that? Well, the way they compete with it is they got to have a better product. They got a product that's 10 times better. You know, they- >> Right. >> they're not going to compete based on where the lowest cost, Microsoft's got that locked up, or where the easiest to, you know, Microsoft basically give it away for free, and that's their playbook. So that's, you know, up to UiPath. UiPath brought on Rob Ensslin, I've interviewed him. Very, very capable individual, is now Co-CEO. So he's kind of bringing that adult supervision in, and really tightening up the go to market. So, you know, we know this company has been a rocket ship, and so getting some control on that and really getting focused like a laser, you know, could be good things ahead there for that company. Okay. >> One of the problems, if I could real quick Dave, is what the use cases are. When we first came out with RPA, everyone was super excited about like, "No, UiPath is going to be great for super powerful "projects, use cases." That's not what RPA is being used for. As you mentioned, it's being used for mundane tasks, so it's not automating complex things, which I think UiPath was built for. So if you were going to get UiPath, and choose that over Microsoft, it's going to be 'cause you're doing it for more powerful use case, where it is better. But the problem is that's not where the enterprise is using it. The enterprise are using this for base rote tasks, and simply, Microsoft Power Automate can do that. >> Yeah, it's interesting. I've had people on theCube that are both Microsoft Power Automate customers and UiPath customers, and I've asked them, "Well you know, "how do you differentiate between the two?" And they've said to me, "Look, our users and personal productivity users, "they like Power Automate, "they can use it themselves, and you know, "it doesn't take a lot of, you know, support on our end." The flip side is you could do that with UiPath, but like you said, there's more of a focus now on end-to-end enterprise automation and building out those capabilities. So it's increasingly a value play, and that's going to be obviously the challenge going forward. Okay, my last one, and then I think you've got some bonus ones. Number 10, hybrid events are the new category. Look it, if I can get a thousand inbounds that are largely self-serving, I can do my own here, 'cause we're in the events business. (Eric chuckling) Here's the prediction though, and this is a trend we're seeing, the number of physical events is going to dramatically increase. That might surprise people, but most of the big giant events are going to get smaller. The exception is AWS with Reinvent, I think Snowflake's going to continue to grow. So there are examples of physical events that are growing, but generally, most of the big ones are getting smaller, and there's going to be many more smaller intimate regional events and road shows. These micro-events, they're going to be stitched together. Digital is becoming a first class citizen, so people really got to get their digital acts together, and brands are prioritizing earned media, and they're beginning to build their own news networks, going direct to their customers. And so that's a trend we see, and I, you know, we're right in the middle of it, Eric, so you know we're going to, you mentioned RSA, I think that's perhaps going to be one of those crazy ones that continues to grow. It's shrunk, and then it, you know, 'cause last year- >> Yeah, it did shrink. >> right, it was the last one before the pandemic, and then they sort of made another run at it last year. It was smaller but it was very vibrant, and I think this year's going to be huge. Global World Congress is another one, we're going to be there end of Feb. That's obviously a big big show, but in general, the brands and the technology vendors, even Oracle is going to scale down. I don't know about Salesforce. We'll see. You had a couple of bonus predictions. Quantum and maybe some others? Bring us home. >> Yeah, sure. I got a few more. I think we touched upon one, but I definitely think the data prep tools are facing extinction, unfortunately, you know, the Talons Informatica is some of those names. The problem there is that the BI tools are kind of including data prep into it already. You know, an example of that is Tableau Prep Builder, and then in addition, Advanced NLP is being worked in as well. ThoughtSpot, Intelius, both often say that as their selling point, Tableau has Ask Data, Click has Insight Bot, so you don't have to really be intelligent on data prep anymore. A regular business user can just self-query, using either the search bar, or even just speaking into what it needs, and these tools are kind of doing the data prep for it. I don't think that's a, you know, an out in left field type of prediction, but it's the time is nigh. The other one I would also state is that I think knowledge graphs are going to break through this year. Neo4j in our survey is growing in pervasion in Mindshare. So more and more people are citing it, AWS Neptune's getting its act together, and we're seeing that spending intentions are growing there. Tiger Graph is also growing in our survey sample. I just think that the time is now for knowledge graphs to break through, and if I had to do one more, I'd say real-time streaming analytics moves from the very, very rich big enterprises to downstream, to more people are actually going to be moving towards real-time streaming, again, because the data prep tools and the data pipelines have gotten easier to use, and I think the ROI on real-time streaming is obviously there. So those are three that didn't make the cut, but I thought deserved an honorable mention. >> Yeah, I'm glad you did. Several weeks ago, we did an analyst prediction roundtable, if you will, a cube session power panel with a number of data analysts and that, you know, streaming, real-time streaming was top of mind. So glad you brought that up. Eric, as always, thank you very much. I appreciate the time you put in beforehand. I know it's been crazy, because you guys are wrapping up, you know, the last quarter survey in- >> Been a nuts three weeks for us. (laughing) >> job. I love the fact that you're doing, you know, the ETS survey now, I think it's quarterly now, right? Is that right? >> Yep. >> Yep. So that's phenomenal. >> Four times a year. I'll be happy to jump on with you when we get that done. I know you were really impressed with that last time. >> It's unbelievable. This is so much data at ETR. Okay. Hey, that's a wrap. Thanks again. >> Take care Dave. Good seeing you. >> All right, many thanks to our team here, Alex Myerson as production, he manages the podcast force. Ken Schiffman as well is a critical component of our East Coast studio. Kristen Martin and Cheryl Knight help get the word out on social media and in our newsletters. And Rob Hoof is our editor-in-chief. He's at siliconangle.com. He's just a great editing for us. Thank you all. Remember all these episodes that are available as podcasts, wherever you listen, podcast is doing great. Just search "Breaking analysis podcast." Really appreciate you guys listening. I publish each week on wikibon.com and siliconangle.com, or you can email me directly if you want to get in touch, david.vellante@siliconangle.com. That's how I got all these. I really appreciate it. I went through every single one with a yellow highlighter. It took some time, (laughing) but I appreciate it. You could DM me at dvellante, or comment on our LinkedIn post and please check out etr.ai. Its data is amazing. Best survey data in the enterprise tech business. This is Dave Vellante for theCube Insights, powered by ETR. Thanks for watching, and we'll see you next time on "Breaking Analysis." (upbeat music beginning) (upbeat music ending)
SUMMARY :
insights from the Cube and ETR, do for the community, Dave, good to see you. actually come back to me if you would. It just stays at the top. the most aggressive to cut. that have the most to lose What's the primary method still leads the way, you know, So in addition to what we're seeing here, And so I actually thank you I went through it for you. I'm going to ask you to explain and they're certainly not going to get it to you in a zero trust way. So all of that is the One is just the number of So come back to me in 12 So 52% of the ETR survey amount of money on the Metaverse and also in the data prep tools. the cloud expands to the biggest shock to me "Ah, it's, you know, really and Fastly is their really the folks said, you know, for a home in the enterprise, Yeah, and I got to be honest, in the community, you know, and I don't know if that's the right move and the vertical axis is shared net score. So that's really what you want Well, the way they compete So that's, you know, One of the problems, if and that's going to be obviously even Oracle is going to scale down. and the data pipelines and that, you know, Been a nuts three I love the fact I know you were really is so much data at ETR. and we'll see you next time
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Alex Myerson | PERSON | 0.99+ |
Eric | PERSON | 0.99+ |
Eric Bradley | PERSON | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Rob Hoof | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
10 | QUANTITY | 0.99+ |
Ravi Mayuram | PERSON | 0.99+ |
Cheryl Knight | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Ken Schiffman | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Tristan Handy | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Atif Kahn | PERSON | 0.99+ |
November | DATE | 0.99+ |
Frank Slootman | PERSON | 0.99+ |
APAC | ORGANIZATION | 0.99+ |
Zscaler | ORGANIZATION | 0.99+ |
Palo | ORGANIZATION | 0.99+ |
David Foyer | PERSON | 0.99+ |
February | DATE | 0.99+ |
January 2023 | DATE | 0.99+ |
DBT Labs | ORGANIZATION | 0.99+ |
October | DATE | 0.99+ |
Rob Ensslin | PERSON | 0.99+ |
Scott Stevenson | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
69% | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
CrowdStrike | ORGANIZATION | 0.99+ |
4.6% | QUANTITY | 0.99+ |
10 times | QUANTITY | 0.99+ |
2023 | DATE | 0.99+ |
Scott | PERSON | 0.99+ |
1,181 responses | QUANTITY | 0.99+ |
Palo Alto | ORGANIZATION | 0.99+ |
third year | QUANTITY | 0.99+ |
Boston | LOCATION | 0.99+ |
Alex | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
OneTrust | ORGANIZATION | 0.99+ |
45% | QUANTITY | 0.99+ |
33% | QUANTITY | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
two reasons | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
last year | DATE | 0.99+ |
BeyondTrust | ORGANIZATION | 0.99+ |
7% | QUANTITY | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Jack Greenfield, Walmart | A Dive into Walmart's Retail Supercloud
>> Welcome back to SuperCloud2. This is Dave Vellante, and we're here with Jack Greenfield. He's the Vice President of Enterprise Architecture and the Chief Architect for the global technology platform at Walmart. Jack, I want to thank you for coming on the program. Really appreciate your time. >> Glad to be here, Dave. Thanks for inviting me and appreciate the opportunity to chat with you. >> Yeah, it's our pleasure. Now we call what you've built a SuperCloud. That's our term, not yours, but how would you describe the Walmart Cloud Native Platform? >> So WCNP, as the acronym goes, is essentially an implementation of Kubernetes for the Walmart ecosystem. And what that means is that we've taken Kubernetes off the shelf as open source, and we have integrated it with a number of foundational services that provide other aspects of our computational environment. So Kubernetes off the shelf doesn't do everything. It does a lot. In particular the orchestration of containers, but it delegates through API a lot of key functions. So for example, secret management, traffic management, there's a need for telemetry and observability at a scale beyond what you get from raw Kubernetes. That is to say, harvesting the metrics that are coming out of Kubernetes and processing them, storing them in time series databases, dashboarding them, and so on. There's also an angle to Kubernetes that gets a lot of attention in the daily DevOps routine, that's not really part of the open source deliverable itself, and that is the DevOps sort of CICD pipeline-oriented lifecycle. And that is something else that we've added and integrated nicely. And then one more piece of this picture is that within a Kubernetes cluster, there's a function that is critical to allowing services to discover each other and integrate with each other securely and with proper configuration provided by the concept of a service mesh. So Istio, Linkerd, these are examples of service mesh technologies. And we have gone ahead and integrated actually those two. There's more than those two, but we've integrated those two with Kubernetes. So the net effect is that when a developer within Walmart is going to build an application, they don't have to think about all those other capabilities where they come from or how they're provided. Those are already present, and the way the CICD pipelines are set up, it's already sort of in the picture, and there are configuration points that they can take advantage of in the primary YAML and a couple of other pieces of config that we supply where they can tune it. But at the end of the day, it offloads an awful lot of work for them, having to stand up and operate those services, fail them over properly, and make them robust. All of that's provided for. >> Yeah, you know, developers often complain they spend too much time wrangling and doing things that aren't productive. So I wonder if you could talk about the high level business goals of the initiative in terms of the hardcore benefits. Was the real impetus to tap into best of breed cloud services? Were you trying to cut costs? Maybe gain negotiating leverage with the cloud guys? Resiliency, you know, I know was a major theme. Maybe you could give us a sense of kind of the anatomy of the decision making process that went in. >> Sure, and in the course of answering your question, I think I'm going to introduce the concept of our triplet architecture which we haven't yet touched on in the interview here. First off, just to sort of wrap up the motivation for WCNP itself which is kind of orthogonal to the triplet architecture. It can exist with or without it. Currently does exist with it, which is key, and I'll get to that in a moment. The key drivers, business drivers for WCNP were developer productivity by offloading the kinds of concerns that we've just discussed. Number two, improving resiliency, that is to say reducing opportunity for human error. One of the challenges you tend to run into in a large enterprise is what we call snowflakes, lots of gratuitously different workloads, projects, configurations to the extent that by developing and using WCNP and continuing to evolve it as we have, we end up with cookie cutter like consistency across our workloads which is super valuable when it comes to building tools or building services to automate operations that would otherwise be manual. When everything is pretty much done the same way, that becomes much simpler. Another key motivation for WCNP was the ability to abstract from the underlying cloud provider. And this is going to lead to a discussion of our triplet architecture. At the end of the day, when one works directly with an underlying cloud provider, one ends up taking a lot of dependencies on that particular cloud provider. Those dependencies can be valuable. For example, there are best of breed services like say Cloud Spanner offered by Google or say Cosmos DB offered by Microsoft that one wants to use and one is willing to take the dependency on the cloud provider to get that functionality because it's unique and valuable. On the other hand, one doesn't want to take dependencies on a cloud provider that don't add a lot of value. And with Kubernetes, we have the opportunity, and this is a large part of how Kubernetes was designed and why it is the way it is, we have the opportunity to sort of abstract from the underlying cloud provider for stateless workloads on compute. And so what this lets us do is build container-based applications that can run without change on different cloud provider infrastructure. So the same applications can run on WCNP over Azure, WCNP over GCP, or WCNP over the Walmart private cloud. And we have a private cloud. Our private cloud is OpenStack based and it gives us some significant cost advantages as well as control advantages. So to your point, in terms of business motivation, there's a key cost driver here, which is that we can use our own private cloud when it's advantageous and then use the public cloud provider capabilities when we need to. A key place with this comes into play is with elasticity. So while the private cloud is much more cost effective for us to run and use, it isn't as elastic as what the cloud providers offer, right? We don't have essentially unlimited scale. We have large scale, but the public cloud providers are elastic in the extreme which is a very powerful capability. So what we're able to do is burst, and we use this term bursting workloads into the public cloud from the private cloud to take advantage of the elasticity they offer and then fall back into the private cloud when the traffic load diminishes to the point where we don't need that elastic capability, elastic capacity at low cost. And this is a very important paradigm that I think is going to be very commonplace ultimately as the industry evolves. Private cloud is easier to operate and less expensive, and yet the public cloud provider capabilities are difficult to match. >> And the triplet, the tri is your on-prem private cloud and the two public clouds that you mentioned, is that right? >> That is correct. And we actually have an architecture in which we operate all three of those cloud platforms in close proximity with one another in three different major regions in the US. So we have east, west, and central. And in each of those regions, we have all three cloud providers. And the way it's configured, those data centers are within 10 milliseconds of each other, meaning that it's of negligible cost to interact between them. And this allows us to be fairly agnostic to where a particular workload is running. >> Does a human make that decision, Jack or is there some intelligence in the system that determines that? >> That's a really great question, Dave. And it's a great question because we're at the cusp of that transition. So currently humans make that decision. Humans choose to deploy workloads into a particular region and a particular provider within that region. That said, we're actively developing patterns and practices that will allow us to automate the placement of the workloads for a variety of criteria. For example, if in a particular region, a particular provider is heavily overloaded and is unable to provide the level of service that's expected through our SLAs, we could choose to fail workloads over from that cloud provider to a different one within the same region. But that's manual today. We do that, but people do it. Okay, we'd like to get to where that happens automatically. In the same way, we'd like to be able to automate the failovers, both for high availability and sort of the heavier disaster recovery model between, within a region between providers and even within a provider between the availability zones that are there, but also between regions for the sort of heavier disaster recovery or maintenance driven realignment of workload placement. Today, that's all manual. So we have people moving workloads from region A to region B or data center A to data center B. It's clean because of the abstraction. The workloads don't have to know or care, but there are latency considerations that come into play, and the humans have to be cognizant of those. And automating that can help ensure that we get the best performance and the best reliability. >> But you're developing the dataset to actually, I would imagine, be able to make those decisions in an automated fashion over time anyway. Is that a fair assumption? >> It is, and that's what we're actively developing right now. So if you were to look at us today, we have these nice abstractions and APIs in place, but people run that machine, if you will, moving toward a world where that machine is fully automated. >> What exactly are you abstracting? Is it sort of the deployment model or, you know, are you able to abstract, I'm just making this up like Azure functions and GCP functions so that you can sort of run them, you know, with a consistent experience. What exactly are you abstracting and how difficult was it to achieve that objective technically? >> that's a good question. What we're abstracting is the Kubernetes node construct. That is to say a cluster of Kubernetes nodes which are typically VMs, although they can run bare metal in certain contexts, is something that typically to stand up requires knowledge of the underlying cloud provider. So for example, with GCP, you would use GKE to set up a Kubernetes cluster, and in Azure, you'd use AKS. We are actually abstracting that aspect of things so that the developers standing up applications don't have to know what the underlying cluster management provider is. They don't have to know if it's GCP, AKS or our own Walmart private cloud. Now, in terms of functions like Azure functions that you've mentioned there, we haven't done that yet. That's another piece that we have sort of on our radar screen that, we'd like to get to is serverless approach, and the Knative work from Google and the Azure functions, those are things that we see good opportunity to use for a whole variety of use cases. But right now we're not doing much with that. We're strictly container based right now, and we do have some VMs that are running in sort of more of a traditional model. So our stateful workloads are primarily VM based, but for serverless, that's an opportunity for us to take some of these stateless workloads and turn them into cloud functions. >> Well, and that's another cost lever that you can pull down the road that's going to drop right to the bottom line. Do you see a day or maybe you're doing it today, but I'd be surprised, but where you build applications that actually span multiple clouds or is there, in your view, always going to be a direct one-to-one mapping between where an application runs and the specific cloud platform? >> That's a really great question. Well, yes and no. So today, application development teams choose a cloud provider to deploy to and a location to deploy to, and they have to get involved in moving an application like we talked about today. That said, the bursting capability that I mentioned previously is something that is a step in the direction of automatic migration. That is to say we're migrating workload to different locations automatically. Currently, the prototypes we've been developing and that we think are going to eventually make their way into production are leveraging Istio to assess the load incoming on a particular cluster and start shedding that load into a different location. Right now, the configuration of that is still manual, but there's another opportunity for automation there. And I think a key piece of this is that down the road, well, that's a, sort of a small step in the direction of an application being multi provider. We expect to see really an abstraction of the fact that there is a triplet even. So the workloads are moving around according to whatever the control plane decides is necessary based on a whole variety of inputs. And at that point, you will have true multi-cloud applications, applications that are distributed across the different providers and in a way that application developers don't have to think about. >> So Walmart's been a leader, Jack, in using data for competitive advantages for decades. It's kind of been a poster child for that. You've got a mountain of IP in the form of data, tools, applications best practices that until the cloud came out was all On Prem. But I'm really interested in this idea of building a Walmart ecosystem, which obviously you have. Do you see a day or maybe you're even doing it today where you take what we call the Walmart SuperCloud, WCNP in your words, and point or turn that toward an external world or your ecosystem, you know, supporting those partners or customers that could drive new revenue streams, you know directly from the platform? >> Great question, Steve. So there's really two things to say here. The first is that with respect to data, our data workloads are primarily VM basis. I've mentioned before some VMware, some straight open stack. But the key here is that WCNP and Kubernetes are very powerful for stateless workloads, but for stateful workloads tend to be still climbing a bit of a growth curve in the industry. So our data workloads are not primarily based on WCNP. They're VM based. Now that said, there is opportunity to make some progress there, and we are looking at ways to move things into containers that are currently running in VMs which are stateful. The other question you asked is related to how we expose data to third parties and also functionality. Right now we do have in-house, for our own use, a very robust data architecture, and we have followed the sort of domain-oriented data architecture guidance from Martin Fowler. And we have data lakes in which we collect data from all the transactional systems and which we can then use and do use to build models which are then used in our applications. But right now we're not exposing the data directly to customers as a product. That's an interesting direction that's been talked about and may happen at some point, but right now that's internal. What we are exposing to customers is applications. So we're offering our global integrated fulfillment capabilities, our order picking and curbside pickup capabilities, and our cloud powered checkout capabilities to third parties. And this means we're standing up our own internal applications as externally facing SaaS applications which can serve our partners' customers. >> Yeah, of course, Martin Fowler really first introduced to the world Zhamak Dehghani's data mesh concept and this whole idea of data products and domain oriented thinking. Zhamak Dehghani, by the way, is a speaker at our event as well. Last question I had is edge, and how you think about the edge? You know, the stores are an edge. Are you putting resources there that sort of mirror this this triplet model? Or is it better to consolidate things in the cloud? I know there are trade-offs in terms of latency. How are you thinking about that? >> All really good questions. It's a challenging area as you can imagine because edges are subject to disconnection, right? Or reduced connection. So we do place the same architecture at the edge. So WCNP runs at the edge, and an application that's designed to run at WCNP can run at the edge. That said, there are a number of very specific considerations that come up when running at the edge, such as the possibility of disconnection or degraded connectivity. And so one of the challenges we have faced and have grappled with and done a good job of I think is dealing with the fact that applications go offline and come back online and have to reconnect and resynchronize, the sort of online offline capability is something that can be quite challenging. And we have a couple of application architectures that sort of form the two core sets of patterns that we use. One is an offline/online synchronization architecture where we discover that we've come back online, and we understand the differences between the online dataset and the offline dataset and how they have to be reconciled. The other is a message-based architecture. And here in our health and wellness domain, we've developed applications that are queue based. So they're essentially business processes that consist of multiple steps where each step has its own queue. And what that allows us to do is devote whatever bandwidth we do have to those pieces of the process that are most latency sensitive and allow the queue lengths to increase in parts of the process that are not latency sensitive, knowing that they will eventually catch up when the bandwidth is restored. And to put that in a little bit of context, we have fiber lengths to all of our locations, and we have I'll just use a round number, 10-ish thousand locations. It's larger than that, but that's the ballpark, and we have fiber to all of them, but when the fiber is disconnected, and it does get disconnected on a regular basis. In fact, I forget the exact number, but some several dozen locations get disconnected daily just by virtue of the fact that there's construction going on and things are happening in the real world. When the disconnection happens, we're able to fall back to 5G and to Starlink. Starlink is preferred. It's a higher bandwidth. 5G if that fails. But in each of those cases, the bandwidth drops significantly. And so the applications have to be intelligent about throttling back the traffic that isn't essential, so that it can push the essential traffic in those lower bandwidth scenarios. >> So much technology to support this amazing business which started in the early 1960s. Jack, unfortunately, we're out of time. I would love to have you back or some members of your team and drill into how you're using open source, but really thank you so much for explaining the approach that you've taken and participating in SuperCloud2. >> You're very welcome, Dave, and we're happy to come back and talk about other aspects of what we do. For example, we could talk more about the data lakes and the data mesh that we have in place. We could talk more about the directions we might go with serverless. So please look us up again. Happy to chat. >> I'm going to take you up on that, Jack. All right. This is Dave Vellante for John Furrier and the Cube community. Keep it right there for more action from SuperCloud2. (upbeat music)
SUMMARY :
and the Chief Architect for and appreciate the the Walmart Cloud Native Platform? and that is the DevOps Was the real impetus to tap into Sure, and in the course And the way it's configured, and the humans have to the dataset to actually, but people run that machine, if you will, Is it sort of the deployment so that the developers and the specific cloud platform? and that we think are going in the form of data, tools, applications a bit of a growth curve in the industry. and how you think about the edge? and allow the queue lengths to increase for explaining the and the data mesh that we have in place. and the Cube community.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Steve | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Jack Greenfield | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Jack | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
Martin Fowler | PERSON | 0.99+ |
US | LOCATION | 0.99+ |
Zhamak Dehghani | PERSON | 0.99+ |
Today | DATE | 0.99+ |
each | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
Starlink | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
two things | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
each step | QUANTITY | 0.99+ |
First | QUANTITY | 0.99+ |
early 1960s | DATE | 0.98+ |
one | QUANTITY | 0.98+ |
a day | QUANTITY | 0.98+ |
GCP | TITLE | 0.97+ |
Azure | TITLE | 0.96+ |
WCNP | TITLE | 0.96+ |
10 milliseconds | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
Kubernetes | TITLE | 0.94+ |
Cloud Spanner | TITLE | 0.94+ |
Linkerd | ORGANIZATION | 0.93+ |
Cube | ORGANIZATION | 0.93+ |
triplet | QUANTITY | 0.92+ |
three cloud providers | QUANTITY | 0.91+ |
two core sets | QUANTITY | 0.88+ |
John Furrier | PERSON | 0.86+ |
one more piece | QUANTITY | 0.86+ |
SuperCloud2 | ORGANIZATION | 0.86+ |
two public clouds | QUANTITY | 0.86+ |
thousand locations | QUANTITY | 0.83+ |
Vice President | PERSON | 0.8+ |
10-ish | QUANTITY | 0.79+ |
WCNP | ORGANIZATION | 0.75+ |
decades | QUANTITY | 0.75+ |
three different major regions | QUANTITY | 0.74+ |
Bob Muglia, George Gilbert & Tristan Handy | How Supercloud will Support a new Class of Data Apps
(upbeat music) >> Hello, everybody. This is Dave Vellante. Welcome back to Supercloud2, where we're exploring the intersection of data analytics and the future of cloud. In this segment, we're going to look at how the Supercloud will support a new class of applications, not just work that runs on multiple clouds, but rather a new breed of apps that can orchestrate things in the real world. Think Uber for many types of businesses. These applications, they're not about codifying forms or business processes. They're about orchestrating people, places, and things in a business ecosystem. And I'm pleased to welcome my colleague and friend, George Gilbert, former Gartner Analyst, Wiki Bond market analyst, former equities analyst as my co-host. And we're thrilled to have Tristan Handy, who's the founder and CEO of DBT Labs and Bob Muglia, who's the former President of Microsoft's Enterprise business and former CEO of Snowflake. Welcome all, gentlemen. Thank you for coming on the program. >> Good to be here. >> Thanks for having us. >> Hey, look, I'm going to start actually with the SuperCloud because both Tristan and Bob, you've read the definition. Thank you for doing that. And Bob, you have some really good input, some thoughts on maybe some of the drawbacks and how we can advance this. So what are your thoughts in reading that definition around SuperCloud? >> Well, I thought first of all that you did a very good job of laying out all of the characteristics of it and helping to define it overall. But I do think it can be tightened a bit, and I think it's helpful to do it in as short a way as possible. And so in the last day I've spent a little time thinking about how to take it and write a crisp definition. And here's my go at it. This is one day old, so gimme a break if it's going to change. And of course we have to follow the industry, and so that, and whatever the industry decides, but let's give this a try. So in the way I think you're defining it, what I would say is a SuperCloud is a platform that provides programmatically consistent services hosted on heterogeneous cloud providers. >> Boom. Nice. Okay, great. I'm going to go back and read the script on that one and tighten that up a bit. Thank you for spending the time thinking about that. Tristan, would you add anything to that or what are your thoughts on the whole SuperCloud concept? >> So as I read through this, I fully realize that we need a word for this thing because I have experienced the inability to talk about it as well. But for many of us who have been living in the Confluence, Snowflake, you know, this world of like new infrastructure, this seems fairly uncontroversial. Like I read through this, and I'm just like, yeah, this is like the world I've been living in for years now. And I noticed that you called out Snowflake for being an example of this, but I think that there are like many folks, myself included, for whom this world like fully exists today. >> Yeah, I think that's a fair, I dunno if it's criticism, but people observe, well, what's the big deal here? It's just kind of what we're living in today. It reminds me of, you know, Tim Burns Lee saying, well, this is what the internet was supposed to be. It was supposed to be Web 2.0, so maybe this is what multi-cloud was supposed to be. Let's turn our attention to apps. Bob first and then go to Tristan. Bob, what are data apps to you? When people talk about data products, is that what they mean? Are we talking about something more, different? What are data apps to you? >> Well, to understand data apps, it's useful to contrast them to something, and I just use the simple term people apps. I know that's a little bit awkward, but it's clear. And almost everything we work with, almost every application that we're familiar with, be it email or Salesforce or any consumer app, those are applications that are targeted at responding to people. You know, in contrast, a data application reacts to changes in data and uses some set of analytic services to autonomously take action. So where applications that we're familiar with respond to people, data apps respond to changes in data. And they both do something, but they do it for different reasons. >> Got it. You know, George, you and I were talking about, you know, it comes back to SuperCloud, broad definition, narrow definition. Tristan, how do you see it? Do you see it the same way? Do you have a different take on data apps? >> Oh, geez. This is like a conversation that I don't know has an end. It's like been, I write a substack, and there's like this little community of people who all write substack. We argue with each other about these kinds of things. Like, you know, as many different takes on this question as you can find, but the way that I think about it is that data products are atomic units of functionality that are fundamentally data driven in nature. So a data product can be as simple as an interactive dashboard that is like actually had design thinking put into it and serves a particular user group and has like actually gone through kind of a product development life cycle. And then a data app or data application is a kind of cohesive end-to-end experience that often encompasses like many different data products. So from my perspective there, this is very, very related to the way that these things are produced, the kinds of experiences that they're provided, that like data innovates every product that we've been building in, you know, software engineering for, you know, as long as there have been computers. >> You know, Jamak Dagani oftentimes uses the, you know, she doesn't name Spotify, but I think it's Spotify as that kind of example she uses. But I wonder if we can maybe try to take some examples. If you take, like George, if you take a CRM system today, you're inputting leads, you got opportunities, it's driven by humans, they're really inputting the data, and then you got this system that kind of orchestrates the business process, like runs a forecast. But in this data driven future, are we talking about the app itself pulling data in and automatically looking at data from the transaction systems, the call center, the supply chain and then actually building a plan? George, is that how you see it? >> I go back to the example of Uber, may not be the most sophisticated data app that we build now, but it was like one of the first where you do have users interacting with their devices as riders trying to call a car or driver. But the app then looks at the location of all the drivers in proximity, and it matches a driver to a rider. It calculates an ETA to the rider. It calculates an ETA then to the destination, and it calculates a price. Those are all activities that are done sort of autonomously that don't require a human to type something into a form. The application is using changes in data to calculate an analytic product and then to operationalize that, to assign the driver to, you know, calculate a price. Those are, that's an example of what I would think of as a data app. And my question then I guess for Tristan is if we don't have all the pieces in place for sort of mainstream companies to build those sorts of apps easily yet, like how would we get started? What's the role of a semantic layer in making that easier for mainstream companies to build? And how do we get started, you know, say with metrics? How does that, how does that take us down that path? >> So what we've seen in the past, I dunno, decade or so, is that one of the most successful business models in infrastructure is taking hard things and rolling 'em up behind APIs. You take messaging, you take payments, and you all of a sudden increase the capability of kind of your median application developer. And you say, you know, previously you were spending all your time being focused on how do you accept credit cards, how do you send SMS payments, and now you can focus on your business logic, and just create the thing. One of, interestingly, one of the things that we still don't know how to API-ify is concepts that live inside of your data warehouse, inside of your data lake. These are core concepts that, you know, you would imagine that the business would be able to create applications around very easily, but in fact that's not the case. It's actually quite challenging to, and involves a lot of data engineering pipeline and all this work to make these available. And so if you really want to make it very easy to create some of these data experiences for users, you need to have an ability to describe these metrics and then to turn them into APIs to make them accessible to application developers who have literally no idea how they're calculated behind the scenes, and they don't need to. >> So how rich can that API layer grow if you start with metric definitions that you've defined? And DBT has, you know, the metric, the dimensions, the time grain, things like that, that's a well scoped sort of API that people can work within. How much can you extend that to say non-calculated business rules or governance information like data reliability rules, things like that, or even, you know, features for an AIML feature store. In other words, it starts, you started pragmatically, but how far can you grow? >> Bob is waiting with bated breath to answer this question. I'm, just really quickly, I think that we as a company and DBT as a product tend to be very pragmatic. We try to release the simplest possible version of a thing, get it out there, and see if people use it. But the idea that, the concept of a metric is really just a first landing pad. The really, there is a physical manifestation of the data and then there's a logical manifestation of the data. And what we're trying to do here is make it very easy to access the logical manifestation of the data, and metric is a way to look at that. Maybe an entity, a customer, a user is another way to look at that. And I'm sure that there will be more kind of logical structures as well. >> So, Bob, chime in on this. You know, what's your thoughts on the right architecture behind this, and how do we get there? >> Yeah, well first of all, I think one of the ways we get there is by what companies like DBT Labs and Tristan is doing, which is incrementally taking and building on the modern data stack and extending that to add a semantic layer that describes the data. Now the way I tend to think about this is a fairly major shift in the way we think about writing applications, which is today a code first approach to moving to a world that is model driven. And I think that's what the big change will be is that where today we think about data, we think about writing code, and we use that to produce APIs as Tristan said, which encapsulates those things together in some form of services that are useful for organizations. And that idea of that encapsulation is never going to go away. It's very, that concept of an API is incredibly useful and will exist well into the future. But what I think will happen is that in the next 10 years, we're going to move to a world where organizations are defining models first of their data, but then ultimately of their business process, their entire business process. Now the concept of a model driven world is a very old concept. I mean, I first started thinking about this and playing around with some early model driven tools, probably before Tristan was born in the early 1980s. And those tools didn't work because the semantics associated with executing the model were too complex to be written in anything other than a procedural language. We're now reaching a time where that is changing, and you see it everywhere. You see it first of all in the world of machine learning and machine learning models, which are taking over more and more of what applications are doing. And I think that's an incredibly important step. And learned models are an important part of what people will do. But if you look at the world today, I will claim that we've always been modeling. Modeling has existed in computers since there have been integrated circuits and any form of computers. But what we do is what I would call implicit modeling, which means that it's the model is written on a whiteboard. It's in a bunch of Slack messages. It's on a set of napkins in conversations that happen and during Zoom. That's where the model gets defined today. It's implicit. There is one in the system. It is hard coded inside application logic that exists across many applications with humans being the glue that connects those models together. And really there is no central place you can go to understand the full attributes of the business, all of the business rules, all of the business logic, the business data. That's going to change in the next 10 years. And we'll start to have a world where we can define models about what we're doing. Now in the short run, the most important models to build are data models and to describe all of the attributes of the data and their relationships. And that's work that DBT Labs is doing. A number of other companies are doing that. We're taking steps along that way with catalogs. People are trying to build more complete ontologies associated with that. The underlying infrastructure is still super, super nascent. But what I think we'll see is this infrastructure that exists today that's building learned models in the form of machine learning programs. You know, some of these incredible machine learning programs in foundation models like GPT and DALL-E and all of the things that are happening in these global scale models, but also all of that needs to get applied to the domains that are appropriate for a business. And I think we'll see the infrastructure developing for that, that can take this concept of learned models and put it together with more explicitly defined models. And this is where the concept of knowledge graphs come in and then the technology that underlies that to actually implement and execute that, which I believe are relational knowledge graphs. >> Oh, oh wow. There's a lot to unpack there. So let me ask the Colombo question, Tristan, we've been making fun of your youth. We're just, we're just jealous. Colombo, I'll explain it offline maybe. >> I watch Colombo. >> Okay. All right, good. So but today if you think about the application stack and the data stack, which is largely an analytics pipeline. They're separate. Do they, those worlds, do they have to come together in order to achieve Bob's vision? When I talk to practitioners about that, they're like, well, I don't want to complexify the application stack cause the data stack today is so, you know, hard to manage. But but do those worlds have to come together? And you know, through that model, I guess abstraction or translation that Bob was just describing, how do you guys think about that? Who wants to take that? >> I think it's inevitable that data and AI are going to become closer together? I think that the infrastructure there has been moving in that direction for a long time. Whether you want to use the Lakehouse portmanteau or not. There's also, there's a next generation of data tech that is still in the like early stage of being developed. There's a company that I love that is essentially Cross Cloud Lambda, and it's just a wonderful abstraction for computing. So I think that, you know, people have been predicting that these worlds are going to come together for awhile. A16Z wrote a great post on this back in I think 2020, predicting this, and I've been predicting this since since 2020. But what's not clear is the timeline, but I think that this is still just as inevitable as it's been. >> Who's that that does Cross Cloud? >> Let me follow up on. >> Who's that, Tristan, that does Cross Cloud Lambda? Can you name names? >> Oh, they're called Modal Labs. >> Modal Labs, yeah, of course. All right, go ahead, George. >> Let me ask about this vision of trying to put the semantics or the code that represents the business with the data. It gets us to a world that's sort of more data centric, where data's not locked inside or behind the APIs of different applications so that we don't have silos. But at the same time, Bob, I've heard you talk about building the semantics gradually on top of, into a knowledge graph that maybe grows out of a data catalog. And the vision of getting to that point, essentially the enterprise's metadata and then the semantics you're going to add onto it are really stored in something that's separate from the underlying operational and analytic data. So at the same time then why couldn't we gradually build semantics beyond the metric definitions that DBT has today? In other words, you build more and more of the semantics in some layer that DBT defines and that sits above the data management layer, but any requests for data have to go through the DBT layer. Is that a workable alternative? Or where, what type of limitations would you face? >> Well, I think that it is the way the world will evolve is to start with the modern data stack and, you know, which is operational applications going through a data pipeline into some form of data lake, data warehouse, the Lakehouse, whatever you want to call it. And then, you know, this wide variety of analytics services that are built together. To the point that Tristan made about machine learning and data coming together, you see that in every major data cloud provider. Snowflake certainly now supports Python and Java. Databricks is of course building their data warehouse. Certainly Google, Microsoft and Amazon are doing very, very similar things in terms of building complete solutions that bring together an analytics stack that typically supports languages like Python together with the data stack and the data warehouse. I mean, all of those things are going to evolve, and they're not going to go away because that infrastructure is relatively new. It's just being deployed by companies, and it solves the problem of working with petabytes of data if you need to work with petabytes of data, and nothing will do that for a long time. What's missing is a layer that understands and can model the semantics of all of this. And if you need to, if you want to model all, if you want to talk about all the semantics of even data, you need to think about all of the relationships. You need to think about how these things connect together. And unfortunately, there really is no platform today. None of our existing platforms are ultimately sufficient for this. It was interesting, I was just talking to a customer yesterday, you know, a large financial organization that is building out these semantic layers. They're further along than many companies are. And you know, I asked what they're building it on, and you know, it's not surprising they're using a, they're using combinations of some form of search together with, you know, textual based search together with a document oriented database. In this case it was Cosmos. And that really is kind of the state of the art right now. And yet those products were not built for this. They don't really, they can't manage the complicated relationships that are required. They can't issue the queries that are required. And so a new generation of database needs to be developed. And fortunately, you know, that is happening. The world is developing a new set of relational algorithms that will be able to work with hundreds of different relations. If you look at a SQL database like Snowflake or a big query, you know, you get tens of different joins coming together, and that query is going to take a really long time. Well, fortunately, technology is evolving, and it's possible with new join algorithms, worst case, optimal join algorithms they're called, where you can join hundreds of different relations together and run semantic queries that you simply couldn't run. Now that technology is nascent, but it's really important, and I think that will be a requirement to have this semantically reach its full potential. In the meantime, Tristan can do a lot of great things by building up on what he's got today and solve some problems that are very real. But in the long run I think we'll see a new set of databases to support these models. >> So Tristan, you got to respond to that, right? You got to, so take the example of Snowflake. We know it doesn't deal well with complex joins, but they're, they've got big aspirations. They're building an ecosystem to really solve some of these problems. Tristan, you guys are part of that ecosystem, and others, but please, your thoughts on what Bob just shared. >> Bob, I'm curious if, I would have no idea what you were talking about except that you introduced me to somebody who gave me a demo of a thing and do you not want to go there right now? >> No, I can talk about it. I mean, we can talk about it. Look, the company I've been working with is Relational AI, and they're doing this work to actually first of all work across the industry with academics and research, you know, across many, many different, over 20 different research institutions across the world to develop this new set of algorithms. They're all fully published, just like SQL, the underlying algorithms that are used by SQL databases are. If you look today, every single SQL database uses a similar set of relational algorithms underneath that. And those algorithms actually go back to system R and what IBM developed in the 1970s. We're just, there's an opportunity for us to build something new that allows you to take, for example, instead of taking data and grouping it together in tables, treat all data as individual relations, you know, a key and a set of values and then be able to perform purely relational operations on it. If you go back to what, to Codd, and what he wrote, he defined two things. He defined a relational calculus and relational algebra. And essentially SQL is a query language that is translated by the query processor into relational algebra. But however, the calculus of SQL is not even close to the full semantics of the relational mathematics. And it's possible to have systems that can do everything and that can store all of the attributes of the data model or ultimately the business model in a form that is much more natural to work with. >> So here's like my short answer to this. I think that we're dealing in different time scales. I think that there is actually a tremendous amount of work to do in the semantic layer using the kind of technology that we have on the ground today. And I think that there's, I don't know, let's say five years of like really solid work that there is to do for the entire industry, if not more. But the wonderful thing about DBT is that it's independent of what the compute substrate is beneath it. And so if we develop new platforms, new capabilities to describe semantic models in more fine grain detail, more procedural, then we're going to support that too. And so I'm excited about all of it. >> Yeah, so interpreting that short answer, you're basically saying, cause Bob was just kind of pointing to you as incremental, but you're saying, yeah, okay, we're applying it for incremental use cases today, but we can accommodate a much broader set of examples in the future. Is that correct, Tristan? >> I think you're using the word incremental as if it's not good, but I think that incremental is great. We have always been about applying incremental improvement on top of what exists today, but allowing practitioners to like use different workflows to actually make use of that technology. So yeah, yeah, we are a very incremental company. We're going to continue being that way. >> Well, I think Bob was using incremental as a pejorative. I mean, I, but to your point, a lot. >> No, I don't think so. I want to stop that. No, I don't think it's pejorative at all. I think incremental, incremental is usually the most successful path. >> Yes, of course. >> In my experience. >> We agree, we agree on that. >> Having tried many, many moonshot things in my Microsoft days, I can tell you that being incremental is a good thing. And I'm a very big believer that that's the way the world's going to go. I just think that there is a need for us to build something new and that ultimately that will be the solution. Now you can argue whether it's two years, three years, five years, or 10 years, but I'd be shocked if it didn't happen in 10 years. >> Yeah, so we all agree that incremental is less disruptive. Boom, but Tristan, you're, I think I'm inferring that you believe you have the architecture to accommodate Bob's vision, and then Bob, and I'm inferring from Bob's comments that maybe you don't think that's the case, but please. >> No, no, no. I think that, so Bob, let me put words into your mouth and you tell me if you disagree, DBT is completely useless in a world where a large scale cloud data warehouse doesn't exist. We were not able to bring the power of Python to our users until these platforms started supporting Python. Like DBT is a layer on top of large scale computing platforms. And to the extent that those platforms extend their functionality to bring more capabilities, we will also service those capabilities. >> Let me try and bridge the two. >> Yeah, yeah, so Bob, Bob, Bob, do you concur with what Tristan just said? >> Absolutely, I mean there's nothing to argue with in what Tristan just said. >> I wanted. >> And it's what he's doing. It'll continue to, I believe he'll continue to do it, and I think it's a very good thing for the industry. You know, I'm just simply saying that on top of that, I would like to provide Tristan and all of those who are following similar paths to him with a new type of database that can actually solve these problems in a much more architected way. And when I talk about Cosmos with something like Mongo or Cosmos together with Elastic, you're using Elastic as the join engine, okay. That's the purpose of it. It becomes a poor man's join engine. And I kind of go, I know there's a better answer than that. I know there is, but that's kind of where we are state of the art right now. >> George, we got to wrap it. So give us the last word here. Go ahead, George. >> Okay, I just, I think there's a way to tie together what Tristan and Bob are both talking about, and I want them to validate it, which is for five years we're going to be adding or some number of years more and more semantics to the operational and analytic data that we have, starting with metric definitions. My question is for Bob, as DBT accumulates more and more of those semantics for different enterprises, can that layer not run on top of a relational knowledge graph? And what would we lose by not having, by having the knowledge graph store sort of the joins, all the complex relationships among the data, but having the semantics in the DBT layer? >> Well, I think this, okay, I think first of all that DBT will be an environment where many of these semantics are defined. The question we're asking is how are they stored and how are they processed? And what I predict will happen is that over time, as companies like DBT begin to build more and more richness into their semantic layer, they will begin to experience challenges that customers want to run queries, they want to ask questions, they want to use this for things where the underlying infrastructure becomes an obstacle. I mean, this has happened in always in the history, right? I mean, you see major advances in computer science when the data model changes. And I think we're on the verge of a very significant change in the way data is stored and structured, or at least metadata is stored and structured. Again, I'm not saying that anytime in the next 10 years, SQL is going to go away. In fact, more SQL will be written in the future than has been written in the past. And those platforms will mature to become the engines, the slicer dicers of data. I mean that's what they are today. They're incredibly powerful at working with large amounts of data, and that infrastructure is maturing very rapidly. What is not maturing is the infrastructure to handle all of the metadata and the semantics that that requires. And that's where I say knowledge graphs are what I believe will be the solution to that. >> But Tristan, bring us home here. It sounds like, let me put pause at this, is that whatever happens in the future, we're going to leverage the vast system that has become cloud that we're talking about a supercloud, sort of where data lives irrespective of physical location. We're going to have to tap that data. It's not necessarily going to be in one place, but give us your final thoughts, please. >> 100% agree. I think that the data is going to live everywhere. It is the responsibility for both the metadata systems and the data processing engines themselves to make sure that we can join data across cloud providers, that we can join data across different physical regions and that we as practitioners are going to kind of start forgetting about details like that. And we're going to start thinking more about how we want to arrange our teams, how does the tooling that we use support our team structures? And that's when data mesh I think really starts to get very, very critical as a concept. >> Guys, great conversation. It was really awesome to have you. I can't thank you enough for spending time with us. Really appreciate it. >> Thanks a lot. >> All right. This is Dave Vellante for George Gilbert, John Furrier, and the entire Cube community. Keep it right there for more content. You're watching SuperCloud2. (upbeat music)
SUMMARY :
and the future of cloud. And Bob, you have some really and I think it's helpful to do it I'm going to go back and And I noticed that you is that what they mean? that we're familiar with, you know, it comes back to SuperCloud, is that data products are George, is that how you see it? that don't require a human to is that one of the most And DBT has, you know, the And I'm sure that there will be more on the right architecture is that in the next 10 years, So let me ask the Colombo and the data stack, which is that is still in the like Modal Labs, yeah, of course. and that sits above the and that query is going to So Tristan, you got to and that can store all of the that there is to do for the pointing to you as incremental, but allowing practitioners to I mean, I, but to your point, a lot. the most successful path. that that's the way the that you believe you have the architecture and you tell me if you disagree, there's nothing to argue with And I kind of go, I know there's George, we got to wrap it. and more of those semantics and the semantics that that requires. is that whatever happens in the future, and that we as practitioners I can't thank you enough John Furrier, and the
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Tristan | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
John | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Steve Mullaney | PERSON | 0.99+ |
Katie | PERSON | 0.99+ |
David Floyer | PERSON | 0.99+ |
Charles | PERSON | 0.99+ |
Mike Dooley | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Tristan Handy | PERSON | 0.99+ |
Bob | PERSON | 0.99+ |
Maribel Lopez | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Mike Wolf | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Merim | PERSON | 0.99+ |
Adrian Cockcroft | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Brian | PERSON | 0.99+ |
Brian Rossi | PERSON | 0.99+ |
Jeff Frick | PERSON | 0.99+ |
Chris Wegmann | PERSON | 0.99+ |
Whole Foods | ORGANIZATION | 0.99+ |
Eric | PERSON | 0.99+ |
Chris Hoff | PERSON | 0.99+ |
Jamak Dagani | PERSON | 0.99+ |
Jerry Chen | PERSON | 0.99+ |
Caterpillar | ORGANIZATION | 0.99+ |
John Walls | PERSON | 0.99+ |
Marianna Tessel | PERSON | 0.99+ |
Josh | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Jerome | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Lori MacVittie | PERSON | 0.99+ |
2007 | DATE | 0.99+ |
Seattle | LOCATION | 0.99+ |
10 | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
Ali Ghodsi | PERSON | 0.99+ |
Peter McKee | PERSON | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
Eric Herzog | PERSON | 0.99+ |
India | LOCATION | 0.99+ |
Mike | PERSON | 0.99+ |
Walmart | ORGANIZATION | 0.99+ |
five years | QUANTITY | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Kit Colbert | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Tanuja Randery | PERSON | 0.99+ |
Video Exclusive: Oracle Lures MongoDB Devs With New API for ADB
(upbeat music) >> Oracle continues to pursue a multi-mode converged database strategy. The premise of this all in one approach is to make life easier for practitioners and developers. And the most recent example is the Oracle database API for MongoDB, which was announced today. Now, Oracle, they're not the first to come out with a MongoDB compatible API, but Oracle hopes to use its autonomous database as a differentiator and further build a moat around OCI, Oracle Cloud Infrastructure. And with us to talk about Oracle's MongoDB compatible API is Gerald Venzl, who's a distinguished Product Manager at Oracle. Gerald was a guest along with Maria Colgan on the CUBE a while back, and we talked about Oracle's converge database and the kind of Swiss army knife strategy, I called it, of databases. This is dramatically different. It's an approach that we see at the opposite end of the the spectrum, for instance, from AWS, who, for example, goes after the world of developers with a different database for every use case. So, kind of picking up from there, Gerald, I wonder if you could talk about how this new MongoDB API adds to your converged model and the whole strategy there. Where does it fit? >> Yeah, thank you very much, Dave and, by the way, thanks for having me on the CUBE again. A pleasure to be here. So, essentially the MongoDB API to build the compatibility that we used with this API is a continuation of the converge database story, as you said before. Which is essentially bringing the many features of the many single purpose databases that people often like and use, together into one technology so that everybody can benefit from it. So as such, this is just a continuation that we have from so many other APIs or standards that we support. Since a long time, we already, of course to SQL because we are relational database from the get go. Also other standard like GraphQL, Sparkle, et cetera that we have. And the MongoDB API, is now essentially just the next step forward to give the developers this API that they've gotten to love and use. >> I wonder if you could talk about from the developer angle, what do they get out of it? Obviously you're appealing to the Mongo developers out there, but you've got this Mongo compatible API you're pouting the autonomous database on OCI. Why aren't they just going to use MongoDB Atlas on whatever cloud, Azure or AWS or Google Cloud platform? >> That's a very good question. We believe that the majority of developers want to just worry about their application, writing the application, and not so much about the database backend that they're using. And especially in cloud with cloud services, the reason why developers choose these services is so that they don't have to manage them. Now, autonomous database brings many topnotch advanced capabilities to database cloud services. We firmly believe that autonomous database is essentially the next generation of cloud services with all the self-driving features built in, and MongoDB developers writing applications against the MongoDB API, should not have to hold out on these capabilities either. It's like no developer likes to tune the database. No developer likes to take a downtime when they have to rescale their database to accommodate a bigger workload. And this is really where we see the benefit here, so for the developer, ideally nothing will change. You have MongoDB compatible API so they can keep on using their tools. They can build the applications the way that they do, but the benefit from the best cloud database service out there not having to worry about any of these package things anymore, that even MongoDB Atlas has a lot of shortcomings still today, as we find. >> Of cos, this is always a moving target The technology business, that's why we love it. So everybody's moving fast and investing and shaking and jiving. But, I want to ask you about, well, by the way, that's so you're hiding the underlying complexity, That's really the big takeaway there. So that's you huge for developers. But take, I was talking before about, the Amazon's approach, right tool for the right job. You got document DB, you got Microsoft with Cosmos, they compete with Mongo and they've been doing so for some time. How does Oracle's API for Mongo different from those offerings and how you going to attract their users to your JSON offering. >> So, you know, for first of all we have to kind of separate slightly document DB and AWS and Cosmos DB in Azure, they have slightly different approaches there. Document DB essentially is, a document store owned by and built by AWS, nothing different to Mongo DB, it's a head to head comparison. It's like use my document store versus the other document store. So you don't get any of the benefits of a converge database. If you ever want to do a different data model, run analytics over, etc. You still have to use the many other services that AWS provides you to. You cannot all do it into one database. Now Cosmos DB it's more in interesting because they claim to be a multi-model database. And I say claim because what we understand as multi-model database is different to what they understand as multimodel database. And also one of the reasons why we start differentiating with converge database. So what we mean is you should be able to regardless what data format you want to store in the database leverage all the functionality of the database over that data format, with no trade offs. Cosmos DB when you look at it, it essentially gives you mode of operation. When you connect as the application or the user, you have to decide at connection time, how you want, how this database should be treated. Should it be a document store? Should it be a graph store? Should it be a relational store? Once you make that choice, you are locked into that. As long as you establish that connection. So it's like, if you say, I want a document store, all you get is a document store. There's no way for you to crossly analyze with the relational data sitting in the same service. There's no for you to break these boundaries. If you ever want to add some graph data and graph analytics, you essentially have to disconnect and now treat it as a graph store. So you get multiple data models in it, but really you still get, one trick pony the moment you connect to it that you have to choose to. And that is where we see a huge differentiation again with our converge database, because we essentially say, look, one database cloud service on Oracle cloud, where it allows you to do anything, if you wish to do so. You can start as a document store if you wish to do so. If you want to write some SQL queries on top, you can do so. If you want to add some graph data, you can do so. But there's no way for you to have to rewrite your application, use different libraries and frameworks now to connect et cetera, et cetera. >> Got it. Thank you for that. Do you have any data when you talk to customers? Like I'm interested in the diversity of deployments, like for instance, how many customers are using more than one data model? Do for instance, do JSON users need support for other data types or are they happy to stay kind of in their own little sandbox? Do you have any data on that? >> So what we see from the majority of our customers, there is no such thing as one data model fits everything. So, and it's like, there again we have to differentiate the developer that builds a certain microservice, that makes happy to stay in the JSON world or relational world, or the company that's trying to derive value from the data. So it's like the relational model has not gone away since 40 years of it existence. It's still kicking strong. It's still really good at what it does. The JSON data model is really good in what it does. The graph model is really good at what it does. But all these models have been built for different purposes. Try to do graph analytics on relational or JSON data. It's like, it's really tricky, but that's why you use a graph model to begin with. Try to shield yourself from the organization of the data, how it's structured, that's really easy in the relational world, not so much when you get into a document store world. And so what we see about our customers is like as they accumulate more data, is they have many different applications to run their enterprises. The question always comes back, as we have predicted since about six, seven years now, where they say, hey, we have all this different data and different data formats. We want to bring it all together, analyze it together, get value out of the data together. We have seen a whole trend of big data emerge and disappear to answer the question and didn't quite do the trick. And we are basically now back to where we were in the early 2000's when XML databases have faded away, because everybody just allowed you to store XML in the database. >> Got it. So let's make this real for people. So maybe you could give us some examples. You got this new API from Mongo, you have your multi model database. How, take a, paint a picture of how customers are going to benefit in real world use cases. How does it kind of change the customer's world before and after if you will? >> Yeah, absolutely. So, you know the API essentially we are going to use it to accept before, you know, make the lives of the developers easier, but also of course to assist our customers with migrations from Mongo DB over to Oracle Autonomous Database. One customer that we have, for example, that would've benefited of the API several a couple of years ago, two, three years ago, it's one of the largest logistics company on the planet. They track every package that is being sent in JSON documents. So every track package is entries resembled in a JSON document, and they very early on came in with the next question of like, hey, we track all these packages and document in JSON documents. It will be really nice to know actually which packages are stuck, or anywhere where we have to intervene. It's like, can we do this? Can we analyze just how many packages get stuck, didn't get delivered on, the end of a day or whatever. And they found this struggle with this question a lot, they found this was really tricky to do back then, in that case in MongoDB. So they actually approached Oracle, they came over, they migrated over and they rewrote their applications to accommodate that. And there are happy JSON users in Oracle database, but if we were having this API already for them then they wouldn't have had to rewrite their applications or would we often see like worry about the rewriting the application later on. Usually migration use cases, we want to get kind of the migration done, get the data over be running, and then worry about everything else. So this would be one where they would've greatly benefited to shorten this migration time window. If we had already demo the Mongo API back then or this compatibility layer. >> That's a good use case. I mean, it's, one of the most prominent and painful, so anything you could do to help that is key. I remember like the early days of big data, NoSQL, of course was the big thing. There was a lot of confusion. No, people thought was none or not only SQL, which is kind of the more widely accepted interpretation today. But really, it's talking about data that's stored in a non-relational format. So, some people, again they thought that SQL was going to fade away, some people probably still believe that. And, we saw the rise of NoSQL and document databases, but if I understand it correctly, a premise for your Mongo DB API is you really see SQL as a main contributor over Mongo DB's document collections for analytics for example. Can you make, add some color here? Are you seeing, what are you seeing in terms of resurgence of SQL or the momentum in SQL? Has it ever really waned? What's your take? >> Yeah, no, it's a very good point. So I think there as well, we see to some extent history repeating itself from, this all has been tried beforehand with object databases, XML database, et cetera. But if we stay with the NoSQL databases, I think it speaks at length that every NoSQL database that as you write for the sensor you started with NoSQL, and then while actually we always meant, not only SQL, everybody has introduced a SQL like engine or interface. The last two actually join this family is MongoDB. Now they have just recently introduced a SQL compatibility for the aggregation pipelines, something where you can put in a SQL statement and that essentially will then work with aggregation pipeline. So they all acknowledge that SQL is powerful, for us this was always clear. SQL is a declarative language. Some argue it's the only true 4GL language out there. You don't have to code how to get the data, but you just ask the question and the rest is done for you. And, we think that as we, basically, has SQL ever diminished as you said before, if you look out there? SQL has always been a demand. Look at the various developer surveys, etc. The various top skills that are asked for SQL has never gone away. Everybody loves and likes and you wants to use SQL. And so, yeah, we don't think this has ever been, going away. It has maybe just been, put in the shadow by some hypes. But again, we had the same discussion in the 2000's with XML databases, with the same discussions in the 90's with object databases. And we have just frankly, all forgotten about it. >> I love when you guys come on and and let me do my thing and I can pretty much ask any question I want, because, I got to say, when Oracle starts talking about another company I know that company's doing well. So I like, I see Mongo in the marketplace and I love that you guys are calling it out and making some moves there. So here's the thing, you guys have a large install base and that can be an advantage, but it can also be a weight in your shoulder. These specialized cloud databases they don't have that legacy. So they can just kind of move freely about, less friction. Now, all the cloud database services they're going to have more and more automation. I mean, I think that's pretty clear and inevitable. And most if not all of the database vendors they're going to provide support for these kind of converged data models. However they choose to do that. They might do it through the ecosystem, like what Snowflake's trying to do, or bring it in the house themselves, like a watch maker that brings an in-house movement, if you will. But it's like death and taxes, you can't avoid it. It's got to happen. That's what customers want. So with all that being said, how do you see the capabilities that you have today with automation and converge capabilities, How do you see that, that playing out? What's, do you think it gives you enough of an advantage? And obviously it's an advantage, but is it enough of an advantage over the specialized cloud database vendors, where there's clearly a lot of momentum today? >> I mean, honestly yes, absolutely. I mean, we are with some of these databases 20 years ahead. And I give you concrete examples. It's like Oracle had transaction support asset transactions since forever. NoSQL players all said, oh, we don't need assets transactions, base transactions is fine. Yada, yada, yada. Mongo DB started introducing some transaction support. It comes with some limits, cannot be longer than 60 seconds, cannot touch more than a thousand documents as well, et cetera. They still will have to do some catching up there. I mean, it took us a while to get there, let's be honest. Glad We have been around for a long time. Same thing, now that happened with version five, is like we started some simple version of multi version concurrency control that comes along with asset transactions. The interesting part here is like, we've introduced this also an Oracle five, which was somewhere in the 80's before I even started using Oracle Database. So there's a lot of catching up to do. And then you look at the cloud services as well, there's actually certain, a lot of things that we kind of gotten take, we've kind of, we Oracle people have taken for granted and we kind of keep forgetting. For example, our elastic scale, you want to add one CPU, you add one CPU. Should you take downtime for that? Absolutely not. It's like, this is ridiculous. Why would you, you cannot take it downtime in a 24/7 backend system that runs the world. Take any of our customers. If you look at most of these cloud services or you want to reshape, you want to scale your cloud service, that's fine. It's just the VM under the covers, we just shut everything down, give you a VM with more CPU, and you boot it up again, downtown right there. So it's like, there's a lot of these things where we go like, well, we solved this frankly decades ago, that these cloud vendors will run into. And just to add one more point here, so it's like one thing that we see with all these migrations happening is exactly in that field. It's like people essentially started building on whether it's Mongo DB or other of these NoSQL databases or cloud databases. And eventually as these systems grow, as they ask more difficult questions, their use cases expand, they find shortcomings. Whether it's the scalability, whether it's the security aspects, the functionalities that we have, and this is essentially what drives them back to Oracle. And this is why we see essentially this popularity now of pendulum swimming towards our direction again, where people actually happily come over back and they come over to us, to get their workloads enterprise grade if you like. >> Well, It's true. I mean, I just reported on this recently, the momentum that you guys have in cloud because it is, 'cause you got the best mission critical database. You're all about maps. I got to tell you a quick story. I was at a vertical conference one time, I was on stage with Kurt Monash. I don't know if you know Kurt, but he knows this space really well. He's probably forgot and more about database than I'll ever know. But, and I was kind of busting his chops. He was talking about asset transactions. I'm like, well with NoSQL, who needs asset transactions, just to poke him. And he was like, "Are you out of your mind?" And, and he said, look it's everybody is going to head in this direction. It turned out, it's true. So I got to give him props for that. And so, my last question, if you had a message for, let's say there's a skeptical developer out there that's using Mongo DB and Atlas, what would you say to them? >> I would say go try it for yourself. If you don't believe us, we have an always free cloud tier out there. You just go to oracle.com/cloud/free. You sign up for an always free tier, spin up an autonomous database, go try it for yourself. See what's actually possible today. Don't just follow your trends on Hackernews and use a set study here or there. Go try it for yourself and see what's capable of >> All right, Gerald. Hey, thanks for coming into my firing line today. I really appreciate your time. >> Thank you for having me again. >> Good luck with the announcement. You're very welcome, and thank you for watching this CUBE conversation. This is Dave Vellante, We'll see you next time. (gentle music)
SUMMARY :
the first to come out the next step forward to I wonder if you could talk is so that they don't have to manage them. and how you going to attract their users the moment you connect to it you talk to customers? So it's like the relational So maybe you could give us some examples. to accept before, you know, make API is you really see SQL that as you write for the and I love that you And I give you concrete examples. the momentum that you guys have in cloud If you don't believe us, I really appreciate your time. and thank you for watching
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Maria Colgan | PERSON | 0.99+ |
Gerald Venzl | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Gerald | PERSON | 0.99+ |
Kurt | PERSON | 0.99+ |
NoSQL | TITLE | 0.99+ |
MongoDB | TITLE | 0.99+ |
JSON | TITLE | 0.99+ |
SQL | TITLE | 0.99+ |
MongoDB Atlas | TITLE | 0.99+ |
40 years | QUANTITY | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
One customer | QUANTITY | 0.99+ |
oracle.com/cloud/free | OTHER | 0.98+ |
first | QUANTITY | 0.98+ |
Kurt Monash | PERSON | 0.98+ |
more than a thousand documents | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
one time | QUANTITY | 0.97+ |
two | DATE | 0.97+ |
one database | QUANTITY | 0.97+ |
more than one data model | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.97+ |
90's | DATE | 0.97+ |
one technology | QUANTITY | 0.96+ |
20 years | QUANTITY | 0.96+ |
80's | DATE | 0.96+ |
one more point | QUANTITY | 0.95+ |
decades ago | DATE | 0.95+ |
one data model | QUANTITY | 0.95+ |
Azure | TITLE | 0.94+ |
three years ago | DATE | 0.93+ |
seven years | QUANTITY | 0.93+ |
version five | OTHER | 0.92+ |
one approach | QUANTITY | 0.92+ |
Survey Data Shows no Slowdown in AWS & Cloud Momentum
from the cube studios in palo alto in boston bringing you data-driven insights from the cube and etr this is breaking analysis with dave vellante despite all the chatter about cloud repatriation and the exorbitant cost of cloud computing customer spending momentum continues to accelerate in the post-isolation economy if the pandemic was good for the cloud it seems that the benefits of cloud migration remain lasting in the late stages of covid and beyond and we believe this stickiness is going to continue for quite some time we expect i asked revenue for the big four hyperscalers to surpass 115 billion dollars in 2021 moreover the strength of aws specifically as well as microsoft azure remain notable such large organizations showing elevated spending momentum as shown in the etr survey results is perhaps unprecedented in the technology sector hello everyone and welcome to this week's wikibon cube insights powered by etr in this breaking analysis we'll share some some fresh july survey data that indicates accelerating momentum for the largest cloud computing firms importantly not only is the momentum broad-based but it's also notable in key strategic sectors namely ai and database there seems to be no stopping the cloud momentum there's certainly plenty of buzz about the cloud tax so-called cloud tax but other than wildly assumptive valuation models and some pockets of anecdotal evidence you don't really see the supposed backlash impacting cloud momentum our forecast calls for the big four hyperscalers aws azure alibaba and gcp to surpass 115 billion as we said in is revenue this year the latest etr survey results show that aws lambda has retaken the lead among all major cloud services tracked in the data set as measured in spending momentum this is the service with the most elevated scores azure overall azure functions vmware cloud on aws and aws overall also demonstrate very highly elevated performance all above that of gcp now impressively aws momentum in the all-important fortune 500 where it has always showed strength is also accelerating one concern in the most recent survey data is that the on-prem clouds and so-called hybrid platforms which we had previously reported as showing an upward spending trajectory seem to have cooled off a bit but the data is mixed and it's a little bit too early to draw firm conclusions nonetheless while hyperscalers are holding steady the spending data appears to be somewhat tepid for the on-prem players you know particularly for their cloud we'll study that further after etr drops its full results on july 23rd now turning our attention back to aws the aws cloud is showing strength across its entire portfolio and we're going to show you that shortly in particular we see notable strength relative to others in analytics ai and the all-important database category aurora and redshift are particularly strong but several other aws database services are showing elevated spending velocity which we'll quantify in a moment all that said snowflake continues to lead all database suppliers in spending momentum by a wide margin which again will quantify in this episode but before we dig into the survey let's take a look at our latest projections for the big four hyperscalers in is as you know we track quarterly revenues for the hyperscalers remember aws and alibaba ias data is pretty clean and reported in their respective earnings reports azure and gcp we have to extrapolate and strip out all a lot of the the apps and other certain revenue to make an apples-to-apples comparison with aws and alibaba and as you can see we have the 2021 market exceeding 115 billion dollars worldwide that's a torrid 35 growth rate on top of 41 in 2020 relative to 2019. aggressive yes but the data continues to point us in this direction until we see some clearer headwinds for the cloud players this is the call we're making aws is perhaps losing a sharepoint or so but it's also is so large that its annual incremental revenue is comparable to alibaba's and google's respective cloud business in total is business in total the big three u.s cloud companies all report at the end of july while alibaba is mid mid-august so we'll update these figures at that time okay let's move on and dig into the survey data we don't have the data yet on alibaba and we're limited as to what we can share until etr drops its research update on on the 23rd but here's a look at the net score timeline in the fortune 500 specifically so we filter the fortune 500 for cloud computing you got azure and the yellow aws and the black and gcp in blue so two points here stand out first is that aws and microsoft are converging and remember the customers who respond to the survey they probably include a fair amount of application software spending in their cloud answers so it favors microsoft in that respect and gcp second point is showing notable deceleration relative to the two leaders and the green call out is because this cut is from an aws point of view so in other words gcp declines are a positive for aws so that's how it should be interpreted now let's take a moment to better understand the idea of net score this is one of the fundamental metrics of the etr methodology here's the data for aws so we use that as a as a reference point net score is calculated by asking customers if they're adding a platform new that's the lime green bar that you see here in the current survey they're asking are you spending six percent or more in the second half relative to the first half of the year that's the forest green they're also asking is spending flat that's the gray or are you spending less that's the pink or are you replacing the platform i.e repatriating so not much spending going on in replacements now in fairness one percent of aws is half a billion dollars so i can see where some folks would get excited about that but in the grand scheme of things it's a sliver so again we don't see repatriation in the numbers okay back to net score subtract the reds from the greens and you get net score which in the case of aws is 61 now just for reference my personal subjective elevated net score level is 40 so anything above that is really impressive based on my experience and to have a company of this size be so elevated is meaningful same for microsoft by the way which is consistently well above the 50 mark in net score in the etr surveys so that's you can think about it that's even more impressive perhaps than aws because it's triple the revenue okay let's stay with aws and take a look at the portfolio and the strength across the board this chart shows net score for the past three surveys serverless is on fire by the way not just aws but azure and gcp functions as well but look at the aws portfolio every category is well above the 40 percent elevated red line the only exception is chime and even chime is showing an uptick and chime is meh if you've ever used chime every other category is well above 50 percent next net score very very strong for aws now as we've frequently reported ai is one of the four biggest focus areas from a spending standpoint along with cloud containers and rpa so it stands to reason that the company with the best ai and ml and the greatest momentum in that space has an advantage because ai is being embedded into apps data processes machines everywhere this chart compares the ai players on two dimensions net score on the vertical axis and market share or presence in the data set on the horizontal axis for companies with more than 15 citations in the survey aws has the highest net score and what's notable is the presence on the horizontal axis databricks is a company where high on also shows elevated scores above both google and microsoft who are showing strength in their own right and then you can see data iq data robot anaconda and salesforce with einstein all above that 40 percent mark and then below you can see the position of sap with leonardo ibm watson and oracle which is well below the 40 line all right let's look at at the all-important database category for a moment and we'll first take a look at the aws database portfolio this chart shows the database services in aws's arsenal and breaks down the net score components with the total net score superimposed on top of the bars point one is aurora is highly elevated with a net score above 70 percent that's due to heavy new adoptions redshift is also very strong as are virtually all aws database offerings with the exception of neptune which is the graph database rds dynamodb elastic document db time stream and quantum ledger database all show momentum above that all important 40 line so while a lot of people criticize the fragmentation of the aws data portfolio and their right tool for the right job approach the spending spending metrics tell a story and that that the strategy is working now let's take a look at the microsoft database portfolio there's a story here similar similar to that of aws azure sql and cosmos db microsoft's nosql distributed database are both very highly elevated as are azure database for mysql and mariadb azure cash for redis and azure for cassandra also microsoft is giving look at microsoft's giving customers a lot of options which is kind of interesting you know we've often said that oracle's strategy because we think about oracle they're building the oracle database cloud we've said oracle strategy should be to not just be the cloud for oracle databases but be the cloud for all databases i mean oracle's got a lot of specialty capability there but it looks like microsoft is beating oracle to that punch not that oracle is necessarily going there but we think it should to expand the appeal of its cloud okay last data chart that we'll show and then and then this one looks at database disruption the chart shows how the cloud database companies are doing in ibm oracle teradata in cloudera accounts the bars show the net score granularity as we described earlier and the etr callouts are interesting so first remember this is an aws this is in an aws context so with 47 responses etr rightly indicates that aws is very well positioned in these accounts with the 68 net score but look at snowflake it has an 81 percent net score which is just incredible and you can see google database is also very strong and the high 50 percent range while microsoft even though it's above the 40 percent mark is noticeably lower than the others as is mongodb with presumably atlas which is surprisingly low frankly but back to snowflake so the etr callout stresses that snowflake doesn't have a strong as strong a presence in the legacy database vendor accounts yet now i'm not sure i would put cloudair in the legacy database category but okay whatever cloudera they're positioning cdp is a hybrid platform as are all the on-prem players with their respective products and platforms but it's going to be interesting to see because snowflake has flat out said it's not straddling the cloud and on-prem rather it's all in on cloud but there is a big opportunity to connect on-prem to the cloud and across clouds which snowflake is pursuing that that ladder the cross-cloud the multi-cloud and snowflake is betting on incremental use cases that involve data sharing and federated governance while traditional players they're protecting their turf at the same time trying to compete in cloud native and of course across cloud i think there's room for both but clearly as we've shown cloud has the spending velocity and a tailwind at its back and aws along with microsoft seem to be getting stronger especially in the all-important categories related to machine intelligence ai and database now to be an essential infrastructure technology player in the data era it would seem obvious that you have to have database and or data management intellectual property in your portfolio or you're going to be less valuable to customers and investors okay we're going to leave it there for today remember these episodes they're all available as podcasts wherever you listen all you do is search breaking analysis podcast and please subscribe to the series check out etr's website at etr dot plus plus etr plus we also publish a full report every week on wikibon.com and siliconangle.com you can get in touch with me david.velante at siliconangle.com you can dm me at d vallante or you can hit hit me up on our linkedin post this is dave vellante for the cube insights powered by etr have a great week stay safe be well and we'll see you next time you
SUMMARY :
that the company with the best ai and ml
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
alibaba | ORGANIZATION | 0.99+ |
six percent | QUANTITY | 0.99+ |
81 percent | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
2021 | DATE | 0.99+ |
2019 | DATE | 0.99+ |
40 percent | QUANTITY | 0.99+ |
july 23rd | DATE | 0.99+ |
microsoft | ORGANIZATION | 0.99+ |
115 billion | QUANTITY | 0.99+ |
dave vellante | PERSON | 0.99+ |
50 percent | QUANTITY | 0.99+ |
41 | QUANTITY | 0.99+ |
61 | QUANTITY | 0.99+ |
47 responses | QUANTITY | 0.99+ |
boston | LOCATION | 0.99+ |
one percent | QUANTITY | 0.99+ |
second half | QUANTITY | 0.99+ |
aws | ORGANIZATION | 0.99+ |
40 | QUANTITY | 0.99+ |
two leaders | QUANTITY | 0.99+ |
second point | QUANTITY | 0.99+ |
115 billion dollars | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
half a billion dollars | QUANTITY | 0.99+ |
more than 15 citations | QUANTITY | 0.98+ |
mid mid-august | DATE | 0.98+ |
two points | QUANTITY | 0.98+ |
ORGANIZATION | 0.98+ | |
siliconangle.com | OTHER | 0.98+ |
end of july | DATE | 0.98+ |
david.velante | PERSON | 0.97+ |
july | DATE | 0.97+ |
50 | QUANTITY | 0.97+ |
40 percent | QUANTITY | 0.97+ |
this year | DATE | 0.97+ |
both | QUANTITY | 0.96+ |
oracle | ORGANIZATION | 0.95+ |
sql | TITLE | 0.95+ |
mysql | TITLE | 0.95+ |
first half | QUANTITY | 0.95+ |
palo alto | ORGANIZATION | 0.95+ |
pandemic | EVENT | 0.95+ |
35 | QUANTITY | 0.94+ |
this week | DATE | 0.93+ |
etr | ORGANIZATION | 0.93+ |
four biggest focus areas | QUANTITY | 0.91+ |
aws azure | ORGANIZATION | 0.91+ |
azure | ORGANIZATION | 0.91+ |
one | QUANTITY | 0.91+ |
23rd | DATE | 0.9+ |
40 line | QUANTITY | 0.89+ |
Breaking Analysis: Best of theCUBE on Cloud
>> Narrator: From theCUBE Studios in Palo Alto, in Boston bringing you data-driven insights from theCUBE and ETR. This is "Breaking Analysis" with Dave Vellante. >> The next 10 years of cloud, they're going to differ dramatically from the past decade. The early days of cloud, deployed virtualization of standard off-the-shelf components, X86 microprocessors, disk drives et cetera, to then scale out and build a large distributed system. The coming decade is going to see a much more data-centric, real-time, intelligent, call it even hyper-decentralized cloud that will comprise on-prem, hybrid, cross-cloud and edge workloads with a services layer that will obstruct the underlying complexity of the infrastructure which will also comprise much more custom and varied components. This was a key takeaway of the guests from theCUBE on Cloud, an event hosted by SiliconANGLE on theCUBE. Welcome to this week's Wikibon CUBE Insights Powered by ETR. In this episode, we'll summarize the findings of our recent event and extract the signal from our great guests with a couple of series and comments and clips from the show. CUBE on Cloud is our very first virtual editorial event. It was designed to bring together our community in an open forum. We ran the day on our 365 software platform and had a great lineup of CEOs, CIOs, data practitioners technologists. We had cloud experts, analysts and many opinion leaders all brought together in a day long series of sessions that we developed in order to unpack the future of cloud computing in the coming decade. Let me briefly frame up the conversation and then turn it over to some of our guests. First, we put forth our view of how modern cloud has evolved and where it's headed. This graphic that we're showing here, talks about the progression of cloud innovation over time. A cloud like many innovations, it started as a novelty. When AWS announced S3 in March of 2006, nobody in the vendor or user communities really even in the trade press really paid too much attention to it. Then later that year, Amazon announced EC2 and people started to think about a new model of computing. But it was largely tire kickers, bleeding-edge developers that took notice and really leaned in. Now the financial crisis of 2007 to 2009, really created what we call a cloud awakening and it put cloud on the radar of many CFOs. Shadow IT emerged within departments that wanted to take IT in bite-sized chunks and along with the CFO wanted to take it as OPEX versus CAPEX. And then I teach transformation that really took hold. We came out of the financial crisis and we've been on an 11-year cloud boom. And it doesn't look like it's going to stop anytime soon, cloud has really disrupted the on-prem model as we've reported and completely transformed IT. Ironically, the pandemic hit at the beginning of this decade, and created a mandate to go digital. And so it accelerated the industry transformation that we're highlighting here, which probably would have taken several more years to mature but overnight the forced March to digital happened. And it looks like it's here to stay. Now the next wave, we think we'll be much more about business or industry transformation. We're seeing the first glimpses of that. Holger Mueller of Constellation Research summed it up at our event very well I thought, he basically said the cloud is the big winner of COVID. Of course we know that now normally we talk about seven-year economic cycles. He said he was talking about for planning and investment cycles. Now we operate in seven-day cycles. The examples he gave where do we open or close the store? How do we pivot to support remote workers without the burden of CAPEX? And we think that the things listed on this chart are going to be front and center in the coming years, data AI, a fully digitized and intelligence stack that will support next gen disruptions in autos, manufacturing, finance, farming and virtually every industry where the system will expand to the edge. And the underlying infrastructure across physical locations will be hidden. Many issues remain, not the least of which is latency which we talked about at the event in quite some detail. So let's talk about how the Big 3 cloud players are going to participate in this next era. Well, in short, the consensus from the event was that the rich get richer. Let's take a look at some data. This chart shows our most recent estimates of IaaS and PaaS spending for the Big 3. And we're going to update this after earning season but there's a couple of points stand out. First, we want to make the point that combined the Big 3 now account for almost $80 billion of infrastructure spend last year. That $80 billion, was not all incremental (laughs) No it's caused consolidation and disruption in the on-prem data center business and within IT shops companies like Dell, HPE, IBM, Oracle many others have felt the heat and have had to respond with hybrid and cross cloud strategies. Second while it's true that Azure and GCP they appear to be growing faster than AWS. We don't know really the exact numbers, of course because only AWS provides a clean view of IaaS and passwords, Microsoft and Google. They kind of hide them all ball on their numbers which by the way, I don't blame them but they do leave breadcrumbs and clues on growth rates. And we have other means of estimating through surveys and the like, but it's undeniable Azure is closing the revenue gap on AWS. The third is that I like the fact that Azure and Google are growing faster than AWS. AWS is the only company by our estimates to grow its business sequentially last quarter. And in and of itself, that's not really enough important. What is significant is that because AWS is so large now at 45 billion, even at their slower growth rates it grows much more in absolute terms than its competitors. So we think AWS is going to keep its lead for some time. We think Microsoft and AWS will continue to lead the pack. You know, they might converge maybe it will be a 200 just race in terms of who's first who's second in terms of cloud revenue and how it's counted depending on what they count in their numbers. And Google look with its balance sheet and global network. It's going to play the long game and virtually everyone else with the exception of perhaps Alibaba is going to be secondary players on these platforms. Now this next graphic underscores that reality and kind of lays out the competitive landscape. What we're showing here is survey data from ETR of more than 1400 CIOs and IT buyers and on the vertical axis is Net Score which measures spending momentum on the horizontal axis is so-called Market Share which is a measure of pervasiveness in the data set. The key points are AWS and Microsoft look at it. They stand alone so far ahead of the pack. I mean, they really literally, it would have to fall down to lose their lead high spending velocity and large share of the market or the hallmarks of these two companies. And we don't think that's going to change anytime soon. Now, Google, even though it's far behind they have the financial strength to continue to position themselves as an alternative to AWS. And of course, an analytics specialist. So it will continue to grow, but it will be challenged. We think to catch up to the leaders. Now take a look at the hybrid zone where the field is playing. These are companies that have a large on-prem presence and have been forced to initiate a coherent cloud strategy. And of course, including multicloud. And we include Google in this so pack because they're behind and they have to take a differentiated approach relative to AWS, and maybe cozy up to some of these traditional enterprise vendors to help Google get to the enterprise. And you can see from the on-prem crowd, VMware Cloud on AWS is stands out as having some, some momentum as does Red Hat OpenShift, which is it's cloudy, but it's really sort of an ingredient it's not really broad IaaS specifically but it's a component of cloud VMware cloud which includes VCF or VMware Cloud Foundation. And even Dell's cloud. We would expect HPE with its GreenLake strategy. Its financials is shoring up, should be picking up momentum in the future in terms of what the customers of this survey consider cloud. And then of course you could see IBM and Oracle you're in the game, but they don't have the spending momentum and they don't have the CAPEX chops to compete with the hyperscalers IBM's cloud revenue actually dropped 7% last quarter. So that highlights the challenges that that company facing Oracle's cloud business is growing in the single digits. It's kind of up and down, but again underscores these two companies are really about migrating their software install basis to their captive clouds and as well for IBM, for example it's launched a financial cloud as a way to differentiate and not take AWS head-on an infrastructure as a service. The bottom line is that other than the Big 3 in Alibaba the rest of the pack will be plugging into hybridizing and cross-clouding those platforms. And there are definitely opportunities there specifically related to creating that abstraction layer that we talked about earlier and hiding that underlying complexity and importantly creating incremental value good examples, snowfallLike what snowflake is doing with its data cloud, what the data protection guys are doing. A company like Loomio is headed in that direction as are others. So, you keep an eye on that and think about where the white space is and where the value can be across-clouds. That's where the opportunity is. So let's see, what is this all going to look like? How does the cube community think it's going to unfold? Let's hear from theCUBE Guests and theCUBE on Cloud speakers and some of those highlights. Now, unfortunately we don't have time to show you clips from every speaker. We are like 10-plus hours of video content but we've tried to pull together some comments that summarize the sentiment from the community. So I'm going to have John Furrier briefly explain what theCUBE on Cloud is all about and then let the guests speak for themselves. After John, Pradeep Sindhu is going to give a nice technical overview of how the cloud was built out and what's changing in the future. I'll give you a hint it has to do with data. And then speaking of data, Mai-Lan Bukovec, who heads up AWS is storage portfolio. She'll explain how she views the coming changes in cloud and how they look at storage. Again, no surprise, it's all about data. Now, one of the themes that you'll hear from guests is the notion of a distributed cloud model. And Zhamak Deghani, he was a data architect. She'll explain her view of the future of data architectures. We also have thoughts from analysts like Zeus Karavalla and Maribel Lopez, and some comments from both Microsoft and Google to compliment AWS's view of the world. In fact, we asked JG Chirapurath from Microsoft to comment on the common narrative that Microsoft products are not best-to-breed. They put out a one dot O and then they get better, or sometimes people say, well, they're just good enough. So we'll see what his response is to that. And Paul Gillin asks, Amit Zavery of Google his thoughts on the cloud leaderboard and how Google thinks about their third-place position. Dheeraj Pandey gives his perspective on how technology has progressed and been miniaturized over time. And what's coming in the future. And then Simon Crosby gives us a framework to think about the edge as the most logical opportunity to process data not necessarily a physical place. And this was echoed by John Roese, and Chris Wolf to experience CTOs who went into some great depth on this topic. Unfortunately, I don't have the clips of those two but their comments can be found on the CTO power panel the technical edge it's called that's the segment at theCUBE on Cloud events site which we'll share the URL later. Now, the highlight reel ends with CEO Joni Klippert she talks about the changes in securing the cloud from a developer angle. And finally, we wrap up with a CIO perspective, Dan Sheehan. He provides some practical advice on building on his experience as a CIO, COO and CTO specifically how do you as a business technology leader deal with the rapid pace of change and still be able to drive business results? Okay, so let's now hear from the community please run the highlights. >> Well, I think one of the things we talked about COVID is the personal impact to me but other people as well one of the things that people are craving right now is information, factual information, truth, textures that we call it. But here this event for us Dave is our first inaugural editorial event. Rob, both Kristen Nicole the entire cube team, SiliconANGLE on theCUBE we're really trying to put together more of a cadence. We're going to do more of these events where we can put out and feature the best people in our community that have great fresh voices. You know, we do interview the big names Andy Jassy, Michael Dell, the billionaires of people making things happen, but it's often the people under them that are the real Newsmakers. >> If you look at the architecture of cloud data centers the single most important invention was scale-out. Scale-out of identical or near identical servers all connected to a standard IP ethernet network. That's the architecture. Now the building blocks of this architecture is ethernet switches which make up the network, IP ethernet switches. And then the server is all built using general purpose x86 CPU's with DRAM, with SSD, with hard drives all connected to inside the CPU. Now, the fact that you scale these server nodes as they're called out was very, very important in addressing the problem of how do you build very large scale infrastructure using general purpose compute but this architecture, Dave is a compute centric architecture. And the reason it's a compute centric architecture is if you open this, is server node. What you see is a connection to the network typically with a simple network interface card. And then you have CPU's which are in the middle of the action. Not only are the CPU's processing the application workload but they're processing all of the IO workload what we call data centric workload. And so when you connect SSDs and hard drives and GPU is everything to the CPU, as well as to the network you can now imagine that the CPU is doing two functions. It's running the applications but it's also playing traffic cop for the IO. So every IO has to go to the CPU and you're executing instructions typically in the operating system. And you're interrupting the CPU many many millions of times a second. Now general purpose CPU and the architecture of the CPU's was never designed to play traffic cop because the traffic cop function is a function that requires you to be interrupted very, very frequently. So it's critical that in this new architecture where does a lot of data, a lot of these stress traffic the percentage of workload, which is data centric has gone from maybe one to 2% to 30 to 40%. >> The path to innovation is paved by data. If you don't have data, you don't have machine learning you don't have the next generation of analytics applications that helps you chart a path forward into a world that seems to be changing every week. And so in order to have that insight in order to have that predictive forecasting that every company needs, regardless of what industry that you're in today, it all starts from data. And I think the key shift that I've seen is how customers are thinking about that data, about being instantly usable. Whereas in the past, it might've been a backup. Now it's part of a data Lake. And if you can bring that data into a data lake you can have not just analytics or machine learning or auditing applications it's really what does your application do for your business and how can it take advantage of that vast amount of shared data set in your business? >> We are actually moving towards decentralization if we think today, like if it let's move data aside if we said is the only way web would work the only way we get access to various applications on the web or pages to centralize it We would laugh at that idea. But for some reason we don't question that when it comes to data, right? So I think it's time to embrace the complexity that comes with the growth of number of sources, the proliferation of sources and consumptions models, embrace the distribution of sources of data that they're not just within one part of organization. They're not just within even bounds of organizations that are beyond the bounds of organization. And then look back and say, okay, if that's the trend of our industry in general, given the fabric of compensation and data that we put in, you know, globally in place then how the architecture and technology and organizational structure incentives need to move to embrace that complexity. And to me that requires a paradigm shift a full stack from how we organize our organizations how we organize our teams, how we put a technology in place to look at it from a decentralized angle. >> I actually think we're in the midst of the transition to what's called a distributed cloud, where if you look at modernized cloud apps today they're actually made up of services from different clouds. And also distributed edge locations. And that's going to have a pretty profound impact on the way we go vast. >> We wake up every day, worrying about our customer and worrying about the customer condition and to absolutely make sure we dealt with the best in the first attempt that we do. So when you take the plethora of products we've dealt with in Azure, be it Azure SQL be it Azure cosmos DB, Synapse, Azure Databricks, which we did in partnership with Databricks Azure machine learning. And recently when we sort of offered the world's first comprehensive data governance solution and Azure overview, I would, I would humbly submit to you that we are leading the way. >> How important are rankings within the Google cloud team or are you focused mainly more on growth and just consistency? >> No, I don't think again, I'm not worried about we are not focused on ranking or any of that stuff. Typically I think we are worried about making sure customers are satisfied and the adding more and more customers. So if you look at the volume of customers we are signing up a lot of the large deals we did doing. If you look at the announcement we've made over the last year has been tremendous momentum around that. >> The thing that is really interesting about where we have been versus where we're going is we spend a lot of time talking about virtualizing hardware and moving that around. And what does that look like? And creating that as more of a software paradigm. And the thing we're talking about now is what does cloud as an operating model look like? What is the manageability of that? What is the security of that? What, you know, we've talked a lot about containers and moving into different, DevSecOps and all those different trends that we've been talking about. Like now we're doing them. So we've only gotten to the first crank of that. And I think every technology vendor we talked to now has to address how are they are going to do a highly distributed management insecurity landscape? Like, what are they going to layer on top of that? Because it's not just about, oh, I've taken a rack of something, server storage, compute, and virtualized it. I know have to create a new operating model around it in a way we're almost redoing what the OSI stack looks like and what the software and solutions are for that. >> And the whole idea of we in every recession we make things smaller. You know, in 91 we said we're going to go away from mainframes into Unix servers. And we made the unit of compute smaller. Then in the year, 2000 windows the next bubble burst and the recession afterwards we moved from Unix servers to Wintel windows and Intel x86 and eventually Linux as well. Again, we made things smaller going from million dollar servers to $5,000 servers, shorter lib servers. And that's what we did in 2008, 2009. I said, look, we don't even need to buy servers. We can do things with virtual machines which are servers that are an incarnation in the digital world. There's nothing in the physical world that actually even lives but we made it even smaller. And now with cloud in the last three, four years and what will happen in this coming decade. They're going to make it even smaller not just in space, which is size, with functions and containers and virtual machines, but also in time. >> So I think the right way to think about edges where can you reasonably process the data? And it obviously makes sense to process data at the first opportunity you have but much data is encrypted between the original device say and the application. And so edge as a place doesn't make as much sense as edge as an opportunity to decrypt and analyze it in the care. >> When I think of Shift-left, I think of that Mobius that we all look at all of the time and how we deliver and like plan, write code, deliver software, and then manage it, monitor it, right like that entire DevOps workflow. And today, when we think about where security lives, it either is a blocker to deploying production or most commonly it lives long after code has been deployed to production. And there's a security team constantly playing catch up trying to ensure that the development team whose job is to deliver value to their customers quickly, right? Deploy as fast as we can as many great customer facing features. They're then looking at it months after software has been deployed and then hurrying and trying to assess where the bugs are and trying to get that information back to software developers so that they can fix those issues. Shifting left to me means software engineers are finding those bugs as they're writing code or in the CIC CD pipeline long before code has been deployed to production. >> During this for quite a while now, it still comes down to the people. I can get the technology to do what it needs to do as long as they have the right requirements. So that goes back to people making sure we have the partnership that goes back to leadership and the people and then the change management aspects right out of the gate, you should be worrying about how this change is going to be how it's going to affect, and then the adoption and an engagement, because adoption is critical because you can go create the best thing you think from a technology perspective. But if it doesn't get used correctly, it's not worth the investment. So I agree, what is a digital transformation or innovation? It still comes down to understand the business model and injecting and utilizing technology to grow our reduce costs, grow the business or reduce costs. >> Okay, so look, there's so much other content on theCUBE on Cloud events site we'll put the link in the description below. We have other CEOs like Kathy Southwick and Ellen Nance. We have the CIO of UI path. Daniel Dienes talks about automation in the cloud and Appenzell from Anaplan. And a plan is not her company. By the way, Dave Humphrey from Bain also talks about his $750 million investment in Nutanix. Interesting, Rachel Stevens from red monk talks about the future of software development in the cloud and CTO, Hillary Hunter talks about the cloud going vertical into financial services. And of course, John Furrier and I along with special guests like Sergeant Joe Hall share our take on key trends, data and perspectives. So right here, you see the coupon cloud. There's a URL, check it out again. We'll, we'll pop this URL in the description of the video. So there's some great content there. I want to thank everybody who participated and thank you for watching this special episode of theCUBE Insights Powered by ETR. This is Dave Vellante and I'd appreciate any feedback you might have on how we can deliver better event content for you in the future. We'll be doing a number of these and we look forward to your participation and feedback. Thank you, all right, take care, we'll see you next time. (upbeat music)
SUMMARY :
bringing you data-driven and kind of lays out the about COVID is the personal impact to me and GPU is everything to the Whereas in the past, it the only way we get access on the way we go vast. and to absolutely make sure we dealt and the adding more and more customers. And the thing we're talking And the whole idea and analyze it in the care. or in the CIC CD pipeline long before code I can get the technology to of software development in the cloud
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
IBM | ORGANIZATION | 0.99+ |
Daniel Dienes | PERSON | 0.99+ |
Zhamak Deghani | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
John Roese | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Paul Gillin | PERSON | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Rachel Stevens | PERSON | 0.99+ |
Maribel Lopez | PERSON | 0.99+ |
Michael Dell | PERSON | 0.99+ |
$5,000 | QUANTITY | 0.99+ |
Chris Wolf | PERSON | 0.99+ |
2008 | DATE | 0.99+ |
Joni Klippert | PERSON | 0.99+ |
seven-day | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Dan Sheehan | PERSON | 0.99+ |
Pradeep Sindhu | PERSON | 0.99+ |
Dheeraj Pandey | PERSON | 0.99+ |
March of 2006 | DATE | 0.99+ |
Rob | PERSON | 0.99+ |
Hillary Hunter | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Amit Zavery | PERSON | 0.99+ |
Ellen Nance | PERSON | 0.99+ |
JG Chirapurath | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Dave Humphrey | PERSON | 0.99+ |
Simon Crosby | PERSON | 0.99+ |
Mai-Lan Bukovec | PERSON | 0.99+ |
2009 | DATE | 0.99+ |
$80 billion | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
John | PERSON | 0.99+ |
11-year | QUANTITY | 0.99+ |
Kristen Nicole | PERSON | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
Loomio | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
10-plus hours | QUANTITY | 0.99+ |
45 billion | QUANTITY | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
$750 million | QUANTITY | 0.99+ |
7% | QUANTITY | 0.99+ |
Holger Mueller | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
First | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
third | QUANTITY | 0.99+ |
two companies | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
first | QUANTITY | 0.99+ |
Zeus Karavalla | PERSON | 0.99+ |
last year | DATE | 0.99+ |
Kathy Southwick | PERSON | 0.99+ |
second | QUANTITY | 0.99+ |
Constellation Research | ORGANIZATION | 0.99+ |
JG Chirapurath, Microsoft | theCUBE on Cloud 2021
>>from around the globe. It's the Cube presenting Cuban cloud brought to you by silicon angle. Okay, >>we're now going to explore the vision of the future of cloud computing From the perspective of one of the leaders in the field, J G >>Share >>a pure off is the vice president of As Your Data ai and Edge at Microsoft G. Welcome to the Cuban cloud. Thanks so much for participating. >>Well, thank you, Dave, and it's a real pleasure to be here with you. And I just wanna welcome the audience as well. >>Well, jg judging from your title, we have a lot of ground to cover, and our audience is definitely interested in all the topics that are implied there. So let's get right into it. You know, we've said many times in the Cube that the new innovation cocktail comprises machine intelligence or a I applied to troves of data. With the scale of the cloud. It's it's no longer, you know, we're driven by Moore's law. It's really those three factors, and those ingredients are gonna power the next wave of value creation and the economy. So, first, do you buy into that premise? >>Yes, absolutely. we do buy into it. And I think, you know, one of the reasons why we put Data Analytics and Ai together is because all of that really begins with the collection of data and managing it and governing it, unlocking analytics in it. And we tend to see things like AI, the value creation that comes from a I as being on that continues off, having started off with really things like analytics and proceeding toe. You know, machine learning and the use of data. Interesting breaks. Yes. >>I'd like to get some more thoughts around a data and how you see the future data and the role of cloud and maybe how >>Microsoft, you >>know, strategy fits in there. I mean, you, your portfolio, you got you got sequel server, Azure, Azure sequel. You got arc, which is kinda azure everywhere for people that aren't familiar with that. You've got synapse. Which course that's all the integration a data warehouse, and get things ready for B I and consumption by the business and and the whole data pipeline and a lot of other services as your data bricks you got You got cosmos in their, uh, Blockchain. You've got open source services like Post Dress and my sequel. So lots of choices there. And I'm wondering, you know, how do you think about the future of Of of Cloud data platforms? It looks like your strategies, right tool for the right job? Is that fair? >>It is fair, but it's also just to step back and look at it. It's fundamentally what we see in this market today is that customer was the Sikh really a comprehensive proposition? And when I say a comprehensive proposition, it is sometimes not just about saying that. Hey, listen way No, you're a sequel server company. We absolutely trust that you have the best Azure sequel database in the cloud, but tell us more. We've got data that's sitting in her group systems. We've got data that's sitting in Post Press in things like mongo DB, right? So that open source proposition today and data and data management and database management has become front and center, so are really sort of push. There is when it comes to migration management, modernization of data to present the broadest possible choice to our customers so we can meet them where they are. However, when it comes to analytics. One of the things they asked for is give us a lot more convergence use. You know it, really, it isn't about having 50 different services. It's really about having that one comprehensive service that is converged. That's where things like synapse Fitzer, where in just land any kind of data in the leg and then use any compute engine on top of it to drive insights from it. So, fundamentally, you know, it is that flexibility that we really sort of focus on to meet our customers where they are and really not pushing our dogma and our beliefs on it. But to meet our customers according to the way they have deployed stuff like this. >>So that's great. I want to stick on this for a minute because, you know, I know when when I have guests on like yourself, do you never want to talk about you know, the competition? But that's all we ever talk about. That's all your customers ever talk about, because because the counter to that right tool for the right job and that I would say, is really kind of Amazon's approach is is that you got the single unified data platform, the mega database that does it all. And that's kind of Oracle's approach. It sounds like you wanna have your cake and eat it, too, so you you got the right tool for the right job approach. But you've got an integration layer that allows you to have that converge database. I wonder if you could add color to that and you confirm or deny what I just said. >>No, that's a That's a very fair observation, but I I say there's a nuance in what I sort of describe when it comes to data management. When it comes to APS, we have them customers with the broadest choice. Even in that, even in that perspective, we also offer convergence. So, case in point, when you think about Cosmos TV under that one sort of service, you get multiple engines, but with the same properties, right global distribution, the five nines availability. It gives customers the ability to basically choose when they have to build that new cloud native AB toe, adopt cosmos Davey and adopted in a way that it's and choose an engine that is most flexible. Tow them, however you know when it comes to say, you know, writing a sequel server, for example from organizing it you know you want. Sometimes you just want to lift and shift it into things. Like I asked In other cases, you want to completely rewrite it, so you need to have the flexibility of choice there that is presented by a legacy off What's its on premises? When it moved into things like analytics, we absolutely believe in convergence, right? So we don't believe that look, you need to have a relation of data warehouse that is separate from a loop system that is separate from, say, a B I system. That is just, you know, it's a bolt on for us. We love the proposition off, really building things that are so integrated that once you land data, once you prep it inside the lake, you can use it for analytics. You can use it for being. You can use it for machine learning. So I think you know, are sort of differentiated. Approach speaks for itself there. Well, >>that's that's interesting, because essentially, again, you're not saying it's an either or, and you're seeing a lot of that in the marketplace. You got some companies say no, it's the Data Lake and others saying No, no put in the data warehouse and that causes confusion and complexity around the data pipeline and a lot of calls. And I'd love to get your thoughts on this. Ah, lot of customers struggled to get value out of data and and specifically data product builders of frustrated that it takes too long to go from. You know, this idea of Hey, I have an idea for a data service and it could drive monetization, but to get there, you gotta go through this complex data lifecycle on pipeline and beg people to add new data sources. And do you do you feel like we have to rethink the way that we approach data architectures? >>Look, I think we do in the cloud, and I think what's happening today and I think the place where I see the most amount of rethink the most amount of push from our customers to really rethink is the area of analytics in a I. It's almost as if what worked in the past will not work going forward. Right? So when you think about analytics on in the Enterprise today, you have relational systems, you have produced systems. You've got data marts. You've got data warehouses. You've got enterprise data warehouses. You know, those large honking databases that you use, uh, to close your books with right? But when you start to modernize it, what deep you are saying is that we don't want to simply take all of that complexity that we've built over say, you know, 34 decades and simply migrated on mass exactly as they are into the cloud. What they really want is a completely different way of looking at things. And I think this is where services like synapse completely provide a differentiated proposition to our customers. What we say there is land the data in any way you see shape or form inside the lake. Once you landed inside the lake, you can essentially use a synapse studio toe. Prep it in the way that you like, use any compute engine of your choice and and operate on this data in any way that you see fit. So, case in point, if you want to hydrate relation all data warehouse, you can do so if you want to do ad hoc analytics using something like spark. You can do so if you want to invoke power. Bi I on that data or b i on that data you can do so if you want to bring in a machine learning model on this breath data you can do so, so inherently. So when customers buy into this proposition, what it solves for them and what it gives them is complete simplicity, right? One way to land the data, multiple ways to use it. And it's all eso. >>Should we think of synapse as an abstraction layer that abstracts away the complexity of the underlying technology? Is that a fair way toe? Think about it. >>Yeah, you can think of it that way. It abstracts away, Dave a couple of things. It takes away the type of data, you know, sort of the complexities related to the type of data. It takes away the complexity related to the size of data. It takes away the complexity related to creating pipelines around all these different types of data and fundamentally puts it in a place where it can be now consumed by any sort of entity inside the actual proposition. And by that token, even data breaks. You know, you can, in fact, use data breaks in in sort off an integrated way with a synapse, Right, >>Well, so that leads me to this notion of and then wonder if you buy into it s Oh, my inference is that a data warehouse or a data lake >>could >>just be a node in inside of a global data >>mesh on. >>Then it's synapses sort of managing, uh, that technology on top. Do you buy into that that global data mesh concept >>we do. And we actually do see our customers using synapse and the value proposition that it brings together in that way. Now it's not where they start. Often times when a customer comes and says, Look, I've got an enterprise data warehouse, I want to migrate it or I have a group system. I want to migrate it. But from there, the evolution is absolutely interesting to see. I give you an example. You know, one of the customers that we're very proud off his FedEx And what FedEx is doing is it's completely reimagining its's logistics system that basically the system that delivers What is it? The three million packages a day on in doing so in this covert times, with the view of basically delivering our covert vaccines. One of the ways they're doing it is basically using synapse. Synapse is essentially that analytic hub where they can get complete view into their logistic processes. Way things are moving, understand things like delays and really put all that together in a way that they can essentially get our packages and these vaccines delivered as quickly as possible. Another example, you know, is one of my favorite, uh, we see once customers buy into it, they essentially can do other things with it. So an example of this is, uh is really my favorite story is Peace Parks Initiative. It is the premier Air White Rhino Conservancy in the world. They essentially are using data that has landed in azure images in particular. So, basically, you know, use drones over the vast area that they patrol and use machine learning on this data to really figure out where is an issue and where there isn't an issue so that this part with about 200 rangers can scramble surgically versus having to read range across the last area that they cover. So What do you see here is you know, the importance is really getting your data in order. Landed consistently. Whatever the kind of data ideas build the right pipelines and then the possibilities of transformation are just endless. >>Yeah, that's very nice how you worked in some of the customer examples. I appreciate that. I wanna ask you, though, that that some people might say that putting in that layer while it clearly adds simplification and e think a great thing that they're begins over time to be be a gap, if you will, between the ability of that layer to integrate all the primitives and all the peace parts on that, that you lose some of that fine grain control and it slows you down. What would you say to that? >>Look, I think that's what we excel at, and that's what we completely sort of buy into on. It's our job to basically provide that level off integration that granularity in the way that so it's an art, absolutely admit it's an art. There are areas where people create simplicity and not a lot of you know, sort of knobs and dials and things like that. But there are areas where customers want flexibility, right? So I think just to give you an example of both of them in landing the data inconsistency in building pipelines, they want simplicity. They don't want complexity. They don't want 50 different places to do this. Just 100 to do it. When it comes to computing and reducing this data analyzing this data, they want flexibility. This is one of the reasons why we say, Hey, listen, you want to use data breaks? If you're you're buying into that proposition and you're absolutely happy with them, you can plug plug it into it. You want to use B I and no, essentially do a small data mart. You can use B I If you say that. Look, I've landed in the lake. I really only want to use em melt, bringing your animal models and party on. So that's where the flexibility comes in. So that's sort of really sort of think about it. Well, >>I like the strategy because, you know, my one of our guest, Jim Octagon, e E. I think one of the foremost thinkers on this notion of off the data mesh and her premises that that that data builders, data product and service builders air frustrated because the the big data system is generic to context. There's no context in there. But by having context in the big data architecture and system, you could get products to market much, much, much faster. So but that seems to be your philosophy. But I'm gonna jump ahead to do my ecosystem question. You've mentioned data breaks a couple of times. There's another partner that you have, which is snowflake. They're kind of trying to build out their own, uh, data cloud, if you will, on global mesh in and the one hand, their partner. On the other hand, there are competitors. How do you sort of balance and square that circle? >>Look, when I see snowflake, I actually see a partner. You know that when we essentially you know, we are. When you think about as you know, this is where I sort of step back and look at Azure as a whole and in azure as a whole. Companies like snowflakes are vital in our ecosystem, right? I mean, there are places we compete, but you know, effectively by helping them build the best snowflake service on Asia. We essentially are able toe, you know, differentiate and offer a differentiated value proposition compared to, say, a Google or on AWS. In fact, that's being our approach with data breaks as well, where you know they are effectively on multiple club, and our opportunity with data breaks is toe essentially integrate them in a way where we offer the best experience. The best integrations on Azure Barna That's always been a focus. >>That's hard to argue with. Strategy. Our data with our data partner eat er, shows Microsoft is both pervasive and impressively having a lot of momentum spending velocity within the budget cycles. I wanna come back thio ai a little bit. It's obviously one of the fastest growing areas in our in our survey data. As I said, clearly, Microsoft is a leader in this space. What's your what's your vision of the future of machine intelligence and how Microsoft will will participate in that opportunity? >>Yeah, so fundamentally, you know, we've built on decades of research around, you know, around, you know, essentially, you know, vision, speech and language that's being the three core building blocks and for the for a really focused period of time we focused on essentially ensuring human parody. So if you ever wondered what the keys to the kingdom are it, czar, it's the most we built in ensuring that the research posture that we've taken there, what we then done is essentially a couple of things we focused on, essentially looking at the spectrum. That is a I both from saying that, Hollis and you know it's gotta work for data. Analysts were looking toe basically use machine learning techniques, toe developers who are essentially, you know, coding and building a machine learning models from scratch. So for that select proposition manifesto us, as you know, really a. I focused on all skill levels. The other court thing we've done is that we've also said, Look, it will only work as long as people trust their data and they can trust their AI models. So there's a tremendous body of work and research we do in things like responsibility. So if you ask me where we sort of push on is fundamentally to make sure that we never lose sight of the fact that the spectrum off a I, and you can sort of come together for any skill level, and we keep that responsibly. I proposition. Absolutely strong now against that canvas, Dave. I'll also tell you that you know, as edge devices get way more capable, right where they can input on the edge, see a camera or a mike or something like that, you will see us pushing a lot more of that capability onto the edge as well. But to me, that's sort of a modality. But the core really is all skill levels and that responsible denia. >>Yeah, So that that brings me to this notion of wanna bring an edge and and hybrid cloud Understand how you're thinking about hybrid cloud multi cloud. Obviously one of your competitors, Amazon won't even say the word multi cloud you guys have, Ah, you know, different approach there. But what's the strategy with regard? Toe, toe hybrid. You know, Do you see the cloud you bringing azure to the edge? Maybe you could talk about that and talk about how you're different from the competition. >>Yeah, I think in the edge from Annette, you know, I live in I'll be the first one to say that the word nge itself is conflated. Okay, It's, uh but I will tell you, just focusing on hybrid. This is one of the places where you know I would say the 2020 if I would have looked back from a corporate perspective. In particular, it has Bean the most informative because we absolutely saw customers digitizing moving to the cloud. And we really saw hybrid in action. 2020 was the year that hybrid sort of really became really from a cloud computing perspective and an example of this is we understood it's not all or nothing. So sometimes customers want azure consistency in their data centers. This is where things like Azure stack comes in. Sometimes they basically come to us and say, We want the flexibility of adopting flexible pattern, you know, platforms like, say, containers orchestra, Cuban Pettis, so that we can essentially deployed wherever you want. And so when we design things like art, it was built for that flexibility in mind. So here is the beauty of what's something like our can do for you. If you have a kubernetes endpoint anywhere we can deploy and as your service onto it, that is the promise, which means if for some reason, the customer says that. Hey, I've got this kubernetes endpoint in AWS and I love as your sequel. You will be able to run as your sequel inside AWS. There's nothing that stops you from doing it so inherently you remember. Our first principle is always to meet our customers where they are. So from that perspective, multi cloud is here to stay. You know, we're never going to be the people that says, I'm sorry, we will never see a But it is a reality for our customers. >>So I wonder if we could close. Thank you for that by looking, looking back and then and then ahead. And I wanna e wanna put forth. Maybe it's, Ah criticism, but maybe not. Maybe it's an art of Microsoft, but But first you know, you get Microsoft an incredible job of transitioning. It's business as your nominee president Azzawi said. Our data shows that so two part question First, Microsoft got there by investing in the cloud, really changing its mind set, I think, in leveraging its huge software state and customer base to put Azure at the center of its strategy, and many have said me included that you got there by creating products that air Good enough. You know, we do a 1.0, it's not that great. And the two Dato, and maybe not the best, but acceptable for your customers. And that's allowed you to grow very rapidly expanding market. >>How >>do you respond to that? Is that is that a fair comment? Ume or than good enough? I wonder if you could share your >>thoughts, gave you? You hurt my feelings with that question. I don't hate me, g getting >>it out there. >>So there was. First of all, thank you for asking me that. You know, I am absolutely the biggest cheerleader. You'll find a Microsoft. I absolutely believe you know that I represent the work off almost 9000 engineers and we wake up every day worrying about our customer and worrying about the customer condition and toe. Absolutely. Make sure we deliver the best in the first time that we do. So when you take the platter off products we've delivered in nausea, be it as your sequel, be it as your cosmos TV synapse as your data breaks, which we did in partnership with data breaks, a za machine learning and recently when we prevail, we sort off, you know, sort of offered the world's first comprehensive data government solution in azure purview. I would humbly submit to you that we're leading the way and we're essentially showing how the future off data ai and the actual work in the cloud. >>I'd be disappointed if you if you had If you didn't, if you capitulated in any way J g So so thank you for that. And the kind of last question is, is looking forward and how you're thinking about the future of cloud last decade. A lot about your cloud migration simplifying infrastructure management, deployment SAS if eyeing my enterprise, lot of simplification and cost savings. And, of course, the redeployment of resource is toward digital transformation. Other other other valuable activities. How >>do >>you think this coming decade will will be defined? Will it be sort of more of the same? Or is there Is there something else out there? >>I think I think that the coming decade will be one where customers start one law outside value out of this. You know what happened in the last decade when people leave the foundation and people essentially looked at the world and said, Look, we've got to make the move, you know, the largely hybrid, but we're going to start making steps to basically digitize and modernize our platforms. I would tell you that with the amount of data that people are moving to the cloud just as an example, you're going to see use of analytics ai for business outcomes explode. You're also going to see a huge sort of focus on things like governance. You know, people need to know where the data is, what the data catalog continues, how to govern it, how to trust this data and given all other privacy and compliance regulations out there. Essentially, they're complying this posture. So I think the unlocking of outcomes versus simply Hey, I've saved money Second, really putting this comprehensive sort off, you know, governance, regime in place. And then, finally, security and trust. It's going to be more paramount than ever before. Yeah, >>nobody's gonna use the data if they don't trust it. I'm glad you brought up your security. It's It's a topic that hits number one on the CEO list. J G. Great conversation. Obviously the strategy is working, and thanks so much for participating in Cuba on cloud. >>Thank you. Thank you, David. I appreciate it and thank you to. Everybody was tuning in today. >>All right? And keep it right there. I'll be back with our next guest right after this short break.
SUMMARY :
cloud brought to you by silicon angle. a pure off is the vice president of As Your Data ai and Edge at Microsoft And I just wanna welcome the audience as you know, we're driven by Moore's law. And I think, you know, one of the reasons why And I'm wondering, you know, how do you think about the future of Of So, fundamentally, you know, it is that flexibility that we really sort of focus I want to stick on this for a minute because, you know, I know when when I have guests So I think you know, are sort of differentiated. but to get there, you gotta go through this complex data lifecycle on pipeline and beg people to in the Enterprise today, you have relational systems, you have produced systems. Is that a fair way toe? It takes away the type of data, you know, sort of the complexities related Do you buy into that that global data mesh concept is you know, the importance is really getting your data in order. that you lose some of that fine grain control and it slows you down. So I think just to give you an example of both I like the strategy because, you know, my one of our guest, Jim Octagon, I mean, there are places we compete, but you know, effectively by helping them build It's obviously one of the fastest growing areas in our So for that select proposition manifesto us, as you know, really a. You know, Do you see the cloud you bringing azure to the edge? Cuban Pettis, so that we can essentially deployed wherever you want. Maybe it's an art of Microsoft, but But first you know, you get Microsoft You hurt my feelings with that question. when we prevail, we sort off, you know, sort of offered the world's I'd be disappointed if you if you had If you didn't, if you capitulated in any way J g So Look, we've got to make the move, you know, the largely hybrid, I'm glad you brought up your security. I appreciate it and thank you to. And keep it right there.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Annette | PERSON | 0.99+ |
Hollis | PERSON | 0.99+ |
FedEx | ORGANIZATION | 0.99+ |
JG Chirapurath | PERSON | 0.99+ |
Asia | LOCATION | 0.99+ |
Jim Octagon | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
100 | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
50 different services | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
Azzawi | PERSON | 0.99+ |
First | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
34 decades | QUANTITY | 0.99+ |
Cuba | LOCATION | 0.99+ |
single | QUANTITY | 0.99+ |
J G. | PERSON | 0.99+ |
first time | QUANTITY | 0.98+ |
Second | QUANTITY | 0.98+ |
first one | QUANTITY | 0.98+ |
first principle | QUANTITY | 0.98+ |
last decade | DATE | 0.98+ |
Cosmos TV | ORGANIZATION | 0.98+ |
Sikh | ORGANIZATION | 0.98+ |
about 200 rangers | QUANTITY | 0.97+ |
J G | PERSON | 0.96+ |
three factors | QUANTITY | 0.96+ |
two part | QUANTITY | 0.96+ |
50 different | QUANTITY | 0.96+ |
Azure | TITLE | 0.96+ |
decades | QUANTITY | 0.96+ |
president | PERSON | 0.96+ |
Air White Rhino Conservancy | ORGANIZATION | 0.95+ |
Cuban | OTHER | 0.94+ |
almost 9000 engineers | QUANTITY | 0.91+ |
Post Press | ORGANIZATION | 0.89+ |
As Your Data ai and Edge | ORGANIZATION | 0.88+ |
Moore | PERSON | 0.88+ |
cosmos Davey | ORGANIZATION | 0.87+ |
Peace Parks Initiative | ORGANIZATION | 0.86+ |
three million packages a day | QUANTITY | 0.85+ |
Dress | TITLE | 0.85+ |
wave | EVENT | 0.84+ |
theCUBE | ORGANIZATION | 0.83+ |
synapse | ORGANIZATION | 0.8+ |
Cube | COMMERCIAL_ITEM | 0.79+ |
three core building blocks | QUANTITY | 0.78+ |
one comprehensive service | QUANTITY | 0.77+ |
Data Lake | ORGANIZATION | 0.77+ |
JG Chirapurath, Microsoft CLEAN
>> Okay, we're now going to explore the vision of the future of cloud computing from the perspective of one of the leaders in the field, JG Chirapurath is the Vice President of Azure Data AI and Edge at Microsoft. JG, welcome to theCUBE on Cloud, thanks so much for participating. >> Well, thank you, Dave. And it's a real pleasure to be here with you and just want to welcome the audience as well. >> Well, JG, judging from your title, we have a lot of ground to cover and our audience is definitely interested in all the topics that are implied there. So let's get right into it. We've said many times in theCUBE that the new innovation cocktail comprises machine intelligence or AI applied to troves of data with the scale of the cloud. It's no longer we're driven by Moore's law. It's really those three factors and those ingredients are going to power the next wave of value creation in the economy. So first, do you buy into that premise? >> Yes, absolutely. We do buy into it and I think one of the reasons why we put data analytics and AI together, is because all of that really begins with the collection of data and managing it and governing it, unlocking analytics in it. And we tend to see things like AI, the value creation that comes from AI as being on that continuum of having started off with really things like analytics and proceeding to be machine learning and the use of data in interesting ways. >> Yes, I'd like to get some more thoughts around data and how you see the future of data and the role of cloud and maybe how Microsoft strategy fits in there. I mean, your portfolio, you've got SQL Server, Azure SQL, you got Arc which is kind of Azure everywhere for people that aren't familiar with that you got Synapse which course does all the integration, the data warehouse and it gets things ready for BI and consumption by the business and the whole data pipeline. And then all the other services, Azure Databricks, you got you got Cosmos in there, you got Blockchain, you've got Open Source services like PostgreSQL and MySQL. So lots of choices there. And I'm wondering, how do you think about the future of cloud data platforms? It looks like your strategy is right tool for the right job. Is that fair? >> It is fair, but it's also just to step back and look at it. It's fundamentally what we see in this market today, is that customers they seek really a comprehensive proposition. And when I say a comprehensive proposition it is sometimes not just about saying that, "Hey, listen "we know you're a sequence of a company, "we absolutely trust that you have the best "Azure SQL database in the cloud. "But tell us more." We've got data that is sitting in Hadoop systems. We've got data that is sitting in PostgreSQL, in things like MongoDB. So that open source proposition today in data and data management and database management has become front and center. So our real sort of push there is when it comes to migration management modernization of data to present the broadest possible choice to our customers, so we can meet them where they are. However, when it comes to analytics, one of the things they ask for is give us lot more convergence use. It really, it isn't about having 50 different services. It's really about having that one comprehensive service that is converged. That's where things like Synapse fits in where you can just land any kind of data in the lake and then use any compute engine on top of it to drive insights from it. So fundamentally, it is that flexibility that we really sort of focus on to meet our customers where they are. And really not pushing our dogma and our beliefs on it but to meet our customers according to the way they've deployed stuff like this. >> So that's great. I want to stick on this for a minute because when I have guests on like yourself they never want to talk about the competition but that's all we ever talk about. And that's all your customers ever talk about. Because the counter to that right tool for the right job and that I would say is really kind of Amazon's approach is that you got the single unified data platform, the mega database. So it does it all. And that's kind of Oracle's approach. It sounds like you want to have your cake and eat it too. So you got the right tool with the right job approach but you've got an integration layer that allows you to have that converged database. I wonder if you could add color to that and confirm or deny what I just said. >> No, that's a very fair observation but I'd say there's a nuance in what I sort of described. When it comes to data management, when it comes to apps, we have then customers with the broadest choice. Even in that perspective, we also offer convergence. So case in point, when you think about cosmos DB under that one sort of service, you get multiple engines but with the same properties. Right, global distribution, the five nines availability. It gives customers the ability to basically choose when they have to build that new cloud native app to adopt cosmos DB and adopt it in a way that is an choose an engine that is most flexible to them. However, when it comes to say, writing a SequenceServer for example, if modernizing it, you want sometimes, you just want to lift and shift it into things like IS. In other cases, you want to completely rewrite it. So you need to have the flexibility of choice there that is presented by a legacy of what sits on premises. When you move into things like analytics, we absolutely believe in convergence. So we don't believe that look, you need to have a relational data warehouse that is separate from a Hadoop system that is separate from say a BI system that is just, it's a bolt-on. For us, we love the proposition of really building things that are so integrated that once you land data, once you prep it inside the Lake you can use it for analytics, you can use it for BI, you can use it for machine learning. So I think, our sort of differentiated approach speaks for itself there. >> Well, that's interesting because essentially again you're not saying it's an either or, and you see a lot of that in the marketplace. You got some companies you say, "No, it's the data lake." And others say "No, no, put it in the data warehouse." And that causes confusion and complexity around the data pipeline and a lot of cutting. And I'd love to get your thoughts on this. A lot of customers struggle to get value out of data and specifically data product builders are frustrated that it takes them too long to go from, this idea of, hey, I have an idea for a data service and it can drive monetization, but to get there you got to go through this complex data life cycle and pipeline and beg people to add new data sources and do you feel like we have to rethink the way that we approach data architecture? >> Look, I think we do in the cloud. And I think what's happening today and I think the place where I see the most amount of rethink and the most amount of push from our customers to really rethink is the area of analytics and AI. It's almost as if what worked in the past will not work going forward. So when you think about analytics only in the enterprise today, you have relational systems, you have Hadoop systems, you've got data marts, you've got data warehouses you've got enterprise data warehouse. So those large honking databases that you use to close your books with. But when you start to modernize it, what people are saying is that we don't want to simply take all of that complexity that we've built over, say three, four decades and simply migrate it en masse exactly as they are into the cloud. What they really want is a completely different way of looking at things. And I think this is where services like Synapse completely provide a differentiated proposition to our customers. What we say there is land the data in any way you see, shape or form inside the lake. Once you landed inside the lake, you can essentially use a Synapse Studio to prep it in the way that you like. Use any compute engine of your choice and operate on this data in any way that you see fit. So case in point, if you want to hydrate a relational data warehouse, you can do so. If you want to do ad hoc analytics using something like Spark, you can do so. If you want to invoke Power BI on that data or BI on that data, you can do so. If you want to bring in a machine learning model on this prep data, you can do so. So inherently, so when customers buy into this proposition, what it solves for them and what it gives to them is complete simplicity. One way to land the data multiple ways to use it. And it's all integrated. >> So should we think of Synapse as an abstraction layer that abstracts away the complexity of the underlying technology? Is that a fair way to think about it? >> Yeah, you can think of it that way. It abstracts away Dave, a couple of things. It takes away that type of data. Sort of complexities related to the type of data. It takes away the complexity related to the size of data. It takes away the complexity related to creating pipelines around all these different types of data. And fundamentally puts it in a place where it can be now consumed by any sort of entity inside the Azure proposition. And by that token, even Databricks. You can in fact use Databricks in sort of an integrated way with the Azure Synapse >> Right, well, so that leads me to this notion of and I wonder if you buy into it. So my inference is that a data warehouse or a data lake could just be a node inside of a global data mesh. And then it's Synapse is sort of managing that technology on top. Do you buy into that? That global data mesh concept? >> We do and we actually do see our customers using Synapse and the value proposition that it brings together in that way. Now it's not where they start, oftentimes when a customer comes and says, "Look, I've got an enterprise data warehouse, "I want to migrate it." Or "I have a Hadoop system, I want to migrate it." But from there, the evolution is absolutely interesting to see. I'll give you an example. One of the customers that we're very proud of is FedEx. And what FedEx is doing is it's completely re-imagining its logistics system. That basically the system that delivers, what is it? The 3 million packages a day. And in doing so, in this COVID times, with the view of basically delivering on COVID vaccines. One of the ways they're doing it, is basically using Synapse. Synapse is essentially that analytic hub where they can get complete view into the logistic processes, way things are moving, understand things like delays and really put all of that together in a way that they can essentially get our packages and these vaccines delivered as quickly as possible. Another example, it's one of my favorite. We see once customers buy into it, they essentially can do other things with it. So an example of this is really my favorite story is Peace Parks initiative. It is the premier of white rhino conservancy in the world. They essentially are using data that has landed in Azure, images in particular to basically use drones over the vast area that they patrol and use machine learning on this data to really figure out where is an issue and where there isn't an issue. So that this part with about 200 radios can scramble surgically versus having to range across the vast area that they cover. So, what you see here is, the importance is really getting your data in order, landing consistently whatever the kind of data it is, build the right pipelines, and then the possibilities of transformation are just endless. >> Yeah, that's very nice how you worked in some of the customer examples and I appreciate that. I want to ask you though that some people might say that putting in that layer while you clearly add simplification and is I think a great thing that there begins over time to be a gap, if you will, between the ability of that layer to integrate all the primitives and all the piece parts, and that you lose some of that fine grain control and it slows you down. What would you say to that? >> Look, I think that's what we excel at and that's what we completely sort of buy into. And it's our job to basically provide that level of integration and that granularity in the way that it's an art. I absolutely admit it's an art. There are areas where people crave simplicity and not a lot of sort of knobs and dials and things like that. But there are areas where customers want flexibility. And so I think just to give you an example of both of them, in landing the data, in consistency in building pipelines, they want simplicity. They don't want complexity. They don't want 50 different places to do this. There's one way to do it. When it comes to computing and reducing this data, analyzing this data, they want flexibility. This is one of the reasons why we say, "Hey, listen you want to use Databricks. "If you're buying into that proposition. "And you're absolutely happy with them, "you can plug it into it." You want to use BI and essentially do a small data model, you can use BI. If you say that, "Look, I've landed into the lake, "I really only want to use ML." Bring in your ML models and party on. So that's where the flexibility comes in. So that's sort of that we sort of think about it. >> Well, I like the strategy because one of our guests, Jumark Dehghani is I think one of the foremost thinkers on this notion of of the data mesh And her premise is that the data builders, data product and service builders are frustrated because the big data system is generic to context. There's no context in there. But by having context in the big data architecture and system you can get products to market much, much, much faster. So, and that seems to be your philosophy but I'm going to jump ahead to my ecosystem question. You've mentioned Databricks a couple of times. There's another partner that you have, which is Snowflake. They're kind of trying to build out their own DataCloud, if you will and GlobalMesh, and the one hand they're a partner on the other hand they're a competitor. How do you sort of balance and square that circle? >> Look, when I see Snowflake, I actually see a partner. When we see essentially we are when you think about Azure now this is where I sort of step back and look at Azure as a whole. And in Azure as a whole, companies like Snowflake are vital in our ecosystem. I mean, there are places we compete, but effectively by helping them build the best Snowflake service on Azure, we essentially are able to differentiate and offer a differentiated value proposition compared to say a Google or an AWS. In fact, that's been our approach with Databricks as well. Where they are effectively on multiple clouds and our opportunity with Databricks is to essentially integrate them in a way where we offer the best experience the best integrations on Azure Berna. That's always been our focus. >> Yeah, it's hard to argue with the strategy or data with our data partner and ETR shows Microsoft is both pervasive and impressively having a lot of momentum spending velocity within the budget cycles. I want to come back to AI a little bit. It's obviously one of the fastest growing areas in our survey data. As I said, clearly Microsoft is a leader in this space. What's your vision of the future of machine intelligence and how Microsoft will participate in that opportunity? >> Yeah, so fundamentally, we've built on decades of research around essentially vision, speech and language. That's been the three core building blocks and for a really focused period of time, we focused on essentially ensuring human parity. So if you ever wonder what the keys to the kingdom are, it's the boat we built in ensuring that the research or posture that we've taken there. What we've then done is essentially a couple of things. We've focused on essentially looking at the spectrum that is AI. Both from saying that, "Hey, listen, "it's got to work for data analysts." We're looking to basically use machine learning techniques to developers who are essentially, coding and building machine learning models from scratch. So for that select proposition manifest to us as really AI focused on all skill levels. The other core thing we've done is that we've also said, "Look, it'll only work as long "as people trust their data "and they can trust their AI models." So there's a tremendous body of work and research we do and things like responsible AI. So if you asked me where we sort of push on is fundamentally to make sure that we never lose sight of the fact that the spectrum of AI can sort of come together for any skill level. And we keep that responsible AI proposition absolutely strong. Now against that canvas Dave, I'll also tell you that as Edge devices get way more capable, where they can input on the Edge, say a camera or a mic or something like that. You will see us pushing a lot more of that capability onto the edge as well. But to me, that's sort of a modality but the core really is all skill levels and that responsibility in AI. >> Yeah, so that brings me to this notion of, I want to bring an Edge and hybrid cloud, understand how you're thinking about hybrid cloud, multicloud obviously one of your competitors Amazon won't even say the word multicloud. You guys have a different approach there but what's the strategy with regard to hybrid? Do you see the cloud, you're bringing Azure to the edge maybe you could talk about that and talk about how you're different from the competition. >> Yeah, I think in the Edge from an Edge and I even I'll be the first one to say that the word Edge itself is conflated. Okay, a little bit it's but I will tell you just focusing on hybrid, this is one of the places where, I would say 2020 if I were to look back from a COVID perspective in particular, it has been the most informative. Because we absolutely saw customers digitizing, moving to the cloud. And we really saw hybrid in action. 2020 was the year that hybrid sort of really became real from a cloud computing perspective. And an example of this is we understood that it's not all or nothing. So sometimes customers want Azure consistency in their data centers. This is where things like Azure Stack comes in. Sometimes they basically come to us and say, "We want the flexibility of adopting "flexible button of platforms let's say containers, "orchestrating Kubernetes "so that we can essentially deploy it wherever you want." And so when we designed things like Arc, it was built for that flexibility in mind. So, here's the beauty of what something like Arc can do for you. If you have a Kubernetes endpoint anywhere, we can deploy an Azure service onto it. That is the promise. Which means, if for some reason the customer says that, "Hey, I've got "this Kubernetes endpoint in AWS. And I love Azure SQL. You will be able to run Azure SQL inside AWS. There's nothing that stops you from doing it. So inherently, remember our first principle is always to meet our customers where they are. So from that perspective, multicloud is here to stay. We are never going to be the people that says, "I'm sorry." We will never say (speaks indistinctly) multicloud but it is a reality for our customers. >> So I wonder if we could close, thank you for that. By looking back and then ahead and I want to put forth, maybe it's a criticism, but maybe not. Maybe it's an art of Microsoft. But first, you did Microsoft an incredible job at transitioning its business. Azure is omnipresent, as we said our data shows that. So two-part question first, Microsoft got there by investing in the cloud, really changing its mindset, I think and leveraging its huge software estate and customer base to put Azure at the center of it's strategy. And many have said, me included, that you got there by creating products that are good enough. We do a one Datto, it's still not that great, then a two Datto and maybe not the best, but acceptable for your customers. And that's allowed you to grow very rapidly expand your market. How do you respond to that? Is that a fair comment? Are you more than good enough? I wonder if you could share your thoughts. >> Dave, you hurt my feelings with that question. >> Don't hate me JG. (both laugh) We're getting it out there all right, so. >> First of all, thank you for asking me that. I am absolutely the biggest cheerleader you'll find at Microsoft. I absolutely believe that I represent the work of almost 9,000 engineers. And we wake up every day worrying about our customer and worrying about the customer condition and to absolutely make sure we deliver the best in the first attempt that we do. So when you take the plethora of products we deliver in Azure, be it Azure SQL, be it Azure Cosmos DB, Synapse, Azure Databricks, which we did in partnership with Databricks, Azure Machine Learning. And recently when we premiered, we sort of offered the world's first comprehensive data governance solution in Azure Purview. I would humbly submit it to you that we are leading the way and we're essentially showing how the future of data, AI and the Edge should work in the cloud. >> Yeah, I'd be disappointed if you capitulated in any way, JG. So, thank you for that. And that's kind of last question is looking forward and how you're thinking about the future of cloud. Last decade, a lot about cloud migration, simplifying infrastructure to management and deployment. SaaSifying My Enterprise, a lot of simplification and cost savings and of course redeployment of resources toward digital transformation, other valuable activities. How do you think this coming decade will be defined? Will it be sort of more of the same or is there something else out there? >> I think that the coming decade will be one where customers start to unlock outsize value out of this. What happened to the last decade where people laid the foundation? And people essentially looked at the world and said, "Look, we've got to make a move. "They're largely hybrid, but you're going to start making "steps to basically digitize and modernize our platforms. I will tell you that with the amount of data that people are moving to the cloud, just as an example, you're going to see use of analytics, AI or business outcomes explode. You're also going to see a huge sort of focus on things like governance. People need to know where the data is, what the data catalog continues, how to govern it, how to trust this data and given all of the privacy and compliance regulations out there essentially their compliance posture. So I think the unlocking of outcomes versus simply, Hey, I've saved money. Second, really putting this comprehensive sort of governance regime in place and then finally security and trust. It's going to be more paramount than ever before. >> Yeah, nobody's going to use the data if they don't trust it, I'm glad you brought up security. It's a topic that is at number one on the CIO list. JG, great conversation. Obviously the strategy is working and thanks so much for participating in Cube on Cloud. >> Thank you, thank you, Dave and I appreciate it and thank you to everybody who's tuning into today. >> All right then keep it right there, I'll be back with our next guest right after this short break.
SUMMARY :
of one of the leaders in the field, to be here with you that the new innovation cocktail comprises and the use of data in interesting ways. and how you see the future that you have the best is that you got the single that once you land data, but to get there you got to go in the way that you like. Yeah, you can think of it that way. of and I wonder if you buy into it. and the value proposition and that you lose some of And so I think just to give you an example So, and that seems to be your philosophy when you think about Azure Yeah, it's hard to argue the keys to the kingdom are, Do you see the cloud, you're and I even I'll be the first one to say that you got there by creating products Dave, you hurt my We're getting it out there all right, so. that I represent the work Will it be sort of more of the same and given all of the privacy the data if they don't trust it, thank you to everybody I'll be back with our next guest
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave | PERSON | 0.99+ |
JG | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
Microsoft | ORGANIZATION | 0.99+ |
FedEx | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Jumark Dehghani | PERSON | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
JG Chirapurath | PERSON | 0.99+ |
first | QUANTITY | 0.99+ |
50 different services | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
50 different places | QUANTITY | 0.99+ |
MySQL | TITLE | 0.99+ |
one | QUANTITY | 0.99+ |
GlobalMesh | ORGANIZATION | 0.99+ |
Both | QUANTITY | 0.99+ |
first attempt | QUANTITY | 0.99+ |
Second | QUANTITY | 0.99+ |
Last decade | DATE | 0.99+ |
three | QUANTITY | 0.99+ |
three factors | QUANTITY | 0.99+ |
Synapse | ORGANIZATION | 0.99+ |
one way | QUANTITY | 0.99+ |
COVID | OTHER | 0.99+ |
One | QUANTITY | 0.98+ |
first one | QUANTITY | 0.98+ |
first principle | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
Azure Stack | TITLE | 0.98+ |
Azure SQL | TITLE | 0.98+ |
Spark | TITLE | 0.98+ |
First | QUANTITY | 0.98+ |
MongoDB | TITLE | 0.98+ |
2020 | DATE | 0.98+ |
about 200 radios | QUANTITY | 0.98+ |
Moore | PERSON | 0.97+ |
PostgreSQL | TITLE | 0.97+ |
four decades | QUANTITY | 0.97+ |
Arc | TITLE | 0.97+ |
single | QUANTITY | 0.96+ |
Snowflake | ORGANIZATION | 0.96+ |
last decade | DATE | 0.96+ |
Azure Purview | TITLE | 0.95+ |
3 million packages a day | QUANTITY | 0.95+ |
One way | QUANTITY | 0.94+ |
three core | QUANTITY | 0.94+ |
Sagar Kadakia | CUBE Conversation, December 2020
>> From The Cube Studios in Palo Alto and Boston connecting with thought-leaders all around the world, this is a Cube Conversation. >> Hello, everyone, and welcome to this Cube Conversation, I'm Dave Vellante. Now, you know I love data, and today we're going to introduce you to a new data and analytical platform, and we're going to take it to the world of cloud database and data warehouses. And with me is Sagar Kadakia who's the head of Enterprise IT (indistinct) 7Park Data. Sagar, welcome back to the Cube. Good to see you. >> Thank you so much, David. I appreciate you having me back on. >> Hey, so new gig for you, how's it going? Tell us about 7Park Data. >> Yeah. Look, things are going well. It started at about two months ago, just a, you know, busy. I had a chance last, you know a few months to kind of really dig into the dataset. We have a tremendous amount of research coming out in Q4 Q1 around kind of the public cloud database market public cloud analytics market. So, you know, really looking forward to that. >> Okay, good. Well, let's bring up the first slide. Let's talk about where this data comes from. Tell us a little bit more about the platform. Where's the insight. >> Yeah, absolutely. So I'll talk a little about 7Park and then we'd kind of jump into the data a little bit. So 7Park was founded in 2012 in terms of differentiator, you know with other alternative data firms, you know we use NLP machine learning, you know AI to really kind of, you know, structure like noisy and unstructured data sets really kind of generate insight from that. And so, because a lot of that know how we ended up being acquired by Vista back in 2018. And really like for us, you know the mandate there is to really, you know look across all their different portfolio companies and try to generate insight from all the data assets you know, that these portfolio companies have. So, you know, today we're going to be talking about you know, one of the data sets from those companies it's that cloud infrastructure data set. We get it from one of the portfolio companies that you know, helps organizations kind of manage and optimize their cloud spend. It's real time data. We essentially get this aggregated daily. So this certainly different than, you know your traditional providers maybe giving you quarterly or kind of by annual data. This is incredibly granular, real time all the way down to the invoice level. So within this cloud infrastructure dataset we're tracking several billion dollars worth of spend across AWS, Azure and GCP. Something like 350 services across like 20 plus markets. So, you know, security machine learning analytics database which we're going to talk about today. And again like the granularity of the KPIs I think is kind of really what kind of you know, differentiates this dataset you know, with just within database itself, you know we're tracking over 20 services. So, you know, lots to kind of look forward to kind of into Q4 and Q1. >> So, okay. So the main spring of your data is if I'm a customer and I there's a service out there there are many services like this that can help me optimize my spend and the way they do that is I basically connect their APIs. So they have visibility on what the transactions that I'm making my usage statistics et cetera. And then you take that and then extrapolate that and report on that. Is that right? >> Exactly. Yeah. We're seeing just on this one data set that we're going to talk about today, it's something like six 700 million rows worth of data. And so kind of what we do is, you know we kind of have the insight layer on top of that or the analytics layer on top of all that unstructured data, so that we can get a feel for, you know a whole host of different kind of KPIs spend, adoption rates, market share, you know product size, retention rates, spend, you know, net price all that type of stuff. So, yeah, that's exactly what we're doing. >> Love it, there's more transparency the better. Okay. So, so right, because this whole world of market sizing has been very opaque you know, over the years, and it's like you know, backroom conversations, whether it's IDC, Gartner who's got what don't take, you know and the estimations and it's very, very, you know it's not very transparent so I'm excited to see what you guys have. Okay. So, so you have some data on the public cloud and specifically the database market that you want to share with our audience. Let's bring up the next graphic here. What are we looking at here Sagar? What are these blue lines and red lines what's this all about? >> Yeah. So and look, we can kind of start at the kind of the 10,000 foot view kind of level here. And so what we're looking at here is our estimates for the entire kind of cloud database market, including data warehousing. If you look all the way over to the right I'll kind of explain some of these bars in a minute but just high level, you know we're forecasting for this year, $11.8 billion. Now something to kind of remember about that is that's just AWS, Azure and GCP, right? So that's not the entire cloud database market. It's just specific to those three providers. What you're looking at here is the breakout and blue and purple is SQL databases and then no SQL databases. And so, you know, to no one's surprise here and you can see, you know SQL database is obviously much larger from a revenue standpoint. And so you can see just from this time last year, you know the database market has grown 40% among these three cloud providers. And, you know, though, we're not showing it here, you know from like a PI perspective, you know database is playing a larger and larger role for all three of these providers. And so obviously this is a really hot market, which is why, you know we're kind of discussing a lot of the dynamics. You don't need to Q and Q Q4 and Q1 >> So, okay. Let's get into some of the specific firm-level data. You have numbers that you want to share on Amazon Redshift and Google BigQuery, and some comments on Snowflake let's bring up the next graphic. So tell us, it says public cloud data, warehousing growth tempered by Snowflake, what's the data showing. And let's talk about some of the implications there. >> Yeah, no problem. So yeah, this is kind of one of the markets, you know that we kind of did a deep dive in tomorrow and we'll kind of get this, you know, get to this in a few minutes, we're kind of doing a big CIO panel kind of covering data, warehousing, RDBMS documents store key value, graph all these different database markets but I thought it'd be great, you know just cause obviously what's occurring here and with snowflake to kind of talk about, you know the data warehousing market, you know, look if you look here, these are some of the KPIs that we have you know, and I'll kind of start from the left. Here are some of the orange bars, the darker orange bars. Those are our estimates for AWS Redshift. And so you can see here, you know we're projecting about 667 million in revenue for Redshift. But if you look at the lighter arm bars, you can see that the service went from representing about 2% of you know, AWS revenue to about 1.5%. And we think some of that is because of Snowflake. And if we kind of, take a look at some of these KPIs you know, below those bar charts here, you know one of the things that we've been looking at is, you know how are longer-term customer spending and how are let's just say like newer customers spending, so to speak. So kind of just like organic growth or kind of net expansion analysis. And if you look at on the bottom there, you'll see, you know customers in our dataset that we looked at, you know that were there 3Q20 as well as 3Q19 their spend on AWS Redshift is 23%. Right? And then look at the bifurcation, right? When we include essentially all the new customers that onboard it, right after 3Q19, look at how much they're bringing down the spend increase. And it's because, you know a lot of spend that was perhaps meant for Redshift is now going to Snowflake. And look, you would expect longer-term customers to spend more than newer customers. But really what we're doing is here is really highlighting the stark contrast because you have kind of back to back KPIs here, you know between organic spend versus total spend and obviously the deceleration in market share kind of coming down. So, you know, something that's interesting here and we'll kind of continue tracking that. >> Okay. So let's maybe come back to this mass Colombo questions here. So the start with the orange side. So we're talking about Snowflake being 667 million. These are your estimates extrapolated based on what we talked about earlier, 1.5% of the AWS portfolio of course you see things like, they continue to grow. Amazon made a bunch of storage announcements last week at the first week of re-invent (indistinct) I mean just name all kinds of databases. And so it's competing with a lot of other services in the portfolio and then, but it's interesting to see Google BigQuery a much larger percentage of the portfolio, which again to me, makes sense people like BigQuery. They like the data science components that are built in the machine learning components that are built in. But then if you look at Snowflake's last quarter and just on a run rate basis, it's over there over $600 million. Now, if you just multiply their last quarter by four from a revenue standpoint. So they got Redshift in their sites, you know if this is, you know to the extent this is the correct number and I know it's an estimate but I haven't seen any better numbers out there. Interesting Sagar, I mean Snowflake surpassed the value of snowflakes or past service now last Friday, it's probably just in trading today you know, on Monday it's maybe Snowflake is about a billion dollars less than the in value than IBM. So you're saying snowflake in a lot of attention, post IPO the thing is even exploded more. I mean, it's crazy. And I presume that's rippled into the customer interest areas. Now the ironic thing here of course, is that that snowflake most of its revenue comes from AWS running on AWS at the same time, AWS and or Redshift and snowflake compete. So you have this interesting dynamic going on. >> Yeah. You know, we've spoken to so many CIOs about kind of the dynamics here with Redshift and BigQuery and Snowflake, you know as it kind of pertains to, you know, Redshift and Snowflake. I think, you know, what I've heard the most is, look if you're using Redshift, you're going to keep using it. But if you're new to data warehousing kind of, so to speak you're going to move to Snowflake, or you're going to start with Snowflake, you know, that and I think, you know when it comes to data warehousing, you're seeing a lot of decisions kind of coming from, you know, bottom up now. So a lot of developers and so obviously their preference is going to be Snowflake. And then when you kind of look at BigQuery here over to the right again, like look you're seeing revenue growth, but again, as a as a percentage of total, you know, GCP revenue you're seeing it come down and look, we don't show it here. But another dynamic that we're seeing amongst BigQuery is that we are seeing adoption rates fall versus this time last year. So we think, again, that could be because of Snowflake. Now, one thing to kind of highlight here with BigQuery look it's kind of the low cost alternative, you know, so to speak, you know once Redshift gets too expensive, so to speak, you know you kind of move over to, to BigQuery and we kind of put some price KPIs down here all the way at the bottom of the chart, you know kind of for both of them, you know when you kind of think about the net price per kind of TB scan, you know, Redshift does it pro rate right? It's five bucks or whatever you, you know whatever you scan in, whereas, you know GCP and get the first terabyte for free. And then everything is prorated after that. And so you can see the net price, right? So that's the price that people actually pay. You can see it's significantly lower that than Redshift. And again, you know it's a lower cost alternative. And so when you think about, you know organizations or CIO's that want to save some money certainly BigQuery, you know, is an option. But certainly I think just overall, you know, Snowflake is is certainly having, you know, an impact here and you can see it from, you know the percentage of total revenue for both these coming down. You know, if we look at other AWS database services or you mentioned a few other services, you know we're not seeing that trend, we're seeing, you know percentage of total revenue hang in or accelerate. And so that's kind of why we want to point this out as this is something unique, you know for AWS and GCP where even though you're seeing growth, it's decelerating. And then of course you can kind of see the percentage of revenue represents coming down. >> I think it's interesting to look at these two companies and then of course Snowflake. So if you think about Snowflake and BigQuery both of those started in the cloud they were true born in the cloud databases. Whereas Redshift was a deal that Amazon did, you know with parxl back in the day, one time license fee and then they re-engineered it to be kind of cloud based. And so there is some of that historical o6n-prem baggage in there. I know that AWS did a tremendous job in rearchitecting that but nonetheless, so I'll give you a couple of examples. If you go back to last year's reinvent 2019 of course Snowflake was really the first to popularize this idea of separating compute from storage and even compute from compute, which is kind of nuance. So I won't go into that, but the idea being you can dial up or dial down compute as you need it you can even turn off compute in the world of Snowflake and just, you know, you're paying an S3 for storage charges. What Amazon did last reinvent was they announced the separation of compute and storage, but what the way they did it was they did it with a tiering architecture. So you can't ever actually fully turn off the compute, but it's great. I mean, it's customers I've talked to say, yes I'm saving a lot of money, you know, with this approach. But again, there's these little nuances. So what Snowflake announced this year was their data cloud and what the data cloud is as a whole new architecture. It's based on this global mesh. It lives across both AWS and Azure and GCP. And what Snowflake has done is they've taken they've abstracted the complexity of the clouds. So you don't even necessarily have to know what you're running on. You have to worry about it any Snowflake user inside of that data cloud if given access can share data with any other user. So it's a very powerful concept that they're doing. AWS at reinvent this year announced something called AWS glue elastic views which basically allows you to take data across their entire database portfolio. And I'm going to put, share in quotes. And I put it in quotes because it's essentially doing copying from a source pushing to a target AWS database and then doing a change data management capture and pushes that over time. So it, it feels like kind of an attempt to do their own data cloud. The advantages of AWS is that they've got way more data stores than just Snowflake cause it's one data store. So was AWS says Aurora dynamo DB Redshift on and on and on streaming databases, et cetera where Snowflake is just Snowflake. And so it's going to be interesting to see, you know these two juxtaposing philosophies but I want it to sort of lay that out because this is just it's setting up as a really interesting dynamic. Then you can bring in Azure as well with Microsoft and what they're doing. And I think this is going to be really fascinating to see how this plays out over the next decade. >> Yeah. I think some of the points you brought up maybe a little bit earlier were just around like the functional limits of a Redshift. Right. And I think that's where, you know Snowflake obviously does it does very, very well you know, you kind of have these, you know kind of to come, you know, you kind of have these, you know if you kind of think about like the market drivers right? Like, let's think about even like the prior slide that we showed, where we saw overall you know, database growth, like what's driving all of that what's driving Redshift, right. Obviously proximity application, interdependencies, right. Costs. You get all the credits or people are already working with the big three providers. And so there's so many reasons to continue spending with them, obviously, you know, COVID-19 right. Obviously all these apps being developed right in the cloud versus data centers and things of that nature. So you have all of these market drivers, you know for the cloud database services for Redshift. And so from that perspective, you know you kind of think, well why are people even to go to a third party vendor? And I think, you know, at that point it has to be the functional superiority. And so again, like a lot of times it depends on, you know, where decisions are coming from you know, top down or bottom up obviously at the engineering at the developer level they're going to want better functionality. Maybe, you know, top-down sometimes, you know it's like, look, we have a lot of credits, you know we're trying to save money, you know from a security perspective it could just be easier to spin something up you know, in AWS, so to speak. So, yeah, I think these are all the dynamics that, you know organizations have to figure out every day, but at least within the data warehousing space, you are seeing spend go towards Snowflake and it's going away to an extent as we kind of see, you know growth decelerate for both of these vendors, right. It's not that revenue's not going out there is growth which is that growth is, it's just not the same as it used to be, you know, so to speak. So yeah, this is a interesting area to kind of watch and I think across all the other markets as well, you know when you think about document store, right you have AWS document DB, right. What are the impacts there with with Mongo and some of these other kind of third party data warehousing vendors, right. Having to compete with all the, you know all the different services offered by AWS Azure like the cosmos and all that stuff. So, yeah, it's definitely kind of turning into a battle Royal, you know as we kind of head into, into 2021. And so I think having all these KPIs is really helping us kind of break down and figure out, you know which areas like data warehousing are slowing down. But then what other areas in database where they're seeing a tremendous amount of acceleration, like as we said, database revenue is driving. Like it's becoming a bigger part of their overall revenue. And so they are doing well. It just, you know, there's obviously snowflake they have to compete with here. >> Well, and I think maybe to your point I infer from your point, it's not necessarily a zero sum game. And as I was discussing before, I think Snowflake's really trying to create a new market. It's not just trying to steal share from the Terra datas and the Redshifts and the PCPs of the world, big queries and and Azure SQL server and Oracle and so forth. They're trying to create a whole new concept called the data cloud, which to me is really important because my prediction is what Snowflake is doing. And they don't even really talk a ton about this but they sort of do, if you squint through the lines I think what they're doing is first of all, simplicity is there, what they're doing. And then they're putting data in the hands of business people, business line people who have domain context, that's a whole new way of thinking about a data architecture versus the prevalent way to do a data pipeline is you got data engineers and data scientists, and you ingest data. It's goes to the beginning of the pipeline and that's kind of a traditional way to do it. And kind of how I think most of the AWS customers do it. I think over time, because of the simplicity of Snowflake you're going to see people begin to look at new ways to architect data. Anyway, we're almost out of time here but I want to bring up the next slide which is a graphic, which talks about a database discussion that you guys are having on 12/8 at 2:00 PM Eastern time with Bain and Verizon who what's this all about. >> Yeah. So, you know, one of the things we wanted to do is we kind of kick off a lot of the, you know Q4 Q1 research or putting on the database spark. It is just like kind of, we did, you know we did today, which obviously, you know we're really going to expand on tomorrow at a at 2:00 PM is discuss all the different KPIs. You know, we track something like 20 plus database services. So we're going to be going through a lot more than just kind of Redshift and BigQuery. Look at all the dynamics there, look at, you know how they're very against some of the third party vendors like the Snowflake, like a Mongo DB, as an example we got some really great, you know, thought leaders you know, Michael Delzer and Praveen from verizon they're going to kind of help, or they're going to opine on all the dynamics that we're seeing. And so it's going to be a very kind of, you know structured wise, it's going to be very quantitative but then you're going to have this beautiful qualitative discussion to kind of help support a lot of the data points that we're capturing. And so, yeah, we're really excited about the panel you know, from, you know, why you should join standpoint. Look, it's just, it's great, competitive Intel. If you're a third party, you know, database, data warehousing vendor, this is the type of information that you're going to want to know, you know, adoption rates market sizing, retention rates, you know net price reservers, on demand dynamics. You know, we're going through a lot that tomorrow. So I'm really excited about that. I'm just in general, really excited about a lot of the research that we're kind of putting out. So >> That's interesting. I mean, and we were talking earlier about AWS glue elastic views. I'd love to see your view of all the database services from Amazon. Cause that's where it's really designed to do is leverage those across those. And you know, you listen to Andrew, Jesse talk they've got a completely different philosophy than say Oracle, which says, Hey we've got one database to do all things Amazon saying we need that fine granularity. So it's going to be again. And to the extent that you're providing market context they're very excited to see that data Sagar and see how that evolves over time. Really appreciate you coming back in the cube and look forward to working with you. >> Appreciate Dave. Thank you so much. >> All right. Welcome. Thank you everybody for watching. This is Dave Vellante for the cube. We'll see you next time. (upbeat music)
SUMMARY :
all around the world, and today we're going to introduce you I appreciate you having me back on. Hey, so new gig for I had a chance last, you know more about the platform. the mandate there is to really, you know And then you take that so that we can get a feel for, you know and it's like you know, And so, you know, to You have numbers that you want one of the markets, you know if this is, you know of the chart, you know interesting to see, you know kind of to come, you know, you and you ingest data. It is just like kind of, we did, you know And you know, you listen Thank you so much. Thank you everybody for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David | PERSON | 0.99+ |
Andrew | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
2012 | DATE | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
7Park | ORGANIZATION | 0.99+ |
Monday | DATE | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
40% | QUANTITY | 0.99+ |
Dave | PERSON | 0.99+ |
$11.8 billion | QUANTITY | 0.99+ |
2018 | DATE | 0.99+ |
Jesse | PERSON | 0.99+ |
Verizon | ORGANIZATION | 0.99+ |
December 2020 | DATE | 0.99+ |
23% | QUANTITY | 0.99+ |
five bucks | QUANTITY | 0.99+ |
Sagar Kadakia | PERSON | 0.99+ |
Sagar | PERSON | 0.99+ |
1.5% | QUANTITY | 0.99+ |
10,000 foot | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Boston | LOCATION | 0.99+ |
7Park Data | ORGANIZATION | 0.99+ |
verizon | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
last week | DATE | 0.99+ |
last quarter | DATE | 0.99+ |
today | DATE | 0.99+ |
two companies | QUANTITY | 0.99+ |
350 services | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
2021 | DATE | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
over $600 million | QUANTITY | 0.99+ |
last Friday | DATE | 0.99+ |
Bain | ORGANIZATION | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
first slide | QUANTITY | 0.99+ |
667 million | QUANTITY | 0.99+ |
Praveen | PERSON | 0.99+ |
Cube | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
three providers | QUANTITY | 0.98+ |
tomorrow | DATE | 0.98+ |
about 2% | QUANTITY | 0.98+ |
Redshift | TITLE | 0.98+ |
about 1.5% | QUANTITY | 0.98+ |
20 plus markets | QUANTITY | 0.98+ |
six 700 million rows | QUANTITY | 0.98+ |
first terabyte | QUANTITY | 0.98+ |
Michael Delzer | PERSON | 0.98+ |
2:00 PM | DATE | 0.98+ |
Snowflake | TITLE | 0.98+ |
two | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
three | QUANTITY | 0.98+ |
last year | DATE | 0.98+ |
Michael Woodacre, HPE | Micron Insight 2019
>>live from San Francisco. It's the Q covering Micron Insight 2019. Brought to you by Micron. >>Welcome back to Pier 27 sentences. You're beautiful day here. You're watching the Cube, the leader in live tech coverage recovering micron inside 2019 hashtag micron in sight. My co host, David Floy er and I are pleased to welcome Michael Wood, Acre Cube alum and a fellow at Hewlett Packard Enterprise. Michael, good to see you again. Thanks. Coming on. >>Thanks for having me. >>So you're welcome? So you're talking about HBC on a panel today? But of course, your role inside of HP is is a wider scope. Talk about that a little bit. >>She also I'm the lead technologists in our Compute Solutions business unit that pack out Enterprise. So I've come from the group that worked on in memory computing the Superdome flex platform around things like traditional enterprise computing s it, Hannah. But I'm now responsible not only for that mission critical solutions platform, but also looking at our blades and edge line businesses. Well said broader technology. >>Okay. And then, of course, today we're talking a lot about data, the growth of data and As you say, you're sitting on a panel talking about high performance computing and the impact on science. What are you seeing? One of the big trends in terms of the intersection between data in the collision with H. P. C and science. >>So what we're seeing is just this explosion of data and this really move from traditionally science of space around how you put equations into supercomputers. Run simulations. You test your theories out, look at results. >>Come back in a couple weeks, >>exactly a potential years. Now. We're seeing a lot of work around collecting data from instruments or whether it's genomic analysis, satellite observations of the planner or of the universe. These aerial generating data in vast quantities, very high rates. And so we need to rethink how we're doing our science to gain insights from this massive data increase with seeing, >>you know, when we first started covering the 10th year, the Cuban So in 2010 if you could look at the high performance computing market as sort of an indicator of some of the things that were gonna happen in so called big data, and some of those things have played out on I think it probably still is a harbinger. I wonder, how are you seeing machine intelligence applied to all this data? And what can we learn from that? In your opinion, in terms of its commercial applications. >>So a CZ we'll know this massive data explosion is how do we gain insights from this data? And so, as I mentioned, we serve equations of things like computational fluid dynamics. But now things are progressing, so we need to use other techniques to gain understanding. And so we're using artificial intelligence and particularly today, deep learning techniques to basically gain insights from the state of Wei. Don't have equations that we can use to mind this information. So we're using these aye aye techniques to effectively generate the algorithms that can. Then you bring patterns of interest to our you know, focused of them, really understand what is the scientific phenomenon that's driving the things particular pattern we're seeing within the data? So it's just beyond the ability of the number of HPC programmers, we have the sort of traditional equation based methodologies algorithms to gain insight. We're moving into this world where way just have outstripped knowledge and capabilities to gain insight. >>So So how does that? How is that being made possible? What are the differences in the architecture that you've had to put in, for example, to make this sort of thing possible? >>Yeah, it's it's really interesting time, actually, a few years ago seemed like computing was starting to get boring because wears. Now we've got this explosion of new hardware devices being built, basically moving into the more of a hetero genius. Well, because we have this expo exponential growth of data. But traditional computing techniques are slowing down, so people are looking at exaggerate er's to close that gap and all sorts of hatred genius devices. So we've really been thinking. How do we change that? The whole computing infrastructure to move from a compute centric world to a memory centric world? And how can we use memory driven computing techniques to close that gap to gain insight, so kind of rethinking the whole architectural direction basically merge, sort of collapsing down the traditional hierarchy you have, from storage to memory to the CPU to get rid of the legacy bottlenecks in converting protocols from process of memory storage down to just a simple basically memory driven architecture where you have access to the entire data set you're looking at, which could be many terabytes to pad of eyes to exabytes that you can do simple programming. Just directly load store to that huge data set to gain insights. So that's that's really changed. >>Fascinating, isn't it? So it's the Gen Z. The hope of Gen Z is actually taking place now. >>Yes, so Gen Z is an industry led consulting around a memory fabric and the, you know, Hewlett Packard Enterprise Onda whole host of industry partners, a part of the ecosystem looking at building a memory fabric where people can bring different innovations to operate, whether it's processing types, memory types, that having that common infrastructure. I mean, there's other work to in the industry the Compute Express Link Consortium. So there's a lot of interest now in getting memory semantics out of the process, er into a common fabric for people to innovate. >>Do you have some examples of where this is making a difference now, from from the work in the H B and your commercial work? >>Certainly. Yeah, we're working with customers in areas like precision medicine, genomex basically exaggerating the ability to gain insights into you know what medical pathway to go on for a particular disease were working in cybersecurity. Looking at how you know, we're worried about security of our data and things like network intrusion. So we're looking at How can you gain insights not only into known attacking patterns on a network that the unknown patents that just appearing? So we're actually a flying machine learning techniques on sort of graft data to understand those things. So there's there's really a very broad spectrum where you can apply these techniques to Data Analytics >>are all scientists now, data scientists. And what's the relationship between sort of a classic data scientist, where you think of somebody with stats and math and maybe a little bit of voting expertise and a scientist that has much more domain expertise you're seeing? You see, data scientists sort of traversed domains. How are those two worlds coming together? >>It's funny you mentioned I had that exact conversation with one of the members of the Cosmos Group in Cambridge is the Stephen Hawking's cosmology team, and he said, actually, he realized a couple of years ago, maybe he should call himself a day two scientists not cosmologist, because it seemed like what he was doing was exactly what you said. It's all about understanding their case. They're taking their theoretical ideas about the early universe, taking the day to measurements from from surveys of the sky, the background, the cosmic background radiation and trying to pair these together. So I think your data science is tremendously important. Right now. Thio exhilarate you as they are insights into data. But it's not without you can't really do in isolation because a day two scientists in isolation is just pointing out peaks or troughs trends. But how do you relate that to the underlying scientific phenomenon? So you you need experts in whatever the area you're looking at data to work with, data scientists to really reach that gap. >>Well, with all this data and all this performance, computing capacity and almost all its members will be fascinating to see what kind of insights come out in the next 10 years. Michael, thanks so much for coming on. The Cube is great to have you. >>Thank you very much. >>You're welcome. And thank you for watching. Everybody will be right back at Micron Insight 2019 from San Francisco. You're watching the Cube
SUMMARY :
Brought to you by Micron. Michael, good to see you again. So you're talking about HBC on a panel today? So I've come from the As you say, you're sitting on a panel talking about high performance computing and the impact on science. traditionally science of space around how you put equations into supercomputers. to gain insights from this massive data increase with seeing, you know, when we first started covering the 10th year, the Cuban So in 2010 if So it's just beyond the ability of the number merge, sort of collapsing down the traditional hierarchy you have, from storage to memory So it's the Gen Z. The hope of Gen Z is actually a memory fabric and the, you know, to gain insights into you know what medical pathway to go on for a where you think of somebody with stats and math and maybe a little bit of voting expertise and So you you need experts in whatever to see what kind of insights come out in the next 10 years. And thank you for watching.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Michael | PERSON | 0.99+ |
2010 | DATE | 0.99+ |
Michael Wood | PERSON | 0.99+ |
David Floy er | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Cambridge | LOCATION | 0.99+ |
Compute Express Link Consortium | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
HP | ORGANIZATION | 0.99+ |
HBC | ORGANIZATION | 0.99+ |
Cosmos Group | ORGANIZATION | 0.99+ |
Hannah | PERSON | 0.99+ |
Hewlett Packard Enterprise | ORGANIZATION | 0.99+ |
Stephen Hawking | PERSON | 0.99+ |
Michael Woodacre | PERSON | 0.99+ |
today | DATE | 0.98+ |
two worlds | QUANTITY | 0.98+ |
Thio | PERSON | 0.95+ |
Micron | ORGANIZATION | 0.94+ |
2019 | DATE | 0.93+ |
first | QUANTITY | 0.93+ |
10th year | QUANTITY | 0.92+ |
two scientists | QUANTITY | 0.91+ |
Micron Insight | ORGANIZATION | 0.87+ |
next 10 years | DATE | 0.86+ |
couple of years ago | DATE | 0.83+ |
Acre Cube | ORGANIZATION | 0.78+ |
One | QUANTITY | 0.76+ |
HPE | ORGANIZATION | 0.7+ |
Cuban | OTHER | 0.69+ |
a few years ago | DATE | 0.69+ |
micron | ORGANIZATION | 0.68+ |
Micron Insight 2019 | TITLE | 0.67+ |
Cube | PERSON | 0.67+ |
Z. | OTHER | 0.54+ |
Gen Z | OTHER | 0.54+ |
couple | QUANTITY | 0.5+ |
sentences | QUANTITY | 0.47+ |
micron | TITLE | 0.43+ |
27 | OTHER | 0.39+ |
Superdome flex | COMMERCIAL_ITEM | 0.37+ |
Pier | TITLE | 0.36+ |
Ken Xie, Fortinet | Fortinet Accelerate 2019
>> live from Orlando, Florida It's the que covering Accelerate nineteen. Brought to you by Ford. >> Welcome back to the Q. We air live in Orlando, Florida At Fortinet Accelerate twenty nineteen Lisa Martin with Peter Burst. Pleased to welcome back one of our alumni on ly the CEO and founder of Fortinet. Kensi. Ken, thank you so much for joining Peter and me on the Cuban. Thanks for having the Cube back at accelerate. >> Yeah, I love to be here again. Yeah, Thank you. >> So, so quick by the numbers Can Kino. This morning was awesome. Loved the music and all the lights to start four thousand attendees from forty countries. You guys now have about three hundred eighty five thousand customers globally. Your revenue and F eighteen was up twenty percent year on year. I could go on and on. Lots of partners, lots of academies, tremendous growth. Talk to us about in the evolution of security. Where are we today and why is supporting that so well positioned to help customers dramatically transform security >> First world happy to see all the partner of the cosmos were come here. And also we keep him like every year we in this program also is a great program on another side. Like I say, securities of wherever dynamic space you need to keep in landing on We see more and more people come here s o that's we'LL be happy to discuss in the new technology the new market opportunity and also the new trend on DH Also What we see is a the space is so old and I'm making Also we see a lot of people keeping come here for the training for other sins And also I love the music make make us feel young again So But I >> think one of the reasons why security is so dynamic it is you don't for example, in the server world you don't have, you know you know gangs of bad guys running around with baseball bats trying to eat your servers. In the security world, you have people trying to enable the business to be able to do more, but also people constantly trying to tear the business down. And that tension drives a lot of invention and requires a lot of innovation. How is that changing? We're driving some of the key trends and networks and network security >> Yeah, that's where like I presented this morning. Wait, You see, with more device connected, Actually motive, I Some people being connect today and eventually in few years we'LL be calm. Motive eyes on people. There also is all the five G or icy went technology you can make is connected faster, more broadly reached. And then there's a more application More data also come to the Internet. So that's all you quist tax servants. There's all additional risk We'LL have all this connection. We have all these data transfer to all these different diversity on people. So that's all security business, right? Because secure to have the address where they now walking cannot really are dresses above the connection above the speed. So we have our dressing a content layered application layer the device user layer all regionally or country lier s O. That's making the security always keeping foreign faster than the night walk in the night. He spending on the study become the biggest sector United ninety idea spending environment. That's also one time we just feel security also need a study merger convert together is not working because no longer oh now will get only kind of the speed I can activities secure, canniness and bob. They had to be working together to smart rain route. In a data, put a low risk area tow without a polluted like transfer. All this conscience on that way, see, is the two industries that emerged together. That's where Koda security driven that walk are the arson about how this kind of we see today the mobile on cloud started replacing the traditional PC, right? So about going forward, the wearable divine's all the glass and we award study replaced the mobile. You don't have the whole mobile phone the season, while they're probably in your eyes on the same piled. A smart car that's my home, the wise every single connecting way Are you walking? Like if I walking here our sins related my information on power for me so I don't have to carry innocents, so that's going for you. A few years we'LL be happy. First, security will be part of this space. How this will be going forward contrato today The mobile the cloud way also have some discussion about that one. So we need to prepare for all this because that's how fortunate being founded. That's how our culture about generation, about long career advancement. So that's where we want to make sure the technology the part already for this chance. That's what gave the use of the past benefit of leverage of connection. Same time, lower the risk >> organ has taken an approach in the marketplace of Let Me Step Back. Put it this way. We all talk about software to find everything in virtual ization, and that's clearly an important technology and important trend. Ford has taken advantage of that as well, but the stuff doesn't run. All that's offered stuff doesn't run on hamsters. It runs in hardware. Unfortunate has made taking a strategic position, and it's been a feature of your nearly twenty year history to continuously invest in hardware and open up the performance aperture. Increase the size of the bucket of that hardware. How is that? Both altered your ability to add additional functionality, get ahead of the curve relative to competition, but also enabled your ecosystem to do a lot of new and interesting things that we're not seeing on other another network security companies? >> Yeah, that's why I totally agree with you. Israeli howto unable the past ecosystem for everybody playing a space for the partners of his provider, carrier enterprise, on the photo leverage technology benefit. More broadly, Cosmo base is very important. That's where we feel like a sulfur cloud. They do study in kind of a change, a lot of sense. But you also need a balance among clothes. Suffers were important, but also the hardwork also very important. All right, so that's the hybrid. More post the power on the sulfur. Both the cloud at age both have equal equal weight. Equally important, going forward How to leverage all this post is also also kind of very important for the future growth of future trend Another So you also can see like a mission. Uh, will you have the immersive device? We'LL have some, like security applied in tow Storage in that work in small Sadie, you also need a bad lie. Security be part of it. No, just security. I don't cop as a cost of additional Whatever process are all since, But you know, once you make it secure to be part ofthe like we mentioned a security for even that Working security driven like a future like a wearable device or the other since without it will be huge ecosystem going forward. That's where is the chip technology you can. Bad. We just saw Fervor is also additional servants. We can all walk in together. So that's where we want to look at the whole spectrum. There, make sure different part all can walk in together on also different technology. No, just limiting some part of it. I make sure the faux technologists face hole. Attack service can be a poor tag. And also we can leverage for the security of the high table addition. Opinions? You know, this conducted a war. >> This is what you're calling the third generation of Security? >> Yes, there's more. You for structure security. That's the whole security compared tto first dinners and second generation is our security just secured himself right. So you don't involve with other night walking star recharge the infrastructure? No, because Because they view everything you inside the companies secure You only need a guard at the door This Hey, who has come here? Anything inside I'll find But with today all the mobile pouring on Devise all the data everywhere Go outside the company you need to make sure security for all of the data. So that's the new trend. So now the border disappeared. So it doesn't matter. You said the company or not, is no longer secure anymore because you can use the mobile, the access rights o outside. All people can also come here with data also go out. So that's where the infrastructure security neither give or imposing their work inside on points. I under the cloud of the age and all this a different device on the diversity. Why? So you're even your mobile phone? Hi! Still working together. So it's a much bigger before structure. Much bigger are traceable space. Now that's making secure, more exciting. >> Well, we have gotten used over the past twenty years of building applications that operate on somebody else's device, typically a PC or mobile phone. And we've learned how to deal with that. You're suggesting that we're actually going to be integrating our systems with somebody else's systems at their edge or our edge on a deeply intimate level and life and death level. Sometimes on that, obviously, place is a real premium on security and networking whatnot. So how does the edge and the cloud together informed changes and how we think about security, how we think about networking, >> That's where, like I think age and a cloud they each complaint. Different role, because architecture. So the cloud has a good C all the bigger picture. They're very good on the provisioning. There could archiving cloud, also relatively slow, and also you can see most of data generated and age. That's where, whether you're immersive device, all your mobile, whatever ages were we called a digital made physical, and that's all the people in Device Connect. So that's where, like a seven eighty percent data, Carrion a probably never traveled to the club. They need a processed locally. They also need have the privacy and autonomy locally and also even interactive with other eighty vice locally there. So that's what we see is very important. Both the cloud on age security can be addressed together and also celebrity of architecture, that I say the cloud is good for detection so you can see a something wrong. You can cry the information, but the age new market on the provisions, because prevention need to be really time needed back, moreover, quickly because a lot of application they cannot afford a late Nancy like where do the V I. R. Even you slow down in a microsecond. Pickle feet is the famous signals. You also see the also drive a car. If you react too slow, you may hear something right the same scene for a lot of harder. Even you. Commerce, whatever. If you not response picking out within a half second, people may drop the connection. The memos are married, so that's what the late and see the speed on DH that's making the club play there at all into all this management on their age, playing hero in a really kind on Barlow. Ladies, you're really kind reaction there. So what? That's where we see the both side need to play their role on important transposed market. You said that just a one cloud, which I feel a little bit too hard right now. Try to cool down a little bit of our same age. Also, we see a very important even going forward what I been a bad security in age >> with this massive evolution that you've witnessed for a very long time. As the head of forty nine last nearly twenty years EJ cloud. How how dramatically technology changes in such a short period of time. I'm curious. Can How has your customer conversations evolved in terms of, you know, ten years ago were you talk ng more to security professionals? And now are you talking more to the C suite? As security is fundamental? Teo Digital transformation and unlocking tremendous value in both dollars in society impact has that conversation elevated as security has changed in the threat landscape has changed. >> Yeah, they do go to the board level, the CEO level now compared to like a ten, twenty years ago. Probably gaiety people maybe see so level, because security become probably the most important part of it. Now they keep you got a high high percentage that ikey spending there because when we connect everything together, we can make all the people all this business together to be on the connection. That's where security handled up, right? So that's where we see security studying kind of more. You hope me more important now. But another side, also the space also changing over quick. So that's where we always have to learn it. Woman engaged with Cosmo partner here. That's where this event is about way keeping less into what's the issue they have, how we can help the dress. All these security really the usual. Some even be honest security. Go to like a connection you for structure, some other, like architectural design, whatever their penis model there. So that's all we're very important on. Like I said, security space we need to keep in Lenny every day. Even I spent a few hours a day to Lenny. I You don't feel ready? Can K child? Oh, they >> said, It's a very dynamic world security world. >> You have our dynamic, the knowledge base, the technology refreshed quickly. Way always had to be Len have training. That's where he also see Try to position forty Niners lending company. So that's where we all for the because training program and all the train is afraid for partner for customers. All this kind is really it's a big investment. That's where a lot of people say, Oh, how can you? You've asked more in the training. You said of all come better. You must move your marketing. I say journeys of over a long term benefit. When people get trained, they also see Hey, what's the pants technology? So that's where a lot of organization, a lot of investment, really looking for. How five years here come benefit of space can benefit. The car's my partner, so that's all we see. Training's far long time measurement see modern technology. >> So can you've talked in the keynote? You've talked in the Cube about how networking security come together on how, as they move forward, they're going in form. Or they'LL have an impact on business and have an impact and other technologies. There's a lot of technology change when you talk to network in professional or even your own employees. What technologies out there do you think are going to start impacting how security works? Micro services containers? Are there any technologies that Ford that's looking at and saying, We gotta watch that really closely and that networking professionals have to pay more attention to. I >> have to say pretty much all of them, right? So all these Michael, all this contender technology, micro segmentation, according computing, the immersion lending all this is all very important because security has deal with all this different new technology application on like it was all this a huge, competent power raised on the cost lower ball corner computer. And maybe some of the old technology may not really work any more for some additional risks. Like where the equipment can be break by cute from the computing or some moderate eventually can also kind of take over. All this country is always we tryto tryto learn, tryto tried. Okay, chop every day. Hey, that that's what I say is that's so exciting. Keep you wake up, Keep your Lenny everyday, which I enjoy. But at the same time, there's a lot of young people they probably even even better than us to catch the new technology. >> Oh, no. Oh, no, no, no. >> Yeah. Somehow, my kids can play the fool much greater than mere. That's always the way >> we want to thank you so much for joining Peter and me on the kid this afternoon for having the Cube back at forty nine. Accelerate and really kind of talking about how you guys are leading in the space and we're gonna be having more guests on from Fortinet. And your partner's talking about educate ecosystems and technology that you talked about in your keynote. So we thank you again for your time. And we look forward to a very successful day here. >> Oh, thank you. Thank you very much. You enjoy all this programme for many years. Thank you. >> Excellent. We love to hear that. We want to thank you for watching the Cube for Peter Burress. I'm Lisa Martin. You're watching the Cube. >> Thank you.
SUMMARY :
live from Orlando, Florida It's the que covering and me on the Cuban. Yeah, I love to be here again. Loved the music and all the lights to start four thousand attendees from forty a lot of people keeping come here for the training for other sins And also I love the music in the server world you don't have, you know you is all the five G or icy went technology you can make is connected faster, functionality, get ahead of the curve relative to competition, but also enabled your ecosystem All right, so that's the hybrid. You said the company or not, is no longer secure anymore because you can use So how does the edge and the cloud together DH that's making the club play there at all into all this management on their age, security has changed in the threat landscape has changed. be on the connection. You have our dynamic, the knowledge base, the technology refreshed quickly. There's a lot of technology change when you talk to network in professional or even your own And maybe some of the old technology may not really work any more for some additional That's always the way So we thank you again for your time. Thank you very much. We want to thank you for watching the Cube for Peter Burress.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
Ken Xie | PERSON | 0.99+ |
Fortinet | ORGANIZATION | 0.99+ |
Ford | ORGANIZATION | 0.99+ |
Orlando, Florida | LOCATION | 0.99+ |
Peter Burst | PERSON | 0.99+ |
five years | QUANTITY | 0.99+ |
forty countries | QUANTITY | 0.99+ |
second generation | QUANTITY | 0.99+ |
Peter Burress | PERSON | 0.99+ |
First | QUANTITY | 0.99+ |
Ken | PERSON | 0.99+ |
Kensi | PERSON | 0.99+ |
Nancy | PERSON | 0.99+ |
today | DATE | 0.99+ |
Michael | PERSON | 0.99+ |
both side | QUANTITY | 0.98+ |
one time | QUANTITY | 0.98+ |
one | QUANTITY | 0.98+ |
both dollars | QUANTITY | 0.98+ |
third generation | QUANTITY | 0.98+ |
Lenny | PERSON | 0.98+ |
2019 | DATE | 0.98+ |
four thousand attendees | QUANTITY | 0.98+ |
Both | QUANTITY | 0.98+ |
two industries | QUANTITY | 0.97+ |
about three hundred eighty five thousand customers | QUANTITY | 0.97+ |
forty nine | DATE | 0.97+ |
Can Kino | ORGANIZATION | 0.96+ |
both | QUANTITY | 0.96+ |
ten years ago | DATE | 0.96+ |
seven eighty percent | QUANTITY | 0.96+ |
one cloud | QUANTITY | 0.95+ |
Fervor | ORGANIZATION | 0.95+ |
Len | PERSON | 0.95+ |
first dinners | QUANTITY | 0.94+ |
Accelerate nineteen | ORGANIZATION | 0.93+ |
ten, twenty years ago | DATE | 0.92+ |
twenty percent | QUANTITY | 0.92+ |
forty nine | QUANTITY | 0.91+ |
ninety idea | QUANTITY | 0.91+ |
this morning | DATE | 0.91+ |
eighty vice | QUANTITY | 0.89+ |
this afternoon | DATE | 0.86+ |
nearly twenty year | QUANTITY | 0.85+ |
a half second | QUANTITY | 0.84+ |
eighteen | QUANTITY | 0.81+ |
tryto | PERSON | 0.8+ |
each complaint | QUANTITY | 0.8+ |
Koda | ORGANIZATION | 0.79+ |
forty Niners | QUANTITY | 0.79+ |
This morning | DATE | 0.78+ |
Carrion | ORGANIZATION | 0.78+ |
Cosmo | ORGANIZATION | 0.78+ |
Cube | COMMERCIAL_ITEM | 0.78+ |
United | ORGANIZATION | 0.77+ |
partners | QUANTITY | 0.74+ |
Cuban | OTHER | 0.73+ |
a microsecond | QUANTITY | 0.69+ |
nineteen | ORGANIZATION | 0.67+ |
twenty years | QUANTITY | 0.66+ |
few hours a day | QUANTITY | 0.65+ |
Teo | PERSON | 0.64+ |
Sadie | ORGANIZATION | 0.63+ |
Israeli | OTHER | 0.6+ |
G | OTHER | 0.54+ |
twenty | QUANTITY | 0.54+ |
cosmos | LOCATION | 0.51+ |
Barlow | ORGANIZATION | 0.51+ |
EJ cloud | ORGANIZATION | 0.51+ |
Accelerate | ORGANIZATION | 0.5+ |
single | QUANTITY | 0.49+ |
Cube | TITLE | 0.48+ |
past | DATE | 0.44+ |
John Mracek, Imanis Data | Microsoft Ignite 2018
>> Live from Orlando, Florida, it's theCUBE, covering Microsoft Ignite. Brought to you by Cohesity and theCUBE's ecosystem partners. >> Welcome back to theCUBE'S coverage of Microsoft Ignite 2018 here in Orlando. I'm Stu Miniman, and happy to welcome back to the program John Mracek who's the CEO of a Imanis Data. It's our first time at the show, but not your first time on theCube. Thanks so much for joining us and tell us we caught up with you in New York City talking about kind of the AI, analytics, all those things there. what what what brings a Imanis to Microsoft Ignite? >> So this has been a great show for us. And what I really see happening here is there's a vibrancy that probably didn't exist in Microsoft events, maybe four or five years ago. Because Microsoft really getting their act together on the whole how you migrate and bring people to the Azure. Right, because that's their agenda. And so where we fit in there is in our data management platform. We help customers migrate to a Azure. So whether it's moving your Hadoop workloads to Azure, or one of the products that's been featured here that we've gotten a lot of Microsoft support on is our migration tool to move from MongoDB to Cosmo DB. So we play really well into the migration story and it really leverages our platform. >> Yeah, one of the questions we talk about all the time is customers trying to figure out where things live and, well, it's like your cloud strategy. Things are changing over time. Customers have really multi-cloud environments, which really means they're doing a lot of different things and a lot of times they need to move them and sort those out. So what are the challenges you're seeing? How do you help those businesses make decisions today and be able to move things as needed in the future? >> Yeah, what we see and what we're playing into is really this evolution. You know, solutions really drive technologies. So in a large enterprise, you might have a division or a particular group that says, I need this BI or analytics tool and I need a big data platform to do it. So they build this. They build on top of some either NoSQL or Hadoop and then they've got this great solution. Well, that happens four or five times across the enterprise, and at some point in the enterprise, the CIO or somebody says, "you know, "we kind of got all these distributed data systems, "and like, who's managing them? "How is that data being moved "to your point about cloud migration? "Well, these are on-prem, these are in the cloud. "We want to put them all in the cloud, how do we do that?" And so that's where we're seeing as kind of the call for our product, which is, okay, I need a central way to manage and manipulate this data, as a fundamental problem. >> Yeah, so we all know that data is fundamental to a business. It's one of the most important things. We can use all the tropes of, it's the new oil or anything like that. But when you dig down, it's a lot of complexity into how, how do I get data? How do I manage data? How do I share data? We're sitting here in Cohesity, is the where we are in the booth. Can you help us understand, what are the solutions that you complement in the data space? What are the solutions that you replace? or a modern version or compete against in this space? >> So the way to look at us, we're at our most general, we're a platform for moving data from one platform to another. Okay, and that has many different use cases. But where we're getting a lot of customer uptake is on the backup recovery. It's like, I've got it here, I want to make a backup. We also see a lot in terms of migration, whether it's the Mongo to DB or I want to move from on-prem to cloud or cloud to cloud. And where we fit is if you look there's a legacy providers who don't traditionally go after the NoSQL and Hadoop space. And so where were a perfect complement to either those companies or folks like Cohesity. We have partnerships with Cohesity, Veem and others where they get in RFP or they're talking to a customer and the customer has a specific request for data management solution for NoSQL or Hadoop platforms. And that's where we come in. Because that's what we focused on exclusively from day one. >> Yeah, well, being at a Microsoft show, I mean, applications are central to so much and Microsoft does. Everything from Office, but on the data side, we spend a lot of time this week talking about SQL. Talking about Cosmos DB and cool new things they're doing. And of course, Microsoft's playing in a lot of the modern areas. We see them, big developers base here, even more of it at the Microsoft Build show, what do you see in the Microsoft space on the application modernization? Sounds like that would tie in quite a bit to what you're helping customers with. >> Yeah, so we have customers across all the cloud providers. But what we see in the Microsoft case is really people looking for maybe global easy deployment, customer facing as typical examples. So people who are really pushing the envelope, frankly. And there's almost like a bi-modal distribution. There's kind of some folks who are still trying to retrofit the old world and then others who are really embracing some of the new platforms. >> I'm not sure if you were at the keynote on Monday, Satya Nadella unveiled the Open Data Initiative. We've got Adobe and SAP and Microsoft there. I was talking to one analyst and reading some reports, and I'm like, well, it's not a coincidence that this was launched the week of Salesforce. Salesforce has a lot of data. Maybe that's a little bit of an attack there. But data across these big providers is important. I want to be able to share and leverage my data. You're in the data business. But what viewpoint you have of some of these really big providers of the application as they're going through their digital transformation, and making how do customers get the best value out of their data? >> So, my background, most recent background, I was in an ad tech company, where we're all big data. And the whole play there, is how do you manage your audiences, right? How do you have a unifying way to look at audiences? And so this is what's playing out on a more higher level, a more general level of how do I normalize and create a unified view of the customer and consistent data so that I can then manage it. And so that's an essential requirement to get the maximum value at out of that. Once you have that and you're in your data repositories, it's incredibly critical to protect them, to be able to orchestrate and move around. Where we fit in and how we see it is, these things are data, to reuse the term is the new oil and the new gold. And companies are realizing that it's really time to protect this data. I put all this investment into getting unified view of data. Wow, what are we doing about how do we back it up, restore it and move it? >> It's interesting, I've watched the space long enough. You go back kind of BI and DW days, go through big data. Now, we talk about a lot more of the analytics in the intelligence there. Help us as to, what are we actually realizing today that we were been talking about for years, and what what are still some of the stumbling blocks as to what we need to mature as an industry to really help unlock data. >> So, I mean, there's clearly the, what's driven a lot of the machine learning AI is the availability of data. It wasn't so much algorithms change dramatically, it was, we have a, so all the machine learning applications are really benefiting from this. But what we see as you know, some of the immediate things with our customers, is they're using big data as they create their front ends, engage with their customers. So how do they have the most up to date, real-time information to whether it's present an offer to a customer, provide customer service. So a lot of the use case we see is in that really bread and butter customer-level interactions and having an appropriate database to front end that process. >> Alright, so one of the biggest challenges of our time is really talking about distributed architecture. When I talk to companies scale comes on a lot, but it means very different things to different people. Can maybe talk about what you're hearing from customers, and how your solution helps customers for a variety of implementations. >> Yeah, so, we typically are targeting and working with customers in the 10s to 100s of terabytes. Up to, and our system handles up into into the petabytes. Typically, what we see is an evolution is, as I said earlier, somebody will develop a solution in a particular division, and then realize we've got this asset to protect. And then so IT starts to get involved and basically look at it holistically. So, we had one of our prospects, we went in and pitched at an SVP level and said, "what are the problems you're facing?" and it was basically this, I have all these silos of data. To get the maximum value out of them, and have a uniform look, whether it's look at our customers, the market, I need a uniform view to do my BI and AI. And so they brought us in and said, "Okay, paint a picture of how I can continue to have "these groups run autonomously and run their solutions, "yet at the same time, give me a unified view "and make me feel comfortable "that I've been able to protect the data, "move the data, massage the data." >> Great. Talk to me, when I look at this show, I see a lot of customers are still doing things, I'm trying to think how to say it nicely. Kind of the old way, it's like, if you look at them five years ago, is like, okay, Windows 2019's there great. I'll get there in five years, you play with a lot of more modern applications. What do you hear from customers? What, what is the profile of a customer that is, taking advantage and being competitive in the world? And what do you advise companies that maybe are a little bit behind the eight ball. >> So, you're right, and there's a really big spectrum of where people are in the adoption curve. And the way we look at it, if people were waiting for it, you know, when somebody goes, "Yeah, we're looking at setting up a big data system", it's like, okay, we'll talk to you in a year once you get the basics set up. But I see kind of two types of things. There's, say, the smaller, more aggressive companies, who are willing to move forward and say, "I just got to create a product, I don't care how I do it, "I don't have legacy issues." And they've moved ahead, and they're starting to get to the point where they're like, "Okay, we're mature enough where we actually need to spend on data management." The more typical case though is, as I said earlier. It's like these these new apps, that larger companies might have bleeding edge groups. So it's not being driven centrally. And so my, you asked about advice, right? So if you're sitting in the top of large enterprise and say, "Well, how do we get there. "There's kind of the tops down, "I need somebody to help me figure out." But there's also, let 1,000 flowers and let there be some kind of anarchy, if you will. Breaking the model, breaking the mold. Let people go build stuff and then over time start to figure out how to assimilate. So that'd be the biggest single biggest advice is, Yeah, you want to do the top down, but you really want to do the bottoms up. Because those people really know how to use the technology to provide a solution. >> Yeah, absolutely. Guy Kawasaki let 1,000 flowers bloom out there and everything. All right. Help bring this in. What kind of customer conversations are you having this week? We talked to the top about, there's real good energy to this show. Definitely, I felt that. What would you share with your peers that haven't been at the show? >> So the topics here are typically around the migration. Whether it's like to like, moving an existing workload into Azure, or the transformation. We also announced the show cooperation with Microsoft on moving any of your NoSQL workloads to Cosmos DB. So most of the conversations here have been related to migration. Either of, if you will, within the same Hadoop family, or, you know, like to unlike. Going from something to Cosmos DB. And that goes back to your earlier point about people trying to figure out what to do. They know there's this imperative to move to the cloud, and they're trying to figure out how they do it in bite-sized chunks. Right and protect their business at the same time. >> Yeah, so you mentioned Cosmos DB. We had an interview earlier this week about Cosmos DB. I definitely heard some good buzz at the show, What is it about that is drawing customers to it and what's that enable for them? >> Two things that I'm aware of, that I've seen is, again, the global nature and the ability to just kind of deploy anywhere. But also, I've seen a little bit around the dynamic schemas and the ability to map between them as a very quick way to ingest data. So you can get up and running quickly, instead of doing a lot of manual work to start using it. So those are things that are going to win developers 'cause it makes their life easier. >> Alright, John I want to give you the final word. What should we look to see from Imanis over the next six to 12 months. >> So we're going to continue to push forward with our platform around data management. You've seen in some recent announcements that, where leveraging machine learning in a very concrete way to do anomaly detection around ransomware. And also for administrators to be able to basically set rules or set goals and have the software do it. And that really steams from the fact that we're using a big data platform and machine learning to solve the problem of well, if you're running a big data platform, how do you manage the data? So the whole DNA of the company is built around that, and from a go-to-market standpoint, you know, partnering with folks like Cohesity and others where you've already got people in market selling a broad solution but they're missing a piece. So the other thing you'll see from us, is more partner announcements as we go forward. Alright, well, John Mracek really appreciate all the updates on a Imanis Data. Congrats on the progress so far. And look forward to catching with you up at future show. >> Great, thank you. >> Alright, we'll be back with more coverage here. Day three of three days live coverage. Microsoft Ignite here in Orlando. I'm Stu Miniman and thanks for watching theCUBE.
SUMMARY :
Brought to you by Cohesity about kind of the AI, on the whole how you migrate and a lot of times they need to move them and at some point in the enterprise, What are the solutions that you replace? So the way to look at us, a lot of the modern areas. some of the new platforms. You're in the data business. And the whole play there, more of the analytics So a lot of the use case we see Alright, so one of the the 10s to 100s of terabytes. Kind of the old way, it's like, And the way we look at it, if that haven't been at the show? So most of the conversations here good buzz at the show, and the ability to map between them over the next six to 12 months. And look forward to catching with you up I'm Stu Miniman and thanks
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
John | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
John Mracek | PERSON | 0.99+ |
John Mracek | PERSON | 0.99+ |
Satya Nadella | PERSON | 0.99+ |
Orlando | LOCATION | 0.99+ |
New York City | LOCATION | 0.99+ |
Monday | DATE | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
first time | QUANTITY | 0.99+ |
1,000 flowers | QUANTITY | 0.99+ |
10s | QUANTITY | 0.99+ |
Guy Kawasaki | PERSON | 0.99+ |
Cohesity | ORGANIZATION | 0.99+ |
four | QUANTITY | 0.99+ |
one | QUANTITY | 0.99+ |
five times | QUANTITY | 0.99+ |
three days | QUANTITY | 0.99+ |
four | DATE | 0.99+ |
SAP | ORGANIZATION | 0.99+ |
Orlando, Florida | LOCATION | 0.98+ |
two types | QUANTITY | 0.98+ |
Cosmos DB | TITLE | 0.98+ |
NoSQL | TITLE | 0.98+ |
Two things | QUANTITY | 0.98+ |
one platform | QUANTITY | 0.98+ |
five years ago | DATE | 0.98+ |
Windows 2019 | TITLE | 0.97+ |
today | DATE | 0.97+ |
this week | DATE | 0.97+ |
this week | DATE | 0.97+ |
MongoDB | TITLE | 0.97+ |
single | QUANTITY | 0.97+ |
Imanis | ORGANIZATION | 0.96+ |
earlier this week | DATE | 0.96+ |
SQL | TITLE | 0.96+ |
theCUBE | ORGANIZATION | 0.96+ |
five years ago | DATE | 0.96+ |
five years | QUANTITY | 0.94+ |
Azure | TITLE | 0.94+ |
Day three | QUANTITY | 0.93+ |
Veem | ORGANIZATION | 0.93+ |
Salesforce | ORGANIZATION | 0.88+ |
100s of terabytes | QUANTITY | 0.87+ |
Microsoft Build | EVENT | 0.86+ |
Office | TITLE | 0.86+ |
Cosmo DB | TITLE | 0.86+ |
Cosmos | TITLE | 0.83+ |
Microsoft Ignite 2018 | EVENT | 0.82+ |
day one | QUANTITY | 0.82+ |
one analyst | QUANTITY | 0.81+ |
a year | QUANTITY | 0.8+ |
Imanis Data | ORGANIZATION | 0.72+ |
Hadoop | TITLE | 0.71+ |
theCube | ORGANIZATION | 0.69+ |
Data | PERSON | 0.67+ |
12 months | QUANTITY | 0.65+ |
Mongo | TITLE | 0.6+ |
theCUBE Insights | Microsoft Ignite 2018
>> Live from Orlando, Florida, it's theCUBE covering Microsoft Ignite. Brought to you by Cohesity and theCUBE's ecosystem partners. >> Welcome back everyone, we are wrapping up day three of Microsoft Ignite here in Orlando, Florida. CUBE's live coverage, I'm your host Rebecca Knight, along with Stu Miniman, my esteemed cohost for these past three days, it's been fun working with you, Stu. >> Rebecca, it's been a great show, real excited. Our first time at a Microsoft show and it's a big one. I mean, the crowds are phenomenal. Great energy at the show and yeah, it's been great breaking down this ecosystem with you. >> So, three days, what do we know, what did you learn, what is your big takeaway, what are you going to to go back to Boston with? >> You know, it's interesting, we've been all talking and people that I know that have been here a couple of years, I've talked to people that have been at this show for decades, this is a different show. There's actually a friend of mine said, he's like, "Well look, Windows pays the bills for a lot of companies." There's a lot of people that all the Windows components, that's their job. I mean, I think back through my career when I was on the vendor side, how many rollouts of Exchange and SharePoint and all these things we've done over the years. Office 365 been a massive wave that we watched. So Microsoft has a broad portfolio and they've got three anchor shows. I was talking with one of the partners here and he's like, "You know, there's not a lot of channel people "at this event, at VMworld there's a lot of channel people." I'm like, "Well yeah because there's a separate show "that Microsoft has for them." You and I were talking at an earlier analytics session with Patrick Moorhead and he said, "You know when I look at the buy versus build, "a lot of these people are buying and I don't "feel I have as many builders." Oh wait, what's that other show that they have in the Spring, it's called Microsoft Build. A lot of the developers have moved there so it's a big ecosystem, Microsoft has a lot of products. Everything from, my son's excited about a lot of the Xbox stuff that they have here. Heck, a bunch of our crew was pickin' up Xbox sweatshirts while they're here. But a lot has changed, as Tim Crawford said, this is a very, it feels like a different Microsoft, than it even was 12 or 24 months ago. They're innovating, so look at how fast Microsoft moves and some of these things. There's good energy, people are happy and it's still trying to, you know. It's interesting, I definitely learned a lot at this show even though it wasn't the most sparkly or shiny but that's not necessarily a bad thing. >> Right, I mean, I think as you made a great point about just how integral Microsoft is to all of our lives as consumers, as enterprise, the Xbox, the Windows, the data storage, there's just so much that Microsoft does that if we were to take away Microsoft, I can't even imagine what life would be like. What have been your favorite guests? I mean, we've had so many really, really interesting people. Customers, we've had partners, we're going to have a VC. What are some of the most exciting things you've heard? >> Yeah, it's interesting, we've had Jeffrey Snover on the program a couple of years ago and obviously a very smart person here. But at this show, in his ecosystem, I mean, he created PowerShell. And so many people is like, I built my career off of what he did and this product that he launched back in 2001. But we talked a little bit about PowerShell with him but then we were talking about Edge and the Edge Boxes and AI and all those things, it's like this is really awesome stuff. And help connecting the dots to where we hid. So obviously, big name guest star, always, and I always love talking to the customers. The thing I've been looking at the last couple of years is how all of these players fit into a multicloud world. And Microsoft, if you talk about digital transformation, and you talk about who will customers turn to to help them in this multicloud world. Well, I don't think there's any company that is closer to companies applications across the spectrum of options. Office 365 and other options in SaaS, all the private cloud things, you start with Windows Server, you've got Windows on the desktop, Windows on the server. Virtualization, they're starting to do hyperconversion everything, even deeper. As well as all the public cloud with Azure and developers. I talked to the Azure functions team while I was here. Such breadth and depth of offering that Microsoft is uniquely positioned to play in a lot of those areas even if, as I said, certain areas if the latest in data there might be some other company, Google, Amazon, well positioned there. We had a good discussion Bernard Golden, who's with Capital One, gave us some good commentary on where Alibaba fits in the global scheme. So, nice broad ecosystem, and I learned a lot and I know resonated with both of us, the "you want to be a learn it all, not a know it all." And I think people that are in that mindset, this was a great show for them. >> Well, you bring up the mindset, and that is something that Satya Nadella is really such a proponent of. He says that we need to have a growth mindset. This is off of the Carol Dweck and Angela Duckworth research that talks about how important that is, how important continual learning is for success. And that is success in life and success on the job and organization success and I think that that is something that we are also really picked up on. This is the vibe of Microsoft, this is a company, Satya Nadella's leadership, talking to so many of the employees, and these are employees who've been there for decades, these are people who are really making their career, and they said, "Yeah, I been here 20 years, if I had my way, "I'll be here another 30." But the point is that people have really recommitted to Microsoft, I feel. And that's really something interesting to see, especially in the tech industry where people, millennials especially, stay a couple years and then move on to the next shiny, new thing. >> Yeah, there was one of our first guests on for Microsoft said that, "Been there 20 years and what is different about "the Satya Nadella Microsoft to the others is "we're closer and listening even more to our customers." We talk about co-creation, talk about how do we engage. Microsoft is focusing even deeper on industries. So that's really interesting. An area that I wanted to learn a little bit more about is we've been talking about Azure Stack for a number of years, we've been talking about how people are modernizing their data center. I actually had something click with me this week because when I look at Azure Stack, it reminds me of solutions I helped build with converged infrastructure and I was a big proponent of the hyper-converged infrastructure wave. And what you heard over and over again, especially from Microsoft people, is I shouldn't think of Azure Stack in that continuum. Really, Azure Stack is not from the modernization out but really from the cloud in. This is the operating model of Azure. And of course it's in the name, it's Azure, but when I looked at it and said, "Oh, well I've got partners like "Lenovo and Dell and HPE and Sysco." Building this isn't this just the next generation of platform there? But really, it's the Azure model, it's the Azure operating stack, and that is what it has. And it's more, WSSD is their solution for the converged and then what they're doing with Windows Server 2019 is the hyper-converged. Those the models that we just simplify what was happening in the data center and it's similar but a little bit different when we go to things like Azure and Azure Stack and leads to something that I wanted to get your feedback on. You talk business productivity because when we talk to companies like Nutanix, we talk to companies like Cohesity who we really appreciate their support bringing us here, giving us this great thing right in the center of it, they talk about giving people back their nights and weekends, giving them back time, because they're an easy button for a lot of things, they help make the infrastructure invisible and allow that. Microsoft says we're going to try to give you five to ten percent back of your business productivity, going to allow you to focus on things like AI and your data rather than all the kind of underlying spaghetti underneath. What's your take on the business productivity piece of things? >> I mean, I'm in favor of it; it is a laudable goal. If I can have five to ten percent of my day back of just sort of not doing the boring admin stuff, I would love that. Is it going to work, I don't know. I mean, the fact of the matter is I really applaud what Cohesity said and the customers and the fact that people are getting, yes, time back in their day to focus on the more creative projects, the more stimulating challenges that they face, but also just time back in their lives to spend with their children and their spouse and doing whatever they want to do. So those are really critical things, and those are critical things to employee satisfaction. We know, a vast body of research shows, how much work life balance is important to employees coming to their office or working remotely and doing their best work. They need time to recharge and rest and so if Microsoft can pull that off, wow, more power to them. >> And the other thing I'll add to that is if you, say, if you want that work life balance and you want to be fulfilled in your job, a lot of times what we're getting rid of is some of those underlying, those menial tasks the stuff that you didn't love doing in the first place. And what you're going to have more time to do, and every end user that we talked to says, "By the way, I'm not getting put out of a job, "I've got plenty of other tasks I could do." And those new tasks are really tying back to what the business needs. Because business and IT, they need to tie together, they need to work together, it is a partnership there. Because if IT can't deliver what the business needs, there's other alternatives, that's what Stealth IT was and the public cloud could be. And Microsoft really positions things as we're going to help you work through that transition and get there to work on these environments. >> I want to bring up another priority of Microsoft's and that is diversity. So that is another track here, there's a lot of participants who are learning about diversity in tech. It's not a good place right now, we know that. The tech industry is way too male, way too white. And Satya Nadella, along with a lot of other tech industry leaders, has said we need more underrepresented minorities, we need more women, not only as employees but also in leadership positions. Bev Crair, who was on here yesterday, she's from Lenovo. She said that things are starting to change because women are buying a lot of the tech and so that is going to force changes. What do you think, do you buy it? >> And I do, and here's where I'd say companies like Lenovo and Microsoft, when you talk about who makes decisions and how are decisions made, these are global companies. Big difference between a multi-national company or a company that's headquartered in Silicon Valley or Seattle or anything versus a global company. You look at both of those companies, they have, they are working not just to localize but have development around the world, have their teams that are listening to requirements, understand what is needed in those environments. Going back to what we talked about before, different industries, different geographies, and different cultures, we need to be able to fit and work and have products that work in those environments, everything. I think it was Bev that talked about, even when we think about what color lights. Well, you know, oh well default will use green and red. Well, in different cultures, those have different meanings. So yeah, it is, it's something that definitely I've heard the last five to ten years of my career that people understand that, it's not just, in the United States, it can't just be the US or Silicon Valley creating great technology and delivering that device all the way around the world. It needs to be something that is globally developed, that co-creation, and more, and hopefully we're making progress on the diversity front. We definitely try to do all we can to bring in diverse voices. I was glad we had a gentleman from Italy shouting back to his daughters that were watching it. We had a number of diverse guests from a geography, from a gender, from ethnicity, on the program and always trying to give those various viewpoints on theCUBE. >> I want to ask you about the show itself: the 30,000 people from 5,000 different organizations around the globe have convened here at the Orange County Convention Center, what do you think? >> Yeah, so it was impressive. We go to a lot of shows, I've been to bigger shows. Amazon Reinvent was almost 50,000 last year. I've been to Oracle OpenWorld, it's like takes over San Francisco, 60 or 70,000. This convention center is so sprawling, it's not my favorite convention center, but at least the humidity is to make sure I don't get dried out like Las Vegas. But logistics have run really well, the food has not been a complaint, it's been good, the show floor has been bustling and sessions are going well. I was talking to a guy at breakfast this morning that was like, "Oh yeah, I'm a speaker, "I'm doing a session 12 times." I'm like, "You're not speaking on the same thing 12 times?" He's like, "No, no it's a demo and hands on lab." I'm like, "Oh, of course." So they make sure that you have lots of different times to be able to do what you want. There is so much that people want to see. The good news is that they can go watch the replays of almost all of them online. Even the demos are usually something that they're cloud enabled and they get on live. And of course we help to bring a lot of this back to them to give them a taste of what's there. All of our stuff's always available on the website of thecube.net. This one, actually, this interview goes up on a podcast we call theCUBE Insights. So please, our audience, we ask you, whether it's iTunes or your favorite podcast reader, go to Spotify, theCUBE Insights. You can get a key analysis from every show that we do, we put that up there and that's kind of a tease to let you go to thecube.net and see the hundreds and thousands of interviews that we do across all of our shows. >> Great, and I want to give a final, second shout out to Cohesity, it's been so fun having them, being in the Cohesity booth, and having a lot of great Cohesity people around. >> Yeah, absolutely, I mean, so much I wish we could spend a little more time even. AI, if we go back to the keynote analysis then, but you can watch that, I can talk about the research we've done, and said how the end user information that Microsoft can get access to to help people when you talk about what they have, the TouchPoint to Microsoft Office. And even things like Xbox, down to the consumer side, to understand, have a position in the marketplace that really is unparalleled if you look at kind of the breadth and depth that Microsoft has. So yeah, big thanks to Cohesity, our other sponsors of the program that help allow us to bring this great content out to our community, and big shout out I have to give out to the community too. First time we've done this show, I reached out to all my connections and the community reached back, helped bring us a lot of great guests. I learned a lot: Cosmos DB, all the SQL stuff, all the Office and Microsoft 365, so much. My brain's full leaving this show and it's been a real pleasure. >> Great, I agree, Stu, and thank you so much to Microsoft, thank you to the crew, this has been a really fun time. We will have more coming up from the Orange County Civic Center, Microsoft Ignite. I'm Rebecca Knight for Stu Miniman, we will see you in just a little bit. (digital music)
SUMMARY :
Brought to you by Cohesity and of Microsoft Ignite here I mean, the crowds are phenomenal. There's a lot of people that all the Microsoft is to all of our lives about Edge and the Edge Boxes and then move on to the Azure Stack and leads to I mean, the fact of the and get there to work that is going to force changes. that device all the way around the world. but at least the humidity is to make sure being in the Cohesity the TouchPoint to Microsoft Office. the Orange County Civic
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lenovo | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Satya Nadella | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Sysco | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Patrick Moorhead | PERSON | 0.99+ |
Tim Crawford | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
2001 | DATE | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
12 times | QUANTITY | 0.99+ |
Rebecca | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Seattle | LOCATION | 0.99+ |
Bernard Golden | PERSON | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
Boston | LOCATION | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
Jeffrey Snover | PERSON | 0.99+ |
Bev Crair | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
Nutanix | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
Office 365 | TITLE | 0.99+ |
Cohesity | ORGANIZATION | 0.99+ |
Italy | LOCATION | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
United States | LOCATION | 0.99+ |
Stu | PERSON | 0.99+ |
20 years | QUANTITY | 0.99+ |
Angela Duckworth | PERSON | 0.99+ |
Orlando, Florida | LOCATION | 0.99+ |
Windows | TITLE | 0.99+ |
San Francisco | LOCATION | 0.99+ |
thecube.net | OTHER | 0.99+ |
SharePoint | TITLE | 0.99+ |
one | QUANTITY | 0.99+ |
Orange County Convention Center | LOCATION | 0.99+ |
30,000 people | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Azure Stack | TITLE | 0.99+ |
ten years | QUANTITY | 0.99+ |
Windows Server 2019 | TITLE | 0.99+ |
ten percent | QUANTITY | 0.99+ |
CUBE | ORGANIZATION | 0.99+ |
60 | QUANTITY | 0.99+ |
Capital One | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
first time | QUANTITY | 0.99+ |
Azure | TITLE | 0.99+ |
Xbox | COMMERCIAL_ITEM | 0.98+ |
Exchange | TITLE | 0.98+ |
12 | DATE | 0.98+ |
30 | QUANTITY | 0.98+ |
Joachim Hammer, Microsoft | Microsoft Ignite 2018
>> Live from Orlando, Florida. It's theCUBE. Covering Microsoft Ignite. Brought to you by Cohesity, and theCUBE's ecosystem partners. >> Welcome back everyone to theCUBE's live coverage of Microsoft Ignite here in Orlando, Florida. I'm your host, Rebecca Knight along with my cohost Stu Miniman. We're joined by Joachim Hammer, he is the Principal Product Manager at Microsoft. Thanks so much for coming on the show. >> Sure, you're welcome. Happy to be here. >> So there's been a lot of news and announcements with Azure SQL, can you sort of walk our viewers through a little bit about what's happened here at Ignite this week? >> Oh sure thing, so first of all I think it's a great time to be a customer of Azure SQL Database. We have a lot of innovations, and the latest one that we're really proud of, and we're just announced GA is SQL Managed Instance. So our family of database offers had so far a single database and then a pool of databases where you could do resource sharing. What was missing was this one ability for enterprise customers to migrate their workloads into Azure and take advantage of Azure without having to do any rewriting or refactoring and Managed Instance does exactly this. It's a way for enterprise customers to take their workloads, migrate them, it has all the features that they are used to from sequel server on-prem including all the security, which is of course as you can imagine always a concern in the cloud where you need to have the same or better security that customers are used to from on-prem, and with Managed Instance we have the security isolation, we have private IPV nets, we have all the intelligent protection that we have in Azure so it's a real package. And so this is a big deal for us, and the general purpose went GA yesterday actually, so I heard. >> Security's really interesting 'cause of course database is at the core of so many customer's businesses. You've been in this industry for a while, what do you see from customers as to the drivers and the differences of going to public cloud deployments versus really owning their database in-house and are security meeting the needs of what customers need now? >> Yeah sure, so, you're right, security is probably the most important topic or one of the most important topics that comes up when you discuss the cloud. And what customers want is they want a trust, they want this trust relationship that we do the right thing and doing the right thing means we have all the compliances, we adhere to all the privacy standards, but then we also offer them state of the art security so that they can rely on Microsoft on Azure for the next however many years they want to use the cloud to develop customer leading-edge security. And we do this for example with our encryption technology with Always Encrypted. This is one of those technologies that helps you protect your database against attacks by encrypting sensitive data and the data remains encrypted even though we process queries against it. So we protect against third-party attacks on the database, so Always Encrypted is one of those technologies that may not be for everybody today but customers get the sense that yes, Microsoft is thinking ahead, they're developing this security offering, and I can trust them that they continue to do this, keep my data safe and secure. >> Trust is so fundamental to this whole entire enterprise. How do you build trust with your customers, I mean you have the reputation, but how do you really go about getting your customers to say "Okay, I'm going to board your train?" >> That's a good question, Rebecca. I think as I said it starts with the portfolio of compliance requirements that we have and that we provide for Azure's SQL Database and all the other Azure services as well. But it also goes beyond that, it goes, for example, we have this right to audit capability in Azure where a company can come to us and says we want to look behind the scenes, we want to see what auditors see so that we can really believe that you are doing all the things you're saying. You're updating your virus protection, you're patching and you have all the right administrative workflows. So this is one way for us to say our doors are open if you want to come and see what we do, then you can come and peek behind the scenes so to speak. And then the other, the third part is by developing features like we do that help customers, first of all make it easy to secure the database, and help them understand vulnerabilities, and help them understand the configurations of their database and then implement the security strategy that they feel comfortable with and then letting them move that strategy into the cloud and implement it, and I think that's what we do in Azure, and that's why we've had so much success so far. >> Earlier this week we interviewed one of your peers, talked about Cosmos DB. >> Okay. >> There's a certain type of scale we talk about there. Scale means different things to different sized customers. What does scale mean in your space? >> Yeah so you're right, scale can mean a lot of different things, and actually thank you for bringing this up so we have another announcement that we made on namely Hyper-Scale architecture. So far in Azure SQL DB, we were pretty much constrained in terms of space by the underlying hardware, how much storage comes on these VMs, and thanks to our re-architectured hardware, sorry software, we now have the ability to scale way beyond four terabytes which is the current scale of Azure SQL DB. So we can go to 64 terabytes, 100 terabytes. And we can, not only does that free up, free us from the limitations, but it also keeps it simple for customers. So customers don't have to go and build a complicated scale out architecture to take advantage of this. They can just turn a knob in a portal, and then we give them as much horsepower as they need to include in the storage. And in order for this to happen, we had to do a lot of work. So it doesn't just mean, we didn't just re-architect storage but we also have to make fail-over's faster. We have to continue to invest in online operations like online index rebuild and create to make those resumable, pause and resumable, so that with bigger and bigger databases, you can actually do all those activities that you used to do ya know, without getting in the way of your workloads. So lot of work, but we have Hyper-Scale now in Azure SQL DB and so I think this is another sort of something that customers will be really excited about. >> Sounds like that could have been a real pain point for a lot of DBA's out there, and I'm wondering, I'm sure, as a PM, you get lots of feedback from customers. What are the biggest challenges they're facing? What are some of the things they're excited about that Microsoft's helping them with these days? >> So you're right, this was a big pain point, because if you go to a big enterprise customer and say, hey bring your workload to Azure, and then they say oh yeah great, we've got this big telemetry database, what's your size limit? And you have to say four terabytes, that doesn't go too well. So that's one thing, we've removed that blocker thankfully. Other pain points I think we have by and large, I think the large pain points are we've removed, I think we have small ones where we're still working on making our deployments less painful for some customers. There's customers who are really, really sensitive to disconnects or latent variations in latency. And sometimes when we do deployments, worldwide deployments, we are impacting somebody's customer, so this is a pain point that we're currently working on. Security, as you said, is always a pain point, so this is something that will stay with us, and we just have to make sure that we're keeping up with the security demands from customers. And then, another pain point, or has been a pain point for customers, especially customers sequel server on-prem is the performance tuning. When you have to be a really, really good DBA to tune your workloads well, and so this is something that we are working on in Azure SQL DB with our intelligence performance tuning. This is a paint point that we are removing. We've removed a lot of it already. There's still, occasionally, there's still customers who complaining about performance and that's understood. And this is something that we're also trying to help them with, make it easier, give 'em insights into what their workload is doing, where are the weights, where are the slow queries, and then help them diffuse that. >> So thinking about these announcements and the changes that you've made to improve functionality and increase, not have size limits be such a road block, when you're thinking ahead to making the database more intelligent, what are some of the things you're most excited about that are still in progress right now, still in development, that we'll be talking about at next year's Ignite? >> Yeah, so personally for me on the security side, what's really exciting to me is the, so security's a very complicated topic, and not all of our customers are fully comfortable figuring out what is my security strategy and how do I implement it, and is my data really secure. So understanding threats, understanding all this technology, so I think one of the visions that gets me excited about the potential of the cloud, is that we can make security in the future hopefully as easy as we were able to make query processing with the invention of the relational model, where we made this leap from having to write code to access your data to basically a declarative SQL type language where you say this is what I want and I don't care how to database system returns it to me. If you translate that to security, what would be ideal the sort of the North Star, is to tell it to have customers in some sort of declarative policy based manner, say I have some data that I don't trust to the cloud please find the sensitive information here, and then protect it so that I'm meeting ISO or I'm meeting HIPPA requirements or that I'm meeting my internal ya know, every company has internal policies about how data needs to be secured and handled. And so if you could translate that into a declarative policy and then upload that to us, and we figure out behind the scenes these are the things we need, you need to turn on auditing, these are where the audit events have to go, and this is where the data has to be protected. But before all that, we actually identify all the sensitive data for you, we'll tag it and so forth. That to me has been a tremendous, sort of untapped potential of the cloud. That's where I think this intelligence could go potentially. >> Yeah, great. >> Who knows, maybe. >> (laughs) Well, we shall see at next year's Ignite. >> We are making handholds there. We have a classification engine that helps customers find sensitive data. We have a vulnerability assessment, a rules engine that allows you to basically test the configuration of your database against potential vulnerabilities, and we have threat detection. So we have a lot of the pieces, and I think the next step for us is to put these all together into something that can then be much more automated so that a customer doesn't have to think technology anymore. They can they business. They can think about the kinds of compliances they have to meet. They can think about, based on these compliances, this data can go this month, this data can go maybe next year, or ya know, in that kind of terms. So I think, that to me is exciting. >> Well Joachim, thank you so much for coming on theCUBE. It was a pleasure having you here. >> It was my pleasure too. Thank you. >> I'm Rebecca Knight for Stu Miniman, we'll have more from theCUBE's live coverage of Microsoft Ignite coming up in just a little bit. (upbeat music)
SUMMARY :
Brought to you by Cohesity, Thanks so much for coming on the show. Happy to be here. we have all the intelligent protection that and the differences of going to public cloud deployments And we do this for example with our encryption Trust is so fundamental to this whole entire enterprise. so that we can really believe that you are Earlier this week we interviewed one of your peers, There's a certain type of scale we talk about there. And in order for this to happen, we had to do a lot of work. What are some of the things they're excited about and so this is something that we are working on in these are the things we need, you need to turn on auditing, and we have threat detection. It was a pleasure having you here. It was my pleasure too. of Microsoft Ignite coming up in just a little bit.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Joachim | PERSON | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
Joachim Hammer | PERSON | 0.99+ |
Rebecca | PERSON | 0.99+ |
100 terabytes | QUANTITY | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
next year | DATE | 0.99+ |
64 terabytes | QUANTITY | 0.99+ |
Orlando, Florida | LOCATION | 0.99+ |
yesterday | DATE | 0.99+ |
third part | QUANTITY | 0.99+ |
theCUBE | ORGANIZATION | 0.98+ |
SQL | TITLE | 0.98+ |
Earlier this week | DATE | 0.98+ |
one way | QUANTITY | 0.98+ |
Cohesity | ORGANIZATION | 0.98+ |
Azure | TITLE | 0.97+ |
Azure SQL | TITLE | 0.97+ |
today | DATE | 0.97+ |
one | QUANTITY | 0.97+ |
North Star | ORGANIZATION | 0.97+ |
four terabytes | QUANTITY | 0.96+ |
Hyper-Scale | TITLE | 0.96+ |
this week | DATE | 0.95+ |
Azure SQL DB | TITLE | 0.94+ |
one thing | QUANTITY | 0.94+ |
this month | DATE | 0.93+ |
HIPPA | ORGANIZATION | 0.93+ |
Azure SQL Database | TITLE | 0.92+ |
Azure | ORGANIZATION | 0.9+ |
ISO | ORGANIZATION | 0.89+ |
single database | QUANTITY | 0.89+ |
first | QUANTITY | 0.89+ |
Microsoft Ignite | ORGANIZATION | 0.84+ |
Cosmos DB | TITLE | 0.77+ |
SQL DB | TITLE | 0.75+ |
SQL Managed Instance | TITLE | 0.68+ |
Scale | TITLE | 0.67+ |
Hyper | TITLE | 0.67+ |
Ignite | TITLE | 0.63+ |
Ignite | ORGANIZATION | 0.58+ |
Always Encrypted | TITLE | 0.56+ |
Microsoft Ignite 2018 | EVENT | 0.48+ |
nd | QUANTITY | 0.47+ |
Instance | ORGANIZATION | 0.45+ |
Ignite | EVENT | 0.45+ |
Managed | TITLE | 0.35+ |
Bernard Golden, Capital One | Microsoft Ignite 2018
>> Live, from Orlando, Florida, it's theCUBE, covering Microsoft Ignite. Brought to you by Cohesity and theCUBE's ecosystem partners. >> Welcome back, everyone, to theCUBE's live coverage of Microsoft Ignite. I'm your host, Rebecca Knight. Joined, of course, by my esteemed co-host, Stu Miniman. We have one guest for this segment, Bernard Golden. He is the vice president of Cloud Strategy at Capital One. Thank you so much for coming on the show Bernard. >> Well, thank you so much for inviting me to be on. >> You are famous in the world of cloud. You're named by wired.com as one of the most ten influential people in cloud computing. I'd love to just ask you a very broad question to start, and that is where are we right now with the cloud? Where do you see, what are sort of the biggest issue, the biggest challenges that you see with companies adopting and embracing the cloud right now? >> Well, unlike a lot of people, I think we're still a lot earlier in cloud adoption than other people do, maybe. If it were a baseball game, I'd say, maybe, the pitcher's coming out for the top of the second inning. And I think the barriers tend to be two-fold. One is, for traditional enterprises, there's still a lot of, we have a lot of embedded, a lot of legacy, a lot of investment, sunk costs, how can we step away from that, should we step away from that? So you hear a lot of discussion around what's the right role, hybrid clouds, so forth, and so on. For companies like Capital One that have said, "we're going all in on cloud," and Capital One has announced it's going all in on public cloud. Then the challenge becomes how do I adopt the practices of the organizations around the frontier of cloud? Because you have to really adapt a whole range of things. It's not just ... a lot of people treat cloud computing as kind of like it's a data center at the end of a wire. I have my traditional practices, my traditional models, my traditional tools, my traditional cost models. All of those things have to change. And so, I think for companies like Capital One, and we certainly have faced those things I would say, but for those companies that make the break to say "yeah, we're going to go all out on cloud," then it's how do I restructure the entire way I do information technology? >> Yeah, and Bernard, I agree with a lot of what you were saying. You and I have had conversations about cloud over the years. I've read lots of what you've written. Amazon would agree it's still day one, right? >> It's their phrase. >> Microsoft, I want to get you, not as Capital One, but just as a watcher of the industry, I remember a few years back, Microsoft put out TV ads like "to the cloud." At least made me cringe when I saw some of it. When I look at Microsoft today, they play strongly in SAS, they've got public cloud, they've got all the virtualization in various business products for private cloud. So they play a lot of places, they have a lot of strengths, they understand application, they understand data, they're well positioned. They might not be number one in many of these areas, but a strong customer base. And they're doing good, but I'd like to see them do even more. I'm curious of your viewpoint. >> Well I guess what I'd say is, if you look at the universe or the aggregate of cloud providers, Gartner says there's three that really matter, up in the upper right-hand quadrant of the Magic Quadrant. And that's what I call AMG, Amazon, Microsoft, Google. If you look at Gartner, they've said these are the three that really have both a vision and the ability to execute. >> We believe Alibaba might be making its way in there at this point. >> You know, Alibaba's quite interesting. I've had some interactions with Alibaba, before I joined Capital One, and they're tremendously capable technically, they have huge ambition, so I wouldn't dismiss them or write them off. They don't have much presence in the U.S., at least Capital One is primarily a U.S.-based company. But also, because of the fact that Alibaba doesn't have much presence in U.S., and not that much in Europe, they tend to not be so present, but I would definitely follow them going forward, for sure. >> Sorry, I took you off track, talking about Microsoft's positioning in the marketplace. >> Well, so they're clearly one of the three players. I would say they've had a pretty dramatic turnaround from where they were, say, four or five years ago. You can track that, maybe, to their CEO. I think they're making a strong play in this space, and obviously are committed to it. >> I think Capital One is an adopter and pushing on many of the disruptive technologies. I remember the first Echo Dot that I got, it was actually a Capital One giveaway at a conference, I bumped into you at the Serverless conference. A lot of this show is talking about the business productivity, the applications. There's lots of Azure, but I haven't heard as much about Azure, there's some announcements around Kubernetes. I had a great conversation at the Serverless booth, but if you look at the cloud piece, I want to get your viewpoint as to how Microsoft's doing, where customers are. I know we're in, especially, Serverless' very early days. But get your viewpoints on how those fit into the overall position, and anything you could say about Capital One there would be great, too. Capital One, as I said, is all-in on public cloud computing. It's announced it's going to close all of its data centers. And, as I said, the second challenge that organizations face is really when they go all-in, they go "now I have to really adapt all my practices." So, Capital One is looking at things like containers, serverless, it sponsors the serverless conference, so it's very much engaged with those kinds of things. This conference, I mean, unlike AWS that basically says "all we do is AWS," Microsoft has a very broad range of products, and they have to represent all of them at their conference. So, it's certainly not an only-Azure conference, and that's to be expected. I've said in a number of the sessions, and it's part of my job, I have to track what's going on with all these providers. And so I've tracked what's gone on in the sessions. I've been pretty impressed with some of the stuff that Azure has put forward. But there's other sessions as well. And they have to cover all the rest of their stuff. >> As you said, Capital One is all-in on the public cloud, but it is a multi-cloud world. And a lot of companies are still sort of struggling to figure out "how do I make this work, where do I go?" Can you walk us through your decision process at Capital One, and then also maybe tease out some best practices about how other organizations should make decisions? >> Well, I can't say a ton about Capital One, and about how we've looked at it, other than what we publicly announced, which is "we're all in on public cloud." Our CIO has been up on stage at AWS, very strong adapter of AWS. What I would say, is that, for most organizations, there's sort of two factors you might think about in terms of looking at using multiple clouds. One is from a risk communication strategy. Do you want to have all your eggs in one basket? And that's probably for most enterprises it's not that much of a problem in the sense that they own something of everything, no matter what. You'll never find any enterprise that only uses one thing. In any technology place, and even if they do, then they buy another company that's on a different one. But, from a risk communication strategy you might want that, and then, you might also be looking at opportunistic deployment of workloads if you want to take advantage of superior functionality available from one cloud provider or another. So, do you really like the machine learning that comes out of Azure, well you might decide to put workloads based on that. Or if you like something about certain kinds of database offerings, you might look at that. If you want a certain breadth of services, you might look at AWS, so there's a criteria you have to establish about what you want to accomplish with your applications, or what you want to do around risk management. >> Great, Bernard, what other things have you been seeing at the conferences, what's exciting you? Any takeaways for people that haven't been at the show? Or any things you'd recommend people go poke at? >> As I said, I attended a number of sessions yesterday. I was pretty impressed with the Cosmos DB multi-master. I used to run engineering groups at a database company, and I'll tell you, there's a huge revolution going on in databases, from all the providers, and having some domain experience, there's stuff that gets announced and I go, "how do they that?" I mean, that's amazing. So that was pretty impressive. There were a couple of announcements around Express Route. One, they've announced the 100 Gig Connectivity, which is pretty amazing, I think. And the second thing, this didn't get a lot of coverage and all that, is they announced that basically, let's say you're a corporation with stuff in Argentina and Switzerland. You can basically put Express Route connections into the Microsoft fiber backbone and then just transit your data across their private fiber backbone, which is pretty, pretty interesting. So, I thought that was pretty interesting. I think the rest of it is slipping my mind at the moment. >> I tell you, that is fascinating, because I remember, I've been watching since when AWS came out it had Direct Connect. It was, well, this is really interesting, there's some use cases, but Amazon, Azure, and Google, all of those versions, just hearing massive adoption as people go to a hosting colo service provider, and that can get them, I have the stuff that I'm going to own, and then I'm going to have the stuff, the public cloud in it, physics still exist, but I'm going to get them closer with high band with low-latency connections, so it's a real game-changer as to how I build my applications, and build that ... The hybrid cloud, or multi-cloud, which is something we've been kind of looking at as it's a challenging thing to do, over time. >> Yeah, it's interesting, because there was a time when the huge challenge was the skinny straw. If you had 100 megs, that was a pretty skinny straw. And now, that's really opened up a lot. And these direct connects are pretty good cross-connect performance. That was the pretty interesting era, I thought. >> Great, Bernard, thank you so much for coming on theCUBE. It was a pleasure having you. >> Thank you so much for inviting me. >> I'm Rebecca Knight for Stu Miniman, we will have more from theCUBE's live coverage of Microsoft's Ignite coming up in just a little bit. (techno music)
SUMMARY :
Brought to you by Cohesity Thank you so much for coming on the show Bernard. the biggest challenges that you see of the organizations around the frontier of cloud? of what you were saying. put out TV ads like "to the cloud." if you look at the universe or We believe Alibaba might be making But also, because of the fact that Sorry, I took you off track, talking about and obviously are committed to it. of the stuff that Azure has put forward. And a lot of companies are still sort of struggling of workloads if you want to take advantage And the second thing, this didn't get Azure, and Google, all of those versions, If you had 100 megs, that was a pretty skinny straw. Great, Bernard, thank you so of Microsoft's Ignite coming up
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Amazon | ORGANIZATION | 0.99+ |
Rebecca Knight | PERSON | 0.99+ |
AMG | ORGANIZATION | 0.99+ |
Alibaba | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Argentina | LOCATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Bernard Golden | PERSON | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Bernard | PERSON | 0.99+ |
Switzerland | LOCATION | 0.99+ |
Capital One | ORGANIZATION | 0.99+ |
Europe | LOCATION | 0.99+ |
U.S. | LOCATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Gartner | ORGANIZATION | 0.99+ |
three | QUANTITY | 0.99+ |
three players | QUANTITY | 0.99+ |
second challenge | QUANTITY | 0.99+ |
theCUBE | ORGANIZATION | 0.99+ |
second inning | QUANTITY | 0.99+ |
Echo Dot | COMMERCIAL_ITEM | 0.99+ |
yesterday | DATE | 0.99+ |
today | DATE | 0.99+ |
second thing | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
100 megs | QUANTITY | 0.98+ |
Orlando, Florida | LOCATION | 0.98+ |
one basket | QUANTITY | 0.98+ |
Cohesity | ORGANIZATION | 0.98+ |
four | DATE | 0.98+ |
one guest | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
first | QUANTITY | 0.97+ |
one thing | QUANTITY | 0.96+ |
Azure | ORGANIZATION | 0.95+ |
Azure | TITLE | 0.94+ |
five years ago | DATE | 0.94+ |
two-fold | QUANTITY | 0.94+ |
wired.com | ORGANIZATION | 0.91+ |
ten influential people | QUANTITY | 0.89+ |
two factors | QUANTITY | 0.89+ |
Express Route | ORGANIZATION | 0.88+ |
SAS | ORGANIZATION | 0.87+ |
One | ORGANIZATION | 0.81+ |
Microsoft Ignite | ORGANIZATION | 0.81+ |
Cloud Strategy | ORGANIZATION | 0.81+ |
few years back | DATE | 0.78+ |
Cosmos | ORGANIZATION | 0.76+ |
Direct Connect | TITLE | 0.75+ |
Andrew Liu, Microsoft | Microsoft Ignite 2018
>> Live from Orlando, Florida. It's theCUBE. Covering Microsoft Ignite. Brought to you by Cohesity, and theCUBE's ecosystem partners. >> Welcome back to the CUBE's live coverage of Microsoft Ignite here in Orlando, Florida. I'm your host, Rebecca Knight. Along with my co-host Stu Miniman. We're joined by Andrew Liu. He is the senior product manager at Azure Cosmos DB. Thanks so much for coming on the show Andrew. >> Oh, thank you for hosting. >> You're a first timer, so this will be a lot of fun. So, talk to me a little bit. Azure Cosmos DB is a database for building blazing fast planet scale applications. Can you tell our viewers a little bit about what you do and about the history of Azure Cosmos? >> Sure, so Azure Cosmos DB started with, about eight years ago, where we were also outgrowing a lot of our own database needs with what we had previously built. And a lot of the challenges that we had was really around partitioning, replication, and resource governance. So, I'll talk a little bit about each one. Partitioning is really about solving the problem of scale. Right? I have so much data, doesn't fit on a single machine, and I have so many requests per second. Also doesn't, can't be served out of a single machine. So how do I go and build a system, a database that can elastically scale over a cluster of machines, so I don't have to manually shard, and as a user have to shard a database across many, many instances. This way I really want to be able to scale just seamlessly. The velocity problem is, we also wanted to build something that, can respond in a very fast manner, in terms of latency. So, it's great and all that we can serve lots of request per second, but, what is the response time of each one of those requests? And the resource governance was there to really actually build this as a cloud native database in which we wanted to exploit the properties of our cloud. We wanted to use the economies of scale that we can have basically data centers built all around the world, and build this as a multi, truly multi-tenant service. And by doing so we can also afford the total cost of ownership for us, as well as, a guaranteed predictable performance for the tenants. Now we did this, for initially our first party tenants at Microsoft, where we have made a bet on everything from our Microsoft live platform, to Office, to Azure itself as built on Azure Cosmos DB. And about four years ago we found that hey, this is not really just a Microsoft problem that we're solving, but it's an everybody problem, it's become universal, and so we've launched it out to the open. >> Yeah, Andrew that's, great point, and I want you to help unpack that for us a little bit because you know, we've been saying on theCUBE for many years, distributed architectures are some of the toughest challenges of our time, but, if I'm a Facebook, or a Google, or a Microsoft, I understand some of the challenges, and I understand why I need it, but, when you talk about scale, well, scale means a lot of different things to a lot of different people. So, how does Cosmos? What does that mean to your users, end users, why do they need this? You know, haven't they just felt some microservices architecture? And they'll just leverage, ya know what's in Azure. And things like that. How does this global scale impact the typical user? >> So I'm actually seeing this come in different types of patterns for different types of industries. So for example, in manufacturing we're commonly seeing Cosmos DB used really for that scalability for the write scalability, and having many, many concurrent writes per second. Typically this is done in an IoT telemetry, or an IoT device registry case. So let's use one of our customers for example, Toyota. Each year they're shipping millions of vehicles on the road, and they're building a big connected car platform. The connected car platform allows you to do things like, whenever it alerts an airbag gets deployed, they can go and make sure and call their driver, hey, I saw the airbag was deployed are you okay? And if the user doesn't pick up their phone, immediately notify emergency services. But the challenge here is if each year I'm shipping millions of vehicles on the road, and each of 'em has a heartbeat every second, I'm dealing with millions of writes per second, and I need a database that can scale to that. In contrast, in retail I'm actually seeing very different use cases. They're using more of the replication side of our stock where they have a global user base, and they're trying to expand an eCommerce shop. So for example ASOS is a big fashion retailer, they ship to 200 different countries globally, and they want to make sure that they can deliver real-time experiences like real-time personalization, and based off of who the user is recommended set of products that is tailored to that user. Well now what I need is a data set that can expand to my shoppers across two different hundred, 200 countries around the globe, and deliver that with very, very low latency so that my web experience is also very robust. So what they use is our global distribution, and our multi-mastering technology. Where we can actually have a database presence, similar to like what a CDN does for static content, we're doing for our dynamic evolving content. So in a database your work load, typically your data set is evolving, and you want to be able to run queries with consistency over that. As opposed to in CDN you're typically serving static assets. Well here we can actually support those dynamic content, and then build these low latency experiences to users all around the globe. The other area we see a lot of usage is in ISV's for mission critical workloads. And the replication actually gets us two awesome properties, right? One is the low latency by shipping data closer to where the user is, but the other property you get is a lot of redundancy, and so we actually also offer industry leading SLA's where we guarantee five nines of availability, and the way we're able to do so is, with a highly redundant architecture you don't care if let's say a machine were to bomb out at any given time, because we have multiple redundant copies in different parts of the globe. You're guaranteed that your workload is always online. >> So my question for you is, when you have these, you just described some really, really interesting customer use cases in manufacturing, in retail, do you then create products and services for each of these industries? Or do you say hey other retail customers, we've noticed this really works for this customer over here, how do you go out to the community with what you're selling? >> Ah, got it. So we actually have found that this can be a challenging space for some of our customers today, 'cause we have so many products. The way we kind of view it is we want to have a portfolio, so that you can always choose the right tool for the right job. And I think a lot of how Microsoft has evolved as a business actually is around this. Previously we would sell a hammer, and we'd tell you don't worry everything's a nail, even if it looks like a screw let's just pretend it's a nail and whack it down. But today we've built this big vast toolbox, and you can think of Cosmos DB as just one of many tools in our vast toolbox. So if you have a screw maybe you pickup a screwdriver, and screw that in. And the way Azure works is then if we have a very comprehensive toolbox, depending on what precise scenario you have, you can kind of mix and match the tools that fit your problem. So think of them as like individual Lego blocks, and whether you're building like a death star, or an x-wing, you can go, and assemble the right pieces for your application. >> Andrew, some news at the show around Cosmos DB. Share us what the updates are. >> Oh sure, so we're really excited to launch a few new features. The highlights are multi-master, and Cassandra API. So multi-master really exploits the replicated nature of our database. Before multi-master what we would do is, we would allow you to have a globally distributed database in which you can have write requests go to single region, and reads being served out of any of these other locations. With multi-master we've actually made it so that each of those replicas we've deployed around the globe can also accept write requests. What that translates to from a user point of view is number one, your write requests are a lot faster, they're super low latency, single-digit millisecond latency in fact. No matter where the user is around the globe. And number two, you also get much higher write availability. So even if let's say, we're having a natural disaster, we had a nasty hurricane as you know pass through on the east coast last week, but with a globally distributed database the nice thing is even if you have, let's say, a power disruption in one region of the world, it doesn't matter cause you can then just fail over, and talk to another data center, where you have a live replica already located. So we just came out with multi-master. The short summary is low latency writes, as well as high available writes. The other feature that we launched is Cassandra API, and as you know this is a multi-model, multi-API database. What that means is, what we're trying to do is also meet our users where they are. As opposed to pushing our proprietary software on them, and we take the whole concept of vendor lock-in very, very seriously. Which is why we make such a big bet on the open source ecosystem. If you already have, let's say a MongoDB application, or a Cassandra application, but you'd really love to be able to take advantage of some of the novel properties that we've built with building a fully managed multi-master database. Well, what we've done is we've implemented this as a wire level protocol on the server side. So it can take an existing application, not change a single line of code, and point it to Cosmos DB as a back-end, and then take advantage of Cosmos DB as your database. >> One of the interesting things if you look at the kind of changing face of databases, it's how users are being able to leverage their data. You talk about everything from you know, I think Cassandra back, and some of the big data discussions, today everything's AI which I know is near and dear to Microsoft's heart. Satya Nadella I'm talking about, how do you think of the role of data in this solution set? >> Sorry, can you say that one more time? >> So, how customers think about leveraging data, how things like Cosmos allow them to really extract the value out of data, not just be some database that kind of stuck in the back-end somewhere. >> Yeah, yeah. I mean a lot of it is the new novel experiences people are building. So for example, like the connected car platform, I'm seeing people actually build this, and take advantage of new novel territories that a traditional automobile manufacturer used to not do. Not only are they building experiences around, how do they provide value to their end users? Like the air bag scenario, but they're also using this as a way of building value for their business, and how to make sure that, hey when, next time you're up for an oil change that they can send a helpful reminder, and say hey I noticed you're due for an oil change in terms of mileage. Why don't I just go set up an appointment, just up for you, as well as other experiences for things, like when they want to do fleet management, and do partnerships with either ride sharing companies like Uber, and Lyft, or rental car companies like Avis, Hertz, et cetera. I've also seen people take advantage of, taking kind of new novel experiences through databases, through AI, and machine learning. So for example, the product recommendations. This was something that historically, when I wanted to do recommendations a decade ago, maybe I have some big beefy data lake running somewhere in the back-end, it might take a week to munch through that data, but that's okay, a week later once I'm ready, I'll send out some mail, maybe some email to you, but today when I want to actually show live right when the user is browsing my website, my website has to load fast right? If my goal is to increase conversions on sales, having a slow running website is the fastest way for my user to click the back button. But if I want to build real-time personalization, and want to generate let's say a recommendation within 200 millisecond latency, well now that I have databases that can guarantee me single-digit millisecond latency, it gives me ample time to actually improve the business logic for those recommendations. >> I want to ask you a question about culture, because you are based at the mothership in Redmond, Washington. So we heard Satya Nadella on the main stage today talk about tech intensiveness, tech intensity, sorry, this idea that we need to not only be adopting technology, but also building the latest, and greatest. I'm curious about, how that translates at Microsoft's campus, and sort of how, how this idea is, infuses how you work with your colleagues, and then also how you work with your customers and partners? >> I think some of the biggest positive changes I've seen over the last decade has been how much more of a customer focus we have today then ever. And i think a lot of things have led to that. One is, just the ability to ship much faster. As we move to Cloud services we're no longer in these big box product release cycles of building a product, and waiting like one or two years to ship it to our users. But now we can actually get some real-time feedback. So as we go, and ship, and deploy software, we actually deploy even on a weekly cadence over here. What that allows us to do is actually experiment a lot more, and get real-time feedback, so if we have an idea, and rather than having to go through a long lengthy vetting process, spending years building, and hoping that it really pays off. What we can do is we can just go talk to our users, and say hey, ya know, we have an idea for our future. We'd love to get your feedback, or a lot of times honestly our customers actually come to us, where we're so tightly engaged these days, that when, users even come to us, and say like hey, what do you think about this idea? It would really add a lot of value to my scenario. We go, and try to root cause that, really get an idea of what exactly that they need. But then we can turn that around in blazing fast time. And I think a lot of the shift to Cloud services, and being able to avoid the overhead of well we got to wait for this ship train, and then wait for the right operation personnel to go and deploy the updates. Now that we can control our own destiny, and just ship on a very, very fast cadence, we're closer to our users, and we experiment a lot more, and I think it's a beautiful thing. >> Great, well Andrew thank you so much for coming on theCUBE, it was fun talking to you. >> Oh yeah, thank you for hosting. >> I'm Rebecca Knight, for Stu Miniman, we will have more from theCUBE's live coverage of Microsoft Ignite coming up just after this. (techno music)
SUMMARY :
Brought to you by Cohesity, Thanks so much for coming on the show Andrew. what you do and about the history of Azure Cosmos? And a lot of the challenges that we had was and I want you to help unpack that and I need a database that can scale to that. and you can think of Cosmos DB as just one Andrew, some news at the show around Cosmos DB. and as you know this is a multi-model, One of the interesting things if you look that kind of stuck in the back-end somewhere. So for example, like the connected car platform, and then also how you work with your customers and partners? and say like hey, what do you think about this idea? Great, well Andrew thank you so much we will have more from theCUBE's live coverage
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Rebecca Knight | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
Infinidat | ORGANIZATION | 0.99+ |
Andrew | PERSON | 0.99+ |
Satya Nadella | PERSON | 0.99+ |
Andrew Liu | PERSON | 0.99+ |
Uber | ORGANIZATION | 0.99+ |
Stu Miniman | PERSON | 0.99+ |
Hertz | ORGANIZATION | 0.99+ |
Toyota | ORGANIZATION | 0.99+ |
Lyft | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Avis | ORGANIZATION | 0.99+ |
Tuesday March 27th | DATE | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
March 2018 | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
ASOS | ORGANIZATION | 0.99+ |
one | QUANTITY | 0.99+ |
last week | DATE | 0.99+ |
ORGANIZATION | 0.99+ | |
Orlando, Florida | LOCATION | 0.99+ |
Crowdchat.net/infinichat | OTHER | 0.99+ |
CUBE | ORGANIZATION | 0.99+ |
Moshe Yanai | PERSON | 0.99+ |
a week later | DATE | 0.99+ |
two years | QUANTITY | 0.99+ |
next week | DATE | 0.99+ |
Each year | QUANTITY | 0.99+ |
each | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Azure Cosmos DB | TITLE | 0.98+ |
millions of vehicles | QUANTITY | 0.98+ |
a decade ago | DATE | 0.98+ |
Cohesity | ORGANIZATION | 0.98+ |
200 different countries | QUANTITY | 0.98+ |
single machine | QUANTITY | 0.98+ |
Cassandra | TITLE | 0.98+ |
Redmond, Washington | LOCATION | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
hundred, 200 countries | QUANTITY | 0.97+ |
each year | QUANTITY | 0.97+ |
Office | TITLE | 0.97+ |
each one | QUANTITY | 0.97+ |
a week | QUANTITY | 0.97+ |
single-digit | QUANTITY | 0.96+ |
Lego | ORGANIZATION | 0.95+ |
about four years ago | DATE | 0.95+ |
200 millisecond | QUANTITY | 0.95+ |
about eight years ago | DATE | 0.94+ |
MongoDB | TITLE | 0.93+ |
two awesome properties | QUANTITY | 0.93+ |
Microsoft Ignite | ORGANIZATION | 0.93+ |
single line | QUANTITY | 0.92+ |
first party | QUANTITY | 0.92+ |
Azure Cosmos | TITLE | 0.92+ |
one region | QUANTITY | 0.92+ |
Cosmos | TITLE | 0.91+ |
Azure | TITLE | 0.89+ |
millions of writes per second | QUANTITY | 0.89+ |
John Mracek & Peter Smails, Imanis Data | theCUBE NYC 2018
live from New York it's the cube covering the cube New York City 2018 brought to you by silicon angle media and its ecosystem partners i'm jeff workday Villante we're here nine years our nine years of coverage two days live in New York City and our next two guests shot Mrazek CEO amana stayed at fiendish males CMO mystic good to see you again welcome back thank you bad to be here guys so obviously this show we've been here nine years we were the first original Hadoop world we've seen a change Hadoop was gonna change the world it kind of didn't but we get the idea of it did not it did didn't but it would change our world it brought open source and the notion of low-cost Hardware into the big data game and then the big data became so much more powerful around these new tools but then the cloud comes in full throttles and while they can get horsepower that compute you can stand up infrastructure for analytics all this data goodness starts to change machine learning then becomes the the real utility that's showing this demand for using data right now not the set up using data this is a fundamental big trend so I don't get you guys reaction what do you see this evolving more cloud like how do you guys see the trend in this as data science certainly becoming more mainstream and productivity users to hardcore users and then you got cloud native developers doing things like kubernetes we've heard kubernetes here it's like a cloud is a data science what's going on what's your view of the market so I came from a company that was in an tech and we were built on big data and in looking at how big data is evolved and the movement towards analytics and machine learning it really being enabled by Big Data people have rushed to build these solutions and they've done a great job but it was always about what's the solution to my problem how do i leverage this data and they built out these platforms and in our context what we've seen is that enterprises get to a certain point where they say okay i've got all these different stacks that have been built these apps that have been built to solve my bi and analytics problems but what do I do about how do I manage all these and that's what I encounter my last company where we built everything ourselves and then so wait a minute but what we see at an enterprise level is fascinating because when I go to a large company I go you know we work with no sequel databases and Hadoop and you know how much Couchbase do you have how much Mongo etc the inevitable answer is yes and five of each right and they're cutting to this point where I've got all this distributed data distributed across my organization how am I going to actually manage it and make sure that that data is protected that I can migrate to the cloud or in a hybrid cloud environment and all these questions start to come up at an enterprise level we actually have had some very high-level discussions at a large financial institution here in New York where they literally brought 26 people to the meeting the initial meeting this was literally a second call where we were presenting our capability because they're they're now at the point where it's like this is mission-critical data this is not just some cool stuff somebody built off in one of our divisions it matters to the whole enterprise how do we make sure that data is protected backed up how do we move data around and that's really the the trend that we're tapping into and that the founders of our company saw many years ago and said I need to I need to we need to build a solution around this it's interesting you know you think about network data as a concept or data in general it's kind of got the same concepts we've seen in networking and/or cloud a control plane of some sorts out there and you know we're networking kind of went wrong as the management plane was different than the control plane so management and control or huge issues I mean you bring up this sprawl of data these companies are data full it's not like hey we might have data in the future right they got data now they're like bursting with data one what's the control plane look like what's the management plane look like these are all there's a technical concepts but with that with that in mind this is a big problem what our company is doing right now what are what are some of the steps that are taking now to get a handle on the management the data management it's not just your grandfather's data management so we anymore it's different it looks different your thoughts on on this chain of management so they're approaching the problem now and that's our sweet spot but I don't think they have in their minds yet come to exactly how to solve it it's there's this realization about we need to do this at this point and and and in fact doing it right is something that our founders when they built Lee said look if this problem of data management across big data needs to be solved by a data we're platform built on big data so let's use big data techniques to solve the problem all right so let's before getting some of the solution you guys are doing take a minute to explain what you guys are doing for the company the mission you know the value proposition status what do you guys do how are people gonna consume your product I mean take a particular type gen simple elevator pitch and we were enterprise data management focused specific than had you been no sequel so everyone's familiar with the traditional space of data management in the relational space relational world very large market very mature market well we're tapping into is what John was just saying which is you've got this proliferation but Dupin no sequel and people are hitting the wall they're hitting the ceiling because they don't have the same level of operational tools that they need to be able to mainstream these deployments whether it's data protection whether it's orchestration whether it's migration whatever the case may be so what we do that's essentially our value prophecy at a management for a Dupin no sequel we help organizations essentially drive that control plane really around three buckets data protection if it's business critical I got to protect it okay disaster recovery falls into protection bucket good old stuff everyone's familiar with but not in Hadoop in no single space orchestrations the second big bucket for us which is I'm moving to an agile development model how do i do things like automated test dev how do i do things like GD are the compliance management how do i do things like cloud migration you tut you know john touched on this one before a really interesting trend that we're seeing is you said what are customers doing they're trying to create a unified taxonomy they're trying to create a unified data strategy which is why 26 people end up in the but in lieu of that there's this huge opportunity because of what they need they know that it's got to be protected and they have 12 different platforms and they also want to be able to do things like one Cosmo I'm on go today but I'll be cosmos tomorrow I'm a dupe today but I might be HD inside tomorrow I want to just move from one to the other I want to be able to do intelligence so essentially the problem that we solve is we give them that agility and we give them that protection as they're sort of figuring this all out so we have this right you basically come in and say look it you can have whatever platform you want for your day there whether it's Hadoop and with most equals get unstructured and structured data together which makes sense but protections specifically does it have to morph and get swapped out based upon a decision correct make well now we're focused specifically Hadoop and no sequel so we would not be playing like if you we're not the 21st vendor to be helping s AP and Oracle you know customers backup their data it's basically if your Hadoop renewal sequel that's the platform regardless of what Hadoop distribution you're doing or where it's no see you know change out your piece what they do as they evolve and are correct I feel exactly right you're filling white space right because when this whole movement started it was like you were saying commodity Hardware yeah and you had this this idea of pushing code to data and oh hey his life is so easy and all of a sudden there's no governance there's no data protection no business continuity is all his enterprise stuff I didn't you heard for a long time people were gonna bring enterprise grade to Hadoop but they really didn't focus on the data protection space correct or the orchestra either was in those buckets and you touch them just the last piece of that puzzle value wise is on the machine learning piece yeah we do protection we do orchestration and we're bringing machine learning to bear to automate protection what amazing we hear a lot and that's a huge concern because the HDFS clusters need to talk speech out there right so there's a lot of nuances and Hadoop that are great but also can create headache from a user human standpoint because you need exact errors can get folded I gotta write scripts it creates a huge problem on multiple fronts the whole notion of being eventually being clustered in the first base being eventually consistent in the second place it creates a huge opportunity for us because this notion of being a legs we get the question asked the question why well you know there are a lot of traditional vendors they're just getting into the space and then what do that that's actually good because it rises you know rises all boats if you will because we think we've got a pretty significant technology mode around our ability to provide protection orchestration for eventually consistent clustered environments which is radically different than the traditional I love the story about the 26 people showing them me take me through what happened because that's kind of like what your jonquil fishbowl what do they do it they sit in their auditing they take a node so they really raising their hand they peppering you with questions what what happened in that meeting tell us so so it's an interesting microcosm what's happening in these organizations because as the various divisions and kind of like the federated IT structure started building their own stuff and I think the cloud enabled that it's like you know basically giving a the middle finger to central IT and so I can do all this stuff myself and then the organization gets to this realization of like no we need a central way to approach data management so in this meeting basically so we had an initial meeting with a couple of senior people and said we are we are going about consolidating how we manage all this data across all these platforms we want you to come in and present so when we presented there was a lot of engagement a lot of questions you could also see people still though there's an element of I want to protect my world and so this organizational dynamic plays out but you know when you're at a fortune 50 company and data is everything there's the central control starts to assert itself again and that's what we saw in this because the consequences of not addressing it is what is potentially massive data you know data loss loss of millions hundreds of millions of dollars you know data is the gold now right is the new oil so the central organizations are starting to assert that so we say that see that playing out and that's why all these people were in this meeting which is good in a way because then we're not like okay we got to sell ten different groups or ten different organizations it's actually being so there's there's kind of this pull back to the center it's happened in the no sequel world of your perspectives on this I mean early on you had guys like Mongo took off because it was so simple to use and capture unstructured data and now you're hearing everybody's talking about you know acid compliance and enterprise you know great capabilities that's got to be a tailwind for you guys could you bring it in the data protection and orchestration component but yeah what do you see it in that world what do you guys support today and maybe give us a glimpse of the future sure so that what we see as well a couple different things we are we are agnostic to the databases in the sense that we are definitely in Switzerland we were we you know we support all commerce so it's you know it's follow the follow the follow of the market share if you will Cassandra Mongo couch data stacks right on down the line on the no sequel side and what's interesting so they have very there have all varying degrees of maturity in terms of what their enterprise capabilities are some of them offer sort of rudimentary backup type stuff some fancy they have more backup versus others but at the end of the day you know their core differentiation they each it's fascinating to each have sort of a unique value prop in terms of what they're good at so it's a very fragmented market so that's a challenge that's an opportunity for us but it's a challenge from a marketplace networkers they've got to carve out there they all want the biggest slice of the pie but it's very fragmented because each of them is good at doing something slightly different yeah okay and so that like the the situation described before is they've got yes so you got one of everything yeah so they've got 19 different backup and recovery right coordinate processes approach or the or nothing or scripting law so that they do have to they've got a zillion steps associated with that and they're all scripted and so their probability of a failure you know very you drop a mirror that's a human error to is another problem and you use the word tailwind and I think that's very appropriate because with most of these vendors they're there they've got their hands full just moving their database features forward right you know where the engagement so when we can come in and actually help them with a customer who's now like okay great thank you database platform what do you do for backup well we have a rudimentary thing we should belong with it but there is one of our partners a manas who can provide these like robust enterprise it really helps them so with some of those vendors were actually a lot of partner traction because they see it's like that's not what their their strength is and they got to focus on moving their database so I'll give you some stats I'm writing a piece right now a traditional enterprise back in recovery but I wonder if you could comment on how it applies to your world so these are these are research that David flora did and some survey work that we've done on average of global 2000 organizations will have 50 to 80 steps associated with its backup and recovery processes and they're generally automated with scripts which of course a fragile yeah right and their prefer own to era and it's basically because of all this complexity there's a 1 in 4 chance of encountering an error on recovery which is obviously going to lead to longer outages and you know if you look at I mean the average cost the downtime for a typical global global 2000 companies between 75 thousand and two hundred fifteen thousand dollars an hour right now I don't know is your world because it's data it's all digitally the worst built as a source is it probably higher end of the spectrum all those numbers go AHA all those numbers go up and here's why all those metrics tie back to a monolithic architecture the world is now micro services based apps and you're running these applications in clusters and distributor architectures drop a note which is common I mean think you know you're talking about you're talking about commodity hardware to come out of the infrastructure it's completely normal to drop notes drops off you just add one back in everything keeps going on if your script expects five nodes and now there's four everything goes sideways so the probability I would I don't have the same stats back but it's worse because the the likelihood of error based upon configuration changes something as simple as that and you said micro-services was interesting to is is that now is it just a data lake kind of idea of storing data and a new cluster with microservices now you're having data that's an input to another app check so now so that the level of outage 7so mole severity is multiple because there could be a revenue-generating app at good young some sort of recommendation engine for e-commerce or something yeah something that's important like sorry you can't get your bank balance right now can't you any transfers because the hadoo closes down okay this is pretty big yes so it's a little bit different than say oh well to have a guy go out there and add a new server maybe a little bit different yeah and this is the you know this is the type of those are the types of stats that organizations that we're talking to now are caring a lot more it speaks to the market maturity do you run into the problem of you know it's insurance yeah and so they don't want to pay for insurance but a big theme in that you know the traditional enterprises how do we get more out of this data whether it's helping manage you know this I guess where that that's where your orchestration comes in cloud management maybe cloud migration maybe talk about some of the non insurance value add to our components and how that's resonating with with cost yeah yeah I so I'll jump in but the yeah the non protection stuff the orchestration bucket we're actually seeing it comes back to the to the problem sting we just said before which is they don't have it's not a monolithic stack it's a micro services based stack they've got multiple data sources they've got multiple data types it's sort of a it's the it's the byproduct of essentially putting power into into divisions hands to drive these different data strategies so you know the whole cloud let me double click on cloud migrations is a is a huge value problem that we have we talked about this notion of being data where so the ability to I'm here today but I want to be somewhere else tomorrow is a very strong operational argument that we hear from customers that we also also hear from the SI community because they hear it from the other community the other piece of that puzzle is you also hear that from the cloud folks because you've got multiple data for platforms that you're dealing with that you need agility to move around and the second piece is you've got the cloud obviously there's a massive migration to the cloud particularly with the dubidouxs sequel workloads so how do I streamline that process how do I provide the agility to be able to go from point A to point B just from of migration standpoint so that's a very very important use case for us has a lot of strategic value like it's coming it's sort of the markets talking to us like no no no we have this is him but we have to be able to do this and then simple things like not simple but you know automated test step is a big deal for us everybody's moved agile development so they want to spin up you know I don't want it I don't want to basically I want 10% of my data set I want to mask out my PII data I want to spin it up on Azure and I want to do that automatically every hour because I'm gonna run 16 I'm gonna run six builds today clouds certainly accelerates your opportunity big-time it forces everything to the table right yeah everybody's you can't hide anymore right what are you gonna do right you gotta answer the questions these are the questions so okay my final question I want to get on the table is for you in the segment is the product strategy how you guys looking at as an assassin gonna be software on premise cloud how's that look at how people consume the OP the offering and to opportunities because you guys are a young growing company you're kind of good good time you don't have the dog'll or the bagging it's Hadoop has changed a lot certainly there's a use case that neurons getting behind but clouds now a factor that product strategy and then when you're in deal why are you being called in why would someone want to call you rotor signs that would say you know call you guys up when with it when would a customer see signals and what signals would that be and to give you guys a ring or a digital connection product so the primary use cases are talking about recovery there's also data migration and the test step we have a big account right now that we're in final negotiations with where their primary use case is they're they're in health care and it's all about privacy and they need to securely mask and subset the data to your specific question around how are we getting called in basically you've got two things you've got the the administrators either the database architect or the IT or infrastructure people who are saying okay I need a backup solution I'm at a point now where I really need to protect my data as one and then there's this other track which is these higher-level strategic discussions where we're called in like the twenty six person meeting it's like okay we need an enterprise-wide data strategy so we're kind of attacking it both at the use case and at the higher level strategic and and and obviously the more we can drive that strategic discussion and get more of people wanting to talk to us about that that's gonna be better for our business and the stakeholders in that strategic discussion or whomever CIT is involved CIO maybe use their chief data officer and yeah database architect enterprise architecture head of enterprise architecture you know various flavors but you basically it kind of ways comes down to like two polls there's somebody who's kind of owns infrastructure and then there's somebody who kind of owns the data so it could be a chief data officer data architect or whatever depending on the scale of your and they're calling you because they're full they had to move the production workloads or they have production workloads that are from a bond from what uncared-for undershirt or is that the main reason they're in pain or you're the aspirin are you more others like we had a day loss and we didn't have any point in time recovery and that's what you guys provide so we don't want to go through this again so that's that's a huge impetus for us it is all about to your point it is mature its production workloads I mean the simple qualifying are you are you running a duper no sequel yes are you running in production yes you have a backup strategy sort of tip of the spear now to just briefly answer your question before we before we run out of time so it's an it's it's not a SAS basement we're software-defined solution will run in bare mantle running VMs will run in the cloud as your Google whatever you want to run on so we run anywhere you want we're sorry for be fine we use any storage that you want and basically it's an annual subscription base so it's not a SAS consumption model that may come down the road but it's basically in a license that you buy deploy it wherever you want customers choose what to do basically customers can do you know it's complete flexible flexible but back to you so let's go back to something you said you said they didn't have a point in time recovery what their point in time recovery was their last full backup or they just didn't have one or they just didn't have one all of the above you know see we've seen both yeah there's a market maturity issues so it's represented yeah you know that a lot its clustered I you know I just replicate my data and replication is not earth and truth be told my old company that was our approach we had a script but still it was like and the key thing is even if you write that script as you point out before the whole recovery thing so you know having a recovery sandbox is really in thing about this we designed everything exactly extract the value and show the use case prove it out yeah dupes real the history is repeating itself in that regard if you refuel a tional space there's a very in correlation to the Delton between the database platforms of the data mention logical hence they are involved coming in okay let's look at this in the big picture let's dad what's the recovery strategy how we gonna scale this exactly it's just a product Carson so your granularity for a point in time is you offer any point in time any point in time is varying and we'll have more news on that in the next couple weeks okay mantas data here inside the cube hot new startup growing companies really solving a real need need in the marketplace you're kind of an aspirant today but you know growth opportunity for as they scale up so congratulations good luck with the opportunity to secure bringing you live coverage here is part of Cuban YC our ninth year covering the big data ecosystem starting originally 2010 with a dupe world now it's a machine learning Hadoop clusters going at the production guys thanks for coming I really appreciate it this is the cube thanks for watching day one we'll be here all day tomorrow stay with us for more tomorrow be right back tomorrow I'll see you tomorrow
**Summary and Sentiment Analysis are not been shown because of improper transcript**
ENTITIES
Entity | Category | Confidence |
---|---|---|
50 | QUANTITY | 0.99+ |
New York City | LOCATION | 0.99+ |
Switzerland | LOCATION | 0.99+ |
26 people | QUANTITY | 0.99+ |
tomorrow | DATE | 0.99+ |
John | PERSON | 0.99+ |
two days | QUANTITY | 0.99+ |
New York | LOCATION | 0.99+ |
2010 | DATE | 0.99+ |
New York | LOCATION | 0.99+ |
John Mracek | PERSON | 0.99+ |
today | DATE | 0.99+ |
nine years | QUANTITY | 0.99+ |
10% | QUANTITY | 0.99+ |
26 people | QUANTITY | 0.99+ |
Peter Smails | PERSON | 0.99+ |
five | QUANTITY | 0.99+ |
26 people | QUANTITY | 0.99+ |
second piece | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
1 | QUANTITY | 0.99+ |
Lee | PERSON | 0.99+ |
12 different platforms | QUANTITY | 0.99+ |
75 thousand | QUANTITY | 0.99+ |
ninth year | QUANTITY | 0.99+ |
Hadoop | TITLE | 0.98+ |
millions | QUANTITY | 0.98+ |
two things | QUANTITY | 0.98+ |
Cassandra Mongo | PERSON | 0.98+ |
each | QUANTITY | 0.98+ |
twenty six person | QUANTITY | 0.98+ |
second place | QUANTITY | 0.98+ |
New York City | LOCATION | 0.98+ |
ten different groups | QUANTITY | 0.98+ |
ten different organizations | QUANTITY | 0.98+ |
ORGANIZATION | 0.98+ | |
hundreds of millions of dollars | QUANTITY | 0.97+ |
second call | QUANTITY | 0.97+ |
first base | QUANTITY | 0.97+ |
80 steps | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
one | QUANTITY | 0.97+ |
two guests | QUANTITY | 0.97+ |
amana | PERSON | 0.97+ |
David flora | PERSON | 0.96+ |
john | PERSON | 0.96+ |
five nodes | QUANTITY | 0.96+ |
second | QUANTITY | 0.96+ |
two polls | QUANTITY | 0.96+ |
both | QUANTITY | 0.96+ |
two hundred fifteen thousand dollars an hour | QUANTITY | 0.95+ |
a lot of questions | QUANTITY | 0.95+ |
CEO | PERSON | 0.95+ |
Mongo | ORGANIZATION | 0.94+ |
four | QUANTITY | 0.93+ |
anis | PERSON | 0.93+ |
Delton | ORGANIZATION | 0.92+ |
7so | QUANTITY | 0.92+ |
19 different backup | QUANTITY | 0.92+ |
Carson | ORGANIZATION | 0.92+ |
zillion steps | QUANTITY | 0.91+ |
AP | ORGANIZATION | 0.91+ |
2018 | DATE | 0.91+ |
theCUBE | ORGANIZATION | 0.91+ |
day one | QUANTITY | 0.9+ |
jeff workday | PERSON | 0.89+ |
many years ago | DATE | 0.87+ |
every hour | QUANTITY | 0.87+ |
a couple of senior people | QUANTITY | 0.84+ |
4 | QUANTITY | 0.84+ |
next couple weeks | DATE | 0.84+ |
Dupin | TITLE | 0.84+ |
21st vendor | QUANTITY | 0.83+ |
agile | TITLE | 0.8+ |
Day One Afternoon Keynote | Red Hat Summit 2018
[Music] [Music] [Music] [Music] ladies and gentlemen please welcome Red Hat senior vice president of engineering Matt Hicks [Music] welcome back I hope you're enjoying your first day of summit you know for us it is a lot of work throughout the year to get ready to get here but I love the energy walking into someone on that first opening day now this morning we kick off with Paul's keynote and you saw this morning just how evolved every aspect of open hybrid cloud has become based on an open source innovation model that opens source the power and potential of open source so we really brought me to Red Hat but at the end of the day the real value comes when were able to make customers like yourself successful with open source and as much passion and pride as we put into the open source community that requires more than just Red Hat given the complexity of your various businesses the solution set you're building that requires an entire technology ecosystem from system integrators that can provide the skills your domain expertise to software vendors that are going to provide the capabilities for your solutions even to the public cloud providers whether it's on the hosting side or consuming their services you need an entire technological ecosystem to be able to support you and your goals and that is exactly what we are gonna talk about this afternoon the technology ecosystem we work with that's ready to help you on your journey now you know this year's summit we talked about earlier it is about ideas worth exploring and we want to make sure you have all of the expertise you need to make those ideas a reality so with that let's talk about our first partner we have him today and that first partner is IBM when I talk about IBM I have a little bit of a nostalgia and that's because 16 years ago I was at IBM it was during my tenure at IBM where I deployed my first copy of Red Hat Enterprise Linux for a customer it's actually where I did my first professional Linux development as well you and that work on Linux it really was the spark that I had that showed me the potential that open source could have for enterprise customers now iBM has always been a steadfast supporter of Linux and a great Red Hat partner in fact this year we are celebrating 20 years of partnership with IBM but even after 20 years two decades I think we're working on some of the most innovative work that we ever have before so please give a warm welcome to Arvind Krishna from IBM to talk with us about what we are working on Arvind [Applause] hey my pleasure to be here thank you so two decades huh that's uh you know I think anything in this industry to going for two decades is special what would you say that that link is made right Hatton IBM so successful look I got to begin by first seeing something that I've been waiting to say for years it's a long strange trip it's been and for the San Francisco folks they'll get they'll get the connection you know what I was just thinking you said 16 it is strange because I probably met RedHat 20 years ago and so that's a little bit longer than you but that was out in Raleigh it was a much smaller company and when I think about the connection I think look IBM's had a long long investment and a long being a long fan of open source and when I think of Linux Linux really lights up our hardware and I think of the power box that you were showing this morning as well as the mainframe as well as all other hardware Linux really brings that to life and I think that's been at the root of our relationship yeah absolutely now I alluded to a little bit earlier we're working on some new stuff and this time it's a little bit higher in the software stack and we have before so what do you what would you say spearheaded that right so we think of software many people know about some people don't realize a lot of the words are called critical systems you know like reservation systems ATM systems retail banking a lot of the systems run on IBM software and when I say IBM software names such as WebSphere and MQ and db2 all sort of come to mind as being some of that software stack and really when I combine that with some of what you were talking about this morning along hybrid and I think this thing called containers you guys know a little about combining the two we think is going to make magic yeah and I certainly know containers and I think for myself seeing the rise of containers from just the introduction of the technology to customers consuming at mission-critical capacities it's been probably one of the fastest technology cycles I've ever seen before look we completely agree with that when you think back to what Paul talks about this morning on hybrid and we think about it we are made of firm commitment to containers all of our software will run on containers and all of our software runs Rell and you put those two together and this belief on hybrid and containers giving you their hybrid motion so that you can pick where you want to run all the software is really I think what has brought us together now even more than before yeah and the best part I think I've liked we haven't just done the product in downstream alignment we've been so tied in our technology approach we've been aligned all the way to the upstream communities absolutely look participating upstream participating in these projects really bringing all the innovation to bear you know when I hear all of you talk about you can't just be in a single company you got to tap into the world of innovation and everybody should contribute we firmly believe that instead of helping to do that is kind of why we're here yeah absolutely now the best part we're not just going to tell you about what we're doing together we're actually going to show you so how every once you tell the audience a little bit more about what we're doing I will go get the demo team ready in the back so you good okay so look we're doing a lot here together we're taking our software and we are begging to put it on top of Red Hat and openshift and really that's what I'm here to talk about for a few minutes and then we go to show it to you live and the demo guard should be with us so it'll hopefully go go well so when we look at extending our partnership it's really based on three fundamental principles and those principles are the following one it's a hybrid world every enterprise wants the ability to span across public private and their own premise world and we got to go there number two containers are strategic to both of us enterprise needs the agility you need a way to easily port things from place to place to place and containers is more than just wrapping something up containers give you all of the security the automation the deploy ability and we really firmly believe that and innovation is the path forward I mean you got to bring all the innovation to bear whether it's around security whether it's around all of the things we heard this morning around going across multiple infrastructures right the public or private and those are three firm beliefs that both of us have together so then explicitly what I'll be doing here number one all the IBM middleware is going to be certified on top of openshift and rel and through cloud private from IBM so that's number one all the middleware is going to run in rental containers on OpenShift on rail with all the cloud private automation and deployability in there number two we are going to make it so that this is the complete stack when you think about from hardware to hypervisor to os/2 the container platform to all of the middleware it's going to be certified up and down all the way so that you can get comfort that this is certified against all the cyber security attacks that come your way three because we do the certification that means a complete stack can be deployed wherever OpenShift runs so that way you give the complete flexibility and you no longer have to worry about that the development lifecycle is extended all the way from inception to production and the management plane then gives you all of the delivery and operation support needed to lower that cost and lastly professional services through the IBM garages as well as the Red Hat innovation labs and I think that this combination is really speaks to the power of both companies coming together and both of us working together to give all of you that flexibility and deployment capabilities across one can't can't help it one architecture chart and that's the only architecture chart I promise you so if you look at it right from the bottom this speaks to what I'm talking about you begin at the bottom and you have a choice of infrastructure the IBM cloud as well as other infrastructure as a service virtual machines as well as IBM power and IBM mainframe as is the infrastructure choices underneath so you choose what what is best suited for the workload well with the container service with the open shift platform managing all of that environment as well as giving the orchestration that kubernetes gives you up to the platform services from IBM cloud private so it contains the catalog of all middle we're both IBM's as well as open-source it contains all the deployment capability to go deploy that and it contains all the operational management so things like come back up if things go down worry about auto scaling all those features that you want come to you from there and that is why that combination is so so powerful but rather than just hear me talk about it I'm also going to now bring up a couple of people to talk about it and what all are they going to show you they're going to show you how you can deploy an application on this environment so you can think of that as either a cloud native application but you can also think about it as how do you modernize an application using micro services but you don't want to just keep your application always within its walls you also many times want to access different cloud services from this and how do you do that and I'm not going to tell you which ones they're going to come and tell you and how do you tackle the complexity of both hybrid data data that crosses both from the private world to the public world and as well as target the extra workloads that you want so that's kind of the sense of what you're going to see through through the demonstrations but with that I'm going to invite Chris and Michael to come up I'm not going to tell you which one's from IBM which runs from Red Hat hopefully you'll be able to make the right guess so with that Chris and Michael [Music] so so thank you Arvind hopefully people can guess which ones from Red Hat based on the shoes I you know it's some really exciting stuff that we just heard there what I believe that I'm I'm most excited about when I look out upon the audience and the opportunity for customers is with this announcement there are quite literally millions of applications now that can be modernized and made available on any cloud anywhere with the combination of IBM cloud private and OpenShift and I'm most thrilled to have mr. Michael elder a distinguished engineer from IBM here with us today and you know Michael would you maybe describe for the folks what we're actually going to go over today absolutely so when you think about how do I carry forward existing applications how do I build new applications as well you're creating micro services that always need a mixture of data and messaging and caching so this example application shows java-based micro services running on WebSphere Liberty each of which are then leveraging things like IBM MQ for messaging IBM db2 for data operational decision manager all of which is fully containerized and running on top of the Red Hat open chip container platform and in fact we're even gonna enhance stock trader to help it understand how you feel but okay hang on so I'm a little slow to the draw sometimes you said we're gonna have an application tell me how I feel exactly exactly you think about your enterprise apps you want to improve customer service understanding how your clients feel can't help you do that okay well this I'd like to see that in action all right let's do it okay so the first thing we'll do is we'll actually take a look at the catalog and here in the IBM cloud private catalog this is all of the content that's available to deploy now into this hybrid solution so we see workloads for IBM will see workloads for other open source packages etc each of these are packaged up as helm charts that are deploying a set of images that will be certified for Red Hat Linux and in this case we're going to go through and start with a simple example with a node out well click a few actions here we'll give it a name now do you have your console up over there I certainly do all right perfect so we'll deploy this into the new old namespace and will deploy notate okay alright anything happening of course it's come right up and so you know what what I really like about this is regardless of if I'm used to using IBM clout private or if I'm used to working with open shift yeah the experience is well with the tool of whatever I'm you know used to dealing with on a daily basis but I mean you know I got to tell you we we deployed node ourselves all the time what about and what about when was the last time you deployed MQ on open shift you never I maybe never all right let's fix that so MQ obviously is a critical component for messaging for lots of highly transactional systems here we'll deploy this as a container on the platform now I'm going to deploy this one again into new worlds I'm gonna disable persistence and for my application I'm going to need a queue manager so I'm going to have it automatically setup my queue manager as well now this will deploy a couple of things what do you see I see IBM in cube all right so there's your stateful set running MQ and of course there's a couple of other components that get stood up as needed here including things like credentials and secrets and the service etc but all of this is they're out of the box ok so impressive right but that's the what I think you know what I'm really looking at is maybe how a well is this running you know what else does this partnership bring when I look at IBM cloud private windows inches well so that's a key reason about why it's not just about IBM middleware running on open shift but also IBM cloud private because ultimately you need that common management plane when you deploy a container the next thing you have to worry about is how do I get its logs how do I manage its help how do I manage license consumption how do I have a common security plan right so cloud private is that enveloping wrapper around IBM middleware to provide those capabilities in a common way and so here we'll switch over to our dashboard this is our Griffin and Prometheus stack that's deployed also now on cloud private running on OpenShift and we're looking at a different namespace we're looking at the stock trader namespace we'll go back to this app here momentarily and we can see all the different pieces what if you switch over to the stock trader workspace on open shipped yeah I think we might be able to do that here hey there it is alright and so what you're gonna see here all the different pieces of this op right there's d b2 over here I see the portfolio Java microservice running on Webster Liberty I see my Redis cash I see MQ all of these are the components we saw in the architecture picture a minute ago ya know so this is really great I mean so maybe let's take a look at the actual application I see we have a fine stock trader app here now we mentioned understanding how I feel exactly you know well I feel good that this is you know a brand new stock trader app versus the one from ten years ago that don't feel like we used forever so the key thing is this app is actually all of those micro services in addition to things like business rules etc to help understand the loyalty program so one of the things we could do here is actually enhance it with a a AI service from Watson this is tone analyzer it helps me understand how that user actually feels and will be able to go through and submit some feedback to understand that user ok well let's see if we can take a look at that so I tried to click on youth clearly you're not very happy right now here I'll do one quick thing over here go for it we'll clear a cache for our sample lab so look you guys don't actually know as Michael and I just wrote this no js' front end backstage while Arvin was actually talking with Matt and we deployed it real-time using continuous integration and continuous delivery that we have available with openshift well the great thing is it's a live demo right so we're gonna do it all live all the time all right so you mentioned it'll tell me how I'm feeling right so if we look at so right there it looks like they're pretty angry probably because our cache hadn't been cleared before we started the demo maybe well that would make me angry but I should be happy because I mean I have a lot of money well it's it's more than I get today for sure so but you know again I don't want to remain angry so does Watson actually understand southern I know it speaks like eighty different languages but well you know I'm from South Carolina to understand South Carolina southern but I don't know about your North Carolina southern alright well let's give it a go here y'all done a real real know no profanity now this is live I've done a real real nice job on this here fancy demo all right hey all right likes me now all right cool and the key thing is just a quick note right it's showing you've got a free trade so we can integrate those business rules and then decide to I do put one trade if you're angry give me more it's all bringing it together into one platform all running on open show yeah and I can see the possibilities right of we've not only deployed services but getting that feedback from our customers to understand well how well the services are being used and are people really happy with what they have hey listen Michael this was amazing I read you joining us today I hope you guys enjoyed this demo as well so all of you know who this next company is as I look out through the crowd based on what I can actually see with the sun shining down on me right now I can see their influence everywhere you know Sports is in our everyday lives and these guys are equally innovative in that space as they are with hybrid cloud computing and they use that to help maintain and spread their message throughout the world of course I'm talking about Nike I think you'll enjoy this next video about Nike and their brand and then we're going to hear directly from my twitting about what they're doing with Red Hat technology new developments in the top story of the day the world has stopped turning on its axis top scientists are currently racing to come up with a solution everybody going this way [Music] the wrong way [Music] please welcome Nike vice president of infrastructure engineering Mike witig [Music] hi everybody over the last five years at Nike we have transformed our technology landscape to allow us to connect more directly to our consumers through our retail stores through Nike comm and our mobile apps the first step in doing that was redesigning our global network to allow us to have direct connectivity into both Asia and AWS in Europe in Asia and in the Americas having that proximity to those cloud providers allows us to make decisions about application workload placement based on our strategy instead of having design around latency concerns now some of those workloads are very elastic things like our sneakers app for example that needs to burst out during certain hours of the week there's certain moments of the year when we have our high heat product launches and for those type of workloads we write that code ourselves and we use native cloud services but being hybrid has allowed us to not have to write everything that would go into that app but rather just the parts that are in that application consumer facing experience and there are other back-end systems certain core functionalities like order management warehouse management finance ERP and those are workloads that are third-party applications that we host on relevent over the last 18 months we have started to deploy certain elements of those core applications into both Azure and AWS hosted on rel and at first we were pretty cautious that we started with development environments and what we realized after those first successful deployments is that are the impact of those cloud migrations on our operating model was very small and that's because the tools that we use for monitoring for security for performance tuning didn't change even though we moved those core applications into Azure in AWS because of rel under the covers and getting to the point where we have that flexibility is a real enabler as an infrastructure team that allows us to just be in the yes business and really doesn't matter where we want to deploy different workload if either cloud provider or on-prem anywhere on the planet it allows us to move much more quickly and stay much more directed to our consumers and so having rel at the core of our strategy is a huge enabler for that flexibility and allowing us to operate in this hybrid model thanks very much [Applause] what a great example it's really nice to hear an IQ story of using sort of relish that foundation to enable their hybrid clout enable their infrastructure and there's a lot that's the story we spent over ten years making that possible for rel to be that foundation and we've learned a lot in that but let's circle back for a minute to the software vendors and what kicked off the day today with IBM IBM s one of the largest software portfolios on the planet but we learned through our journey on rel that you need thousands of vendors to be able to sport you across all of your different industries solve any challenge that you might have and you need those vendors aligned with your technology direction this is doubly important when the technology direction is changing like with containers we saw that two years ago bread had introduced our container certification program now this program was focused on allowing you to identify vendors that had those shared technology goals but identification by itself wasn't enough in this fast-paced world so last year we introduced trusted content we introduced our container health index publicly grading red hats images that form the foundation for those vendor images and that was great because those of you that are familiar with containers know that you're taking software from vendors you're combining that with software from companies like Red Hat and you are putting those into a single container and for you to run those in a mission-critical capacity you have to know that we can both stand by and support those deployments but even trusted content wasn't enough so this year I'm excited that we are extending once again to introduce trusted operations now last week we announced that cube con kubernetes conference the kubernetes operator SDK the goal of the kubernetes operators is to allow any software provider on kubernetes to encode how that software should run this is a critical part of a container ecosystem not just being able to find the vendors that you want to work with not just knowing that you can trust what's inside the container but knowing that you can efficiently run that software now the exciting part is because this is so closely aligned with the upstream technology that today we already have four partners that have functioning operators specifically Couchbase dynaTrace crunchy and black dot so right out of the gate you have security monitoring data store options available to you these partners are really leading the charge in terms of what it means to run their software on OpenShift but behind these four we have many more in fact this morning we announced over 60 partners that are committed to building operators they're taking their domain expertise and the software that they wrote that they know and extending that into how you are going to run that on containers in environments like OpenShift this really brings the power of being able to find the vendors being able to trust what's inside and know that you can run their software as efficiently as anyone else on the planet but instead of just telling you about this we actually want to show you this in action so why don't we bring back up the demo team to give you a little tour of what's possible with it guys thanks Matt so Matt talked about the concept of operators and when when I think about operators and what they do it's taking OpenShift based services and making them even smarter giving you insight into how they do things for example have we had an operator for the nodejs service that I was running earlier it would have detected the problem and fixed itself but when we look at it what really operators do when I look at it from an ecosystem perspective is for ISVs it's going to be a catalyst that's going to allow them to make their services as manageable and it's flexible and as you know maintainable as any public cloud service no matter where OpenShift is running and to help demonstrate this I've got my buddy Rob here Rob are we ready on the demo front we're ready awesome now I notice this screen looks really familiar to me but you know I think we want to give folks here a dev preview of a couple of things well we want to show you is the first substantial integration of the core OS tectonic technology with OpenShift and then the other thing is we are going to dive in a little bit more into operators and their usefulness so Rob yeah so what we're looking at here is the service catalog that you know and love and openshift and we've got a few new things in here we've actually integrated operators into the Service Catalog and I'm going to take this filter and give you a look at some of them that we have today so you can see we've got a list of operators exposed and this is the same way that your developers are already used to integrating with products they're right in your catalog and so now these are actually smarter services but how can we maybe look at that I mentioned that there's maybe a new view I'm used to seeing this as a developer but I hear we've got some really cool stuff if I'm the administrator of the console yeah so we've got a whole new side of the console for cluster administrators to get a look at under the infrastructure versus this dev focused view that we're looking at today today so let's go take a look at it so the first thing you see here is we've got a really rich set of monitoring and health status so we can see that we've got some alerts firing our control plane is up and we can even do capacity planning anything that you need to do to maintenance your cluster okay so it's it's not only for the the services in the cluster and doing things that you know I may be normally as a human operator would have to do but this this console view also gives me insight into the infrastructure itself right like maybe the nodes and maybe handling the security context is that true yes so these are new capabilities that we're bringing to open shift is the ability to do node management things like drain and unscheduled nodes to do day-to-day maintenance and then as well as having security constraints and things like role bindings for example and the exciting thing about this is this is a view that you've never been able to see before it's cross-cutting across namespaces so here we've got a number of admin bindings and we can see that they're connected to a number of namespaces and these would represent our engineering teams all the groups that are using the cluster and we've never had this view before this is a perfect way to audit your security you know it actually is is pretty exciting I mean I've been fortunate enough to be on the up and shift team since day one and I know that operations view is is something that we've you know strived for and so it's really exciting to see that we can offer that now but you know really this was a we want to get into what operators do and what they can do for us and so maybe you show us what the operator console looks like yeah so let's jump on over and see all the operators that we have installed on the cluster you can see that these mirror what we saw on the Service Catalog earlier now what we care about though is this Couchbase operator and we're gonna jump into the demo namespace as I said you can share a number of different teams on a cluster so it's gonna jump into this namespace okay cool so now what we want to show you guys when we think about operators you know we're gonna have a scenario here where there's going to be multiple replicas of a Couchbase service running in the cluster and then we're going to have a stateful set and what's interesting is those two things are not enough if I'm really trying to run this as a true service where it's highly available in persistent there's things that you know as a DBA that I'm normally going to have to do if there's some sort of node failure and so what we want to demonstrate to you is where operators combined with the power that was already within OpenShift are now coming together to keep this you know particular database service highly available and something that we can continue using so Rob what have you got there yeah so as you can see we've got our couch based demo cluster running here and we can see that it's up and running we've got three members we've got an off secret this is what's controlling access to a UI that we're gonna look at in a second but what really shows the power of the operator is looking at this view of the resources that it's managing you can see that we've got a service that's doing load balancing into the cluster and then like you said we've got our pods that are actually running the software itself okay so that's cool so maybe for everyone's benefit so we can show that this is happening live could we bring up the the Couchbase console please and keep up the openshift console both sides so what we see there we go so what we see on the on the right hand side is obviously the same console Rob was working in on the left-hand side as you can see by the the actual names of the pods that are there the the couch based services that are available and so Rob maybe um let's let's kill something that's always fun to do on stage yeah this is the power of the operator it's going to recover it so let's browse on over here and kill node number two so we're gonna forcefully kill this and kick off the recovery and I see right away that because of the integration that we have with operators the Couchbase console immediately picked up that something has changed in the environment now why is that important normally a human being would have to get that alert right and so with operators now we've taken that capability and we've realized that there has been a new event within the environment this is not something that you know kubernetes or open shipped by itself would be able to understand now I'm presuming we're gonna end up doing something else it's not just seeing that it failed and sure enough there we go remember when you have a stateful application rebalancing that data and making it available is just as important as ensuring that the disk is attached so I mean Rob thank you so much for you know driving this for us today and being here I mean you know not only Couchbase but as was mentioned by matt we also have you know crunchy dynaTrace and black duck I would encourage you all to go visit their booths out on the floor today and understand what they have available which are all you know here with a dev preview and then talk to the many other partners that we have that are also looking at operators so again rub thank you for joining us today Matt come on out okay this is gonna make for an exciting year of just what it means to consume container base content I think containers change how customers can get that I believe operators are gonna change how much they can trust running that content let's circle back to one more partner this next partner we have has changed the landscape of computing specifically with their work on hardware design work on core Linux itself you know in fact I think they've become so ubiquitous with computing that we often overlook the technological marvels that they've been able to overcome now for myself I studied computer engineering so in the late 90s I had the chance to study processor design I actually got to build one of my own processors now in my case it was the most trivial processor that you could imagine it was an 8-bit subtractor which means it can subtract two numbers 256 or smaller but in that process I learned the sheer complexity that goes into processor design things like wire placements that are so close that electrons can cut through the insulation in short and then doing those wire placements across three dimensions to multiple layers jamming in as many logic components as you possibly can and again in my case this was to make a processor that could subtract two numbers but once I was done with this the second part of the course was studying the Pentium processor now remember that moment forever because looking at what the Pentium processor was able to accomplish it was like looking at alien technology and the incredible thing is that Intel our next partner has been able to keep up that alien like pace of innovation twenty years later so we're excited have Doug Fisher here let's hear a little bit more from Intel for business wide open skies an open mind no matter the context the idea of being open almost only suggests the potential of infinite possibilities and that's exactly the power of open source whether it's expanding what's possible in business the science and technology or for the greater good which is why-- open source requires the involvement of a truly diverse community of contributors to scale and succeed creating infinite possibilities for technology and more importantly what we do with it [Music] you know what Intel one of our core values is risk-taking and I'm gonna go just a bit off script for a second and say I was just backstage and I saw a gentleman that looked a lot like Scott Guthrie who runs all of Microsoft's cloud enterprise efforts wearing a red shirt talking to Cormier I'm just saying I don't know maybe I need some more sleep but that's what I saw as we approach Intel's 50th anniversary these words spoken by our co-founder Robert Noyce are as relevant today as they were decades ago don't be encumbered by history this is about breaking boundaries in technology and then go off and do something wonderful is about innovation and driving innovation in our industry and Intel we're constantly looking to break boundaries to advance our technology in the cloud in enterprise space that is no different so I'm going to talk a bit about some of the boundaries we've been breaking and innovations we've been driving at Intel starting with our Intel Xeon platform Orion Xeon scalable platform we launched several months ago which was the biggest and mark the most advanced movement in this technology in over a decade we were able to drive critical performance capabilities unmatched agility and added necessary and sufficient security to that platform I couldn't be happier with the work we do with Red Hat and ensuring that those hero features that we drive into our platform they fully expose to all of you to drive that innovation to go off and do something wonderful well there's taking advantage of the performance features or agility features like our advanced vector extensions or avx-512 or Intel quick exist those technologies are fully embraced by Red Hat Enterprise Linux or whether it's security technologies like txt or trusted execution technology are fully incorporated and we look forward to working with Red Hat on their next release to ensure that our advancements continue to be exposed and their platform and all these workloads that are driving the need for us to break boundaries and our technology are driving more and more need for flexibility and computing and that's why we're excited about Intel's family of FPGAs to help deliver that additional flexibility for you to build those capabilities in your environment we have a broad set of FPGA capabilities from our power fish at Mac's product line all the way to our performance product line on the 6/10 strat exten we have a broad set of bets FPGAs what i've been talking to customers what's really exciting is to see the combination of using our Intel Xeon scalable platform in combination with FPGAs in addition to the acceleration development capabilities we've given to software developers combining all that together to deliver better and better solutions whether it's helping to accelerate data compression well there's pattern recognition or data encryption and decryption one of the things I saw in a data center recently was taking our Intel Xeon scalable platform utilizing the capabilities of FPGA to do data encryption between servers behind the firewall all the while using the FPGA to do that they preserve those precious CPU cycles to ensure they delivered the SLA to the customer yet provided more security for their data in the data center one of the edges in cyber security is innovation and route of trust starts at the hardware we recently renewed our commitment to security with our security first pledge has really three elements to our security first pledge first is customer first urgency we have now completed the release of the micro code updates for protection on our Intel platforms nine plus years since launch to protect against things like the side channel exploits transparent and timely communication we are going to communicate timely and openly on our Intel comm website whether it's about our patches performance or other relevant information and then ongoing security assurance we drive security into every one of our products we redesigned a portion of our processor to add these partition capability which is adding additional walls between applications and user level privileges to further secure that environment from bad actors I want to pause for a second and think everyone in this room involved in helping us work through our security first pledge this isn't something we do on our own it takes everyone in this room to help us do that the partnership and collaboration was next to none it's the most amazing thing I've seen since I've been in this industry so thank you we don't stop there we continue to advance our security capabilities cross-platform solutions we recently had a conference discussion at RSA where we talked about Intel Security Essentials where we deliver a framework of capabilities and the end that are in our silicon available for those to innovate our customers and the security ecosystem to innovate on a platform in a consistent way delivering that assurance that those capabilities will be on that platform we also talked about things like our security threat technology threat detection technology is something that we believe in and we launched that at RSA incorporates several elements one is ability to utilize our internal graphics to accelerate some of the memory scanning capabilities we call this an accelerated memory scanning it allows you to use the integrated graphics to scan memory again preserving those precious cycles on the core processor Microsoft adopted this and are now incorporated into their defender product and are shipping it today we also launched our threat SDK which allows partners like Cisco to utilize telemetry information to further secure their environments for cloud workloads so we'll continue to drive differential experiences into our platform for our ecosystem to innovate and deliver more and more capabilities one of the key aspects you have to protect is data by 2020 the projection is 44 zettabytes of data will be available 44 zettabytes of data by 2025 they project that will grow to a hundred and eighty s data bytes of data massive amount of data and what all you want to do is you want to drive value from that data drive and value from that data is absolutely critical and to do that you need to have that data closer and closer to your computation this is why we've been working Intel to break the boundaries in memory technology with our investment in 3d NAND we're reducing costs and driving up density in that form factor to ensure we get warm data closer to the computing we're also innovating on form factors we have here what we call our ruler form factor this ruler form factor is designed to drive as much dense as you can in a 1u rack we're going to continue to advance the capabilities to drive one petabyte of data at low power consumption into this ruler form factor SSD form factor so our innovation continues the biggest breakthrough and memory technology in the last 25 years in memory media technology was done by Intel we call this our 3d crosspoint technology and our 3d crosspoint technology is now going to be driven into SSDs as well as in a persistent memory form factor to be on the memory bus giving you the speed of memory characteristics of memory as well as the characteristics of storage given a new tier of memory for developers to take full advantage of and as you can see Red Hat is fully committed to integrating this capability into their platform to take full advantage of that new capability so I want to thank Paul and team for engaging with us to make sure that that's available for all of you to innovate on and so we're breaking boundaries and technology across a broad set of elements that we deliver that's what we're about we're going to continue to do that not be encumbered by the past your role is to go off and doing something wonderful with that technology all ecosystems are embracing this and driving it including open source technology open source is a hub of innovation it's been that way for many many years that innovation that's being driven an open source is starting to transform many many businesses it's driving business transformation we're seeing this coming to light in the transformation of 5g driving 5g into the networked environment is a transformational moment an open source is playing a pivotal role in that with OpenStack own out and opie NFV and other open source projects were contributing to and participating in are helping drive that transformation in 5g as you do software-defined networks on our barrier breaking technology we're also seeing this transformation rapidly occurring in the cloud enterprise cloud enterprise are growing rapidly and innovation continues our work with virtualization and KVM continues to be aggressive to adopt technologies to advance and deliver more capabilities in virtualization as we look at this with Red Hat we're now working on Cube vert to help move virtualized workloads onto these platforms so that we can now have them managed at an open platform environment and Cube vert provides that so between Intel and Red Hat and the community we're investing resources to make certain that comes to product as containers a critical feature in Linux becomes more and more prevalent across the industry the growth of container elements continues at a rapid rapid pace one of the things that we wanted to bring to that is the ability to provide isolation without impairing the flexibility the speed and the footprint of a container with our clear container efforts along with hyper run v we were able to combine that and create we call cotta containers we launched this at the end of last year cotta containers is designed to have that container element available and adding elements like isolation both of these events need to have an orchestration and management capability Red Hat's OpenShift provides that capability for these workloads whether containerized or cube vert capabilities with virtual environments Red Hat openshift is designed to take that commercial capability to market and we've been working with Red Hat for several years now to develop what we call our Intel select solution Intel select solutions our Intel technology optimized for downstream workloads as we see a growth in a workload will work with a partner to optimize a solution on Intel technology to deliver the best solution that could be deployed quickly our effort here is to accelerate the adoption of these type of workloads in the market working with Red Hat's so now we're going to be deploying an Intel select solution design and optimized around Red Hat OpenShift we expect the industry's start deploying this capability very rapidly I'm excited to announce today that Lenovo is committed to be the first platform company to deliver this solution to market the Intel select solution to market will be delivered by Lenovo now I talked about what we're doing in industry and how we're transforming businesses our technology is also utilized for greater good there's no better example of this than the worked by dr. Stephen Hawking it was a sad day on March 14th of this year when dr. Stephen Hawking passed away but not before Intel had a 20-year relationship with dr. Hawking driving breakthrough capabilities innovating with him driving those robust capabilities to the rest of the world one of our Intel engineers an Intel fellow which is the highest technical achievement you can reach at Intel got to spend 10 years with dr. Hawking looking at innovative things they could do together with our technology and his breakthrough innovative thinking so I thought it'd be great to bring up our Intel fellow Lema notch Minh to talk about her work with dr. Hawking and what she learned in that experience come on up Elina [Music] great to see you Thanks something going on about the breakthrough breaking boundaries and Intel technology talk about how you use that in your work with dr. Hawking absolutely so the most important part was to really make that technology contextually aware because for people with disability every single interaction takes a long time so whether it was adapting for example the language model of his work predictor to understand whether he's gonna talk to people or whether he's writing a book on black holes or to even understand what specific application he might be using and then making sure that we're surfacing only enough actions that were relevant to reduce that amount of interaction so the tricky part is really to make all of that contextual awareness happen without totally confusing the user because it's constantly changing underneath it so how is that your work involving any open source so you know the problem with assistive technology in general is that it needs to be tailored to the specific disability which really makes it very hard and very expensive because it can't utilize the economies of scale so basically with the system that we built what we wanted to do is really enable unleashing innovation in the world right so you could take that framework you could tailor to a specific sensor for example a brain computer interface or something like that where you could actually then support a different set of users so that makes open-source a perfect fit because you could actually build and tailor and we you spoke with dr. Hawking what was this view of open source is it relevant to him so yeah so Stephen was adamant from the beginning that he wanted a system to benefit the world and not just himself so he spent a lot of time with us to actually build this system and he was adamant from day one that he would only engage with us if we were commit to actually open sourcing the technology that's fantastic and you had the privilege of working with them in 10 years I know you have some amazing stories to share so thank you so much for being here thank you so much in order for us to scale and that's what we're about at Intel is really scaling our capabilities it takes this community it takes this community of diverse capabilities it takes two births thought diverse thought of dr. Hawking couldn't be more relevant but we also are proud at Intel about leading efforts of diverse thought like women and Linux women in big data other areas like that where Intel feels that that diversity of thinking and engagement is critical for our success so as we look at Intel not to be encumbered by the past but break boundaries to deliver the technology that you all will go off and do something wonderful with we're going to remain committed to that and I look forward to continue working with you thank you and have a great conference [Applause] thank God now we have one more customer story for you today when you think about customers challenges in the technology landscape it is hard to ignore the public cloud these days public cloud is introducing capabilities that are driving the fastest rate of innovation that we've ever seen in our industry and our next customer they actually had that same challenge they wanted to tap into that innovation but they were also making bets for the long term they wanted flexibility and providers and they had to integrate to the systems that they already have and they have done a phenomenal job in executing to this so please give a warm welcome to Kerry Pierce from Cathay Pacific Kerry come on thanks very much Matt hi everyone thank you for giving me the opportunity to share a little bit about our our cloud journey let me start by telling you a little bit about Cathay Pacific we're an international airline based in Hong Kong and we serve a passenger and a cargo network to over 200 destinations in 52 countries and territories in the last seventy years and years seventy years we've made substantial investments to develop Hong Kong as one of the world's leading transportation hubs we invest in what matters most to our customers to you focusing on our exemplary service and our great product and it's both on the ground and in the air we're also investing and expanding our network beyond our multiple frequencies to the financial districts such as Tokyo New York and London and we're connecting Asia and Hong Kong with key tech hubs like San Francisco where we have multiple flights daily we're also connecting Asia in Hong Kong to places like Tel Aviv and our upcoming destination of Dublin in fact 2018 is actually going to be one of our biggest years in terms of network expansion and capacity growth and we will be launching in September our longest flight from Hong Kong direct to Washington DC and that'll be using a state-of-the-art Airbus a350 1000 aircraft so that's a little bit about Cathay Pacific let me tell you about our journey through the cloud I'm not going to go into technical details there's far smarter people out in the audience who will be able to do that for you just focus a little bit about what we were trying to achieve and the people side of it that helped us get there we had a couple of years ago no doubt the same issues that many of you do I don't think we're unique we had a traditional on-premise non-standardized fragile infrastructure it didn't meet our infrastructure needs and it didn't meet our development needs it was costly to maintain it was costly to grow and it really inhibited innovation most importantly it slowed the delivery of value to our customers at the same time you had the hype of cloud over the last few years cloud this cloud that clouds going to fix the world we were really keen on making sure we didn't get wound up and that so we focused on what we needed we started bottom up with a strategy we knew we wanted to be clouded Gnostic we wanted to have active active on-premise data centers with a single network and fabric and we wanted public clouds that were trusted and acted as an extension of that environment not independently we wanted to avoid single points of failure and we wanted to reduce inter dependencies by having loosely coupled designs and finally we wanted to be scalable we wanted to be able to cater for sudden surges of demand in a nutshell we kind of just wanted to make everything easier and a management level we wanted to be a broker of services so not one size fits all because that doesn't work but also not one of everything we want to standardize but a pragmatic range of services that met our development and support needs and worked in harmony with our public cloud not against it so we started on a journey with red hat we implemented Red Hat cloud forms and ansible to manage our hybrid cloud we also met implemented Red Hat satellite to maintain a manager environment we built a Red Hat OpenStack on crimson vironment to give us an alternative and at the same time we migrated a number of customer applications to a production public cloud open shift environment but it wasn't all Red Hat you love heard today that the Red Hat fits within an overall ecosystem we looked at a number of third-party tools and services and looked at developing those into our core solution I think at last count we had tried and tested somewhere past eight different tools and at the moment we still have around 62 in our environment that help us through that journey but let me put the technical solution aside a little bit because it doesn't matter how good your technical solution is if you don't have the culture and the people to get it right as a group we needed to be aligned for delivery and we focused on three core behaviors we focused on accountability agility and collaboration now I was really lucky we've got a pretty fantastic team for whom that was actually pretty easy but but again don't underestimate the importance of getting the culture and the people right because all the technology in the world doesn't matter if you don't have that right I asked the team what did we do differently because in our situation we didn't go out and hire a bunch of new people we didn't go out and hire a bunch of consultants we had the staff that had been with us for 10 20 and in some cases 30 years so what did we do differently it was really simple we just empowered and supported our staff we knew they were the smart ones they were the ones that were dealing with a legacy environment and they had the passion to make the change so as a team we encouraged suggestions and contributions from our overall IT community from the bottom up we started small we proved the case we told the story and then we got by him and only did did we implement wider the benefits the benefit through our staff were a huge increase in staff satisfaction reduction and application and platform outage support incidents risk free and failsafe application releases work-life balance no more midnight deployments and our application and infrastructure people could really focus on delivering customer value not on firefighting and for our end customers the people that travel with us it was really really simple we could provide a stable service that allowed for faster releases which meant we could deliver value faster in terms of stats we migrated 16 production b2c applications to a public cloud OpenShift environment in 12 months we decreased provisioning time from weeks or occasionally months we were waiting for hardware two minutes and we had a hundred percent availability of our key customer facing systems but most importantly it was about people we'd built a culture a culture of innovation that was built on a foundation of collaboration agility and accountability and that permeated throughout the IT organization not those just those people that were involved in the project everyone with an IT could see what good looked like and to see what it worked what it looked like in terms of working together and that was a key foundation for us the future for us you will have heard today everything's changing so we're going to continue to develop our open hybrid cloud onboard more public cloud service providers continue to build more modern applications and leverage the emerging technology integrate and automate everything we possibly can and leverage more open source products with the great support from the open source community so there you have it that's our journey I think we succeeded by not being over awed and by starting with the basics the technology was key obviously it's a cool component but most importantly it was a way we approached our transition we had a clear strategy that was actually developed bottom-up by the people that were involved day to day and we empowered those people to deliver and that provided benefits to both our staff and to our customers so thank you for giving the opportunity to share and I hope you enjoy the rest of the summer [Applause] I got one thanks what a great story would a great customer story to close on and we have one more partner to come up and this is a partner that all of you know that's Microsoft Microsoft has gone through an amazing transformation they've we've built an incredibly meaningful partnership with them all the way from our open source collaboration to what we do in the business side we started with support for Red Hat Enterprise Linux on hyper-v and that was truly just the beginning today we're announcing one of the most exciting joint product offerings on the market today let's please give a warm welcome to Paul correr and Scott Scott Guthrie to tell us about it guys come on out you know Scot welcome welcome to the Red Hat summer thanks for coming really appreciate it great to be here you know many surprises a lot of people when we you know published a list of speakers and then you rock you were on it and you and I are on stage here it's really really important and exciting to us exciting new partnership we've worked together a long time from the hypervisor up to common support and now around hybrid hybrid cloud maybe from your perspective a little bit of of what led us here well you know I think the thing that's really led us here is customers and you know Microsoft we've been on kind of a transformation journey the last several years where you know we really try to put customers at the center of everything that we do and you know as part of that you quickly learned from customers in terms of I'm including everyone here just you know you've got a hybrid of state you know both in terms of what you run on premises where it has a lot of Red Hat software a lot of Microsoft software and then really is they take the journey to the cloud looking at a hybrid of state in terms of how do you run that now between on-premises and a public cloud provider and so I think the thing that both of us are recognized and certainly you know our focus here at Microsoft has been you know how do we really meet customers with where they're at and where they want to go and make them successful in that journey and you know it's been fantastic working with Paul and the Red Hat team over the last two years in particular we spend a lot of time together and you know really excited about the journey ahead so um maybe you can share a bit more about the announcement where we're about to make today yeah so it's it's it's a really exciting announcement it's and really kind of I think first of its kind in that we're delivering a Red Hat openshift on Azure service that we're jointly developing and jointly managing together so this is different than sort of traditional offering where it's just running inside VMs and it's sort of two vendors working this is really a jointly managed service that we're providing with full enterprise support with a full SLA where the you know single throat to choke if you will although it's collectively both are choke the throats in terms of making sure that it works well and it's really uniquely designed around this hybrid world and in that it supports will support both Windows and Linux containers and it role you know it's the same open ship that runs both in the public cloud on Azure and on-premises and you know it's something that we hear a lot from customers I know there's a lot of people here that have asked both of us for this and super excited to be able to talk about it today and we're gonna show off the first demo of it just a bit okay well I'm gonna ask you to elaborate a bit more about this how this fits into the bigger Microsoft picture and I'll get out of your way and so thanks again thank you for coming here we go thanks Paul so I thought I'd spend just a few minutes talking about wouldn't you know that some of the work that we're doing with Microsoft Asher and the overall Microsoft cloud I didn't go deeper in terms of the new offering that we're announcing today together with red hat and show demo of it actually in action in a few minutes you know the high level in terms of you know some of the work that we've been doing at Microsoft the last couple years you know it's really been around this this journey to the cloud that we see every organization going on today and specifically the Microsoft Azure we've been providing really a cloud platform that delivers the infrastructure the application and kind of the core computing needs that organizations have as they want to be able to take advantage of what the cloud has to offer and in terms of our focus with Azure you know we've really focused we deliver lots and lots of different services and features but we focused really in particular on kind of four key themes and we see these four key themes aligning very well with the journey Red Hat it's been on and it's partly why you know we think the partnership between the two companies makes so much sense and you know for us the thing that we've been really focused on has been with a or in terms of how do we deliver a really productive cloud meaning how do we enable you to take advantage of cutting-edge technology and how do we kind of accelerate the successful adoption of it whether it's around the integration of managed services that we provide both in terms of the application space in the data space the analytic and AI space but also in terms of just the end-to-end management and development tools and how all those services work together so that teams can basically adopt them and be super successful yeah we deeply believe in hybrid and believe that the world is going to be a multi cloud and a multi distributed world and how do we enable organizations to be able to take the existing investments that they already have and be able to easily integrate them in a public cloud and with a public cloud environment and get immediate ROI on day one without how to rip and replace tons of solutions you know we're moving very aggressively in the AI space and are looking to provide a rich set of AI services both finished AI models things like speech detection vision detection object motion etc that any developer even at non data scientists can integrate to make application smarter and then we provide a rich set of AI tooling that enables organizations to build custom models and be able to integrate them also as part of their applications and with their data and then we invest very very heavily on trust Trust is sort of at the core of a sure and we now have more compliant certifications than any other cloud provider we run in more countries than any other cloud provider and we really focus around unique promises around data residency data sovereignty and privacy that are really differentiated across the industry and terms of where Iser runs today we're in 50 regions around the world so our region for us is typically a cluster of multiple data centers that are grouped together and you can see we're pretty much on every continent with the exception of Antarctica today and the beauty is you're going to be able to take the Red Hat open shift service and run it on ashore in each of these different locations and really have a truly global footprint as you look to build and deploy solutions and you know we've seen kind of this focus on productivity hybrid intelligence and Trust really resonate in the market and about 90 percent of Fortune 500 companies today are deployed on Azure and you heard Nike talked a little bit earlier this afternoon about some of their journeys as they've moved to a dot public cloud this is a small logo of just a couple of the companies that are on ashore today and what I do is actually even before we dive into the open ship demo is actually just show a quick video you know one of the companies thing there are actually several people from that organization here today Deutsche Bank who have been working with both Microsoft and Red Hat for many years Microsoft on the other side Red Hat both on the rel side and then on the OpenShift side and it's just one of these customers that have helped bring the two companies together to deliver this managed openshift service on Azure and so I'm just going to play a quick video of some of the folks that Deutsche Bank talking about their experiences and what they're trying to get out of it so we could roll the video that'd be great technology is at the absolute heart of Deutsche Bank we've recognized that the cost of running our infrastructure was particularly high there was a enormous amount of under utilization we needed a platform which was open to polyglot architecture supporting any kind of application workload across the various business lines of the third we analyzed over 60 different vendor products and we ended up with Red Hat openshift I'm super excited Microsoft or supporting Linux so strongly to adopting a hybrid approach we chose as here because Microsoft was the ideal partner to work with on constructs around security compliance business continuity as you as in all the places geographically that we need to be we have applications now able to go from a proof of concept to production in three weeks that is already breaking records openshift gives us given entities and containers allows us to apply the same sets of processes automation across a wide range of our application landscape on any given day we run between seven and twelve thousand containers across three regions we start see huge levels of cost reduction because of the level of multi-tenancy that we can achieve through containers open ship gives us an abstraction layer which is allows us to move our applications between providers without having to reconfigure or recode those applications what's really exciting for me about this journey is the way they're both Red Hat and Microsoft have embraced not just what we're doing but what each other are doing and have worked together to build open shift as a first-class citizen with Microsoft [Applause] in terms of what we're announcing today is a new fully managed OpenShift service on Azure and it's really the first fully managed service provided end-to-end across any of the cloud providers and it's jointly engineer operated and supported by both Microsoft and Red Hat and that means again sort of one service one SLA and both companies standing for a link firmly behind it really again focusing around how do we make customers successful and as part of that really providing the enterprise-grade not just isolates but also support and integration testing so you can also take advantage of all your rel and linux-based containers and all of your Windows server based containers and how can you run them in a joint way with a common management stack taking the advantage of one service and get maximum density get maximum code reuse and be able to take advantage of a containerized world in a better way than ever before and make this customer focus is very much at the center of what both companies are really centered around and so what if I do be fun is rather than just talk about openshift as actually kind of show off a little bit of a journey in terms of what this move to take advantage of it looks like and so I'd like to invite Brendan and Chris onstage who are actually going to show off a live demo of openshift on Azure in action and really walk through how to provision the service and basically how to start taking advantage of it using the full open ship ecosystem so please welcome Brendan and Chris we're going to join us on stage for a demo thanks God thanks man it's been a good afternoon so you know what we want to get into right now first I'd like to think Brandon burns for joining us from Microsoft build it's a busy week for you I'm sure your own stage there a few times as well you know what I like most about what we just announced is not only the business and technical aspects but it's that operational aspect the uniqueness the expertise that RedHat has for running OpenShift combined with the expertise that Microsoft has within Azure and customers are going to get this joint offering if you will with you know Red Hat OpenShift on Microsoft Azure and so you know kind of with that again Brendan I really appreciate you being here maybe talk to the folks about what we're going to show yeah so we're going to take a look at what it looks like to deploy OpenShift on to Azure via the new OpenShift service and the real selling point the really great part of this is the the deep integration with a cloud native app API so the same tooling that you would use to create virtual machines to create disks trade databases is now the tooling that you're going to use to create an open chip cluster so to show you this first we're going to create a resource group here so we're going to create that resource group in East us using the AZ tool that's the the azure command-line tooling a resource group is sort of a folder on Azure that holds all of your stuff so that's gonna come back into the second I've created my resource group in East us and now we're gonna use that exact same tool calling into into Azure api's to provision an open shift cluster so here we go we have AZ open shift that's our new command line tool putting it into that resource group I'm gonna get into East us alright so it's gonna take a little bit of time to deploy that open shift cluster it's doing a bunch of work behind the scenes provisioning all kinds of resources as well as credentials to access a bunch of different as your API so are we actually able to see this to you yeah so we can cut over to in just a second we can cut over to that resource group in a reload so Brendan while relating the beauty of what you know the teams have been doing together already is the fact that now open shift is a first-class citizen as it were yeah absolutely within the agent so I presume not only can I do a deployment but I can do things like scale and check my credentials and pretty much everything that I could do with any other service with that that's exactly right so we can anything that you you were used to doing via the my computer has locked up there we go the demo gods are totally with me oh there we go oh no I hit reload yeah that was that was just evil timing on the house this is another use for operators as we talked about earlier today that's right my dashboard should be coming up do I do I dare click on something that's awesome that was totally it was there there we go good job so what's really interesting about this I've also heard that it deploys you know in as little as five to six minutes which is really good for customers they want to get up and running with it but all right there we go there it is who managed to make it see that shows that it's real right you see the sweat coming off of me there but there you can see the I feel it you can see the various resources that are being created in order to create this openshift cluster virtual machines disks all of the pieces provision for you automatically via that one single command line call now of course it takes a few minutes to to create the cluster so in order to show the other side of that integration the integration between openshift and Azure I'm going to cut over to an open shipped cluster that I already have created alright so here you can see my open shift cluster that's running on Microsoft Azure I'm gonna actually log in over here and the first sign you're gonna see of the integration is it's actually using my credentials my login and going through Active Directory and any corporate policies that I may have around smart cards two-factor off anything like that authenticate myself to that open chef cluster so I'll accept that it can access my and now we're gonna load up the OpenShift web console so now this looks familiar to me oh yeah so if anybody's used OpenShift out there this is the exact same console and what we're going to show though is how this console via the open service broker and the open service broker implementation for Azure integrates natively with OpenShift all right so we can go down here and we can actually see I want to deploy a database I'm gonna deploy Mongo as my key value store that I'm going to use but you know like as we talk about management and having a OpenShift cluster that's managed for you I don't really want to have to manage my database either so I'm actually going to use cosmos DB it's a native Azure service it's a multilingual database that offers me the ability to access my data in a variety of different formats including MongoDB fully managed replicated around the world a pretty incredible service so I'm going to go ahead and create that so now Brendan what's interesting I think to me is you know we talked about the operational aspects and clearly it's not you and I running the clusters but you do need that way to interface with it and so when customers are able to deploy this all of this is out of the box there's no additional contemporary like this is what you get when you create when you use that tool to create that open chef cluster this is what you get with all of that integration ok great step through here and go ahead don't have any IP ranges there we go all right and we create that binding all right and so now behind the scenes openshift is integrated with the azure api's with all of my credentials to go ahead and create that distributed database once it's done provisioning actually all of the credentials necessary to access the database are going to be automatically populated into kubernetes available for me inside of OpenShift via service discovery to access from my application without any further work so I think that really shows not only the power of integrating openshift with an azure based API but actually the power of integrating a Druze API is inside of OpenShift to make a truly seamless experience for managing and deploying your containers across a variety of different platforms yeah hey you know Brendan this is great I know you've got a flight to catch because I think you're back onstage in a few hours but you know really appreciate you joining us today absolutely I look forward to seeing what else we do yeah absolutely thank you so much thanks guys Matt you want to come back on up thanks a lot guys if you have never had the opportunity to do a live demo in front of 8,000 people it'll give you a new appreciation for standing up there and doing it and that was really good you know every time I get the chance just to take a step back and think about the technology that we have at our command today I'm in awe just the progress over the last 10 or 20 years is incredible on to think about what might come in the next 10 or 20 years really is unthinkable you even forget 10 years what might come in the next five years even the next two years but this can create a lot of uncertainty in the environment of what's going to be to come but I believe I am certain about one thing and that is if ever there was a time when any idea is achievable it is now just think about what you've seen today every aspect of open hybrid cloud you have the world's infrastructure at your fingertips and it's not stopping you've heard about this the innovation of open source how fast that's evolving and improving this capability you've heard this afternoon from an entire technology ecosystem that's ready to help you on this journey and you've heard from customer after customer that's already started their journey in the successes that they've had you're one of the neat parts about this afternoon you will aren't later this week you will actually get to put your hands on all of this technology together in our live audience demo you know this is what some it's all about for us it's a chance to bring together the technology experts that you can work with to help formulate how to pull off those ideas we have the chance to bring together technology experts our customers and our partners and really create an environment where everyone can experience the power of open source that same spark that I talked about when I was at IBM where I understood the but intial that open-source had for enterprise customers we want to create the environment where you can have your own spark you can have that same inspiration let's make this you know in tomorrow's keynote actually you will hear a story about how open-source is changing medicine as we know it and literally saving lives it is a great example of expanding the ideas it might be possible that we came into this event with so let's make this the best summit ever thank you very much for being here let's kick things off right head down to the Welcome Reception in the expo hall and please enjoy the summit thank you all so much [Music] [Music]
SUMMARY :
from the bottom this speaks to what I'm
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Doug Fisher | PERSON | 0.99+ |
Stephen | PERSON | 0.99+ |
Brendan | PERSON | 0.99+ |
Chris | PERSON | 0.99+ |
Deutsche Bank | ORGANIZATION | 0.99+ |
Robert Noyce | PERSON | 0.99+ |
Deutsche Bank | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Michael | PERSON | 0.99+ |
Arvind | PERSON | 0.99+ |
20-year | QUANTITY | 0.99+ |
March 14th | DATE | 0.99+ |
Matt | PERSON | 0.99+ |
San Francisco | LOCATION | 0.99+ |
Nike | ORGANIZATION | 0.99+ |
Paul | PERSON | 0.99+ |
Hong Kong | LOCATION | 0.99+ |
Antarctica | LOCATION | 0.99+ |
Scott Guthrie | PERSON | 0.99+ |
2018 | DATE | 0.99+ |
Asia | LOCATION | 0.99+ |
Washington DC | LOCATION | 0.99+ |
London | LOCATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
10 years | QUANTITY | 0.99+ |
two minutes | QUANTITY | 0.99+ |
Arvin | PERSON | 0.99+ |
Tel Aviv | LOCATION | 0.99+ |
two numbers | QUANTITY | 0.99+ |
two companies | QUANTITY | 0.99+ |
2020 | DATE | 0.99+ |
Paul correr | PERSON | 0.99+ |
September | DATE | 0.99+ |
Kerry Pierce | PERSON | 0.99+ |
30 years | QUANTITY | 0.99+ |
20 years | QUANTITY | 0.99+ |
8-bit | QUANTITY | 0.99+ |
Mike witig | PERSON | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
2025 | DATE | 0.99+ |
five | QUANTITY | 0.99+ |
dr. Hawking | PERSON | 0.99+ |
Linux | TITLE | 0.99+ |
Arvind Krishna | PERSON | 0.99+ |
Dublin | LOCATION | 0.99+ |
first partner | QUANTITY | 0.99+ |
Rob | PERSON | 0.99+ |
first platform | QUANTITY | 0.99+ |
Matt Hicks | PERSON | 0.99+ |
today | DATE | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
last year | DATE | 0.99+ |
OpenShift | TITLE | 0.99+ |
last week | DATE | 0.99+ |
Action Item, Graph DataBases | April 13, 2018
>> Hi, I'm Peter Burris. Welcome to Wikibon's Action Item. (electronic music) Once again, we're broadcasting from our beautiful theCUBE Studios in Palo Alto, California. Here in the studio with me, George Gilbert, and remote, we have Neil Raden, Jim Kobielus, and David Floyer. Welcome, guys! >> Hey. >> Hi, there. >> We've got a really interesting topic today. We're going to be talking about graph databases, which probably just immediately turned off everybody. But we're actually not going to talk so much about it from a technology standpoint. We're really going to spend most of our time talking about it from the standpoint of the business problems that IT and technology are being asked to address, and the degree to which graph databases, in fact, can help us address those problems, and what do we need to do to actually address them. Human beings tend to think in terms of relationships of things to each other. So what the graph community talks about is graphed-shaped problems. And by graph-shaped problem we might mean that someone owns something and someone owns something else, or someone shares an asset, or it could be any number of different things. But we tend to think in terms of things and the relationship that those things have to other things. Now, the relational model has been an extremely successful way of representing data for a lot of different applications over the course of the last 30 years, and it's not likely to go away. But the question is, do these graph-shaped problems actually lend themselves to a new technology that can work with relational technology to accelerate the rate at which we can address new problems, accelerate the performance of those new problems, and ensure the flexibility and plasticity that we need within the application set, so that we can consistently use this as a basis for going out and extending the quality of our applications as we take on even more complex problems in the future. So let's start here. Jim Kobielus, when we think about graph databases, give us a little hint on the technology and where we are today. >> Yeah, well, graph databases have been around for quite a while in various forms, addressing various core-use cases such as social network analysis, recommendation engines, fraud detection, semantic search, and so on. The graph database technology is essentially very closely related to relational, but it's specialized to, when you think about it, Peter, the very heart of a graph-shaped business problem, the entity relationship polygram. And anybody who's studied databases has mastered, at least at a high level, entity relationship diagrams. The more complex these relationships grow among a growing range of entities, the more complex sort of the network structure becomes, in terms of linking them together at a logical level. So graph database technology was developed a while back to be able to support very complex graphs of entities, and relationships, in order to do, a lot of it's analytic. A lot of it's very focused on fast query, they call query traversal, among very large graphs, to find quick answers to questions that might involve who owns which products that they bought at which stores in which cities and are serviced by which support contractors and have which connections or interrelationships with other products they may have bought from us and our partners, so forth and so on. When you have very complex questions of this sort, they lend themselves to graph modeling. And to some degree, to the extent that you need to perform very complex queries of this sort very rapidly, graph databases, and there's a wide range of those on the market, have been optimized for that. But we also have graph abstraction layers over RDBMSes and multi-model databases. You'll find them running in IBM's databases, or Microsoft Cosmos DB, and so forth. You don't need graph-specialized databases in order to do graph queries, in order to manipulate graphs. That's the issue here. When does a specialized graph database serve your needs better than a non-graph-optimized but nonetheless graph-enabling database? That's the core question. >> So, Neil Raden, let's talk a little bit about the classes of business problems that could in fact be served by representing data utilizing a graph model. So these graph-shaped problems, independent of the underlying technology. Let's start there. What kinds of problems can business people start thinking about solving by thinking in terms of graphs of things and relationships amongst things? >> It all comes down to connectedness. That's the basis of a graph database, is how things are connected, either weakly or strongly. And these connected relationships can be very complicated. They can be based on very complex properties. A relational database is not based on, not only is it not based on connectedness, it's not based on connectedness at all. I'd like to say it's based on un-connectedness. And the whole idea in a relational database is that the intelligence about connectedness is buried in the predicate of a query. It's not in the database itself. So I don't know how overlaying graph abstractions on top of a relational database are a good idea. On the other hand, I don't know how stitching a relational database into your existing operation is going to work, either. We're going to have to see. But I can tell you that a major part of data science, machine learning, and AI is going to need to address the issue of causality, not just what's related to each other. And there's a lot of science behind using graphs to get at the causality problem. >> And we've seen, well, let's come back to that. I want to come back to that. But George Gilbert, we've kind of experienced a similar type of thing back in the '90s with the whole concept of object-orientated databases. They were represented as a way of re-conceiving data. The problem was that they had to go from the concept all the way down to the physical thing, and they didn't seem to work. What happened? >> Well it turns out, the big argument was, with object-oriented databases, we can model anything that's so much richer, especially since we're programming with objects. And it turns out, though, that theoretically, especially at that time, you could model anything down at the physical level or even the logical level in a relational database, and so those code bases were able to handle sort of similar, both ends of the use cases, both ends of the spectrum. But now that we have such extreme demands on our data management, rather than look at a whole application or multiple applications even sharing a single relational database, like some of the big enterprise apps, we have workloads within apps like recommendation engines, or a knowledge graph, which explains the relationship between people, places, and things. Or digital twins, or mapping your IT infrastructure and applications, and how they all hold together. You could do that in a relational database, but in a graph database, you can organize it so that you can have really fast analysis of these structures. But, the trade-off is, you're going to be much more restricted in how you can update the stuff. >> Alright, so we think about what happened, then, with some of the object-orientated technology, the original world database, the database was bound to the application, and the developer used the database to tell the application where to go find the data. >> George: Right. >> Relational data allowed us not to tell the applications where to find things, but rather how to find things, and that was persisted, and was very successful for a long time. Object-orientated technologies, in many respects, went back to the idea that the developer had to be very concrete about telling the application where the data was, but we didn't want to do things that way. Now, something's happened, David Floyer. One of the reasons why we had this challenge of representing data in a more abstract way across a lot of different forms without having it also being represented physically, and therefore a lot of different copies and a lot of different representations of the data which broke systems of record and everything else, was that the underlying technology was focused on just persisting data and not necessarily delivering it into these new types of datas, databases, data models, et cetera. But Flash changes that, doesn't it? Can't we imagine a world in which we can have our data in Flash and then, which is a technology that's more focused on delivering data, and then having that data be delivered to a lot of different representations, including things like graph databases, graph models. Is that accurate? >> Absolutely. In a moment I'll take it even further. I think the first point is that when we were designing real-time applications, transactional applications, we were very constrained, indeed, by the amount of data that we could get to. So, as a database administrator, I used to have a rule which you could, database developers could not issue more than 100 database calls. And the reason was that, they could always do more than that, but the applications became very unstable and they became very difficult to maintain. The cost of maintenance went up a lot. The whole area of Flash allows us to do a number of things, and the area of UniGrid enables us to do a number of things very differently. So that we can, for example, share data and have many different views of it. We can use UniGrid to be able to bring far greater amounts of power, compute power, GPUs, et cetera, to bear on specific workloads. I think the most useful thing to think about this is this type of architecture can be used to create systems of intelligence, where you have the traditional relational databases dealing with systems of record, and then you can have the AI systems, graph systems, all the other components there looking at the best way of providing data and making decisions in real time that can be fed back into the systems of record. >> Alright, alright. So let's-- >> George: I want to add to something on this. >> So, Neil, let me come back to you very quickly, sorry, George. Let me come back to Neil. I want to spend, go back to this question of what does a graph-shaped problem look like? Let's kind of run down it. We talked about AI, what about IoT, guys? Is IoT going to help us, is IoT going to drive this notion of looking at the world in terms of graphs more or less? What do you think, Neil? >> I don't know. I hadn't really thought about it, Peter, to tell you the truth. I think that one thing we leave out when we talk about graphs is we talk about, you know, nodes and edges and relationships and so forth, but you can also build a graph with very rich properties. And one thing you can get from a graph query that you can't get from a relational query, unless you write careful predicate, is it can actually do some thinking for you. It can tell you something you don't know. And I think that's important. So, without being too specific about IoT, I have to say that, you know, streaming data and trying to relate it to other data, getting down to, very quickly, what's going on, root-cause analysis, I think graph would be very helpful. >> Great, and, Jim Kobielus, how about you? >> I think, yeah I think that IoT is tailor-made for, or I should say, graph modeling and graph databases are tailor-made for the IoT. Let me explain. I think the IoT, the graph is very much a metadata technology, it's expressing context in a connected universe. Where the IoT is concerned it's all about connectivity, and so graphs, increasingly complex graphs of, say, individuals and the devices and the apps they use and locations and various contexts and so forth, these are increasingly graph-based. They're hierarchical and shifting and changing, and so in order to contextualize and personalize experience in a graph, in an IoT world, I think graph databases will be embedded in the very fabric of these environments. Microsoft has a strategy they announced about a year ago to build more of an intelligent edge around, a distributed graph across all their offerings. So I think graphs will become more important in this era, undoubtedly. >> George, what do you think? Business problems? >> Business problems on IoT. The knowledge graph that holds together digital twin, both of these lend themselves to graph modeling, but to use the object-oriented databases as an example, where object modeling took off was in the applications server, where you had the ability to program, in object-oriented language, and that mapped to a relational database. And that is an option, not the only one, but it's an option for handling graph-model data like a digital twin or IT operations. >> Well that suggests that what we're thinking about here, if we talk about graph as a metadata, and I think, Neil, this partly answers the question that you had about why would anybody want to do this, that we're representing the output of a relational data as a node in a network of data types or data forms so that the data itself may still be relationally structured, but from an application standpoint, the output of that query is, itself, a thing that is then used within the application. >> But to expand on that, if you store it underneath, as fully normalized, in relational language, laid out so that there's no duplicates and things like that, it gives you much faster update performance, but the really complex queries, typical of graph data models, would be very, very slow. So, once we have, say, more in memory technology, or we can manage under the covers the sort of multiple representations of the data-- >> Well that's what Flash is going to allow us to do. >> Okay. >> What David Floyer just talked about. >> George: Okay. >> So we can have a single, persistent, physical storage >> Yeah. >> but it can be represented in a lot of different ways so that we avoid some of the problems that you're starting to raise. If we had to copy the data and have physical, physical copies of the data on disc in a lot of different places then we would run into all kinds of consistency and update. It would probably break the model. We'd probably come back to the notion of a single data store. >> George: (mumbles) >> I want to move on here, guys. One really quick thing, David Floyer, I want to ask you. If there's, you mentioned when you were database administrator and you put restrictions on how many database actions an application or transaction was allowed to generate. When we think about what a business is going to have to do to take advantage of this, are there any particular, like one thing that we need to think about? What's going to change within an IT organization to take advantage of graph database? And we'll do the action items. >> Right. So the key here is the number of database calls can grow by a factor of probably a thousand times what it is now with what we can see is coming as technologies over the next couple of years. >> So let me put that in context, David. That's a single transaction now generating a hundred thousand, >> Correct. >> a hundred thousand database calls. >> Well, access calls to data. >> Right. >> Whatever type of database. And the important thing here is that a lot of that is going to move out, with the discussion of IoT, to where the data is coming in. Because the quicker you can do that, the earlier you can analyze that data, and you talked about IoT with possible different sources coming in, a simple one like traffic lights, for example. The traffic lights are being affected by the traffic lights around them within the city. Those sort of problems are ideal for this sort of graph database. And having all of that data locally and being processed locally in memory very, very close to where those sensors are, is going to be the key to developing solutions in this area. >> So, Neil, I've got one question from you, or one question for you. I'm going to put you on the spot. I just had a thought. And here's the thought. We talk a lot about, in some of the new technologies that could in fact be employed here, whether it be blockchain or even going back to SOA, but when we talk about what a system is going to have the authority to do about the idea of writing contracts that describe very, very discretely, what a system is or is not going to do. I have a feeling those contracts are not going to be written in relational terms. I have a feeling that, like most legal documents, they will be written in what looks more like graph terms. I'm extending that a little bit, but this has rights to do this at this point in time. Is that also, this notion of incorporating more contracts directly to how systems work, to assure that we have the appropriate authorities laid out. What do you think? Is that going to be easier or harder as a consequence of thinking in terms of these graph-shaped models? >> Boy, I don't know. Again, another thing I hadn't really thought about. But I do see some real gaps in thinking. Let me give you an analogy. OLAP databases came on the scene back in the '90s whatever. People in finance departments and whatever they loved OLAP. What they hated was the lack of scalability. And now what we see now is scalability isn't a problem and OLAP solutions are suddenly bursting out all over the place. So I think there's a role for a mental model of how you model your data and how you use it that's different from the relational model. I think the relational model has prominence and has that advantage of, what's it called? Occupancy or something. But I think that the graph is going to show some real capabilities that people are lacking right now. I think some of them are at the very high end, things, like I said, getting to causality. But I think that graph theory itself is so much richer than the simple concept of graphs that's implemented in graph databases today. >> Yeah, I agree with that totally. Okay, let's do the action item round. Jim Kobielus, I want to start with you. Jim, action item. >> Yeah, for data professionals and analytic professionals, focus on what graphs can't do, cannot do, because you hear a lot of hyperbolic, they're not useful for unstructured data or for machine learning in database. They're not as useful for schema on read. What they are useful for is the same core thing that relational is useful for which is schema on write applied to structured data. Number one. Number two, and I'll be quick on this, focus on the core use cases that are already proven out for graph databases. We've already ticked them off here, social network analysis, recommendation engines, influencer analysis, semantic web. There's a rich range of mature use cases for which semantic techniques are suited. And then finally, and I'll be very quick here, bear in mind that relational databases have been supporting graph modeling, graph traversal and so forth, for quite some time, including pretty much all the core mature enterprise databases. If you're using those databases already, and they can perform graph traversals and so forth reasonably well for your intended application, stick with that. No need to investigate the pure play, graph-optimized databases on the market. However, that said, there's plenty of good ones, including AWS is coming out with Neptune. Please explore the other alternatives, but don't feel like you have to go to a graph database first and foremost. >> Alright. David Floyer, action item. >> Action item. You are going to need to move your data center and your applications from the traditional way of thinking about it, of handling things, which is sequential copies going around, usually taking it two or three weeks. You're going to have to move towards a shared data model where the same set of data can have multiple views of it and multiple uses for multiple different types of databases. >> George Gilbert, action item. >> Okay, so when you're looking at, you have a graph-oriented problem, in other words the data is shaped like a graph, question is what type of database do you use? If you have really complex query and analysis use cases, probably best to use a graph database. If you have really complex update requirements, best to use a combination, perhaps of relational and graph or something like multi-model. We can learn from Facebook where, for years, they've built their source of truth for the social graph on a bunch of sharded MySQL databases with some layers on top. That's for analyzing the graph and doing graph searches. I'm sorry, for updating the graph and maintaining it and its integrity. But for reading the graph, they have an entirely different layer for comprehensive queries and manipulating and traversing all those relationships. So, you don't get a free lunch either way. You have to choose your sweet spots and the trade-offs associated with them. >> Alright, Neil Raden, action item. >> Well, first of all, I don't think the graph databases are subject to a lot of hype. I think it's just the opposite. I think they haven't gotten much hype at all. And maybe we're going to see that. But another thing is, a fundamental difference when you're looking at a graph and a graph query, it uses something called open world reasoning. A relational database uses closed world reasoning. I'll give you an example. Country has capital city. Now you have in your graph that China has capital city Beijing, China has capital city Beijing. That doesn't violate the graph. The graph simply understands and intuits that they're different names for the same thing. Now, if you love to write correlated sub-queries for many, many different relationships, I'd say stick to your relational database. I see unique capabilities in a graph that would be difficult to implement in a relational database. >> Alright. Thank you very much, guys. Let's talk about what the action item is for all of us. This week we talked about graph databases. We do believe that they have enormous potential, but we first off have to draw a distinction between graph theory, which is a way of looking at the world and envisioning and conceptualizing solutions to problems, and graph database technology, which has the advantages of being able, for certain classes of data models, to be able to very quickly both write and read data that is based on relationships and hierarchies and network structures that are difficult to represent in a normalized relational database manager. Ultimately, our expectation is that over the next few years, we're going to see an explosion in the class of business problems that lend themselves to a graph-modeling orientation. IoT is an example, very complex analytics systems will be an example, but it is not the only approach or the only way of doing things. But what is interesting, what is especially interesting, is over the last few years, a change in the underlying hardware technology is allowing us to utilize and expand the range of tools that we might use to support these new classes of applications. Specifically, the move to Flash allows us to sustain a single physical copy of data and then have that be represented in a lot of different ways to support a lot of different model forms and a lot of different application types, without undermining the fundamental consistency and integrity of the data itself. So that is going to allow us to utilize new types of technologies in ways that we haven't utilized before, because before, whether it was object-oriented technology or OLAP technology, there was always this problem of having to create new physical copies of data which led to enormous data administrative nightmares. So looking forward, the ability to use Flash as a basis for physically storing the data and delivering it out to a lot of different model and tool forms creates an opportunity for us to use technologies that, in fact, may more naturally map to the way that human beings think about things. Now, where is this likely to really play? We talked about IoT, we talked about other types of technologies. Where it's really likely to play is when the domain expertise of a business person is really pushing the envelope on the nature of the business problem. Historically, applications like accounting or whatnot, were very focused on highly stylized data models, things that didn't necessarily exist in the real world. You don't have double-entry bookkeeping running in the wild. You do have it in the legal code, but for some of the things that we want to build in the future, people, the devices they own, where they are, how they're doing things, that lends itself to a real-world experience and human beings tend to look at those using a graph orientation. And the expectations over the next few years, because of the changes in the physical technology, how we can store data, we will be able to utilize a new set of tools that are going to allow us to more quickly bring up applications, more naturally manage data associated with those applications, and, very important, utilize targeted technology in a broader set of complex application portfolios that are appropriate to solve that particular part of the problem, whether it's a recommendation engine or something else. Alright, so, once again, I want to thank the remote guys, Jim Kobielus, Neil Raden, and David Floyer. Thank you very much for being here. George Gilbert, you're in the studio with me. And, once again, I'm Peter Burris and you've been listening to Action Item. Thank you for joining us and we'll talk to you again soon. (electronic music)
SUMMARY :
Here in the studio with me, George Gilbert, and the degree to which graph databases, And to some degree, to the extent that you need to perform independent of the underlying technology. that the intelligence about connectedness from the concept all the way down both ends of the use cases, both ends of the spectrum. and the developer used the database and a lot of different representations of the data and the area of UniGrid enables us to do a number of things So let's-- So, Neil, let me come back to you very quickly, I have to say that, you know, and so in order to contextualize and personalize experience and that mapped to a relational database. so that the data itself may still be relationally But to expand on that, if you store it underneath, so that we avoid some of the problems What's going to change within an IT organization So the key here is the number of database calls can grow So let me put that in context, David. the earlier you can analyze that data, the authority to do about the idea of writing contracts But I think that the graph is going to show some real Okay, let's do the action item round. focus on the core use cases that are already proven out David Floyer, action item. You are going to need to move your data center and the trade-offs associated with them. are subject to a lot of hype. So looking forward, the ability to use Flash as a basis
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
David Floyer | PERSON | 0.99+ |
Jim Kobielus | PERSON | 0.99+ |
Neil Raden | PERSON | 0.99+ |
George | PERSON | 0.99+ |
Neil | PERSON | 0.99+ |
George Gilbert | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
David | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
April 13, 2018 | DATE | 0.99+ |
Peter | PERSON | 0.99+ |
Jim | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
one question | QUANTITY | 0.99+ |
two | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Palo Alto, California | LOCATION | 0.99+ |
Beijing | LOCATION | 0.99+ |
single | QUANTITY | 0.99+ |
three weeks | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
This week | DATE | 0.99+ |
One | QUANTITY | 0.99+ |
first point | QUANTITY | 0.98+ |
MySQL | TITLE | 0.98+ |
more than 100 database calls | QUANTITY | 0.98+ |
China | LOCATION | 0.98+ |
Flash | TITLE | 0.98+ |
one | QUANTITY | 0.98+ |
today | DATE | 0.97+ |
theCUBE Studios | ORGANIZATION | 0.95+ |
one thing | QUANTITY | 0.94+ |
'90s | DATE | 0.91+ |
single data store | QUANTITY | 0.88+ |
double | QUANTITY | 0.87+ |
both ends | QUANTITY | 0.85+ |
a year ago | DATE | 0.85+ |
first | QUANTITY | 0.84+ |
Number two | QUANTITY | 0.84+ |
next couple of years | DATE | 0.83+ |
years | DATE | 0.82+ |
hundred thousand | QUANTITY | 0.79+ |
last 30 years | DATE | 0.72+ |
hundred thousand database | QUANTITY | 0.72+ |
thousand times | QUANTITY | 0.72+ |
Flash | PERSON | 0.68+ |
Cosmos | TITLE | 0.67+ |
Wikibon | ORGANIZATION | 0.63+ |
database | QUANTITY | 0.57+ |
about | DATE | 0.57+ |
Number one | QUANTITY | 0.56+ |
Randy Meyer, HPE & Paul Shellard, University of Cambridge | HPE Discover 2017 Madrid
>> Announcer: Live from Madrid, Spain, it's the Cube, covering HPE Discover Madrid 2017, brought to you by Hewlett Packard Enterprise. >> Welcome back to Madrid, Spain everybody, this is the Cube, the leader in live tech coverage. We're here covering HPE Discover 2017. I'm Dave Vellante with my cohost for the week, Peter Burris, Randy Meyer is back, he's the vice president and general manager Synergy and Mission Critical Solutions at Hewlett Packard Enterprise and Paul Shellerd is here, the director of the Center for Theoretical Cosmology at Cambridge University, thank you very much for coming on the Cube. >> It's a pleasure. >> Good to see you again. >> Yeah good to be back for the second time this week. I think that's, day stay outlets play too. >> Talking about computing meets the cosmos. >> Well it's exciting, yesterday we talked about Superdome Flex that we announced, we talked about it in the commercial space, where it's taking HANA and Orcale databases to the next level but there's a whole different side to what you can do with in memory compute. It's all in this high performance computing space. You think about the problems people want to solve in fluid dynamics, in forecasting, in all sorts of analytics problems, high performance compute, one of the things it does is it generates massive amounts of data that people then want to do things with. They want to compare that data to what their model said, okay can I run that against, they want to take that data and visualize it, okay how do I go do that. The more you can do that in memory, it means it's just faster to deal with because you're not going and writing this stuff off the disk, you're not moving it to another cluster back and forth, so we're seeing this burgeoning, the HPC guys would call it fat nodes, where you want to put lots of memory and eliminate the IO to go make their jobs easier and Professor Shallard will talk about a lot of that in terms of what they're doing at the Cosmos Institute, but this is a trend, you don't have to be a university. We're seeing this inside of oil and gas companies, aerospace engineering companies, anybody that's solving these complex computational problems that have an analytical element to whether it's comparative model, visualize, do something with that once you've done that. >> Paul, explain more about what it is you do. >> Well in the Cosmos Group, of which I'm the head, we're interested in two things, cosmology, which is trying to understand where the universe comes from, the whole big bang and then we're interested in black holes, particularly their collisions which produce gravitational waves, so they're the two main areas, relativity and cosmology. >> That's a big topic. I don't even know where to start, I just want to know okay what have you learned and can you summarize it for a lay person, where are you today, what can you share with us that we can understand? >> What we do is we take our mathematical models and we make predictions about the real universe and so we try and compare those to the latest observational data. We're in a particularly exciting period of time at the moment because of a flood of new data about the universe and about black holes and in the last two years, gravitational waves were discovered, there's a Nobel prize this year so lots of things are happening. It's a very data driven science so we have to try and keep up with this flood of new data which is getting larger and larger and also with new types of data, because suddenly gravitational waves are the latest thing to look at. >> What are the sources of data and new sources of data that you're tapping? >> Well, in cosmology we're mainly interested in the cosmic microwave background. >> Peter: Yeah the sources of data are the cosmos. >> Yeah right, so this is relic radiation left over from the big bang fireball, it's like a photograph of the universe, a blueprint and then also in the distribution of galaxies, so 3D maps of the universe and we've only, we're in a new age of exploration, we've only got a tiny fraction of the universe mapped so far and we're trying to extract new information about the origin of the universe from that data. In relativity, we've got these gravitational waves, these ripples in space time, they're traversing across the universe, they're essentially earthquakes in the universe and they're sound waves or seismic waves that propagate to us from these very violent events. >> I want to take you to the gravitational waves because in many respects, it's an example of a lot of what's here in action. Here's what I mean, that the experiment and correct me if I'm wrong, but it's basically, you create a, have two lasers perpendicular to each other, shooting a signal about two or three miles in that direction and it is the most precise experiment ever undertaken because what you're doing is you're measuring the time it takes for one laser versus another laser and that time is a function of the slight stretching that comes from the gravitational rays. That is an unbelievable example of edge computing, where you have just the tolerances to do that, that's not something you can send back to the cloud, you gotta do a lot of the compute right there, right? >> That's right, yes so a gravitational wave comes by and you shrink one way and you stretch the other. >> Peter: It distorts the space time. >> Yeah you become thinner and these tiny, tiny changes are what's measured and nobody expected gravitational waves to be discovered in 2015, we all thought, oh another five years, another five years, they've always been saying, we'll discover them, we'll discover them, but it happened. >> And since then, it's been used two or three times to discover new types of things and there's now a whole, I'm sure this is very centric to what you're doing, there's now a whole concept of gravitational information, can in fact becomes an entirely new branch of cosmology, have I got that right? >> Yeah you have, it's called multimessenger astronomy now because you don't just see the universe in electromagnetic waves, in light, you hear the universe. This is qualitatively different, it's sound waves coming across the universe and so combining these two, the latest event was where they heard the event first, then they turned their telescope and they saw it. So much information came out of that, even information about cosmology, because these signals are traveling hundreds of billions of light years across to us, we're getting a picture of the whole universe as they propagate all that way, so we're able to measure the expansion rate of the universe from that point. >> The techniques for the observational, the technology for observation, what is that, how has that evolved? >> Well you've got the wrong guy here. I'm from the theory group, we're doing the predictions and these guys with their incredible technology, are seeing the data, seeing and it's imagined, the whole point is you've gotta get the predictions and then you've gotta look in the data for a needle in the haystack which is this signature of these black holes colliding. >> You think about that, I have a model, I'm looking for the needle in the haystack, that's a different way to describe an in memory analytic search pattern recognition problem, that's really what it is. This is the world's largest pattern recognition problem. >> Most precise, and literally. >> And that's an observation that confirms your theory right? >> Confirms the theory, maybe it was your theory. >> I'm actually a cosmologist, so in my group we have relativists who are actively working on the black hole collisions and making predictions about this stuff. >> But they're dampening vibration from passing trucks and these things and correcting it, it's unbelievable. But coming back to the technology, the technology is, one of the reasons why this becomes so exciting and becomes practical is because for the first time, the technology has gotten to the point where you can assume that the problem you're trying to solve, that you're focused on and you don't have to translate it in technology terms, so talk a little bit about, because in many respects, that's where business is. Business wants to be able to focus on the problem and how to think the problem differently and have the technology to just respond. They don't want to have to start with the technology and then imagine what they can do with it. >> I think from our point of view, it's a very fast moving field, things are changing, new data's coming in. The data's getting bigger and bigger because instruments are getting packed tighter and tighter, there's more information, so we've got a computational problem as well, so we've got to get more computational power but there's new types of data, like suddenly there's gravitational waves. There's new types of analysis that we want to do so we want to be able to look at this data in a very flexible way and ingest it and explore new ideas more quickly because things are happening so fast, so that's why we've adopted this in memory paradigm for a number of years now and the latest incarnation of this is the HP Superdome flex and that's a shared memory system, so you can just pull in all your data and explore it without carefully programming how the memory is distributed around. We find this is very easy for our users to develop data analytic pipelines to develop their new theoretical models and to compare the two on the single system. It's also very easy for new users to use. You don't have to be an advanced programmer to get going, you can just stay with the science in a sense. >> You gotta have a PhD in Physics to do great in Physics, you don't have to have a PhD in Physics and technology. >> That's right, yeah it's a very flexible program. A flexible architecture with which to program so you can more or less take your laptop pipeline, develop your pipeline on a laptop, take it to the Superdome and then scale it up to these huge memory problems. >> And get it done fast and you can iterate. >> You know these are the most brilliant scientists in the world, bar none, I made the analogy the other day. >> Oh, thanks. >> You're supposed to say aw, chucks. >> Peter: Aw, chucks. >> Present company excepted. >> Oh yeah, that's right. >> I made the analogy of, imagine I.M. Pei or Frank Lloyd Wright or someone had to be their own general contractor, right? No, they're brilliant at designing architectures and imagining things that no one else could imagine and then they had people to go do that. This allows the people to focus on the brilliance of the science without having to go become the expert programmer, we see that in business too. Parallel programming techniques are difficult, spoken like an old tandem guy, parallelism is hard but to the extent that you can free yourself up and focus on the problem and not have to mess around with that, it makes life easier. Some problems parallelize well, but a lot of them don't need to be and you can allow the data to shine, you can allow the science to shine. >> Is it correct that the barrier in your ability to reach a conclusion or make a discovery is the ability to find that needle in a haystack or maybe there are many, but. >> Well, if you're talking about obstacles to progress, I would say computational power isn't the obstacle, it's developing the software pipelines and it's the human personnel, the smart people writing the codes that can look for the needle in the haystack who have the efficient algorithms to do that and if they're cobbled by having to think very hard about the hardware and the architecture they're working with and how they've parallelized the problem, our philosophy is much more that you solve the problem, you validate it, it can be quite inefficient if you like, but as long as it's a working program that gets you to where you want, then your second stage you worry about making it efficient, putting it on accelerators, putting it on GPUs, making it go really fast and that's, for many years now we've bought these very flexible shared memory or in memory is the new word for it, in memory architectures which allow new users, graduate students to come straight in without a Master's degree in high performance computing, they can start to tackle problems straight away. >> It's interesting, we hear the same, you talk about it at the outer reaches of the universe, I hear it at the inner reaches of the universe from the life sciences companies, we want to map the genome and we want to understand the interaction of various drug combinations with that genetic structure to say can I tune exactly a vaccine or a drug or something else for that patient's genetic makeup to improve medical outcomes? The same kind of problem, I want to have all this data that I have to run against a complex genome sequence to find the one that gets me to the answer. From the macro to the micro, we hear this problem in all different sorts of languages. >> One of the things we have our clients, mainly in business asking us all the time, is with each, let me step back, as analysts, not the smartest people in the world, as you'll attest I'm sure for real, as analysts, we like to talk about change and we always talked about mainframe being replaced by minicomputer being replaced by this or that. I like to talk in terms of the problems that computing's been able to take on, it's been able to take on increasingly complex, challenging, more difficult problems as a consequence of the advance of technology, very much like you're saying, the advance of technology allows us to focus increasingly on the problem. What kinds of problems do you think physicists are gonna be able to attack in the next five years or so as we think about the combination of increasingly powerful computing and an increasingly simple approach to use it? >> I think the simplification you're indicating here is really going to more memory. Holding your whole workload in memory, so that you, one of the biggest bottlenecks we find is ingesting the data and then writing it out, but if you can do everything at once, then that's the key element, so one of the things we've been working on a great deal is in situ visualization for example, so that you see the black holes coming together and you see that you've set the right parameters, they haven't missed each other or something's gone wrong with your simulation, so that you do the post-processing at the same time, you never need the intermediate data products, so larger and larger memory and the computational power that balances with that large memory. It's all very well to get a fat node, but you don't have the computational power to use all those terrabytes, so that's why this in memory architecture of the Superdome Flex is much more balanced between the two. What are the problems that we're looking forward to in terms of physics? Well, in cosmology we're looking for these hints about the origin of the universe and we've made a lot of progress analyzing the Plank satellite data about the cosmic microwave background. We're honing in on theories of inflation, which is where all the structure in the universe comes from, from Heisenberg's uncertainty principle, rapid period of expansion just like inflation in the financial markets in the very early universe, okay and so we're trying to identify can we distinguish between different types and are they gonna tell us whether the universe comes from a higher dimensional theory, ten dimensions, gets reduced to three plus one or lots of clues like that, we're looking for statistical fingerprints of these different models. In gravitational waves of course, this whole new area, we think of the cosmic microwave background as a photograph of the early universe, well in fact gravitational waves look right back to the earliest moment, fractions of a nanosecond after the big bang and so it may be that the answers, the clues that we're looking for come from gravitational waves and of course there's so much in astrophysics that we'll learn about compact objects, about neutron stars, about the most energetic events there are in the whole universe. >> I never thought about the idea, because cosmic radiation background goes back what, about 300,000 years if that's right. >> Yeah that's right, you're very well informed, 400,000 years because 300 is. >> Not that well informed. >> 370,000. >> I never thought about the idea of gravitational waves as being noise from the big bang and you make sense with that. >> Well with the cosmic microwave background, we're actually looking for a primordial signal from the big bang, from inflation, so it's yeah. Well anyway, what were you gonna say Randy? >> No, I just, it's amazing the frontiers we're heading down, it's kind of an honor to be able to enable some of these things, I've spent 30 years in the technology business and heard customers tell me you transformed by business or you helped me save costs, you helped me enter a new market. Never before in 30 plus years of being in this business have I had somebody tell me the things that you're providing are helping me understand the origins of the universe. It's an honor to be affiliated with you guys. >> Oh no, the honor's mine Randy, you're producing the hardware, the tools that allow us to do this work. >> Well now the honor's ours for coming onto the Cube. >> That's right, how do we learn more about your work and your discoveries, inclusions. >> In terms of looking at. >> Are there popular authors we could read other than Stephen Hawking? >> Well, read Stephen's books, they're very good, he's got a new one called A Briefer History of Time so it's more accessible than the Brief History of Time. >> So your website is. >> Yeah our website is ctc.cam.ac.uk, the center for theoretical cosmology and we've got some popular pages there, we've got some news stories about the latest things that have happened like the HP partnership that we're developing and some nice videos about the work that we're doing actually, very nice videos of that. >> Certainly, there were several videos run here this week that if people haven't seen them, go out, they're available on Youtube, they're available at your website, they're on Stephen's Facebook page also I think. >> Can you share that website again? >> Well, actually you can get the beautiful videos of Stephen and the rest of his group on the Discover website, is that right? >> I believe so. >> So that's at HP Discover website, but your website is? >> Is ctc.cam.ac.uk and we're just about to upload those videos ourselves. >> Can I make a marketing suggestion. >> Yeah. >> Simplify that. >> Ctc.cam.ac.uk. >> Yeah right, thank you. >> We gotta get the Cube at one of these conferences, one of these physics conferences and talk about gravitational waves. >> Bone up a little bit, you're kind of embarrassing us here, 100,000 years off. >> He's better informed than you are. >> You didn't need to remind me sir. Thanks very much for coming on the Cube, great pleasure having you today. >> Thank you. >> Keep it right there everybody, Mr. Universe and I will be back after this short break. (upbeat techno music)
SUMMARY :
brought to you by Hewlett Packard Enterprise. the director of the Center for Theoretical Cosmology Yeah good to be back for the second time this week. to what you can do with in memory compute. Well in the Cosmos Group, of which I'm the head, okay what have you learned and can you summarize it and in the last two years, gravitational waves in the cosmic microwave background. in the universe and they're sound waves or seismic waves and it is the most precise experiment ever undertaken and you shrink one way and you stretch the other. Yeah you become thinner and these tiny, tiny changes of the universe from that point. I'm from the theory group, we're doing the predictions for the needle in the haystack, that's a different way and making predictions about this stuff. the technology has gotten to the point where you can assume to get going, you can just stay with the science in a sense. You gotta have a PhD in Physics to do great so you can more or less take your laptop pipeline, in the world, bar none, I made the analogy the other day. This allows the people to focus on the brilliance is the ability to find that needle in a haystack the problem, our philosophy is much more that you solve From the macro to the micro, we hear this problem One of the things we have our clients, at the same time, you never need the I never thought about the idea, Yeah that's right, you're very well informed, from the big bang and you make sense with that. from the big bang, from inflation, so it's yeah. It's an honor to be affiliated with you guys. the hardware, the tools that allow us to do this work. and your discoveries, inclusions. so it's more accessible than the Brief History of Time. that have happened like the HP partnership they're available at your website, to upload those videos ourselves. We gotta get the Cube at one of these conferences, of embarrassing us here, 100,000 years off. You didn't need to remind me sir. Keep it right there everybody, Mr. Universe and I
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Stephen | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Peter Burris | PERSON | 0.99+ |
2015 | DATE | 0.99+ |
Paul | PERSON | 0.99+ |
Randy Meyer | PERSON | 0.99+ |
Peter | PERSON | 0.99+ |
30 years | QUANTITY | 0.99+ |
Heisenberg | PERSON | 0.99+ |
Frank Lloyd Wright | PERSON | 0.99+ |
Paul Shellerd | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
Hewlett Packard Enterprise | ORGANIZATION | 0.99+ |
Cosmos Institute | ORGANIZATION | 0.99+ |
30 plus years | QUANTITY | 0.99+ |
Center for Theoretical Cosmology | ORGANIZATION | 0.99+ |
A Briefer History of Time | TITLE | 0.99+ |
Cosmos Group | ORGANIZATION | 0.99+ |
Randy | PERSON | 0.99+ |
100,000 years | QUANTITY | 0.99+ |
ten dimensions | QUANTITY | 0.99+ |
three miles | QUANTITY | 0.99+ |
yesterday | DATE | 0.99+ |
five years | QUANTITY | 0.99+ |
second stage | QUANTITY | 0.99+ |
Paul Shellard | PERSON | 0.99+ |
three | QUANTITY | 0.99+ |
ctc.cam.ac.uk | OTHER | 0.99+ |
Shallard | PERSON | 0.99+ |
Stephen Hawking | PERSON | 0.99+ |
three times | QUANTITY | 0.99+ |
Brief History of Time | TITLE | 0.99+ |
HPE | ORGANIZATION | 0.99+ |
today | DATE | 0.98+ |
first time | QUANTITY | 0.98+ |
this week | DATE | 0.98+ |
Ctc.cam.ac.uk | OTHER | 0.98+ |
two lasers | QUANTITY | 0.98+ |
Madrid, Spain | LOCATION | 0.98+ |
400,000 years | QUANTITY | 0.98+ |
hundreds of billions of light years | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
Discover | ORGANIZATION | 0.98+ |
Madrid | LOCATION | 0.98+ |
second time | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
about 300,000 years | QUANTITY | 0.96+ |
two main areas | QUANTITY | 0.96+ |
University of Cambridge | ORGANIZATION | 0.96+ |
Superdome flex | COMMERCIAL_ITEM | 0.96+ |
Nobel prize | TITLE | 0.95+ |
One | QUANTITY | 0.95+ |
about two | QUANTITY | 0.95+ |
one way | QUANTITY | 0.95+ |
one laser | QUANTITY | 0.94+ |
HANA | TITLE | 0.94+ |
single system | QUANTITY | 0.94+ |
HP Discover | ORGANIZATION | 0.94+ |
each | QUANTITY | 0.93+ |
Youtube | ORGANIZATION | 0.93+ |
HP | ORGANIZATION | 0.93+ |
two things | QUANTITY | 0.92+ |
Universe | PERSON | 0.92+ |
first | QUANTITY | 0.92+ |
Professor | PERSON | 0.89+ |
last two years | DATE | 0.88+ |
I.M. Pei | PERSON | 0.88+ |
Cube | COMMERCIAL_ITEM | 0.87+ |
370,000 | QUANTITY | 0.86+ |
Cambridge University | ORGANIZATION | 0.85+ |
Synergy | ORGANIZATION | 0.8+ |
Plank | LOCATION | 0.8+ |
300 | QUANTITY | 0.72+ |
several videos | QUANTITY | 0.65+ |
next five years | DATE | 0.64+ |
HPC | ORGANIZATION | 0.61+ |
a nanosecond | QUANTITY | 0.6+ |
Nenshad Bardoliwalla & Pranav Rastogi | BigData NYC 2017
>> Announcer: Live from Midtown Manhattan it's theCUBE. Covering Big Data New York City 2017. Brought to you by SiliconANGLE Media and its ecosystem sponsors. >> OK, welcome back everyone we're here in New York City it's theCUBE's exclusive coverage of Big Data NYC, in conjunction with Strata Data going on right around the corner. It's out third day talking to all the influencers, CEO's, entrepreneurs, people making it happen in the Big Data world. I'm John Furrier co-host of theCUBE, with my co-host here Jim Kobielus who is the Lead Analyst at Wikibon Big Data. Nenshad Bardoliwalla. >> Bar-do-li-walla. >> Bardo. >> Nenshad Bardoliwalla. >> That guy. >> Okay, done. Of Paxata, Co-Founder & Chief Product Officer it's a tongue twister, third day, being from Jersey, it's hard with our accent, but thanks for being patient with me. >> Happy to be here. >> Pranav Rastogi, Product Manager, Microsoft Azure. Guys, welcome back to theCUBE, good to see you. I apologize for that, third day blues here. So Paxata, we had your partner on Prakash. >> Prakash. >> Prakash. Really a success story, you guys have done really well launching theCUBE fun to watch you guys from launching to the success. Obviously your relationship with Microsoft super important. Talk about the relationship because I think this is really people can start connecting the dots. >> Sure, maybe I'll start and I'LL be happy to get Pranav's point of view as well. Obviously Microsoft is one of the leading brands in the world and there are many aspects of the way that Microsoft has thought about their product development journey that have really been critical to the way that we have thought about Paxata as well. If you look at the number one tool that's used by analysts the world over it's Microsoft Excel. Right, there isn't even anything that's a close second. And if you look at the the evolution of what Microsoft has done in many layers of the stack, whether it's the end user computing paradigm that Excel provides to the world. Whether it's all of their recent innovation in both hybrid cloud technologies as well as the big data technologies that Pranav is part of managing. We just see a very strong synergy between trying to combine the usage by business consumers of being able to take advantage of these big data technologies in a hybrid cloud environment. So there's a very natural resonance between the 2 companies. We're very privileged to have Microsoft Ventures as an investor in Paxata and so the opportunity for us to work with one of the great brands of all time in our industry was really a privilege for us. Yeah, and that's the corporate sides so that wasn't actually part of it. So it's a different part of Microsoft which is great. You have also business opportunity with them. >> Nenshad : We do. >> Obviously data science problem that we're seeing is that they need to get the data faster. All that prep work, seems to be the big issue. >> It does and maybe we can get Pranav's point of view from the Microsoft angle. >> Yeah so to sort of continue what Nenshad was saying, you know the data prep in general is sort of a key core competence which is problematic for lots of users, especially around the knowledge that you need to have in terms of the different tools you can use. Folks who are very proficient will do ETL or data preparation like scenarios using one of the computing engines like Hive or Spark. That's good, but there's this big audience out there who like Excel-like interface, which is easy to use a very visually rich graphical interface where you can drag and drop and can click through. And the idea behind all of this is how quickly can I get insights from my data faster. Because in a big data space, it's volume, variety and velocity. So data is coming at a very fast rate. It's changing it's growing. And if you spend lot of time just doing data prep you're losing the value of data, or the value of data would change over time. So what we're trying to do would sort of enabling Paxata or HDInsight is enabling these users to use Paxata, get insights from data faster by solving key problems of doing data prep. >> So data democracy is a term that we've been kicking around, you guys have been talking about as well. What is actually mean, because we've been teasing out first two days here at theCUBE and BigData NYC is. It's clear the community aspect of data is growing, almost on a similar path as you're seeing with open source software. That genie's out the bottle. Open source software, tier one, it won, it's only growing exponentially. That same paradigm is moving into the data world where the collaboration is super important, in this data democracy, what is that actually mean and how does that relate to you guys? >> So the perspective we have is that first something that one of our customers said, that is there is no democracy without certain degrees of governance. We all live in a in a democracy. And yet we still have rules that we have to abide by. There are still policies that society needs to follow in order for us to be successful citizens. So when when a lot of folks hear the term democracy they really think of the wild wild west, you know. And a lot of the analytic work in the enterprise does have that flavor to it, right, people download stuff to their desktop, they do a little bit of massaging of the data. They email that to their friend, their friend then makes some changes and next thing you know we have what what some folks affectionately call spread mart hell. But if you really want to democratize the technology you have to wrap not only the user experience, like Pranav described, into something that's consumable by a very large number of folks in the enterprise. You have to wrap that with the governance and collaboration capabilities so that multiple people can work off the same data set. That you can apply the permissions so that people, who is allowed to share with each other and under what circumstances are they allowed to share. Under what circumstances are you allowed to promote data from one environment to another? It may be okay for someone like me to work in a sandbox but I cannot push that to a database or HDFS or Azure BLOB storage unless I actually have the right permissions to do so. So I think what you're seeing is that, in general, technology is becoming a, always goes on this trend, towards democratization. Whether it's the phone, whether it's the television, whether it's the personal computer and the same thing is happening with data technologies and certainly companies like. >> Well, Pranav, we're talking about this when you were on theCUBE yesterday. And I want to get your thoughts on this. The old way to solve the governance problem was to put data in silos. That was easy, I'll just put it in a silo and take care of it and access control was different. But now the value of the data is about cross-pollinating and make it freely available, horizontally scalable, so that it can be used. But the same time and you need to have a new governance paradigm. So, you've got to democratize the data by making it available, addressable and use for apps. The same time there's also the concerns on how do you make sure it doesn't get in the wrong hands and so on and so forth. >> Yeah and which is also very sort of common regarding open source projects in the cloud is a how do you ensure that the user authorized to access this open source project or run it has the right credentials is authorized and stuff. So, the benefit that you sort of get in the cloud is there's a centralized authentication system. There's Azure Active Directory, so you know most enterprise would have Active Directory users. Who are then authorized to either access maybe this cluster, or maybe this workload and they can run this job and that sort of further that goes down to the data layer as well. Where we have active policies which then describe what user can access what files and what folders. So if you think about the entrance scenario there is authentication and authorization happening and for the entire system when what user can access what data. And part of what Paxata brings in the picture is like how do you visualize this governance flow as data is coming from various sources, how do you make sure that the person who has access to data does have access data, and the one who doesn't cannot access data. >> Is that the problem with data prep is just that piece of it? What is the big problem with data prep, I mean, that seems to be, everyone keeps coming back to the same problem. What is causing all this data prep. >> People not buying Paxata it's very simple. >> That's a good one. Check out Paxata they're going to solve your problems go. But seriously, there seems to be the same hole people keep digging themselves into. They gather their stuff then next thing they're in the in the same hole they got to prepare all this stuff. >> I think the previous paradigms for doing data preparation tie exactly to the data democracy themes that we're talking about here. If you only have a very silo'd group of people in the organization with very deep technical skills but don't have the business context for what they're actually trying to accomplish, you have this impedance mismatch in the organization between the people who know what they want and the people who have the tools to do it. So what we've tried to do, and again you know taking a page out of the way that Microsoft has approached solving these problems you know both in the past in the present. Is to say look we can actually take the tools that once were only in the hands of the, you know, shamans who know how to utter the right incantations and instead move that into the the common folk who actually. >> The users. >> The users themselves who know what they want to do with the data. Who understand what those data elements mean. So if you were to ask the Paxata point of view, why have we had these data prep problems? Because we've separated the people who had the tools from the people who knew what they wanted to do with it. >> So it sounds to me, correct me if this is the wrong term, that what you offer in your partnership is it basically a broad curational environment for knowledge workers. You know, to sift and sort and annotating shared data with the lineage of the data preserved in essentially a system of record that can follow the data throughout its natural life. Is that a fair characterization? >> Pranav: I would think so yeah. >> You mention, Pranav, the whole issue of how one visualizes or should visualize this entire chain of custody, as it were, for the data, is there is there any special visualization paradigm that you guys offer? Now Microsoft, you've made a fairly significant investment in graph technology throughout your portfolio. I was at Build back in May and Sacha and the others just went to town on all things to do with Microsoft Graph, will that technology be somehow at some point, now or in the future, be reflected in this overall capability that you've established here with your partner here Paxata? >> I am not sure. So far, I think what you've talked about is some Graph capabilities introduced from the Microsoft Graph that's sort of one extreme. The other side of Graph exists today as a developer you can do some Graph based queries. So you can go to Cosmos DB which had a Gremlin API. For Graph based query, so I don't know how. >> I'll get right to the question. What's the Paxata benefits of with HDInsight? How does that, just quickly, explain for the audience. What is that solution, what are the benefits? >> So the the solution is you get a one click install of installing Paxata HDInsight and the benefit is as a benefit for a user persona who's not, sort of, used to big data or Hadoop they can use a very familiar GUI-based experience to get their insights from data faster without having any knowledge of how Spark works or Hadoop works. >> And what does the Microsoft relationship bring to the table for Paxata? >> So I think it's a couple of things. One is Azure is clearly growing at an extremely fast pace. And a lot of the enterprise customers that we work with are moving many of their workloads to Azure and and these cloud based environments. Especially for us, the unique value proposition of a partner who truly understands the hybrid nature of the world. The idea that everything is going to move to the cloud or everything is going to stay on premise is too simplistic. Microsoft understood that from day one. That data would be in it and all of those different places. And they've provided enabling technologies for vendors like us. >> I'll just say it to maybe you're too coy to say it, but the bottom line is you have an Excel-like interface. They have Office 365 they're user's going to instantly love that interface because it's an easy to use interface an Excel-like it's not Excel interface per se. >> Similar. >> Metaphor, graphical user interface. >> Yes it is. >> It's clean and it's targeted at the analyst role or user. >> That's right. >> That's going to resonate in their install base. >> And combined with a lot of these new capabilities that Microsoft is rolling out from a big data perspective. So HDInsight has a very rich portfolio of runtime engines and capabilities. They're introducing new data storage layers whether it's ADLS or Azure BLOB storage, so it's really a nice way of us working together to extract and unlock a lot of the value that Microsoft. >> So, here's the tough question for you, open source projects I see Microsoft, comments were hell froze because LINUX is now part of their DNA, which was a comment I saw at the even this week in Orlando, but they're really getting behind open source. From open compute, it's just clearly new DNA's. They're they're into it. How are you guys working together in open source and what's the impact to developers because now that's only one cloud, there's other clouds out there so data's going to be an important part of it. So open source, together, you guys working together on that and what's the role for the data? >> From an open source perspective, Microsoft plays a big role in embracing open source technologies and making sure that it runs reliably in the cloud. And part of that value prop that we provide in sort of Azure HDInsight is being sure that you can run these open source big data workloads reliably in the cloud. So you can run open source like Apache, Spark, Hive, Storm, Kafka, R Server. And the hard part about running open source technology in the cloud is how do you fine tune it, and how do you configure it, how do you run it reliably. And that's what sort of what we bring in from a cloud perspective. And we also contribute back to the community based on sort of what learned by running these workloads in the cloud. And we believe you know in the broader ecosystem customers will sort of have a mixture of these combinations and their solution They'll be using some of the Microsoft solutions some open source solutions some solutions from ecosystem that's how we see our customer solution sort of being built today. >> What's the big advantage you guys have at Paxata? What's the key differentiator for why someone should work with you guys? Is it the automation? What's the key secret sauce to you guys? >> I think it's a couple of dimensions. One is I think we have come the closest in the industry to getting a user experience that matches the Excel target user. A lot of folks are attempting to do the same but the feedback we consistently get is that when the Excel user uses our solution they just, they get it. >> Was there a design criteria, was that from the beginning how you were going to do this? >> From day one. >> So you engineer everything to make it as simple as like Excel. >> We want people to use our system they shouldn't be coding, they shouldn't be writing scripts. They just need to be able. >> Good Excel you just do good macros though. >> That's right. >> So simple things like that right. >> But the second is being able to interact with the data at scale. There are a lot of solutions out there that make the mistake in our opinion of sampling very tiny amounts of data and then asking you to draw inferences and then publish that to batch jobs. Our whole approach is to smash the batch paradigm and actually bring as much into the interactive world as possible. So end users can actually point and click on 100 million rows of data, instead of the million that you would get in Excel, and get an instantaneous response. Verses designing a job in a batch paradigm and then pushing it through the the batch. >> So it's interactive data profiling over vast corpuses of data in the cloud. >> Nenshad: Correct. >> Nenshad Bardoliwalla thanks for coming on theCUBE appreciate it, congratulations on Paxata and Microsoft Azure, great to have you. Good job on everything you do with Azure. I want to give you guys props, with seeing the growth in the market and the investment's been going well, congratulations. Thanks for sharing, keep coverage here in BigData NYC more coming after this short break.
SUMMARY :
Brought to you by SiliconANGLE Media in the Big Data world. it's hard with our accent, So Paxata, we had your partner on Prakash. launching theCUBE fun to watch you guys has done in many layers of the stack, is that they need to get the data faster. from the Microsoft angle. the different tools you can use. and how does that relate to you guys? have the right permissions to do so. But the same time and you need to have So, the benefit that you sort of get in the cloud What is the big problem with data prep, But seriously, there seems to be the same hole and instead move that into the the common folk from the people who knew what they wanted to do with it. is the wrong term, that what you offer for the data, is there is there So you can go to Cosmos DB which had a Gremlin API. What's the Paxata benefits of with HDInsight? So the the solution is you get a one click install And a lot of the enterprise customers but the bottom line is you have an Excel-like interface. user interface. It's clean and it's targeted at the analyst role to extract and unlock a lot of the value So open source, together, you guys working together and making sure that it runs reliably in the cloud. A lot of folks are attempting to do the same So you engineer everything to make it as simple They just need to be able. Good Excel you just do But the second is being able to interact So it's interactive data profiling and Microsoft Azure, great to have you.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Jim Kobielus | PERSON | 0.99+ |
Jersey | LOCATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
Excel | TITLE | 0.99+ |
2 companies | QUANTITY | 0.99+ |
John Furrier | PERSON | 0.99+ |
New York City | LOCATION | 0.99+ |
Orlando | LOCATION | 0.99+ |
Nenshad | PERSON | 0.99+ |
Bardo | PERSON | 0.99+ |
Nenshad Bardoliwalla | PERSON | 0.99+ |
third day | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Office 365 | TITLE | 0.99+ |
yesterday | DATE | 0.99+ |
SiliconANGLE Media | ORGANIZATION | 0.99+ |
100 million rows | QUANTITY | 0.99+ |
BigData | ORGANIZATION | 0.99+ |
Paxata | ORGANIZATION | 0.99+ |
Microsoft Ventures | ORGANIZATION | 0.99+ |
Pranav Rastogi | PERSON | 0.99+ |
first two days | QUANTITY | 0.99+ |
one | QUANTITY | 0.98+ |
One | QUANTITY | 0.98+ |
million | QUANTITY | 0.98+ |
second | QUANTITY | 0.98+ |
Midtown Manhattan | LOCATION | 0.98+ |
Spark | TITLE | 0.98+ |
this week | DATE | 0.98+ |
first | QUANTITY | 0.97+ |
theCUBE | ORGANIZATION | 0.97+ |
one click | QUANTITY | 0.97+ |
Prakash | PERSON | 0.97+ |
Azure | TITLE | 0.97+ |
May | DATE | 0.97+ |
Wikibon Big Data | ORGANIZATION | 0.96+ |
Hadoop | TITLE | 0.96+ |
Hive | TITLE | 0.94+ |
today | DATE | 0.94+ |
Strata Data | ORGANIZATION | 0.94+ |
Pranav | PERSON | 0.93+ |
NYC | LOCATION | 0.93+ |
one cloud | QUANTITY | 0.93+ |
2017 | DATE | 0.92+ |
Apache | ORGANIZATION | 0.9+ |
Paxata | TITLE | 0.9+ |
Graph | TITLE | 0.89+ |
Pranav | ORGANIZATION | 0.88+ |