Breaking Analysis: Snowflake caught in the storm clouds
>> From the CUBE Studios in Palo Alto in Boston, bringing you data driven insights from the Cube and ETR. This is Breaking Analysis with Dave Vellante. >> A better than expected earnings report in late August got people excited about Snowflake again, but the negative sentiment in the market is weighed heavily on virtually all growth tech stocks and Snowflake is no exception. As we've stressed many times the company's management is on a long term mission to dramatically simplify the way organizations use data. Snowflake is tapping into a multi hundred billion dollar total available market and continues to grow at a rapid pace. In our view, Snowflake is embarking on its third major wave of innovation data apps, while its first and second waves are still bearing significant fruit. Now for short term traders focused on the next 90 or 180 days, that probably doesn't matter. But those taking a longer view are asking, "Should we still be optimistic about the future of this high flyer or is it just another over hyped tech play?" Hello and welcome to this week's Wiki Bond Cube Insights powered by ETR. Snowflake's Quarter just ended. And in this breaking analysis we take a look at the most recent survey data from ETR to see what clues and nuggets we can extract to predict the near term future in the long term outlook for Snowflake which is going to announce its earnings at the end of this month. Okay, so you know the story. If you've been investor in Snowflake this year, it's been painful. We said at IPO, "If you really want to own this stock on day one, just hold your nose and buy it." But like most IPOs we said there will be likely a better entry point in the future, and not surprisingly that's been the case. Snowflake IPOed a price of 120, which you couldn't touch on day one unless you got into a friends and family Delio. And if you did, you're still up 5% or so. So congratulations. But at one point last year you were up well over 200%. That's been the nature of this volatile stock, and I certainly can't help you with the timing of the market. But longer term Snowflake is targeting 10 billion in revenue for fiscal year 2028. A big number. Is it achievable? Is it big enough? Tell you what, let's come back to that. Now shorter term, our expert trader and breaking analysis contributor Chip Simonton said he got out of the stock a while ago after having taken a shot at what turned out to be a bear market rally. He pointed out that the stock had been bouncing around the 150 level for the last few months and broke that to the downside last Friday. So he'd expect 150 is where the stock is going to find resistance on the way back up, but there's no sign of support right now. He said maybe at 120, which was the July low and of course the IPO price that we just talked about. Now, perhaps earnings will be a catalyst, when Snowflake announces on November 30th, but until the mentality toward growth tech changes, nothing's likely to change dramatically according to Simonton. So now that we have that out of the way, let's take a look at the spending data for Snowflake in the ETR survey. Here's a chart that shows the time series breakdown of snowflake's net score going back to the October, 2021 survey. Now at that time, Snowflake's net score stood at a robust 77%. And remember, net score is a measure of spending velocity. It's a proprietary network, and ETR derives it from a quarterly survey of IT buyers and asks the respondents, "Are you adopting the platform new? Are you spending 6% or more? Is you're spending flat? Is you're spending down 6% or worse? Or are you leaving the platform decommissioning?" You subtract the percent of customers that are spending less or churning from those that are spending more and adopting or adopting and you get a net score. And that's expressed as a percentage of customers responding. In this chart we show Snowflake's in out of the total survey which ranges... The total survey ranges between 1,200 and 1,400 each quarter. And the very last column... Oh sorry, very last row, we show the number of Snowflake respondents that are coming in the survey from the Fortune 500 and the Global 2000. Those are two very important Snowflake constituencies. Now what this data tells us is that Snowflake exited 2021 with very strong momentum in a net score of 82%, which is off the charts and it was actually accelerating from the previous survey. Now by April that sentiment had flipped and Snowflake came down to earth with a 68% net score. Still highly elevated relative to its peers, but meaningfully down. Why was that? Because we saw a drop in new ads and an increase in flat spend. Then into the July and most recent October surveys, you saw a significant drop in the percentage of customers that were spending more. Now, notably, the percentage of customers who are contemplating adding the platform is actually staying pretty strong, but it is off a bit this past survey. And combined with a slight uptick in planned churn, net score is now down to 60%. That uptick from 0% and 1% and then 3%, it's still small, but that net score at 60% is still 20 percentage points higher than our highly elevated benchmark of 40% as you recall from listening to earlier breaking analysis. That 40% range is we consider a milestone. Anything above that is actually quite strong. But again, Snowflake is down and coming back to churn, while 3% churn is very low, in previous quarters we've seen Snowflake 0% or 1% decommissions. Now the last thing to note in this chart is the meaningful uptick in survey respondents that are citing, they're using the Snowflake platform. That's up to 212 in the survey. So look, it's hard to imagine that Snowflake doesn't feel the softening in the market like everyone else. Snowflake is guiding for around 60% growth in product revenue against the tough compare from a year ago with a 2% operating margin. So like every company, the reaction of the street is going to come down to how accurate or conservative the guide is from their CFO. Now, earlier this year, Snowflake acquired a company called Streamlit for around $800 million. Streamlit is an open source Python library and it makes it easier to build data apps with machine learning, obviously a huge trend. And like Snowflake, generally its focus is on simplifying the complex, in this case making data science easier to integrate into data apps that business people can use. So we were excited this summer in the July ETR survey to see that they added some nice data and pick on Streamlit, which we're showing here in comparison to Snowflake's core business on the left hand side. That's the data warehousing, the Streamlit pieces on the right hand side. And we show again net score over time from the previous survey for Snowflake's core database and data warehouse offering again on the left as compared to a Streamlit on the right. Snowflake's core product had 194 responses in the October, 22 survey, Streamlit had an end of 73, which is up from 52 in the July survey. So significant uptick of people responding that they're doing business in adopting Streamlit. That was pretty impressive to us. And it's hard to see, but the net scores stayed pretty constant for Streamlit at 51%. It was 52% I think in the previous quarter, well over that magic 40% mark. But when you blend it with Snowflake, it does sort of bring things down a little bit. Now there are two key points here. One is that the acquisition seems to have gained exposure right out of the gate as evidenced by the large number of responses. And two, the spending momentum. Again while it's lower than Snowflake overall, and when you blend it with Snowflake it does pull it down, it's very healthy and steady. Now let's do a little pure comparison with some of our favorite names in this space. This chart shows net score or spending velocity in the Y-axis, an overlap or presence, pervasiveness if you will, in the data set on the X-axis. That red dotted line again is that 40% highly elevated net score that we like to talk about. And that table inserted informs us as to how the companies are plotted, where the dots set up, the net score, the ins. And we're comparing a number of database players, although just a caution, Oracle includes all of Oracle including its apps. But we just put it in there for reference because it is the leader in database. Right off the bat, Snowflake jumps out with a net score of 64%. The 60% from the earlier chart, again included Streamlit. So you can see its core database, data warehouse business actually is higher than the total company average that we showed you before 'cause the Streamlit is blended in. So when you separate it out, Streamlit is right on top of data bricks. Isn't that ironic? Only Snowflake and Databricks in this selection of names are above the 40% level. You see Mongo and Couchbase, they know they're solid and Teradata cloud actually showing pretty well compared to some of the earlier survey results. Now let's isolate on the database data platform sector and see how that shapes up. And for this analysis, same XY dimensions, we've added the big giants, AWS and Microsoft and Google. And notice that those three plus Snowflake are just at or above the 40% line. Snowflake continues to lead by a significant margin in spending momentum and it keeps creeping to the right. That's that end that we talked about earlier. Now here's an interesting tidbit. Snowflake is often asked, and I've asked them myself many times, "How are you faring relative to AWS, Microsoft and Google, these big whales with Redshift and Synapse and Big Query?" And Snowflake has been telling folks that 80% of its business comes from AWS. And when Microsoft heard that, they said, "Whoa, wait a minute, Snowflake, let's partner up." 'Cause Microsoft is smart, and they understand that the market is enormous. And if they could do better with Snowflake, one, they may steal some business from AWS. And two, even if Snowflake is winning against some of the Microsoft database products, if it wins on Azure, Microsoft is going to sell more compute and more storage, more AI tools, more other stuff to these customers. Now AWS is really aggressive from a partnering standpoint with Snowflake. They're openly negotiating, not openly, but they're negotiating better prices. They're realizing that when it comes to data, the cheaper that you make the offering, the more people are going to consume. At scale economies and operating leverage are really powerful things at volume that kick in. Now Microsoft, they're coming along, they obviously get it, but Google is seemingly resistant to that type of go to market partnership. Rather than lean into Snowflake as a great partner Google's field force is kind of fighting fashion. Google itself at Cloud next heavily messaged what they call the open data cloud, which is a direct rip off of Snowflake. So what can we say about Google? They continue to be kind of behind the curve when it comes to go to market. Now just a brief aside on the competitive posture. I've seen Slootman, Frank Slootman, CEO of Snowflake in action with his prior companies and how he depositioned the competition. At Data Domain, he eviscerated a company called Avamar with their, what he called their expensive and slow post process architecture. I think he actually called it garbage, if I recall at one conference I heard him speak at. And that sort of destroyed BMC when he was at ServiceNow, kind of positioning them as the equivalent of the department of motor vehicles. And so it's interesting to hear how Snowflake openly talks about the data platforms of AWS, Microsoft, Google, and data bricks. I'll give you this sort of short bumper sticker. Redshift is just an on-prem database that AWS morphed to the cloud, which by the way is kind of true. They actually did a brilliant job of it, but it's basically a fact. Microsoft Excel, a collection of legacy databases, which also kind of morphed to run in the cloud. And even Big Query, which is considered cloud native by many if not most, is being positioned by Snowflake as originally an on-prem database to support Google's ad business, maybe. And data bricks is for those people smart enough to get it to Berkeley that love complexity. And now Snowflake doesn't, they don't mention Berkeley as far as I know. That's my addition. But you get the point. And the interesting thing about Databricks and Snowflake is a while ago in the cube I said that there was a new workload type emerging around data where you have AWS cloud, Snowflake obviously for the cloud database and Databricks data for the data science and EML, you bring those things together and there's this new workload emerging that's going to be very powerful in the future. And it's interesting to see now the aspirations of all three of these platforms are colliding. That's quite a dynamic, especially when you see both Snowflake and Databricks putting venture money and getting their hooks into the loyalties of the same companies like DBT labs and Calibra. Anyway, Snowflake's posture is that we are the pioneer in cloud native data warehouse, data sharing and now data apps. And our platform is designed for business people that want simplicity. The other guys, yes, they're formidable, but we Snowflake have an architectural lead and of course we run in multiple clouds. So it's pretty strong positioning or depositioning, you have to admit. Now I'm not sure I agree with the big query knockoffs completely. I think that's a bit of a stretch, but snowflake, as we see in the ETR survey data is winning. So in thinking about the longer term future, let's talk about what's different with Snowflake, where it's headed and what the opportunities are for the company. Snowflake put itself on the map by focusing on simplifying data analytics. What's interesting about that is the company's founders are as you probably know from Oracle. And rather than focusing on transactional data, which is Oracle's sweet spot, the stuff they worked on when they were at Oracle, the founder said, "We're going to go somewhere else. We're going to attack the data warehousing problem and the data analytics problem." And they completely re-imagined the database and how it could be applied to solve those challenges and reimagine what was possible if you had virtually unlimited compute and storage capacity. And of course Snowflake became famous for separating the compute from storage and being able to completely shut down compute so you didn't have to pay for it when you're not using it. And the ability to have multiple clusters hit the same data without making endless copies and a consumption/cloud pricing model. And then of course everyone on the planet realized, "Wow, that's a pretty good idea." Every venture capitalist in Silicon Valley has been funding companies to copy that move. And that today has pretty much become mainstream in table stakes. But I would argue that Snowflake not only had the lead, but when you look at how others are approaching this problem, it's not necessarily as clean and as elegant. Some of the startups, the early startups I think get it and maybe had an advantage of starting later, which can be a disadvantage too. But AWS is a good example of what I'm saying here. Is its version of separating compute from storage was an afterthought and it's good, it's... Given what they had it was actually quite clever and customers like it, but it's more of a, "Okay, we're going to tier to storage to lower cost, we're going to sort of dial down the compute not completely, we're not going to shut it off, we're going to minimize the compute required." It's really not true as separation is like for instance Snowflake has. But having said that, we're talking about competitors with lots of resources and cohort offerings. And so I don't want to make this necessarily all about the product, but all things being equal architecture matters, okay? So that's the cloud S-curve, the first one we're showing. Snowflake's still on that S-curve, and in and of itself it's got legs, but it's not what's going to power the company to 10 billion. The next S-curve we denote is the multi-cloud in the middle. And now while 80% of Snowflake's revenue is AWS, Microsoft is ramping up and Google, well, we'll see. But the interesting part of that curve is data sharing, and this idea of data clean rooms. I mean it really should be called the data sharing curve, but I have my reasons for calling it multi-cloud. And this is all about network effects and data gravity, and you're seeing this play out today, especially in industries like financial services and healthcare and government that are highly regulated verticals where folks are super paranoid about compliance. There not going to share data if they're going to get sued for it, if they're going to be in the front page of the Wall Street Journal for some kind of privacy breach. And what Snowflake has done is said, "Put all the data in our cloud." Now, of course now that triggers a lot of people because it's a walled garden, okay? It is. That's the trade off. It's not the Wild West, it's not Windows, it's Mac, it's more controlled. But the idea is that as different parts of the organization or even partners begin to share data that they need, it's got to be governed, it's got to be secure, it's got to be compliant, it's got to be trusted. So Snowflake introduced the idea of, they call these things stable edges. I think that's the term that they use. And they track a metric around stable edges. And so a stable edge, or think of it as a persistent edge is an ongoing relationship between two parties that last for some period of time, more than a month. It's not just a one shot deal, one a done type of, "Oh guys shared it for a day, done." It sent you an FTP, it's done. No, it's got to have trajectory over time. Four weeks or six weeks or some period of time that's meaningful. And that metric is growing. Now I think sort of a different metric that they track. I think around 20% of Snowflake customers are actively sharing data today and then they track the number of those edge relationships that exist. So that's something that's unique. Because again, most data sharing is all about making copies of data. That's great for storage companies, it's bad for auditors, and it's bad for compliance officers. And that trend is just starting out, that middle S-curve, it's going to kind of hit the base of that steep part of the S-curve and it's going to have legs through this decade we think. And then finally the third wave that we show here is what we call super cloud. That's why I called it multi-cloud before, so it could invoke super cloud. The idea that you've built a PAS layer that is purpose built for a specific objective, and in this case it's building data apps that are cloud native, shareable and governed. And is a long-term trend that's going to take some time to develop. I mean, application development platforms can take five to 10 years to mature and gain significant adoption, but this one's unique. This is a critical play for Snowflake. If it's going to compete with the big cloud players, it has to have an app development framework like Snowpark. It has to accommodate new data types like transactional data. That's why it announced this thing called UniStore last June, Snowflake a summit. And the pattern that's forming here is Snowflake is building layer upon layer with its architecture at the core. It's not currently anyway, it's not going out and saying, "All right, we're going to buy a company that's got to another billion dollars in revenue and that's how we're going to get to 10 billion." So it's not buying its way into new markets through revenue. It's actually buying smaller companies that can complement Snowflake and that it can turn into revenue for growth that fit in to the data cloud. Now as to the 10 billion by fiscal year 28, is that achievable? That's the question. Yeah, I think so. Would the momentum resources go to market product and management prowess that Snowflake has? Yes, it's definitely achievable. And one could argue to $10 billion is too conservative. Indeed, Snowflake CFO, Mike Scarpelli will fully admit his forecaster built on existing offerings. He's not including revenue as I understand it from all the new stuff that's in the pipeline because he doesn't know what it's going to look like. He doesn't know what the adoption is going to look like. He doesn't have data on that adoption, not just yet anyway. And now of course things can change quite dramatically. It's possible that is forecast for existing businesses don't materialize or competition picks them off or a company like Databricks actually is able in the longer term replicate the functionality of Snowflake with open source technologies, which would be a very competitive source of innovation. But in our view, there's plenty of room for growth, the market is enormous and the real key is, can and will Snowflake deliver on the promises of simplifying data? Of course we've heard this before from data warehouse, the data mars and data legs and master data management and ETLs and data movers and data copiers and Hadoop and a raft of technologies that have not lived up to expectations. And we've also, by the way, seen some tremendous successes in the software business with the likes of ServiceNow and Salesforce. So will Snowflake be the next great software name and hit that 10 billion magic mark? I think so. Let's reconnect in 2028 and see. Okay, we'll leave it there today. I want to thank Chip Simonton for his input to today's episode. Thanks to Alex Myerson who's on production and manages the podcast. Ken Schiffman as well. Kristin Martin and Cheryl Knight help get the word out on social media and in our newsletters. And Rob Hove is our Editor in Chief over at Silicon Angle. He does some great editing for us. Check it out for all the news. Remember all these episodes are available as podcasts. Wherever you listen, just search Breaking Analysis podcast. I publish each week on wikibon.com and siliconangle.com. Or you can email me to get in touch David.vallante@siliconangle.com. DM me @dvellante or comment on our LinkedIn post. And please do check out etr.ai, they've got the best survey data in the enterprise tech business. This is Dave Vellante for the CUBE Insights, powered by ETR. Thanks for watching, thanks for listening and we'll see you next time on breaking analysis. (upbeat music)
SUMMARY :
insights from the Cube and ETR. And the ability to have multiple
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Alex Myerson | PERSON | 0.99+ |
Mike Scarpelli | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
November 30th | DATE | 0.99+ |
Ken Schiffman | PERSON | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ORGANIZATION | 0.99+ | |
Chip Simonton | PERSON | 0.99+ |
October, 2021 | DATE | 0.99+ |
Rob Hove | PERSON | 0.99+ |
Cheryl Knight | PERSON | 0.99+ |
Frank Slootman | PERSON | 0.99+ |
Four weeks | QUANTITY | 0.99+ |
July | DATE | 0.99+ |
six weeks | QUANTITY | 0.99+ |
10 billion | QUANTITY | 0.99+ |
five | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Slootman | PERSON | 0.99+ |
BMC | ORGANIZATION | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
6% | QUANTITY | 0.99+ |
80% | QUANTITY | 0.99+ |
last year | DATE | 0.99+ |
October | DATE | 0.99+ |
Silicon Valley | LOCATION | 0.99+ |
40% | QUANTITY | 0.99+ |
1,400 | QUANTITY | 0.99+ |
$10 billion | QUANTITY | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
April | DATE | 0.99+ |
3% | QUANTITY | 0.99+ |
77% | QUANTITY | 0.99+ |
64% | QUANTITY | 0.99+ |
60% | QUANTITY | 0.99+ |
194 responses | QUANTITY | 0.99+ |
Kristin Martin | PERSON | 0.99+ |
two parties | QUANTITY | 0.99+ |
51% | QUANTITY | 0.99+ |
2% | QUANTITY | 0.99+ |
Silicon Angle | ORGANIZATION | 0.99+ |
fiscal year 28 | DATE | 0.99+ |
billion dollars | QUANTITY | 0.99+ |
0% | QUANTITY | 0.99+ |
Avamar | ORGANIZATION | 0.99+ |
52% | QUANTITY | 0.99+ |
Berkeley | LOCATION | 0.99+ |
2028 | DATE | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
Data Domain | ORGANIZATION | 0.99+ |
1% | QUANTITY | 0.99+ |
late August | DATE | 0.99+ |
two | QUANTITY | 0.99+ |
three | QUANTITY | 0.99+ |
fiscal year 2028 | DATE | 0.99+ |
Breaking Analysis Further defining Supercloud W/ tech leaders VMware, Snowflake, Databricks & others
from the cube studios in palo alto in boston bringing you data driven insights from the cube and etr this is breaking analysis with dave vellante at our inaugural super cloud 22 event we further refined the concept of a super cloud iterating on the definition the salient attributes and some examples of what is and what is not a super cloud welcome to this week's wikibon cube insights powered by etr you know snowflake has always been what we feel is one of the strongest examples of a super cloud and in this breaking analysis from our studios in palo alto we unpack our interview with benoit de javille co-founder and president of products at snowflake and we test our super cloud definition on the company's data cloud platform and we're really looking forward to your feedback first let's examine how we defl find super cloudant very importantly one of the goals of super cloud 22 was to get the community's input on the definition and iterate on previous work super cloud is an emerging computing architecture that comprises a set of services which are abstracted from the underlying primitives of hyperscale clouds we're talking about services such as compute storage networking security and other native tooling like machine learning and developer tools to create a global system that spans more than one cloud super cloud as shown on this slide has five essential properties x number of deployment models and y number of service models we're looking for community input on x and y and on the first point as well so please weigh in and contribute now we've identified these five essential elements of a super cloud let's talk about these first the super cloud has to run its services on more than one cloud leveraging the cloud native tools offered by each of the cloud providers the builder of the super cloud platform is responsible for optimizing the underlying primitives of each cloud and optimizing for the specific needs be it cost or performance or latency or governance data sharing security etc but those primitives must be abstracted such that a common experience is delivered across the clouds for both users and developers the super cloud has a metadata intelligence layer that can maximize efficiency for the specific purpose of the super cloud i.e the purpose that the super cloud is intended for and it does so in a federated model and it includes what we call a super pass this is a prerequisite that is a purpose-built component and enables ecosystem partners to customize and monetize incremental services while at the same time ensuring that the common experiences exist across clouds now in terms of deployment models we'd really like to get more feedback on this piece but here's where we are so far based on the feedback we got at super cloud 22. we see three deployment models the first is one where a control plane may run on one cloud but supports data plane interactions with more than one other cloud the second model instantiates the super cloud services on each individual cloud and within regions and can support interactions across more than one cloud with a unified interface connecting those instantiations those instances to create a common experience and the third model superimposes its services as a layer or in the case of snowflake they call it a mesh on top of the cloud on top of the cloud providers region or regions with a single global instantiation a single global instantiation of those services which spans multiple cloud providers this is our understanding from a comfort the conversation with benoit dejaville as to how snowflake approaches its solutions and for now we're going to park the service models we need to more time to flesh that out and we'll propose something shortly for you to comment on now we peppered benoit dejaville at super cloud 22 to test how the snowflake data cloud aligns to our concepts and our definition let me also say that snowflake doesn't use the term data cloud they really want to respect and they want to denigrate the importance of their hyperscale partners nor do we but we do think the hyperscalers today anyway are building or not building what we call super clouds but they are but but people who bar are building super clouds are building on top of hyperscale clouds that is a prerequisite so here are the questions that we tested with snowflake first question how does snowflake architect its data cloud and what is its deployment model listen to deja ville talk about how snowflake has architected a single system play the clip there are several ways to do this you know uh super cloud as as you name them the way we we we picked is is to create you know one single system and that's very important right the the the um [Music] there are several ways right you can instantiate you know your solution uh in every region of a cloud and and you know potentially that region could be a ws that region could be gcp so you are indeed a multi-cloud solution but snowflake we did it differently we are really creating cloud regions which are superposed on top of the cloud provider you know region infrastructure region so we are building our regions but but where where it's very different is that each region of snowflake is not one in instantiation of our service our service is global by nature we can move data from one region to the other when you land in snowflake you land into one region but but you can grow from there and you can you know exist in multiple clouds at the same time and that's very important right it's not one single i mean different instantiation of a system is one single instantiation which covers many cloud regions and many cloud providers snowflake chose the most advanced level of our three deployment models dodgeville talked about too presumably so it could maintain maximum control and ensure that common experience like the iphone model next we probed about the technical enablers of the data cloud listen to deja ville talk about snow grid he uses the term mesh and then this can get confusing with the jamaicani's data mesh concept but listen to benoit's explanation well as i said you know first we start by building you know snowflake regions we have today furry region that spawn you know the world so it's a worldwide worldwide system with many regions but all these regions are connected together they are you know meshed together with our technology we name it snow grid and that makes it hard because you know regions you know azure region can talk to a ws region or gcp regions and and as a as a user of our cloud you you don't see really these regional differences that you know regions are in different you know potentially clown when you use snowflake you can exist your your presence as an organization can be in several regions several clouds if you want geographic and and and both geographic and cloud provider so i can share data irrespective of the the cloud and i'm in the snowflake data cloud is that correct i can do that today exactly and and that's very critical right what we wanted is to remove data silos and and when you instantiate a system in one single region and that system is locked in that region you cannot communicate with other parts of the world you are locking the data in one region right and we didn't want to do that we wanted you know data to be distributed the way customer wants it to be distributed across the world and potentially sharing data at world scale now maybe there are many ways to skin the other cat meaning perhaps if a platform does instantiate in multiple places there are ways to share data but this is how snowflake chose to approach the problem next question how do you deal with latency in this big global system this is really important to us because while snowflake has some really smart people working as engineers and and the like we don't think they've solved for the speed of light problem the best people working on it as we often joke listen to benoit deja ville's comments on this topic so yes and no the the way we do it it's very expensive to do that because generally if you want to join you know data which is in which are in different regions and different cloud it's going to be very expensive because you need to move you know data every time you join it so the way we do it is that you replicate the subset of data that you want to access from one region from other regions so you can create this data mesh but data is replicated to make it very cheap and very performant too and is the snow grid does that have the metadata intelligence yes to actually can you describe that a little bit yeah snow grid is both uh a way to to exchange you know metadata about so each region of snowflake knows about all the other regions of snowflake every time we create a new region diary you know the metadata is distributed over our data cloud not only you know region knows all the regions but knows you know every organization that exists in our clouds where this organization is where data can be replicated by this organization and then of course it's it's also used as a way to uh uh exchange data right so you can exchange you know beta by scale of data size and we just had i was just receiving an email from one of our customers who moved more than four petabytes of data cross-region cross you know cloud providers in you know few days and you know it's a lot of data so it takes you know some time to move but they were able to do that online completely online and and switch over you know to the diff to the other region which is failover is very important also so yes and no probably means typically no he says yes and no probably means no so it sounds like snowflake is selectively pulling small amounts of data and replicating it where necessary but you also heard him talk about the metadata layer which is one of the essential aspects of super cloud okay next we dug into security it's one of the most important issues and we think one of the hardest parts related to deploying super cloud so we've talked about how the cloud has become the first line of defense for the cso but now with multi-cloud you have multiple first lines of defense and that means multiple shared responsibility models and multiple tool sets from different cloud providers and an expanded threat surface so listen to benoit's explanation here please play the clip this is a great question uh security has always been the most important aspect of snowflake since day one right this is the question that every customer of ours has you know how you can you guarantee the security of my data and so we secure data really tightly in region we have several layers of security it starts by by encrypting it every data at rest and that's very important a lot of customers are not doing that right you hear these attacks for example on on cloud you know where someone left you know their buckets uh uh open and then you know you can access the data because it's a non-encrypted uh so we are encrypting everything at rest we are encrypting everything in transit so a region is very secure now you know you never from one region you never access data from another region in snowflake that's why also we replicate data now the replication of that data across region or the metadata for that matter is is really highly secure so snow grits ensure that everything is encrypted everything is you know we have multiple you know encryption keys and it's you know stored in hardware you know secure modules so we we we built you know snow grids such that it's secure and it allows very secure movement of data so when we heard this explanation we immediately went to the lowest common denominator question meaning when you think about how aws for instance deals with data in motion or data and rest it might be different from how another cloud provider deals with it so how does aws uh uh uh differences for example in the aws maturity model for various you know cloud capabilities you know let's say they've got a faster nitro or graviton does it do do you have to how does snowflake deal with that do they have to slow everything else down like imagine a caravan cruising you know across the desert so you know every truck can keep up let's listen it's a great question i mean of course our software is abstracting you know all the cloud providers you know infrastructure so that when you run in one region let's say aws or azure it doesn't make any difference as far as the applications are concerned and and this abstraction of course is a lot of work i mean really really a lot of work because it needs to be secure it needs to be performance and you know every cloud and it has you know to expose apis which are uniform and and you know cloud providers even though they have potentially the same concept let's say blob storage apis are completely different the way you know these systems are secure it's completely different the errors that you can get and and the retry you know mechanism is very different from one cloud to the other performance is also different we discovered that when we were starting to port our software and and and you know we had to completely rethink how to leverage blob storage in that cloud versus that cloud because just of performance too so we had you know for example to you know stripe data so all this work is work that's you know you don't need as an application because our vision really is that applications which are running in our data cloud can you know be abstracted of all this difference and and we provide all the services all the workload that this application need whether it's transactional access to data analytical access to data you know managing you know logs managing you know metrics all of these is abstracted too such that they are not you know tied to one you know particular service of one cloud and and distributing this application across you know many regions many cloud is very seamless so from that answer we know that snowflake takes care of everything but we really don't understand the performance implications in you know in that specific case but we feel pretty certain that the promises that snowflake makes around governance and security within their data sharing construct construct will be kept now another criterion that we've proposed for super cloud is a super pass layer to create a common developer experience and an enabler for ecosystem partners to monetize please play the clip let's listen we build it you know a custom build because because as you said you know what exists in one cloud might not exist in another cloud provider right so so we have to build you know on this all these this components that modern application mode and that application need and and and and that you know goes to machine learning as i say transactional uh analytical system and the entire thing so such that they can run in isolation basically and the objective is the developer experience will be identical across those clouds yes right the developers doesn't need to worry about cloud provider and actually our system we have we didn't talk about it but the marketplace that we have which allows actually to deliver we're getting there yeah okay now we're not going to go deep into ecosystem today we've talked about snowflakes strengths in this regard but snowflake they pretty much ticked all the boxes on our super cloud attributes and definition we asked benoit dejaville to confirm that this is all shipping and available today and he also gave us a glimpse of the future play the clip and we are still developing it you know the transactional you know unistore as we call it was announced in last summit so so they are still you know working properly but but but that's the vision right and and and that's important because we talk about the infrastructure right you mentioned a lot about storage and compute but it's not only that right when you think about application they need to use the transactional database they need to use an analytical system they need to use you know machine learning so you need to provide also all these services which are consistent across all the cloud providers so you can hear deja ville talking about expanding beyond taking advantage of the core infrastructure storage and networking et cetera and bringing intelligence to the data through machine learning and ai so of course there's more to come and there better be at this company's valuation despite the recent sharp pullback in a tightening fed environment okay so i know it's cliche but everyone's comparing snowflakes and data bricks databricks has been pretty vocal about its open source posture compared to snowflakes and it just so happens that we had aligotsy on at super cloud 22 as well he wasn't in studio he had to do remote because i guess he's presenting at an investor conference this week so we had to bring him in remotely now i didn't get to do this interview john furrier did but i listened to it and captured this clip about how data bricks sees super cloud and the importance of open source take a listen to goatzee yeah i mean let me start by saying we just we're big fans of open source we think that open source is a force in software that's going to continue for you know decades hundreds of years and it's going to slowly replace all proprietary code in its way we saw that you know it could do that with the most advanced technology windows you know proprietary operating system very complicated got replaced with linux so open source can pretty much do anything and what we're seeing with the data lake house is that slowly the open source community is building a replacement for the proprietary data warehouse you know data lake machine learning real-time stack in open source and we're excited to be part of it for us delta lake is a very important project that really helps you standardize how you lay out your data in the cloud and with it comes a really important protocol called delta sharing that enables you in an open way actually for the first time ever share large data sets between organizations but it uses an open protocol so the great thing about that is you don't need to be a database customer you don't even like databricks you just need to use this open source project and you can now securely share data sets between organizations across clouds and it actually does so really efficiently just one copy of the data so you don't have to copy it if you're within the same cloud so the implication of ellie gotzi's comments is that databricks with delta sharing as john implied is playing a long game now i don't know if enough about the databricks architecture to comment in detail i got to do more research there so i reached out to my two analyst friends tony bear and sanji mohan to see what they thought because they cover these companies pretty closely here's what tony bear said quote i've viewed the divergent lake house strategies of data bricks and snowflake in the context of their roots prior to delta lake databrick's prime focus was the compute not the storage layer and more specifically they were a compute engine not a database snowflake approached from the opposite end of the pool as they originally fit the mold of the classic database company rather than a specific compute engine per se the lake house pushes both companies outside of their original comfort zones data bricks to storage snowflake to compute engine so it makes perfect sense for databricks to embrace the open source narrative at the storage layer and for snowflake to continue its walled garden approach but in the long run their strategies are already overlapping databricks is not a 100 open source company its practitioner experience has always been proprietary and now so is its sql query engine likewise snowflake has had to open up with the support of iceberg for open data lake format the question really becomes how serious snowflake will be in making iceberg a first-class citizen in its environment that is not necessarily officially branding a lake house but effectively is and likewise can databricks deliver the service levels associated with walled gardens through a more brute force approach that relies heavily on the query engine at the end of the day those are the key requirements that will matter to data bricks and snowflake customers end quote that was some deep thought by by tony thank you for that sanjay mohan added the following quote open source is a slippery slope people buy mobile phones based on open source android but it's not fully open similarly databricks delta lake was not originally fully open source and even today its photon execution engine is not we are always going to live in a hybrid world snowflake and databricks will support whatever model works best for them and their customers the big question is do customers care as deeply about which vendor has a higher degree of openness as we technology people do i believe customers evaluation criteria is far more nuanced than just to decipher each vendor's open source claims end quote okay so i had to ask dodgeville about their so-called wall garden approach and what their strategy is with apache iceberg here's what he said iceberg is is very important so just to to give some context iceberg is an open you know table format right which was you know first you know developed by netflix and netflix you know put it open source in the apache community so we embrace that's that open source standard because because it's widely used by by many um many you know companies and also many companies have you know really invested a lot of effort in building you know big data hadoop solution or data like solution and they want to use snowflake and they couldn't really use snowflake because all their data were in open you know formats so we are embracing icebergs to help these companies move through the cloud but why we have been relentless with direct access to data direct access to data is a little bit of a problem for us and and the reason is when you direct access to data now you have direct access to storage now you have to understand for example the specificity of one cloud versus the other so as soon as you start to have direct access to data you lose your you know your cloud diagnostic layer you don't access data with api when you have direct access to data it's very hard to secure data because you need to grant access direct access to tools which are not you know protected and you see a lot of you know hacking of of data you know because of that so so that was not you know direct access to data is not serving well our customers and that's why we have been relented to do that because it's it's cr it's it's not cloud diagnostic it's it's you you have to code that you have to you you you need a lot of intelligence while apis access so we want open apis that's that's i guess the way we embrace you know openness is is by open api versus you know you access directly data here's my take snowflake is hedging its bets because enough people care about open source that they have to have some open data format options and it's good optics and you heard benoit deja ville talk about the risks of directly accessing the data and the complexities it brings now is that maybe a little fud against databricks maybe but same can be said for ollie's comments maybe flooding the proprietaryness of snowflake but as both analysts pointed out open is a spectrum hey i remember unix used to equal open systems okay let's end with some etr spending data and why not compare snowflake and data bricks spending profiles this is an xy graph with net score or spending momentum on the y-axis and pervasiveness or overlap in the data set on the x-axis this is data from the january survey when snowflake was holding above 80 percent net score off the charts databricks was also very strong in the upper 60s now let's fast forward to this next chart and show you the july etr survey data and you can see snowflake has come back down to earth now remember anything above 40 net score is highly elevated so both companies are doing well but snowflake is well off its highs and data bricks has come down somewhat as well databricks is inching to the right snowflake rocketed to the right post its ipo and as we know databricks wasn't able to get to ipo during the covet bubble ali gotzi is at the morgan stanley ceo conference this week they got plenty of cash to withstand a long-term recession i'm told and they've started the message that they're a billion dollars in annualized revenue i'm not sure exactly what that means i've seen some numbers on their gross margins i'm not sure what that means i've seen some numbers on their net retention revenue or net revenue retention again i'll reserve judgment until we see an s1 but it's clear both of these companies have momentum and they're out competing in the market well as always be the ultimate arbiter different philosophies perhaps is it like democrats and republicans well it could be but they're both going after a solving data problem both companies are trying to help customers get more value out of their data and both companies are highly valued so they have to perform for their investors to paraphrase ralph nader the similarities may be greater than the differences okay that's it for today thanks to the team from palo alto for this awesome super cloud studio build alex myerson and ken shiffman are on production in the palo alto studios today kristin martin and sheryl knight get the word out to our community rob hoff is our editor-in-chief over at siliconangle thanks to all please check out etr.ai for all the survey data remember these episodes are all available as podcasts wherever you listen just search breaking analysis podcasts i publish each week on wikibon.com and siliconangle.com and you can email me at david.vellante at siliconangle.com or dm me at devellante or comment on my linkedin posts and please as i say etr has got some of the best survey data in the business we track it every quarter and really excited to be partners with them this is dave vellante for the cube insights powered by etr thanks for watching and we'll see you next time on breaking analysis [Music] you
SUMMARY :
and and the retry you know mechanism is
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
netflix | ORGANIZATION | 0.99+ |
john furrier | PERSON | 0.99+ |
palo alto | ORGANIZATION | 0.99+ |
tony bear | PERSON | 0.99+ |
boston | LOCATION | 0.99+ |
sanji mohan | PERSON | 0.99+ |
ken shiffman | PERSON | 0.99+ |
both | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
ellie gotzi | PERSON | 0.99+ |
VMware | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
siliconangle.com | OTHER | 0.99+ |
more than four petabytes | QUANTITY | 0.99+ |
first point | QUANTITY | 0.99+ |
kristin martin | PERSON | 0.99+ |
both companies | QUANTITY | 0.99+ |
first question | QUANTITY | 0.99+ |
rob hoff | PERSON | 0.99+ |
more than one | QUANTITY | 0.99+ |
second model | QUANTITY | 0.98+ |
alex myerson | PERSON | 0.98+ |
third model | QUANTITY | 0.98+ |
one region | QUANTITY | 0.98+ |
one copy | QUANTITY | 0.98+ |
one region | QUANTITY | 0.98+ |
five essential elements | QUANTITY | 0.98+ |
android | TITLE | 0.98+ |
100 | QUANTITY | 0.98+ |
first line | QUANTITY | 0.98+ |
Databricks | ORGANIZATION | 0.98+ |
sheryl | PERSON | 0.98+ |
more than one cloud | QUANTITY | 0.98+ |
first | QUANTITY | 0.98+ |
iphone | COMMERCIAL_ITEM | 0.98+ |
super cloud 22 | EVENT | 0.98+ |
each cloud | QUANTITY | 0.98+ |
each | QUANTITY | 0.97+ |
sanjay mohan | PERSON | 0.97+ |
john | PERSON | 0.97+ |
republicans | ORGANIZATION | 0.97+ |
this week | DATE | 0.97+ |
hundreds of years | QUANTITY | 0.97+ |
siliconangle | ORGANIZATION | 0.97+ |
each week | QUANTITY | 0.97+ |
data lake house | ORGANIZATION | 0.97+ |
one single region | QUANTITY | 0.97+ |
january | DATE | 0.97+ |
dave vellante | PERSON | 0.96+ |
each region | QUANTITY | 0.96+ |
one | QUANTITY | 0.96+ |
dave vellante | PERSON | 0.96+ |
tony | PERSON | 0.96+ |
above 80 percent | QUANTITY | 0.95+ |
more than one cloud | QUANTITY | 0.95+ |
more than one cloud | QUANTITY | 0.95+ |
data lake | ORGANIZATION | 0.95+ |
five essential properties | QUANTITY | 0.95+ |
democrats | ORGANIZATION | 0.95+ |
first time | QUANTITY | 0.95+ |
july | DATE | 0.94+ |
linux | TITLE | 0.94+ |
etr | ORGANIZATION | 0.94+ |
devellante | ORGANIZATION | 0.93+ |
dodgeville | ORGANIZATION | 0.93+ |
each vendor | QUANTITY | 0.93+ |
super cloud 22 | ORGANIZATION | 0.93+ |
delta lake | ORGANIZATION | 0.92+ |
three deployment models | QUANTITY | 0.92+ |
first lines | QUANTITY | 0.92+ |
dejaville | LOCATION | 0.92+ |
day one | QUANTITY | 0.92+ |
Super Data Cloud | Supercloud22
(electronic music) >> Welcome back to our studios in Palo Alto, California. My name is Dave Vellante, I'm here with John Furrier, who is taking a quick break. You know, in one of the early examples that we used of so called super cloud was Snowflake. We called it a super data cloud. We had, really, a lot of fun with that. And we've started to evolve our thinking. Years ago, we said that data was going to form in the cloud around industries and ecosystems. And Benoit Dogeville is a many time guest of theCube. He's the co-founder and president of products at Snowflake. Benoit, thanks for spending some time with us, at Supercloud 22, good to see you. >> Thank you, thank you, Dave. >> So, you know, like I said, we've had some fun with this meme. But it really is, we heard on the previous panel, everybody's using Snowflake as an example. Somebody how builds on top of hyper scale infrastructure. You're not building your own data centers. And, so, are you building a super data cloud? >> We don't call it exactly that way. We don't like the super word, it's a bit dismissive. >> That's our term. >> About our friends, cloud provider friends. But we call it a data cloud. And the vision, really, for the data cloud is, indeed, it's a cloud which overlays the hyper scaler cloud. But there is a big difference, right? There are several ways to do this super cloud, as you name them. The way we picked is to create one single system, and that's very important, right? There are several ways, right. You can instantiate your solution in every region of the cloud and, you know, potentially that region could be AWS, that region could be GCP. So, you are, indeed, a multi-cloud solution. But Snowflake, we did it differently. We are really creating cloud regions, which are superimposed on top of the cloud provider region, infrastructure region. So, we are building our regions. But where it's very different is that each region of Snowflake is not one instantiation of our service. Our service is global, by nature. We can move data from one region to the other. When you land in Snowflake, you land into one region. But you can grow from there and you can, you know, exist in multiple cloud at the same time. And that's very important, right? It's not different instantiation of a system, it's one single instantiation which covers many cloud regions and many cloud provider. >> So, we used Snowflake as an example. And we're trying to understand what the salient aspects are of your data cloud, what we call super cloud. In fact, you've used the word instantiate. Kit Colbert, just earlier today, laid out, he said, there's sort of three levels. You can run it on one cloud and communicate with the other cloud, you can instantiate on the clouds, or you can have the same service running 24/7 across clouds, that's the hardest example. >> Yeah. >> The most mature. You just described, essentially, doing that. How do you enable that? What are the technical enablers? >> Yeah, so, as I said, first we start by building, you know, Snowflake regions, we have today 30 regions that span the world, so it's a world wide system, with many regions. But all these regions are connected together. They are meshed together with our technology, we name it Snow Grid, and that makes it hard because, you know, Azure region can talk to a WS region, or GCP regions, and as a user for our cloud, you don't see, really, these regional differences, that regions are in different potentially cloud. When you use Snowflake, you can exist, your presence as an organization can be in several regions, several clouds, if you want, geographic, both geographic and cloud provider. >> So, I can share data irrespective of the cloud. And I'm in the Snowflake data cloud, is that correct? I can do that today? >> Exactly, and that's very critical, right? What we wanted is to remove data silos. And when you insociate a system in one single region, and that system is locked in that region, you cannot communicate with other parts of the world, you are locking data in one region. Right, and we didn't want to do that. We wanted data to be distributed the way customer wants it to be distributed across the world. And potentially sharing data at world scales. >> Does that mean if I'm in one region and I want to run a query, if I'm in AWS in one region, and I want to run a query on data that happens to be in an Azure cloud, I can actually execute that? >> So, yes and no. The way we do it is very expensive to do that. Because, generally, if you want to join data which are in different region and different cloud, it's going to be very expensive because you need to move data every time you join it. So, the way we do it is that you replicate the subset of data that you want to access from one region from other region. So, you can create this data mesh, but data is replicated to make it very cheap and very performing too. >> And is the Snow Grid, does that have the metadata intelligence to actually? >> Yes, yes. >> Can you describe that a little? >> Yeah, Snow Grid is both a way to exchange metadata. So, each region of Snowflake knows about all the other regions of Snowflake. Every time we create a new region, the metadata is distributed over our data cloud, not only region knows all the region, but knows every organization that exists in our cloud, where this organization is, where data can be replicated by this organization. And then, of course, it's also used as a way to exchange data, right? So, you can exchange data by scale of data size. And I was just receiving an email from one of our customers who moved more than four petabytes of data, cross region, cross cloud providers in, you know, few days. And it's a lot of data, so it takes some time to move. But they were able to do that online, completely online, and switch over to the other region, which is very important also. >> So, one of the hardest parts about super cloud that I'm still trying to struggling through is the security model. Because you've got the cloud as your sort of first line of defense. And now we've got multiple clouds, with multiple first lines of defense, I've got a shared responsibility model across those clouds, I've got different tools in each of those clouds. Do you take care of that? Where do you pick up from the cloud providers? Do you abstract that security layer? Do you bring in partners? It's a very complicated. >> No, this is a great question. Security has always been the most important aspect of Snowflake sense day one, right? This is the question that every customer of ours has. You know, how can you guarantee the security of my data? And, so, we secure data really tightly in region. We have several layers of security. It starts by creating every data at rest. And that's very important. A lot of customers are not doing that, right? You hear of these attacks, for example, on cloud, where someone left their buckets. And then, you know, you can access the data because it's a non-encrypted. So, we are encrypting everything at rest. We are encrypting everything in transit. So, a region is very secure. Now, you know, from one region, you never access data from another region in Snowflake. That's why, also, we replicate data. Now the replication of that data across region, or the metadata, for that matter, is really our least secure, so Snow Grid ensures that everything is encrypted, everything is, we have multiple encryption keys, and it's stored in hardware secure modules, so, we bit Snow Grid such that it's secure and it allows very secure movement of data. >> Okay, so, I know we kind of, getting into the technology here a lot today, but because super cloud is the future, we actually have to have an architectural foundation on which to build. So, you mentioned a bucket, like an S3 bucket. Okay, that's storage, but you also, for instance, taking advantage of new semi-conductor technology. Like Graviton, as an example, that drives efficiency. You guys talk about how you pass that on to your customers. Even if it means less revenue for you, so, awesome, we love that, you'll make it up in volume. And, so. >> Exactly. >> How do you deal with the lowest common denominator problem? I was talking to somebody the other day and this individual brought up what I thought was a really good point. What if we, let's say, AWS, have the best, silicon. And we can run the fastest and the least expensive, and the lowest power. But another cloud provider hasn't caught up yet. How do you deal with that delta? Do you just take the best of and try to respect that? >> No, it's a great question. I mean, of course, our software is extracting all the cloud providers infrastructure so that when you run in one region, let's say AWS, or Azure, it doesn't make any difference, as far as the applications are concerned. And this abstraction, of course, is a lot of work. I mean, really, a lot of work. Because it needs to be secure, it needs to be performance, and every cloud, and it has to expose APIs which are uniform. And, you know, cloud providers, even though they have potentially the same concept, let's say block storage, APIs are completely different. The way these systems are secure, it's completely different. There errors that you can get. And the retry mechanism is very different from one cloud to the other. The performance is also different. We discovered that when we starting to port our software. And we had to completely rethink how to leverage block storage in that cloud versus that cloud, because just off performance too. And, so, we had, for example, to stripe data. So, all this work is work that you don't need as an application because our vision, really, is that application, which are running in our data cloud, can be abstracted for this difference. And we provide all the services, all the workload that this application need. Whether it's transactional access to data, analytical access to data, managing logs, managing metrics, all of this is abstracted too, so that they are not tied to one particular service of one cloud. And distributing this application across many region, many cloud, is very seamless. >> So, Snowflake has built, your team has built a true abstraction layer across those clouds that's available today? It's actually shipping? >> Yes, and we are still developing it. You know, transactional, Unistore, as we call it, was announced last summit. So, they are still, you know, work in progress. >> You're not done yet. >> But that's the vision, right? And that's important, because we talk about the infrastructure, right. You mention a lot about storage and compute. But it's not only that, right. When you think about application, they need to use the transactional database. They need to use an analytical system. They need to use machine learning. So, you need to provide, also, all these services which are consistent across all the cloud providers. >> So, let's talk developers. Because, you know, you think Snowpark, you guys announced a big application development push at the Snowflake summit recently. And we have said that a criterion of super cloud is a super paz layer, people wince when I say that, but okay, we're just going to go with it. But the point is, it's a purpose built application development layer, specific to your particular agenda, that supports your vision. >> Yes. >> Have you essentially built a purpose built paz layer? Or do you just take them off the shelf, standard paz, and cobble it together? >> No, we build it a custom build. Because, as you said, what exist in one cloud might not exist in another cloud provider, right. So, we have to build in this, all these components that a multi-application need. And that goes to machine learning, as I said, transactional analytical system, and the entire thing. So that it can run in isolation physically. >> And the objective is the developer experience will be identical across those clouds? >> Yes, the developers doesn't need to worry about cloud provider. And, actually, our system will have, we didn't talk about it, but a marketplace that we have, which allows, actually, to deliver. >> We're getting there. >> Yeah, okay. (both laughing) I won't divert. >> No, no, let's go there, because the other aspect of super cloud that we've talked about is the ecosystem. You have to enable an ecosystem to add incremental value, it's not the power of many versus the capabilities of one. So, talk about the challenges of doing that. Not just the business challenges but, again, I'm interested in the technical and architectural challenges. >> Yeah, yeah, so, it's really about, I mean, the way we enable our ecosystem and our partners to create value on top of our data cloud, is via the marketplace. Where you can put shared data on the marketplace. Provide listing on this marketplace, which are data sets. But it goes way beyond data. It's all the way to application. So, you can think of it as the iPhone. A little bit more, all right. Your iPhone is great. Not so much because the hardware is great, or because of the iOS, but because of all the applications that you have. And all these applications are not necessarily developed by Apple, basically. So, we are, it's the same model with our marketplace. We foresee an environment where providers and partners are going to build these applications. We call it native application. And we are going to help them distribute these applications across cloud, everywhere in the world, potentially. And they don't need to worry about that. They don't need to worry about how these applications are going to be instantiated. We are going to help them to monetize these applications. So, that unlocks, you know, really, all the partner ecosystem that you have seen, you know, with something like the iPhone, right? It has created so many new companies that have developed these applications. >> Your detractors have criticized you for being a walled garden. I've actually used that term. I used terms like defacto standard, which are maybe less sensitive to you, but, nonetheless, we've seen defacto standards actually deliver value. I've talked to Frank Slootman about this, and he said, Dave, we deliver value, that's what we're all about. At the same time, he even said to me, and I want your thoughts on this, is, look, we have to embrace open source where it makes sense. You guys announced Apache Iceberg. So, what are your thoughts on that? Is that to enable a developer ecosystem? Why did you do Iceberg? >> Yeah, Iceberg is very important. So, just to give some context, Iceberg is an open table format. >> Right. >> Which was first developed by Netflix. And Netflix put it open source in the Apache community. So, we embraced that open source standard because it's widely used by many companies. And, also, many companies have really invested a lot of effort in building big data, Hadoop Solutions, or DataX Solution, and they want to use Snowflake. And they couldn't really use Snowflake, because all their data were in open format. So, we are embracing Iceberg to help these companies move through the cloud. But why we have been reluctant with direct access to data, direct access to data is a little bit of a problem for us. And the reason is when you direct access to data, now you have direct access to storage. Now you have to understand, for example, the specificity of one cloud versus the other. So, as soon as you start to have direct access to data, you lose your cloud data sync layer. You don't access data with API. When you have direct access to data, it's very hard to sync your data. Because you need to grant access, direct access to tools which are not protected. And you see a lot of hacking of data because of that. So, direct access to data is not serving well our customers, and that's why we have been reluctant to do that. Because it is not cloud diagnostic. You have to code that, you need a lot of intelligence, why APIs access, so we want open APIs. That's, I guess, the way we embrace openness, is by open API versus you access, directly, data. >> iPhone. >> Yeah, yeah, iPhone, APIs, you know. We define a set of APIs because APIs, you know, the implementation of the APIs can change, can improve. You can improve compression of data, for example. If you open direct access to data now, you cannot evolve. >> My point is, you made a promise, from governed, security, data sharing ecosystem. It works the same way, so that's the path that you've chosen. Benoit Dogeville, thank you so much for coming on theCube and participating in Supercloud 22, really appreciate that. >> Thank you, Dave. It was a great pleasure. >> All right, keep it right there, we'll be right back with our next segment, right after this short break. (electronic music)
SUMMARY :
You know, in one of the So, you know, like I said, We don't like the super and you can, you know, or you can have the same How do you enable that? we start by building, you know, And I'm in the Snowflake And when you insociate a So, the way we do it is that you replicate So, you can exchange data So, one of the hardest And then, you know, So, you mentioned a and the least expensive, so that when you run in one So, they are still, you know, So, you need to provide, Because, you know, you think Snowpark, And that goes to machine a marketplace that we have, I won't divert. So, talk about the of all the applications that you have. At the same time, he even said to me, So, just to give some context, You have to code that, you because APIs, you know, so that's the path that you've chosen. It was a great pleasure. with our next segment, right
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Frank Slootman | PERSON | 0.99+ |
Benoit | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Netflix | ORGANIZATION | 0.99+ |
John Furrier | PERSON | 0.99+ |
Kit Colbert | PERSON | 0.99+ |
iPhone | COMMERCIAL_ITEM | 0.99+ |
Palo Alto, California | LOCATION | 0.99+ |
Benoit Dogeville | PERSON | 0.99+ |
one region | QUANTITY | 0.99+ |
iOS | TITLE | 0.99+ |
30 regions | QUANTITY | 0.99+ |
more than four petabytes | QUANTITY | 0.99+ |
Snowflake | EVENT | 0.99+ |
first line | QUANTITY | 0.99+ |
Snowpark | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.98+ |
today | DATE | 0.98+ |
Apache | ORGANIZATION | 0.98+ |
both | QUANTITY | 0.98+ |
each | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
Unistore | ORGANIZATION | 0.97+ |
Supercloud 22 | EVENT | 0.97+ |
first lines | QUANTITY | 0.97+ |
DataX Solution | ORGANIZATION | 0.97+ |
each region | QUANTITY | 0.96+ |
Snow Grid | TITLE | 0.96+ |
one cloud | QUANTITY | 0.96+ |
one single region | QUANTITY | 0.96+ |
first | QUANTITY | 0.96+ |
one region | QUANTITY | 0.96+ |
Hadoop Solutions | ORGANIZATION | 0.95+ |
WS | ORGANIZATION | 0.94+ |
Supercloud 22 | ORGANIZATION | 0.93+ |
three levels | QUANTITY | 0.93+ |
Snowflake | TITLE | 0.93+ |
one single system | QUANTITY | 0.92+ |
Iceberg | TITLE | 0.89+ |
one single instantiation | QUANTITY | 0.89+ |
theCube | ORGANIZATION | 0.86+ |
Azure | ORGANIZATION | 0.85+ |
Years ago | DATE | 0.83+ |
earlier today | DATE | 0.82+ |
one instantiation | QUANTITY | 0.82+ |
Super Data Cloud | ORGANIZATION | 0.81+ |
S3 | COMMERCIAL_ITEM | 0.8+ |
one cloud | QUANTITY | 0.76+ |
delta | ORGANIZATION | 0.76+ |
Azure | TITLE | 0.75+ |
one of our customers | QUANTITY | 0.72+ |
day one | QUANTITY | 0.72+ |
Supercloud22 | EVENT | 0.66+ |
Breaking Analysis: Snowflake Summit 2022...All About Apps & Monetization
>> From theCUBE studios in Palo Alto in Boston, bringing you data driven insights from theCUBE and ETR. This is "Breaking Analysis" with Dave Vellante. >> Snowflake Summit 2022 underscored that the ecosystem excitement which was once forming around Hadoop is being reborn, escalated and coalescing around Snowflake's data cloud. What was once seen as a simpler cloud data warehouse and good marketing with the data cloud is evolving rapidly with new workloads of vertical industry focus, data applications, monetization, and more. The question is, will the promise of data be fulfilled this time around, or is it same wine, new bottle? Hello, and welcome to this week's Wikibon CUBE Insights powered by ETR. In this "Breaking Analysis," we'll talk about the event, the announcements that Snowflake made that are of greatest interest, the major themes of the show, what was hype and what was real, the competition, and some concerns that remain in many parts of the ecosystem and pockets of customers. First let's look at the overall event. It was held at Caesars Forum. Not my favorite venue, but I'll tell you it was packed. Fire Marshall Full, as we sometimes say. Nearly 10,000 people attended the event. Here's Snowflake's CMO Denise Persson on theCUBE describing how this event has evolved. >> Yeah, two, three years ago, we were about 1800 people at a Hilton in San Francisco. We had about 40 partners attending. This week we're close to 10,000 attendees here. Almost 10,000 people online as well, and over over 200 partners here on the show floor. >> Now, those numbers from 2019 remind me of the early days of Hadoop World, which was put on by Cloudera but then Cloudera handed off the event to O'Reilly as this article that we've inserted, if you bring back that slide would say. The headline it almost got it right. Hadoop World was a failure, but it didn't have to be. Snowflake has filled the void created by O'Reilly when it first killed Hadoop World, and killed the name and then killed Strata. Now, ironically, the momentum and excitement from Hadoop's early days, it probably could have stayed with Cloudera but the beginning of the end was when they gave the conference over to O'Reilly. We can't imagine Frank Slootman handing the keys to the kingdom to a third party. Serious business was done at this event. I'm talking substantive deals. Salespeople from a host sponsor and the ecosystems that support these events, they love physical. They really don't like virtual because physical belly to belly means relationship building, pipeline, and deals. And that was blatantly obvious at this show. And in fairness, all theCUBE events that we've done year but this one was more vibrant because of its attendance and the action in the ecosystem. Ecosystem is a hallmark of a cloud company, and that's what Snowflake is. We asked Frank Slootman on theCUBE, was this ecosystem evolution by design or did Snowflake just kind of stumble into it? Here's what he said. >> Well, when you are a data clouding, you have data, people want to do things with that data. They don't want just run data operations, populate dashboards, run reports. Pretty soon they want to build applications and after they build applications, they want build businesses on it. So it goes on and on and on. So it drives your development to enable more and more functionality on that data cloud. Didn't start out that way, you know, we were very, very much focused on data operations. Then it becomes application development and then it becomes, hey, we're developing whole businesses on this platform. So similar to what happened to Facebook in many ways. >> So it sounds like it was maybe a little bit of both. The Facebook analogy is interesting because Facebook is a walled garden, as is Snowflake, but when you come into that garden, you have assurances that things are going to work in a very specific way because a set of standards and protocols is being enforced by a steward, i.e. Snowflake. This means things run better inside of Snowflake than if you try to do all the integration yourself. Now, maybe over time, an open source version of that will come out but if you wait for that, you're going to be left behind. That said, Snowflake has made moves to make its platform more accommodating to open source tooling in many of its announcements this week. Now, I'm not going to do a deep dive on the announcements. Matt Sulkins from Monte Carlo wrote a decent summary of the keynotes and a number of analysts like Sanjeev Mohan, Tony Bear and others are posting some deeper analysis on these innovations, and so we'll point to those. I'll say a few things though. Unistore extends the type of data that can live in the Snowflake data cloud. It's enabled by a new feature called hybrid tables, a new table type in Snowflake. One of the big knocks against Snowflake was it couldn't handle and transaction data. Several database companies are creating this notion of a hybrid where both analytic and transactional workloads can live in the same data store. Oracle's doing this for example, with MySQL HeatWave and there are many others. We saw Mongo earlier this month add an analytics capability to its transaction system. Mongo also added sequel, which was kind of interesting. Here's what Constellation Research analyst Doug Henschen said about Snowflake's moves into transaction data. Play the clip. >> Well with Unistore, they're reaching out and trying to bring transactional data in. Hey, don't limit this to analytical information and there's other ways to do that like CDC and streaming but they're very closely tying that again to that marketplace, with the idea of bring your data over here and you can monetize it. Don't just leave it in that transactional database. So another reach to a broader play across a big community that they're building. >> And you're also seeing Snowflake expand its workload types in its unique way and through Snowpark and its stream lit acquisition, enabling Python so that native apps can be built in the data cloud and benefit from all that structure and the features that Snowflake is built in. Hence that Facebook analogy, or maybe the App Store, the Apple App Store as I propose as well. Python support also widens the aperture for machine intelligence workloads. We asked Snowflake senior VP of product, Christian Kleinerman which announcements he thought were the most impactful. And despite the who's your favorite child nature of the question, he did answer. Here's what he said. >> I think the native applications is the one that looks like, eh, I don't know about it on the surface but he has the biggest potential to change everything. That's create an entire ecosystem of solutions for within a company or across companies that I don't know that we know what's possible. >> Snowflake also announced support for Apache Iceberg, which is a new open table format standard that's emerging. So you're seeing Snowflake respond to these concerns about its lack of openness, and they're building optionality into their cloud. They also showed some cost op optimization tools both from Snowflake itself and from the ecosystem, notably Capital One which launched a software business on top of Snowflake focused on optimizing cost and eventually the rollout data management capabilities, and all kinds of features that Snowflake announced that the show around governance, cross cloud, what we call super cloud, a new security workload, and they reemphasize their ability to read non-native on-prem data into Snowflake through partnerships with Dell and Pure and a lot more. Let's hear from some of the analysts that came on theCUBE this week at Snowflake Summit to see what they said about the announcements and their takeaways from the event. This is Dave Menninger, Sanjeev Mohan, and Tony Bear, roll the clip. >> Our research shows that the majority of organizations, the majority of people do not have access to analytics. And so a couple of the things they've announced I think address those or help to address those issues very directly. So Snowpark and support for Python and other languages is a way for organizations to embed analytics into different business processes. And so I think that'll be really beneficial to try and get analytics into more people's hands. And I also think that the native applications as part of the marketplace is another way to get applications into people's hands rather than just analytical tools. Because most people in the organization are not analysts. They're doing some line of business function. They're HR managers, they're marketing people, they're sales people, they're finance people, right? They're not sitting there mucking around in the data, they're doing a job and they need analytics in that job. >> Primarily, I think it is to contract this whole notion that once you move data into Snowflake, it's a proprietary format. So I think that's how it started but it's usually beneficial to the customers, to the users because now if you have large amount of data in paket files you can leave it on S3, but then you using the Apache Iceberg table format in Snowflake, you get all the benefits of Snowflake's optimizer. So for example, you get the micro partitioning, you get the metadata. And in a single query, you can join, you can do select from a Snowflake table union and select from an iceberg table and you can do store procedure, user defined function. So I think what they've done is extremely interesting. Iceberg by itself still does not have multi-table transactional capabilities. So if I'm running a workload, I might be touching 10 different tables. So if I use Apache Iceberg in a raw format, they don't have it, but Snowflake does. So the way I see it is Snowflake is adding more and more capabilities right into the database. So for example, they've gone ahead and added security and privacy. So you can now create policies and do even cell level masking, dynamic masking, but most organizations have more than Snowflake. So what we are starting to see all around here is that there's a whole series of data catalog companies, a bunch of companies that are doing dynamic data masking, security and governance, data observability which is not a space Snowflake has gone into. So there's a whole ecosystem of companies that is mushrooming. Although, you know, so they're using the native capabilities of Snowflake but they are at a level higher. So if you have a data lake and a cloud data warehouse and you have other like relational databases, you can run these cross platform capabilities in that layer. So that way, you know, Snowflake's done a great job of enabling that ecosystem. >> I think it's like the last mile, essentially. In other words, it's like, okay, you have folks that are basically that are very comfortable with Tableau but you do have developers who don't want to have to shell out to a separate tool. And so this is where Snowflake is essentially working to address that constituency. To Sanjeev's point, and I think part of it, this kind of plays into it is what makes this different from the Hadoop era is the fact that all these capabilities, you know, a lot of vendors are taking it very seriously to put this native. Now, obviously Snowflake acquired Streamlit. So we can expect that the Streamlit capabilities are going to be native. >> I want to share a little bit about the higher level thinking at Snowflake, here's a chart from Frank Slootman's keynote. It's his version of the modern data stack, if you will. Now, Snowflake of course, was built on the public cloud. If there were no AWS, there would be no Snowflake. Now, they're all about bringing data and live data and expanding the types of data, including structured, we just heard about that, unstructured, geospatial, and the list is going to continue on and on. Eventually I think it's going to bleed into the edge if we can figure out what to do with that edge data. Executing on new workloads is a big deal. They started with data sharing and they recently added security and they've essentially created a PaaS layer. We call it a SuperPaaS layer, if you will, to attract application developers. Snowflake has a developer-focused event coming up in November and they've extended the marketplace with 1300 native apps listings. And at the top, that's the holy grail, monetization. We always talk about building data products and we saw a lot of that at this event, very, very impressive and unique. Now here's the thing. There's a lot of talk in the press, in the Wall Street and the broader community about consumption-based pricing and concerns over Snowflake's visibility and its forecast and how analytics may be discretionary. But if you're a company building apps in Snowflake and monetizing like Capital One intends to do, and you're now selling in the marketplace, that is not discretionary, unless of course your costs are greater than your revenue for that service, in which case is going to fail anyway. But the point is we're entering a new error where data apps and data products are beginning to be built and Snowflake is attempting to make the data cloud the defacto place as to where you're going to build them. In our view they're well ahead in that journey. Okay, let's talk about some of the bigger themes that we heard at the event. Bringing apps to the data instead of moving the data to the apps, this was a constant refrain and one that certainly makes sense from a physics point of view. But having a single source of data that is discoverable, sharable and governed with increasingly robust ecosystem options, it doesn't have to be moved. Sometimes it may have to be moved if you're going across regions, but that's unique and a differentiator for Snowflake in our view. I mean, I'm yet to see a data ecosystem that is as rich and growing as fast as the Snowflake ecosystem. Monetization, we talked about that, industry clouds, financial services, healthcare, retail, and media, all front and center at the event. My understanding is that Frank Slootman was a major force behind this shift, this development and go to market focus on verticals. It's really an attempt, and he talked about this in his keynote to align with the customer mission ultimately align with their objectives which not surprisingly, are increasingly monetizing with data as a differentiating ingredient. We heard a ton about data mesh, there were numerous presentations about the topic. And I'll say this, if you map the seven pillars Snowflake talks about, Benoit Dageville talked about this in his keynote, but if you map those into Zhamak Dehghani's data mesh framework and the four principles, they align better than most of the data mesh washing that I've seen. The seven pillars, all data, all workloads, global architecture, self-managed, programmable, marketplace and governance. Those are the seven pillars that he talked about in his keynote. All data, well, maybe with hybrid tables that becomes more of a reality. Global architecture means the data is globally distributed. It's not necessarily physically in one place. Self-managed is key. Self-service infrastructure is one of Zhamak's four principles. And then inherent governance. Zhamak talks about computational, what I'll call automated governance, built in. And with all the talk about monetization, that aligns with the second principle which is data as product. So while it's not a pure hit and to its credit, by the way, Snowflake doesn't use data mesh in its messaging anymore. But by the way, its customers do, several customers talked about it. Geico, JPMC, and a number of other customers and partners are using the term and using it pretty closely to the concepts put forth by Zhamak Dehghani. But back to the point, they essentially, Snowflake that is, is building a proprietary system that substantially addresses some, if not many of the goals of data mesh. Okay, back to the list, supercloud, that's our term. We saw lots of examples of clouds on top of clouds that are architected to spin multiple clouds, not just run on individual clouds as separate services. And this includes Snowflake's data cloud itself but a number of ecosystem partners that are headed in a very similar direction. Snowflake still talks about data sharing but now it uses the term collaboration in its high level messaging, which is I think smart. Data sharing is kind of a geeky term. And also this is an attempt by Snowflake to differentiate from everyone else that's saying, hey, we do data sharing too. And finally Snowflake doesn't say data marketplace anymore. It's now marketplace, accounting for its application market. Okay, let's take a quick look at the competitive landscape via this ETR X-Y graph. Vertical access remembers net score or spending momentum and the x-axis is penetration, pervasiveness in the data center. That's what ETR calls overlap. Snowflake continues to lead on the vertical axis. They guide it conservatively last quarter, remember, so I wouldn't be surprised if that lofty height, even though it's well down from its earlier levels but I wouldn't be surprised if it ticks down again a bit in the July survey, which will be in the field shortly. Databricks is a key competitor obviously at a strong spending momentum, as you can see. We didn't draw it here but we usually draw that 40% line or red line at 40%, anything above that is considered elevated. So you can see Databricks is quite elevated. But it doesn't have the market presence of Snowflake. It didn't get to IPO during the bubble and it doesn't have nearly as deep and capable go-to market machinery. Now, they're getting better and they're getting some attention in the market, nonetheless. But as a private company, you just naturally, more people are aware of Snowflake. Some analysts, Tony Bear in particular, believe Mongo and Snowflake are on a bit of a collision course long term. I actually can see his point. You know, I mean, they're both platforms, they're both about data. It's long ways off, but you can see them sort of in a similar path. They talk about kind of similar aspirations and visions even though they're quite in different markets today but they're definitely participating in similar tam. The cloud players are probably the biggest or definitely the biggest partners and probably the biggest competitors to Snowflake. And then there's always Oracle. Doesn't have the spending velocity of the others but it's got strong market presence. It owns a cloud and it knows a thing about data and it definitely is a go-to market machine. Okay, we're going to end on some of the things that we heard in the ecosystem. 'Cause look, we've heard before how particular technology, enterprise data warehouse, data hubs, MDM, data lakes, Hadoop, et cetera. We're going to solve all of our data problems and of course they didn't. And in fact, sometimes they create more problems that allow vendors to push more incremental technology to solve the problems that they created. Like tools and platforms to clean up the no schema on right nature of data lakes or data swamps. But here are some of the things that I heard firsthand from some customers and partners. First thing is, they said to me that they're having a hard time keeping up sometimes with the pace of Snowflake. It reminds me of AWS in 2014, 2015 timeframe. You remember that fire hose of announcements which causes increased complexity for customers and partners. I talked to several customers that said, well, yeah this is all well and good but I still need skilled people to understand all these tools that I'm integrated in the ecosystem, the catalogs, the machine learning observability. A number of customers said, I just can't use one governance tool, I need multiple governance tools and a lot of other technologies as well, and they're concerned that that's going to drive up their cost and their complexity. I heard other concerns from the ecosystem that it used to be sort of clear as to where they could add value you know, when Snowflake was just a better data warehouse. But to point number one, they're either concerned that they'll be left behind or they're concerned that they'll be subsumed. Look, I mean, just like we tell AWS customers and partners, you got to move fast, you got to keep innovating. If you don't, you're going to be left. Either if your customer you're going to be left behind your competitor, or if you're a partner, somebody else is going to get there or AWS is going to solve the problem for you. Okay, and there were a number of skeptical practitioners, really thoughtful and experienced data pros that suggested that they've seen this movie before. That's hence the same wine, new bottle. Well, this time around I certainly hope not given all the energy and investment that is going into this ecosystem. And the fact is Snowflake is unquestionably making it easier to put data to work. They built on AWS so you didn't have to worry about provisioning, compute and storage and networking and scaling. Snowflake is optimizing its platform to take advantage of things like Graviton so you don't have to, and they're doing some of their own optimization tools. The ecosystem is building optimization tools so that's all good. And firm belief is the less expensive it is, the more data will get brought into the data cloud. And they're building a data platform on which their ecosystem can build and run data applications, aka data products without having to worry about all the hard work that needs to get done to make data discoverable, shareable, and governed. And unlike the last 10 years, you don't have to be a keeper and integrate all the animals in the Hadoop zoo. Okay, that's it for today, thanks for watching. Thanks to my colleague, Stephanie Chan who helps research "Breaking Analysis" topics. Sometimes Alex Myerson is on production and manages the podcasts. Kristin Martin and Cheryl Knight help get the word out on social and in our newsletters, and Rob Hof is our editor in chief over at Silicon, and Hailey does some wonderful editing, thanks to all. Remember, all these episodes are available as podcasts wherever you listen. All you got to do is search Breaking Analysis Podcasts. I publish each week on wikibon.com and siliconangle.com and you can email me at David.Vellante@siliconangle.com or DM me @DVellante. If you got something interesting, I'll respond. If you don't, I'm sorry I won't. Or comment on my LinkedIn post. Please check out etr.ai for the best survey data in the enterprise tech business. This is Dave Vellante for theCUBE Insights powered by ETR. Thanks for watching, and we'll see you next time. (upbeat music)
SUMMARY :
bringing you data driven that the ecosystem excitement here on the show floor. and the action in the ecosystem. Didn't start out that way, you know, One of the big knocks against Snowflake the idea of bring your data of the question, he did answer. is the one that looks like, and from the ecosystem, And so a couple of the So that way, you know, from the Hadoop era is the fact the defacto place as to where
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Frank Slootman | PERSON | 0.99+ |
Frank Slootman | PERSON | 0.99+ |
Doug Henschen | PERSON | 0.99+ |
Stephanie Chan | PERSON | 0.99+ |
Christian Kleinerman | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Rob Hof | PERSON | 0.99+ |
Benoit Dageville | PERSON | 0.99+ |
2014 | DATE | 0.99+ |
Matt Sulkins | PERSON | 0.99+ |
JPMC | ORGANIZATION | 0.99+ |
2019 | DATE | 0.99+ |
Cheryl Knight | PERSON | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Denise Persson | PERSON | 0.99+ |
Alex Myerson | PERSON | 0.99+ |
Tony Bear | PERSON | 0.99+ |
Dave Menninger | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
July | DATE | 0.99+ |
Geico | ORGANIZATION | 0.99+ |
November | DATE | 0.99+ |
Snowflake | TITLE | 0.99+ |
40% | QUANTITY | 0.99+ |
Oracle | ORGANIZATION | 0.99+ |
App Store | TITLE | 0.99+ |
Capital One | ORGANIZATION | 0.99+ |
second principle | QUANTITY | 0.99+ |
Sanjeev Mohan | PERSON | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
1300 native apps | QUANTITY | 0.99+ |
Tony Bear | PERSON | 0.99+ |
David.Vellante@siliconangle.com | OTHER | 0.99+ |
Kristin Martin | PERSON | 0.99+ |
Mongo | ORGANIZATION | 0.99+ |
Databricks | ORGANIZATION | 0.99+ |
Snowflake Summit 2022 | EVENT | 0.99+ |
First | QUANTITY | 0.99+ |
two | DATE | 0.99+ |
Python | TITLE | 0.99+ |
10 different tables | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
ETR | ORGANIZATION | 0.99+ |
both | QUANTITY | 0.99+ |
Snowflake | EVENT | 0.98+ |
one place | QUANTITY | 0.98+ |
each week | QUANTITY | 0.98+ |
O'Reilly | ORGANIZATION | 0.98+ |
This week | DATE | 0.98+ |
Hadoop World | EVENT | 0.98+ |
this week | DATE | 0.98+ |
Pure | ORGANIZATION | 0.98+ |
about 40 partners | QUANTITY | 0.98+ |
theCUBE | ORGANIZATION | 0.98+ |
last quarter | DATE | 0.98+ |
One | QUANTITY | 0.98+ |
S3 | TITLE | 0.97+ |
Hadoop | LOCATION | 0.97+ |
single | QUANTITY | 0.97+ |
Caesars Forum | LOCATION | 0.97+ |
Iceberg | TITLE | 0.97+ |
single source | QUANTITY | 0.97+ |
Silicon | ORGANIZATION | 0.97+ |
Nearly 10,000 people | QUANTITY | 0.97+ |
Apache Iceberg | ORGANIZATION | 0.97+ |
Benoit Dageville, Snowflake | Snowflake Summit 2022
(upbeat music) >> Welcome back everyone, theCUBE's three days of wall to wall coverage of Snowflake Summit '22 is coming to an end, but Dave Vellante and I, Lisa Martin are so pleased to have our final guest as none other than the co-founder and president of products at Snowflake, Benoit Dageville. Benoit, thank you so much for joining us on the program. Welcome. >> Thank you. Thank you, thank you. >> So this is day four, 'cause you guys started on Monday. This is Thursday. The amount of people that are still here speaks volumes. We've had close to 10,000 people here. >> Yeah. >> Could you ever have imagined back in the day, 10 years ago that it would come to something like this in such a short period of time? >> Absolutely not. And I always say if I had imagined that I might not have started Snowflake, right. This is somehow scary. I mean and yeah, it's huge. And you can feel the excitement of everyone. It is like mind boggling and the fact that so many people are still there after four days is great. >> Your keynote on Tuesday was fantastic. Your energy was off the charts. It was standing room only. There were overflow rooms. Like we just mentioned, a lot of people are still here. Talk about the evolution of Snowflake, this week's announcements and what it means for the future of the data cloud. >> Yeah, so evolution, I mean, I will start with the evolution. It's true that that's what we have announced. This week is not where we started necessarily. So we started really very quickly with big data combined with data warehouse as one thing. We saw that the world was moving into fragmented siloing data and we thought with Thierry, we are going to combine big data and data warehouse in one system for the cloud with this elasticity and this service simplicity. So simplicity, amazing elasticity, which is this multi workload architecture that I was explaining during the keynotes and really extreme simplicity with the service. Then we realized that there is one other attribute in the cloud, which is unique, which doesn't exist on-premise, which is collaboration. How you can connect different tenets of the platform together. And Google showed that with Google Docs. I always say to me, it was amazing that you could share document and have direct access to document that you didn't produce and you can collaborate on this document. So we wanted to do the same thing for data and this is where we created the data cloud and the marketplace where you can have all these data sets available and really the next evolution I would say is really about applications that are (indistinct) by that data, but are way simpler to use for all the tenets of the data cloud. And this is the way you can share expertise also, including, ML model, everyone talks about ML and the democratization of ML. How are you going to democratize ML? It's not by making necessary training super easy. Such that everyone can train their ML for themselves. It's by having very specialized application where data and ML is at the core, which are shared, through the marketplace and we shall leverage by many tenets of this marketplace that have no necessary knowledge about building this ML models. So that's where, yeah. >> When you and Thierry started the company, I go back to the improbable rise of Kubernetes and there were other more sophisticated container management systems back then, but they chose to focus on simplicity. And you've told me before, that was our main tenet. We are not going to worry about all the complex database stuff. You knew how to do that, but you chose not to. So my question is, did you envision solving those complex problems over time yourselves or through an ecosystem? Was this by design or did you... As you started to get into it, say let's not even try to go there let's partner to go there. >> Yeah, I mean, it's both. It's a combination of both. Snowflake, the simplicity of the platform is really important because if our partners are struggling to put their solution and build solution on top of Snowflake they will not build it. So it's very important that number one, our platform is really easy to use from day one. And that really has to be built inside the platform. You cannot build simplicity on top. You cannot have a complex solution and all of a sudden realize that, oh, this is complex. I need to build another layer on top of it to make it simpler, that will not work. So it had to be built from day one, but you're right. What is going to be Snowflake? I always say in 10 years from now, we just turn 10 years old or we are going to turn 10 years old in few months. Actually a few months, yes. >> Right. >> So for the next 10 years I really believe that most of Snowflake will not be built by Snowflake. And that's the power of the partners and these applications. When you are going to say I'm using Snowflake, actually, probably you are not going to use directly code developed by Snowflake. That code will leverage our platform, but you will use a solution that has been built on top of Snowflake. And this is the way we are going to decouple, the effort of Snowflake and multiply it. >> It's an interesting balance, isn't it? When I think of what you did with Apache Iceberg, if I use Iceberg and I'm not going to get as much functionality, but I may want that openness, but I'm going to get more functionality inside of the data cloud. And I don't know, but if you know the answer to what's going to happen. >> No, that's a super good question. So to explain what we did with Apache Iceberg, and the fact that now it's a native format for us. So everything that you can do with our internal formats, you can do it with Apache Iceberg, including security, defining masking, data masking all the governors that we have, fine grain security aspects, the replications you can define you can use (indistinct) on top of... >> But there's a but, right? But if I do that with native Snowflake tools, I'm going to get an even greater advantage, am I not? >> Yes. So that's what I'm saying. So that's why we embraced Iceberg, because I think we can bring all the benefit of Snowflake to people who have decided to use Iceberg, I mean open formats. Iceberg is a table format. So and why it was important because people had massive investments in open source in Hadoop. And we had a lot of companies saying, we love Snowflake. We want to be a Snowflake customer, but we cannot really migrate all our data. I mean, it will be really costly. And we have a lot of tools that need access, direct access. So this is why we created Iceberg because we can really... I mean, we really think that we can bring the benefit of Snowflake to this data. >> Gives customers optionality. Okay. I use this term super cloud. You don't use the term, but that's okay. And I get a lot of heat for it. But to me, what you're doing is quite a bit different than multicloud because you're creating that abstraction layer. You're bringing value above it. My question to you is, the most of the heat I get is, oh, that's just SaaS. Are you just SaaS? >> No. I mean, no, absolutely not. I mean, you're right we are a super cloud. I mean it's a much better word than saying we are multicloud. Multicloud is often viewed as oh, I have my system and now I can run this system in the different cloud providers. Snowflake is different. We have one single platform for the world, which happens to have some regions are AWS region, some regions are Azure, some regions are GCP, Google and we merge them together. We have this Snowgrid technology that connects all our regions together so that we have really one platform for the world. And that's very important because when you talk about connections of data and expertise applications you want to have global reach, right. It doesn't exist. We are not siloed by region of the world, right? You have a lot of companies which are multinational that have presence everywhere. And you want to have this global reach. The world is not a independent set of regions and countries, right. And that's the realization. So we had to create this global platform for our customers. >> And now you have people building clouds on top of your data cloud, well that to me is the next signal. In your keynote, you talked about seven pillars, all data, all workloads, global architecture, self-managed, programmable, marketplace, governance, which ones are the most important? >> All of them. It's like when you have kids, you don't want to pick and say, this one is my preferred one, so they are really important. All of them, as I said without data, there is no Snowflake, right? So all data is so important that we can reach every data, wherever it is. And Iceberg is a part of that, but all workload is really important because you don't want to put your data in one platform, if you cannot run all your workloads and workloads are much broader than just data warehousing, there is data engineering, data science, ML engineering, (indistinct) all these workloads applications. So that's critical. Programmable is where we are moving, right. We want to be the place where data applications are built. And we think we have a lot of advantages because data application needs to use many workloads at once, right? It's not that that application will do only data warehousing, they need to store their states, they need to use this new workload that we define, which is Unistore. They need to do data engineering because they need to get data, right. They have to save this data. So they need to combine many workload and if they have to stitch this workload, because the platform was not designed as one single product where everything is consistent and works together, that you have to stitch, it's complicated for this application to make it work. So Snowflake is we believe an ideal platform to run these data applications. So all workloads, programmable, obviously, so that you can program. And programmable has two aspects, which is big part of our announcement. Is both data programmability, which is running Python against petabyte, terabytes of data at scale and doing it scale out. So that's what we call data programmability. So both Java, Python and (indistinct), but also running applications like UI. And we had this acquisition of Streamlit. Streamlit now has been fully integrated in Snowflake. We announced that such that not only you can have this data programmability, but you can expose your data through this nice UIs, interactive UI to business users potentially. So it goes all the way there. Global is super important. As we say, we want to be one platform for the world. And of course, as I said, the last pillar, which is somehow critical for us, because we are cloud, we need to have governance. We need to have security of our data. And why it took us so long to do Python is not because it's out to run Python, right? Everyone can run Python it's because we had to secure it. And I talk about it creating this amazing sandboxing technology, such that when you include third party libraries and third party codes, you are guaranteed that this third party code will not reach to infiltrate your data, right. We control the environment that Snowflake provides. >> Can you share us some of the feedback from the customer? You probably had many customer conversations over the last four days. >> Look at that smile. (interviewer laughing) (Lisa laughing) >> Actually not because I was so busy everywhere. Unfortunately, I didn't speak to many customers. Saying that, I had everyone stopping me and talking about what they heard and yeah, there is a huge excitement about all of this. >> What's been the feedback around the theme of the event? The world of data collaboration. Data collaboration is so critical as every company these days must be a data company to compete, to win. What's been from just some of the feedback that you've had customers really embracing data collaboration, what Snowflake is enabling. >> Yeah. I mean, almost every company which is using Snowflake, is collaborating with data. You have heard, the number of stable edges that we have, and there is a real need for that because your data alone... You cannot make sense of your data if it is just alone. It needs to be connected with other data. You haven't not generated. So all data, when you say the first pillar of Snowflake is all data is not only about your data, but is about all the data that's created around you. That puts perspective on your own data. And that's critical and it's so painful to get. I mean, even your data is difficult to have access to your data, but imagine data that you didn't produce. And so yes, so the data collaboration is critical, and then now we expanded it to application and expertise, sharing models, for example, That's going to have a huge impact. >> All data includes now transaction data, right? >> Yes. >> That's a big part of the announcements that you guys made. >> Yeah. So and that's the motivation for that was really, if we want to run application, full application, we announced native applications, which are fully executed and run inside the (indistinct) data cloud, right. They need all the services that application need and in particular managing their states. And so we created Unistore, which is a new workload, which allows you to combine transactional data, which are generated by this application. And at the same time being able to do analytics directly on this data. So we call it Hybrid Table because it has this hybrid aspect. You can do both transactional access to this data and at the same time analytic here without having data pipeline and moving data and transforming it from the transactional system to the analytical system, right. Snowflake is one system. Again, in the spirit of simplifying everything, this is the Snowflake (indistinct). >> I can ask the same question I ask at first, (indistinct) when was the aha moment that you and Thierry had that said, this is not just a better data warehouse, it's actually more than that. You probably didn't call it a data cloud until later on, but did you know that from the beginning or was that something you kind of stumbled into? >> No. So as I said, we founded Snowflake in 2012 and Thierry and I, we locked in my apartment and we were doing the blueprint of Snowflake and trying to find what is the revolution with the cloud for this data warehouse system and analytical system, both big data and data warehouse. And the aha moment was but of course cloud, okay. What is cloud? It's elasticity, it's service and later collaboration. So in the elasticity aspect, when you ask database people, what is elasticity, they will tell you, oh, you have a cluster of nodes. Like if it is Oracle, it would be a (indistinct) cluster. And the elasticities that you can add one node, two node to this cluster without having too much impact on the existing workload, because you need to shuffle data, right. It's hard and doing it online, right, that's elasticity. If you can do that, you are elastic. We thought that that was not very interesting to do that. What is interesting with elasticity is to plug new workloads. You can plug a workload like that and that workload is running without having any impact on other workloads, which are running on the platform. So elasticity for us was having dedicated computer resources to workloads. And these computer resources could start and be part as soon as the workload starts and will shut down when the workload finishes and they will be sized exactly for the demand of that workload. And we thought the aha moment was, okay if we can do that, now we can run a workload with, let's say 10X more computer resources than what you would have used or 100X more. Okay, let's say 100X more because we paralyzed things. Now this workload can run 100X faster, right? That's assuming we do a good job in the scale, which is our IP. And if we can do that, now the computer resources that you have used, you have used them for 100 times less. So you have used 100 times more resources because you have more nodes, but because you go fast, you use them for less time, right? So if you multiply the two it's constant. So you can run and accelerate workload dramatically 10X, 100X for the same price. Even if we are not better in efficiency than competition, just having that was the magic, right? >> You know how Google founders originally had trouble raising money because who needs another search engine? Did you get from original, like when you started going to raise money, Amazon's got a database, so who needs another cloud database? Did you get that early on or was it just obvious Speiser and companies as well. >> Speiser is a little bit on the crazy side and ambitious and so Speiser is Speiser. And of course he had no doubt, but even him was saying Benoit, Thierry, Hadoop, right. Everyone is saying Hadoop is going to be the revolution. And you guys are betting actually against Hadoop because we told Speiser, Hadoop is a bad system, it's going to fail, but at the time everyone was so bullish about Hadoop, everyone was implementing Hadoop that it didn't look like it was going to fail and we were probably wrong. So there was a lot of skepticism about not leveraging Hadoop and not being an Hadoop. Okay, something being on top of Hadoop. That was number one. There was no cloud warehouse at the time we started. Redshift was not started. It was the pioneer somewhere when Snowflake was founded. So creating a data warehouse in the cloud sounded crazy to people. How am I going to move my data over there? And security and what about security, the cloud is not secure. So that was another... >> So you guys predated that Parexel move by... >> Yes. >> Okay, so that's interesting. And I thought when Redshift... I mean, Amazon announced Redshift, I was sure that Mike Speiser will come and say, guys it's too sad, but they beat you guys and they build something and actually it was the reverse. Mike Speiser was super excited and so it was interesting to me. >> Wow, that's amazing. 'Cause John Furrier and I, we were early with theCUBE. when theCUBE started it was like the beginning of Hadoop. And so we brought theCUBE to, I think it was the second Hadoop World and we was rubbing nickels together at the time. And I was so excited bring compute to storage and it made so much sense. But I remember and I won't say who it was, but an early Hadoop committer told me this is going to fail. And I'm like, what? And he started going age basis crap and all this stuff. And I was sad because I was so excited, but it turned out that you had the same (indistinct). >> Because of complexity. Okay, Hadoop failed for two reasons. One is because they decided that, oh, a lot of this database thing, you don't need transaction, you don't need SQL, you don't necessarily, you don't need to go fast. It'll be batch, normal real time interaction with data, no one needs that. >> Cheap storage. >> So a lot of compromise on the very important technology. And at the same time, extreme complexity and complexity for me was, where I was I knew that it was going to fail big time and we bet Snowflake on the failure of Hadoop indeed. >> And there was no cloud early on in Hadoop. >> And there was no cloud too. >> And that was what killed it. That was like... >> You're right. And the model that Hadoop had for data didn't work on block storage. Block storage is not as efficient as HGFS. So that was also another figure. >> Do you ever sit back and think about... So you think about how much money has poured in to separating compute from storage and cloud databases and you started it all. (interviewer laughing) >> Yeah. No, this is... >> Pretty amazing. >> Yeah. >> Right, so that's good. That means that you're onto a good idea, but a lot of people get confused that again, they think that you're a cloud data warehouse and you're not, I mean, you're much more than that. >> Yeah, I hate that. I have to say, because from day one we were not a cloud data warehouse. As I said, it was all about combining the big data, massive amount of unstructured data, petabytes stored as files. Okay, that's very important, store as files where it's very easy to drop data in the system without... Very low cost to combine with data warehouse, full multi statement transaction when people will tell you today, oh, now we are a data warehouse. They don't have multi statement transaction, right. So we had from day one multi statement transaction really efficient SQL. You could run your dashboard. So combining these two worlds was I think the crazy thing, that's the crazy innovation that Snowflake did initially. >> Yeah. >> And I know it's really easy to build data warehouse somewhere, because if you don't think about big data, petabytes, extremely structured data, you remove a lot of complexity. >> This is why Lisa, when you get excited about technology, but you always have to have a, somebody who really deeply understands technology to stink test it, all right so awesome. Thank you for sharing that story. >> Yeah. >> Fantastic. So over 5,900 customers now. I saw over 500 in the Forbes G2K, over almost 10,000 people here this year. If we think back to 2019, there was about what? Less than 2000 people. >> Yeah. >> What do you think is going to happen next year? >> I don't know. I don't like to think about next year. I mean, I always say, Snowflake is so exciting to me because it is like a TV show, right. Where you wait the next season and we have one season every year. So I'm really excited to know what is going to happen next year. And I don't want to project what I think will happen, but all these movements to the Snowflake being the platform for data application. I want to see what people are going to build on our platform. I mean, that's the excitement. >> Season 11 coming up. >> Yes. Season 11. Yes. >> No binge watching here. Benoit, it's been a pleasure to have you on the program. >> Thank you. >> Congratulations on incredible success, the momentum, the energy is contagious. We love it. (Benoit laughing) >> Thank you so much. >> Thank you. >> Bye bye. >> For Benoit Dageville and Dave Vellante, I'm Lisa Martin. You're watching theCUBE's coverage of Snowflake Summit '22. Dave and I will be right back with a wrap. (upbeat music)
SUMMARY :
is coming to an end, Thank you, thank you. you guys started on Monday. And you can feel the future of the data cloud. and the marketplace where you So my question is, did you envision And that really has to be And that's the power of the and I'm not going to get So everything that you can the benefit of Snowflake to this data. My question to you is, the And that's the realization. And now you have people building clouds And of course, as I said, the last pillar, the feedback from the customer? Look at that smile. I was so busy everywhere. the feedback that you've had but imagine data that you didn't produce. announcements that you guys made. So and that's the motivation I can ask the same question And the elasticities that you can add like when you started at the time we started. So you guys predated and so it was interesting to me. And I was so excited you don't need to go fast. And at the same time, extreme complexity And there was no And that was what killed it. And the model that Hadoop had for data and you started it all. No, this is... but a lot of people get I have to say, because from day one because if you don't think about big data, This is why Lisa, when you I saw over 500 in the Forbes G2K, I mean, that's the excitement. Yes. to have you on the program. the momentum, the energy is contagious. Dave and I will be right back with a wrap.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Mike Speiser | PERSON | 0.99+ |
10X | QUANTITY | 0.99+ |
100X | QUANTITY | 0.99+ |
100 times | QUANTITY | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Mike Speiser | PERSON | 0.99+ |
2012 | DATE | 0.99+ |
Benoit Dageville | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Benoit | PERSON | 0.99+ |
Monday | DATE | 0.99+ |
Thierry | PERSON | 0.99+ |
Thursday | DATE | 0.99+ |
2019 | DATE | 0.99+ |
Tuesday | DATE | 0.99+ |
Snowflake | TITLE | 0.99+ |
ORGANIZATION | 0.99+ | |
next year | DATE | 0.99+ |
two aspects | QUANTITY | 0.99+ |
Lisa | PERSON | 0.99+ |
Python | TITLE | 0.99+ |
This week | DATE | 0.99+ |
one season | QUANTITY | 0.99+ |
two reasons | QUANTITY | 0.99+ |
One | QUANTITY | 0.99+ |
Hadoop | PERSON | 0.99+ |
two | QUANTITY | 0.99+ |
both | QUANTITY | 0.99+ |
Snowflake Summit '22 | EVENT | 0.99+ |
this week | DATE | 0.99+ |
one platform | QUANTITY | 0.99+ |
Streamlit | TITLE | 0.99+ |
Speiser | ORGANIZATION | 0.99+ |
Java | TITLE | 0.99+ |
one platform | QUANTITY | 0.99+ |
10 years | QUANTITY | 0.99+ |
one system | QUANTITY | 0.98+ |
one node | QUANTITY | 0.98+ |
Less than 2000 people | QUANTITY | 0.98+ |
Snowflake | EVENT | 0.98+ |
AWS | ORGANIZATION | 0.98+ |
two node | QUANTITY | 0.98+ |
this year | DATE | 0.98+ |
second | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
John Furrier | PERSON | 0.98+ |
Hadoop | TITLE | 0.97+ |
over 5,900 customers | QUANTITY | 0.97+ |
10 years ago | DATE | 0.97+ |
one single product | QUANTITY | 0.97+ |
first pillar | QUANTITY | 0.97+ |
Google Docs | TITLE | 0.97+ |
Snowflake | ORGANIZATION | 0.97+ |
Multicloud | TITLE | 0.97+ |
over 500 | QUANTITY | 0.97+ |
Parexel | ORGANIZATION | 0.96+ |
Matthew Carroll, Immuta | Snowflake Summit 2022
(Upbeat music) >> Hey everyone. Welcome back to theCUBE's continuing coverage day two Snowflake Summit '22 live from Caesar's forum in Las Vegas. Lisa Martin here with Dave Vellante, bringing you wall to wall coverage yesterday, today, and tomorrow. We're excited to welcome Matthew Carroll to the program. The CEO of Immuta, we're going to be talking about removing barriers to secure data access security. Matthew, welcome. >> Thank you for having me, appreciate it. >> Talk to the audience a little bit about Immuta you're a Snowflake premier technology partner, but give him an overview of Immuta what you guys do, your vision, all that good stuff. >> Yeah, absolutely, thanks. Yeah, if you think about what Immunta at it's core is, we're a data security platform for the modern data stack, right? So what does that mean? It means that we embed natively into a Snowflake and we enforce policies on data, right? So, the rules to be able to use it, to accelerate data access, right? So, that means connecting to the data very easily controlling it with any regulatory or security policy on it as well as contractual policies, and then being able to audit it. So, that way, any corporation of any size can leverage their data and share that data without risking leaking it or potentially violating a regulation. >> What are some of the key as we look at industry by industry challenges that Immuta is helping those customers address and obviously quickly since everything is accelerating. >> Yeah. And it's, you're seeing it 'cause the big guys like Snowflake are verticalizing, right? You're seeing a lot of industry specific, you know, concepts. With us, if you think of, like, where we live obviously policies on data regulated, right? So healthcare, how do we automate HIPAA compliance? How do we redesign clinical trial management post COVID, right? If you're going to have billions of users and you're collecting that data, pharmaceutical companies can't wait to collect that data. They need to remove those barriers. So, they need to be able to collect it, secure it, and be able to share it. Right? So, double and triple blinded studies being redesigned in the cloud. Government organizations, how do we share security information globally with different countries instantaneously? Right? So these are some of the examples where we're helping organizations transform and be able to kind of accelerate their adoption of data. >> Matt, I don't know if you remember, I mean, I know you remember coming to our office. But we had an interesting conversation and I was telling Lisa. Years ago I wrote a piece of you know, how to build on top of, AWS. You know, there's so much opportunity. And we had a conversation, at our office, theCUBE studios in Marlborough, Massachusetts. And we both, sort of, agreed that there was this new workload emerging. We said, okay, there's AWS, there's Snowflake at the time, we were thinking, and you bring machine learning, at time where we were using data bricks, >> Yeah. >> As the example, of course now it's been a little bit- >> Yeah. Careful. >> More of a battle, right, with those guys. But, and so, you see them going in their different directions, but the premise stands is that there's an ecosystem developing, new workloads developing, on top of the hyper scale infrastructure. And you guys play a part in that. So, describe what you're seeing there 'cause you were right on in that conversation. >> Yeah. Yeah. >> It's nice to be, right. >> Yeah. So when you think of this design pattern, right, is you have a data lake, you have a warehouse, and you have an exchange, right? And this architecture is what you're seeing around you now, is this is every single organization in the world is adopting this design pattern. The challenge that where we fit into kind of a sliver of this is, the way we used to do before is application design, right? And we would build lots of applications, and we would build all of our business logic to enforce security controls and policies inside each app. And you'd go through security and get it approved. In this paradigm, any user could potentially access any data. There's just too many data sources, too many users, and too many things that can go wrong. And to scale that is really hard. So, like, with Immuta, what we've done, versus what everyone else has done is we natively embedded into every single one of those compute partners. So ,Snowflake, data breaks, big query, Redshift, synapse on and on. Natively underneath the covers, so that was BI tools, those data science tools hit Snowflake. They don't have to rewrite any of their code, but we automatically enforce policy without them having to do anything. And then we consistently audit that. I call that the separation of policy from platform. So, just like in the world in big data, when we had to separate compute from storage, in this world, because we're global, right? So we're, we have a distributed workforce and our data needs to abide by all these new security rules and regulations. We provide a flexible framework for them to be able to operate at that scale. And we're the only ones in the world doing it. >> Dave Vellante: See the key there is, I mean, Snowflake is obviously building out its data cloud and the functions that it's building in are quite impressive. >> Yeah. >> Dave Vellante: But you know at some point a customer's going to say, look I have other stuff, whether it's in an Oracle database, or data lake or wherever, and that should just be a node on this global, whatever you want to call it, mesh or fabric. And then if I'm hearing you right, you participate in all of that. >> Correct? Yeah We kind of, we were able to just natively inject into each, and then be able to enforce that policy consistently, right? So, hey, can you access HIPAA data? Who are you? Are you authorized to use this? What's the purpose you want to query this data? Is it for fraud? Is it for marketing? So, what we're trying to do as part of this new design paradigm is ensure that we can automate nearly the entire data access process, but with the confidence and de-risk it, that's kind of the key thing. But the one thing I will mention is I think we talk a lot about the core compute, but I think, especially at this summit, data sharing is everything. Right? And this concept of no copy data sharing, because the data is too big and there's too many sets to share, that's the keys to the kingdom. You got to get your lake and your warehouse set with good policy, so you can effectively share it. >> Yeah, so, I wanted to just to follow up, if I may. So, you'd mentioned separating compute from storage and a lot of VC money poured into that. A lot of VC money poured into cloud database. How do you see, do you see Snowflake differentiating substantially from all the other cloud databases? And how so? >> I think it's the ease of use, right? Apple produces a phone that isn't much different than other competitors. Right? But what they do is, end to end, they provide an experience that's very simple. Right? And so yes. Are there other warehouses? Are there other ways to, you know you heard about their analytic workloads now, you know through unistore, where they're going to be able to process analytical workloads as well as their ad hoc queries. I think other vendors are obviously going to have the same capabilities, but I think the user experience of Snowflake right now is top tier. Right? Is I can, whether I'm a small business, I can load my debt in there and build an app really quickly. Or if I'm a JP Morgan or, you know, a West Farmer's I can move legacy, you know monolithic architectures in there in months. I mean, these are six months transitions. When think about 20 years of work is now being transitioned to the cloud in six months. That's the difference. >> So measuring ease of views and time to value, time to market. >> Yeah. That's it's everything is time to value. No one wants to manage the infrastructure. In the Hudup world, no one wants to have expensive customized engineers that are, you know, keeping up your Hudup infrastructure any longer. Those days are completely over. >> Can you share an example of a joint customer, where really the joint value proposition that Immuta and Snowflake bring, are delivering some pretty substantial outcomes? >> Yeah. I, what we're seeing is and we're obviously highly incentivized to get them in there because it's easier on us, right? Because we can leverage their row and com level security. We can leverage their features that they've built in to provide a better experience to our customers. And so when we talk about large banks, they're trying to move Terra data workloads into Snowflake. When we talk about clinical trial management, they're trying to get away from physical copies of data, and leverage the exchanges of mechanism, so you can manage data contracts, right? So like, you know, when we think of even like a company like Latch, right? Like Latch uses us to be able to oversee all of the consumer data they have. Without like a Snowflake, what ends up happening is they end up having to double down and invest on their own people building out all their own infrastructure. And they don't have the capital to invest in third party tools like us that keep them safe, prevent data leaks, allow them to do more and get more value out of their data, which is what they're good at. >> So TCO reduction I'm hearing. >> Matthew Carroll: Yes, exactly. >> Matt, where are you as a company, you've obviously made a lot of progress since we last talked. Maybe give us the update on you know, the headcount, and fundraising, and- >> Yeah, we're just at about 250 people, which scares me every day, but it's awesome. But yeah, we've just raised 100 million dollars- >> Lisa Martin: Saw that, congratulations. >> Series E, thank you, with night dragon leading it. And night dragon was very tactical as well. We are moving, we found that data governance, I think what you're seeing in the market now is the catalog players are really maturing, and they're starting to add a suite of features around governance, right? So quality control, observability, and just traditional asset management around their data. What we are finding is is that there's a new gap in this space, right? So if you think about legacy it's we had infrastructure security we had the four walls and we protect our four walls. Then we moved to network security. We said, oh, the adversary is inside zero trust. So, let's protect all of our endpoints, right? But now we're seeing is data is the security flaw data could be, anyone could potentially access it in this organization. So how do we protect data? And so what we have matured into is a data security company. What we have found is, there's this next generation of data security products that are missing. And it's this blend between authentication like an, an Okta or an AuthO and auth- I'm sorry, authorization. Like Immuta, where we're authorizing certain access. And we have to pair together, with the modern observability, like a data dog, to provide an a layer above this modern data stack, to protect the data to analyze the users, to look for threats. And so Immuta has transformed with this capital. And we brought Dave DeWalt onto our board because he's a cybersecurity expert, he gives us that understanding of what is it like to sell into this modern cyber environment. So now, we have this platform where we can discover data, analyze it, tag it, understand its risk, secure it to author and enforce policies. And then monitor, the key thing is monitoring. Who is using the data? Why are they using the data? What are the risks to that? In order to enforce the security. So, we are a data security platform now with this raise. >> Okay. That, well, that's a new, you know, vector for you guys. I always saw you as an adjacency, but you're saying smack dab in the heart >> Matthew Carroll: Yes. Yeah. We're jumping right in. What we've seen is there is a massive global gap. Data is no longer just in one country. So it is, how do we automate policy enforcement of regulatory oversight, like GDPR or CCPA, which I think got this whole category going. But then we quickly realized is, well we have data jurisdiction. So, where does that data have to live? Where can I send it to? Because from Europe to us, what's the export treaty? We don't have defined laws anymore. So we needed a flexible framework to handle that. And now what we're seeing is data leaks, upon data leaks, and you know, the Snowflakes and the other cloud compute vendors, the last thing they ever want is a data leak out of their ecosystem. So, the security aspects are now becoming more and more important. It's going to be an insider threat. It's someone that already has access to that and has the rights to it. That's going to be the risk. And there is no pattern for a data scientist. There's no zero trust model for data. So we have to create that. >> How are you, last question, how are you going to be using a 100 million raised in series E funding, which you mentioned, how are you going to be leveraging that investment to turn the volume up on data security? >> Well, and we still have also another 80 million still in the bank from our last raise, so 180 million now, and potentially more soon, we'll kind of throw that out there. But, the first thing is M and A I believe in a recessing market, we're going to see these platforms consolidate. Larger customer of ours are driving us to say, Hey, we need less tools. We need to make this easier. So we can go faster. They're, even in a recessing market, these customers are not going to go slower. They're moving in the cloud as fast as possible, but it needs to be easier, right? It's going back to the mid nineties kind of Lego blocks, right? Like the IBM, the SAP, the Informatica, right? So that's number one. Number two is investing globally. Customer success, engineering, support, 24 by seven support globally. Global infrastructure on cloud, moving to true SaaS everywhere in the world. That's where we're going. So sales, engineering, and customer success globally. And the third is, is doubling down on R and D. That monitor capability, we're going to be building software around. How do we monitor and understand risk of users, third parties. So how do you handle data contracts? How do you handle data use agreements? So those are three areas we're focused on. >> Dave Vellante: How are you scaling go to market at this point? I mean, I presume you are. >> Yeah, well, I think as we're leveraging these types of engagements, so like our partners are the big cloud compute vendors, right? Those data clouds. We're injecting as much as we can into them and helping them get more workloads onto their infrastructure because it benefits us. And then obviously we're working with GSIs and then RSIs to kind of help with this transformation, but we're all in, we're actually deprecating support of legacy connectors. And we're all in on cloud compute. >> How did the pivot to all in on security, how did it affect your product portfolio? I mean, is that more positioning or was there other product extensions that where you had to test product market fit? >> Yeah. This comes out of customer drive. So we've been holding customer advisory boards across Europe, Asia and U.S. And what we just saw was a pattern of some of these largest banks and pharmaceutical companies and insurance companies in the world was, hey we need to understand who is actually on our data. We have a better understanding of our data now, but we don't actually understand why they're using our data. Why are they running these types of queries? Is this machine, you know logic, that we're running on this now, we invested all this money in AI. What's the risk? They just don't know. And so, yeah, it's going to change our product portfolio. We modularized our platform to the street components over the past year, specifically now, so we can start building custom applications on top of it, for specific users like the CSO, like, you know, the legal department, and like third party regulators to come in, as well as as going back to data sharing, to build data use agreements between one or many entities, right? So an SMP global can expose their data to third parties and have one consistent digital contract, no more long memo that you have to read the contract, like, Immuta can automate those data contracts between one or many entities. >> Dave Vellante: And make it a checkbox item. >> It's just a checkbox, but then you can audit it all, right? >> The key thing is this, I always tell people, there's negligence and gross negligence. Negligence, you can go back and fix something, gross negligence you don't have anything to put into controls. Regulators want you to be at least negligent, grossly negligent. They get upset. (laughs) >> Matthew, it sounds like great stuff is going on at Immuta, lots of money in the bank. And it sounds like a very clear and strategic vision and direction. We thank you so much for joining us on theCUBE this morning. >> Thank you so much >> For our guest and Dave Vellante, I'm Lisa Martin, you're watching theCUBE's coverage of day two, Snowflake Summit '22, coming at ya live, from the show floor in Las Vegas. Be right back with our next guest. (Soft music)
SUMMARY :
Matthew Carroll to the program. of Immuta what you guys do, your vision, So, the rules to be able to use it, What are some of the key So, they need to be able to collect it, at the time, we were thinking, And you guys play a part in that. of our business logic to Dave Vellante: See the key there is, on this global, whatever you What's the purpose you just to follow up, if I may. they're going to be able to and time to value, time to market. that are, you know, keeping And they don't have the capital to invest Matt, where are you as a company, Yeah, we're just at about 250 people, What are the risks to that? I always saw you That's going to be the risk. but it needs to be easier, right? I mean, I presume you are. and then RSIs to kind of help the CSO, like, you know, Dave Vellante: And Regulators want you to be at Immuta, lots of money in the bank. from the show floor in Las Vegas.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Dave Vellante | PERSON | 0.99+ |
Matthew Carroll | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Matthew | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Europe | LOCATION | 0.99+ |
Dave DeWalt | PERSON | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Lisa | PERSON | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Matt | PERSON | 0.99+ |
Immuta | ORGANIZATION | 0.99+ |
tomorrow | DATE | 0.99+ |
100 million dollars | QUANTITY | 0.99+ |
Asia | LOCATION | 0.99+ |
Apple | ORGANIZATION | 0.99+ |
U.S. | LOCATION | 0.99+ |
Informatica | ORGANIZATION | 0.99+ |
today | DATE | 0.99+ |
six months | QUANTITY | 0.99+ |
Latch | ORGANIZATION | 0.99+ |
100 million | QUANTITY | 0.99+ |
SAP | ORGANIZATION | 0.99+ |
180 million | QUANTITY | 0.99+ |
Hudup | ORGANIZATION | 0.99+ |
JP Morgan | ORGANIZATION | 0.99+ |
third | QUANTITY | 0.99+ |
Immunta | ORGANIZATION | 0.99+ |
yesterday | DATE | 0.99+ |
Marlborough, Massachusetts | LOCATION | 0.99+ |
theCUBE | ORGANIZATION | 0.99+ |
one country | QUANTITY | 0.98+ |
each app | QUANTITY | 0.98+ |
West Farmer | ORGANIZATION | 0.98+ |
Snowflake Summit '22 | EVENT | 0.98+ |
Snowflake | TITLE | 0.98+ |
80 million | QUANTITY | 0.98+ |
one | QUANTITY | 0.97+ |
Oracle | ORGANIZATION | 0.97+ |
both | QUANTITY | 0.97+ |
mid nineties | DATE | 0.96+ |
GDPR | TITLE | 0.96+ |
Snowflakes | EVENT | 0.96+ |
day two | QUANTITY | 0.95+ |
HIPAA | TITLE | 0.95+ |
Snowflake Summit '22 | EVENT | 0.94+ |
24 | QUANTITY | 0.94+ |
about 20 years | QUANTITY | 0.93+ |
first thing | QUANTITY | 0.93+ |
this morning | DATE | 0.93+ |
past year | DATE | 0.92+ |
seven | QUANTITY | 0.92+ |
Snowflake | ORGANIZATION | 0.92+ |
Snowflake Summit 2022 | EVENT | 0.92+ |
about 250 people | QUANTITY | 0.92+ |
each | QUANTITY | 0.91+ |
day | QUANTITY | 0.87+ |
double | QUANTITY | 0.87+ |
SMP | ORGANIZATION | 0.85+ |
unistore | TITLE | 0.84+ |
triple | QUANTITY | 0.84+ |
Snowflake | EVENT | 0.83+ |
Caesar | PERSON | 0.82+ |
three areas | QUANTITY | 0.82+ |
GSIs | ORGANIZATION | 0.81+ |
Years ago | DATE | 0.8+ |
Carl Perry, Snowflake | Snowflake Summit 2022
(calm music) >> Welcome to theCUBE's live coverage of Snowflake Summit '22 from Las Vegas, Caesars Forum. Lisa Martin here with Dave Vellante, we're going to unpack some really cool stuff next, in the next 10 minutes with you, Carl Perry joins us, the Director of Product Management at Snowflake, he's here to talk about Snowflake's new Unistore workloads, how it's driving the next phase of innovation, welcome to the program. >> Oh, thank you so much for having me, this is awesome. >> There's a ton of momentum here, I saw the the numbers from fiscal 23Q1, product revenue 394 million, 85% growth, a lot of customers here, the customer growth is incredible as well, talk to us about Unistore, what is it? Unpack it and how have the customers been influential in it's development? >> Yeah, so Unistore is a way for customers to take their transactional workloads, for their enterprise applications and now have them run on or be built on top of Snowflake and now, you have your transactional data, along with all of your historical data, so now you have a single unified platform for doing anything you need to do with your data, whether it's transactional, single row look-ups, we can do that, whether it's the analytical data across again, transactional and historical data in a single query, our customers are super excited about this. >> So, what are Hybrid Tables? Is that just an extension of external tables? >> Yeah, that's a great question. So, Hybrid Tables are a new table-type that we've added to Snowflake and Hybrid Tables are really kind of just like another table with a couple of key differences, so number one is that Hybrid Tables provide fast, fine-grain read and write operations, so when you do something like a select star from customers where customer ID=832, that's going to return extremely fast, but on top of that same data, your transactional data, you can actually perform amazing analytical queries that return extremely fast and that's what Hybrid Tables at their core are. >> So, what does this mean for, so you're bringing that world of transaction and analytics together, what does it mean for customers? Walk us through Carl, an example of- >> Yeah, so it's great, so Adobe is a customer that is looking at using and leveraging Hybrid Tables today, and then more broadly Unistore, and frankly, Adobe has been an amazing customer since they started their journey, just really quickly, they're in phase three, the first phase was customers had data in Snowflake that they wanted to take advantage of with the Adobe Campaign Platform and so what they did is they built a connector basically into and being able to access customer data, and then they started to look at, "Well, this thing's working really well, let's try to leverage Snowflake for all our analytical needs." And so that was kind of phase two, and now phase three is like, look let's go and reimagine what we can do with the Adobe Campaign Platform by having both the transactional and analytical data in the same platform, so that they can really enable their customers to do personalization, ad campaign management, understanding the ethicacy of those things at a scale that they haven't been able to do before. >> Prior to this capability, they would what? Have to go outside of the Snowflake Data Cloud? And do something else? And then come back in? >> Exactly, right? So, they'd have a transactional system where all of the transactional state for what the customer was doing inside Adobe Campaign, setting up all their campaigns and everything, and that would be stored inside a database, right? And then they would need to ensure that, that data was moved over to Snowflake for further analytical purposes, right? You know you imagine the complexity that our customers have to manage every single day, a separate transactional system, an ETL pipeline to keep that data flowing and then Snowflake, right? And with Unistore, we really believe that customers will be able to remove that complexity from their lives and have that single platform that really makes their lives easier. >> I mean, they'll still have a transactional system, will they not? Or do you see a day where they sort of sunset that? >> I mean, there's a set of workloads that are not going to be the best choice today for Unistore and Hybrid Tables, right? And so we know that customers will continue to have their own transactional systems, right? And there's lots of transactional systems that customers rely and have entire applications, and systems built around, right? Right now with Hybrid Tables and Unistore, customers can take those enterprise applications, not consumer-facing applications and move them over to leverage Snowflake, and then really think about re-imagining how they can use their data that's both realtime transactional, as well as all the historical data without the need to move things between systems or use a ton of different services. >> The Adobe example that you just gave seems like, I loved how you described the phases they're in, they're discovering, it's like peeling the onion and just discovering more, and more, but what it sounds like is that Snowflake has enabled Adobe to transform part of it's business, how is Unistore positioned to be so transformational for your customers? >> Well, I mean I think there's a couple of things, so one, they have this like level of complexity today for a set of applications that they can completely stop worrying about, right? No need to maintain that separate transactional system for that again, enterprise application, no need to maintain that ETL pipeline, that's kind of like one step, the next step is, I mean all your data's in Snowflake, so you can start leveraging that data for insight and action immediately, there's no delay in being able to take advantage of that data, right? And then number three, which I think is the most compelling part is because it's part of Snowflake, you getting the benefit of Snowflake's entire ecosystem, whether it's first party capabilities like easy to manage and enforce really powerful governance, and security policies, right? Being able to take data from the market place and actually join it with my realtime transactional data, this is game-changing and then most importantly is the third-party ecosystem of partners who are building all these incredible solutions on top of Snowflake, I can't even begin to imagine what they're going to do with Hybrid Tables in Unistore. >> So, Carl I have to ask you, so I talked to a lot of customers and I talked to a lot of technology companies, explain, so Snowflake obviously was the first to separate compute from storage and you know the cloud, cloud database and then tons of investment came into that space, kind of follow you on, so that's cool, you reached escape velocity, awesome, but a lot of the companies that I talked to are saying, "We're converging transaction and analytics," I think (speaking softly) calls it HTAP or something, they came up with a name, explain the difference between what you're doing and what everybody else is doing, and why, what customer benefits you're delivering? >> Yeah, so I mean I think that's a really great question and to use the term you used HTAP, right? It's a industry understood term, really when people think about HTAP, what that is about is taking your transactional data that you have and enabling you to do fast analytical capabilities on that, and that's great, but there are a couple of problems that historical HTAP solutions have suffered from, so number one, that acceleration, that colander format of data is all in memory, so you're bound by the total amount of memory that you can use to accelerate the queries that you want to, so that's kind of problem one, this is not the approach that Snowflake is taking, most importantly, it's not just about accelerating queries on transactional data, whether it's a single-row lookup or a complex aggregate, it's about being able to leverage that data within the data cloud, right? I don't want to have a separate dataset on a transactional system or an HTAP system that can give me great analytics on transactional data and then I can't use it with all the other data that I have, it's truly about enabling the transformation with the data cloud and completely taking away silos, so that your data, whether it's realtime, whether it's historical, can be treated as a single dataset, this is the key thing that is different about Unistore, you can take the power of the data cloud, all of it, all of the partners, all the solutions and all the capabilities we continue to add, and leverage your data in ways that nobody's thought of possible before. >> Governance is a huge, huge component of that, right? So, in the press release, you have this statement, "As part of the Unistore Snowflake is introducing Hybrid Tables," you explained that, "Which offer fast, single-row operations and allow customers to build transactional business applications directly on Snowflake"- >> Yep. >> That's a little interesting tidbit, so you expect customers are going to build transactional applications inside the data cloud? And somewhat minimize the work that is going to be required by their existing transactional databases, correct? >> Exactly and I think, so let me say a couple things on this, right? So, first of all, there's a class of applications that will be able to just build on top of Hybrid Tables and run on Snowflake directly, for their transactional needs, I think what's super interesting here though is when you again start to talk about all your data, one example that we're going to walk through tomorrow in our talk is being able to do a transaction that updates data in a Hybrid Table and then updates data in a Standard Snowflake Table, and then either being able to atomically commit, or rollback that transaction, this is a transaction that's spanning multiple different table types inside Snowflake and you'll have consistency of either the rollback or the commit, this type of functionality doesn't exist elsewhere and being able to take, and build transactional applications with these capabilities, we think is transformative- >> And that's all going to happen inside the Snowflake Data Cloud, with all the capabilities and it's not like you know what you're doing with Dell and Pure, it's nice, but it's read-only, you can't you know add and delete, and do all that stuff, this is Native? First class citizen inside the database? >> Yep, just like other table types, you'll be able to take on and leverage the power of the data cloud as a normal table that you'd be able to use elsewhere. >> Got to ask you, your energy in the way that you're talking about this is fantastic, the transformation that it's going to be, how central it is to the product innovations that Snowflake is coming out with, what's been the feedback from customers? As there's so many thousands of folks here today, the keynote was standing in your room only, there was an overflow, what are you hearing on the floor here? >> Well, I mean, I think it was funny in the talk when I announced that primary keys are going to be required and enforced, and we got a standing ovation, I was like, "Wow, I didn't expect people to be so excited about primary key enforcement." I mean, what's been amazing both about the private preview and the feedback we're getting there, and then some of the early feedback we're getting from customers is that they want to understand and they're really thinking about like, "Wait, I can use Snowflake for all of this now?" And honestly I think that people are kind of like, "But wait, what would I do if I could have those applications running on Snowflake and not have to worry about multiple systems? Wait, I can combine it with all my historical data and anything that's in the data cloud, like what can I do?" Is the question they're asking and I think that this is the most fascinating thing, customers are going to build things they haven't been able to build before and I'm super excited to see what they do. >> But more specifically, my takeaway is that customers, actually application builders are going to be able to build applications that have data inherent to those apps, I mean John Furrier years ago said, "You know data is the new development kit." And it never happened the data, the data stack if you will separate from the application development stack, you're bringing those two worlds together, so what do you think the implications are of that? >> Well, I mean I think that we're going to dramatically simplify our customers lives, right? A thing that we focus on at Snowflake is relentless customer innovation, so we can make their lives better, so I mean frankly we talk to customers like, "Wait, I can do all this? Wait, are you sure that I'll be able to do this?" And we walk through what we can do, and what we can't do, and they really are like, "Wow, this could just dramatically simplify our lives and wait, what could we do with our data here?" And so, I think with the announcement of Unistore, and also all the Native app stuff that we're announcing today, I think we're really trying to enable customers and app developers there to think about, and being able to leverage Snowflake as their transactional system, the system of source, so I mean, I'm super excited about this, I came to Snowflake to work on this and I'm like, "Can't believe we get to talk about it." >> How do you, how, how? How does this work? What's the secret sauce behind it? Is it architecture or is it? >> Yeah, so I mean I think a big part of it is the architecture that we chose, so you know number one, a key product philosophy that we have at Snowflake is we have one product, we don't have many, we don't put the onus of complexity onto our customers and so building that into Snowflake is actually really hard, so underlying Hybrid Tables, which is the feature that powers Unistore is a row storage engine, a row-based storage engine, right? And then data is asynchronously copied over into a colander format and what this provides, because it's just another table that's deeply integrated with Snowflake is the compiler's completely aware of this, so you can write a query that spans multiple tables and take advantage of it, and we'll take over all the complexity, whether it needs to be a fast response to a single-row lookup, or it needs to aggregate and scan a ton of data, we'll make sure that we choose the right thing and provide you with the best performance that we have- >> You built that intelligence inside of that? >> Completely built in and amazing, but provided in a very simple fashion. >> You said you came to Snowflake to do this? How long ago was that? >> I came here a little over a year and a half. >> Okay, and had they started working on this obviously beforehand, or at least envisioning it, right? >> Yeah, this I mean, this is absolutely incredible, I have been working on this now for a year and a half, some of the team members have been working on it for more and it's incredible to finally be able to talk to customers and everybody about it, and for them to tell us what they're trying to do. I've already talked to a bunch of customers like, "Well wait, I could do this, or this, what about this scenario?" And it's awesome to hear their requirements, right? The thing that's been most amazing and you'll hear it in the talk tomorrow with Adobe who's been a great customer is like, "Customers give us insanely hard requirements." And what I love about this company is not, "Well, you know it's easier to do it this way." It's like, "No, how can we actually make their life easier?" And so, we really focus on doing that with Snowflake. >> And that's one of the things Frank talked about this morning with that mission alignment being critical there. So, it's in private preview now, when can folks expect to get their hands on it? >> Well, we don't have a date right now we're talking about, but you can go signup to be notified of the public preview when we get there, I think it's like snowflake.com/try-unistore, but we'll publish that later and you know if you're interested in the private preview, talk to your account team and we'll see if we can get you in. >> Carl, thank you so much for joining Dave and me in an action-packed 15 minutes, talking about the power of Unistore, what it's going to enable organizations to do and it sounds like you're tapping the surface, there's just so much more innovation that's to come, you're going to have to come back. >> Yes, that sounds awesome, thank you so much. >> Our pleasure. For Carl and Dave Vellante, I'm Lisa Martin, you're watching theCUBE's live coverage of Snowflake Summit '22 from the show floor in Las Vegas, we're going to be right back with our next guest. (calm music)
SUMMARY :
in the next 10 minutes with you, Oh, thank you so much for having me, and now, you have your transactional data, and that's what Hybrid and then they started to look at, and have that single platform and move them over to leverage Snowflake, and actually join it with my and to use the term you used HTAP, right? and leverage the power of the data cloud and I'm super excited to see what they do. the data stack if you will separate and being able to leverage Snowflake and amazing, and a half. and for them to tell us And that's one of the things and you know if you're interested and it sounds like you're Yes, that sounds awesome, and Dave Vellante,
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Lisa Martin | PERSON | 0.99+ |
Carl | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Carl Perry | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Frank | PERSON | 0.99+ |
Adobe | ORGANIZATION | 0.99+ |
15 minutes | QUANTITY | 0.99+ |
Unistore | ORGANIZATION | 0.99+ |
Snowflake | ORGANIZATION | 0.99+ |
a year and a half | QUANTITY | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
first phase | QUANTITY | 0.99+ |
Snowflake | TITLE | 0.99+ |
tomorrow | DATE | 0.98+ |
fiscal 23Q1 | DATE | 0.98+ |
one product | QUANTITY | 0.98+ |
today | DATE | 0.98+ |
single | QUANTITY | 0.98+ |
Snowflake Summit '22 | EVENT | 0.98+ |
snowflake.com/try-unistore | OTHER | 0.97+ |
two worlds | QUANTITY | 0.97+ |
first | QUANTITY | 0.97+ |
one example | QUANTITY | 0.97+ |
Unistore Snowflake | ORGANIZATION | 0.96+ |
single platform | QUANTITY | 0.96+ |
over a year and a half | QUANTITY | 0.96+ |
both | QUANTITY | 0.95+ |
394 million | QUANTITY | 0.95+ |
this morning | DATE | 0.93+ |
one step | QUANTITY | 0.93+ |
Las Vegas, | LOCATION | 0.92+ |
Hybrid Tables | ORGANIZATION | 0.92+ |
single-row | QUANTITY | 0.91+ |
one | QUANTITY | 0.91+ |
theCUBE | ORGANIZATION | 0.91+ |
Snowflake Summit 2022 | EVENT | 0.89+ |
85% growth | QUANTITY | 0.89+ |
tons | QUANTITY | 0.87+ |
single day | QUANTITY | 0.87+ |
number one | QUANTITY | 0.86+ |
years ago | DATE | 0.85+ |
thousands of folks | QUANTITY | 0.84+ |
Hybrid Tables | TITLE | 0.81+ |
Adobe Campaign Platform | TITLE | 0.79+ |
single row | QUANTITY | 0.79+ |
phase three | QUANTITY | 0.77+ |
Campaign Platform | TITLE | 0.76+ |
First | QUANTITY | 0.75+ |
Caesars Forum | LOCATION | 0.72+ |