Image Title

Search Results for Western:

Harveer Singh, Western Union | Western Union When Data Moves Money Moves


 

(upbeat music) >> Welcome back to Supercloud 2, which is an open industry collaboration between technologists, consultants, analysts, and of course, practitioners, to help shape the future of cloud. And at this event, one of the key areas we're exploring is the intersection of cloud and data, and how building value on top of hyperscale clouds and across clouds is evolving, a concept we call supercloud. And we're pleased to welcome Harvir Singh, who's the chief data architect and global head of data at Western Union. Harvir, it's good to see you again. Thanks for coming on the program. >> Thanks, David, it's always a pleasure to talk to you. >> So many things stand out from when we first met, and one of the most gripping for me was when you said to me, "When data moves, money moves." And that's the world we live in today, and really have for a long time. Money has moved as bits, and when it has to move, we want it to move quickly, securely, and in a governed manner. And the pressure to do so is only growing. So tell us how that trend is evolved over the past decade in the context of your industry generally, and Western Union, specifically. Look, I always say to people that we are probably the first ones to introduce digital currency around the world because, hey, somebody around the world needs money, we move data to make that happen. That trend has actually accelerated quite a bit. If you look at the last 10 years, and you look at all these payment companies, digital companies, credit card companies that have evolved, majority of them are working on the same principle. When data moves, money moves. When data is stale, the money goes away, right? I think that trend is continuing, and it's not just the trend is in this space, it's also continuing in other spaces, specifically around, you know, acquisition of customers, communication with customers. It's all becoming digital, and it's, at the end of the day, it's all data being moved from one place or another. At the end of the day, you're not seeing the customer, but you're looking at, you know, the data that he's consuming, and you're making actionable items on it, and be able to respond to what they need. So I think 10 years, it's really, really evolved. >> Hmm, you operate, Western Union operates in more than 200 countries, and you you have what I would call a pseudo federated organization. You're trying to standardize wherever possible on the infrastructure, and you're curating the tooling and doing the heavy lifting in the data stack, which of course lessens the burden on the developers and the line of business consumers, so my question is, in operating in 200 countries, how do you deal with all the diversity of laws and regulations across those regions? I know you're heavily involved in AWS, but AWS isn't everywhere, you still have some on-prem infrastructure. Can you paint a picture of, you know, what that looks like? >> Yeah, a few years ago , we were primarily mostly on-prem, and one of the biggest pain points has been managing that infrastructure around the world in those countries. Yes, we operate in 200 countries, but we don't have infrastructure in 200 countries, but we do have agent locations in 200 countries. United Nations says we only have like 183 are countries, but there are countries which, you know, declare themselves countries, and we are there as well because somebody wants to send money there, right? Somebody has an agent location down there as well. So that infrastructure is obviously very hard to manage and maintain. We have to comply by numerous laws, you know. And the last few years, specifically with GDPR, CCPA, data localization laws in different countries, it's been a challenge, right? And one of the things that we did a few years ago, we decided that we want to be in the business of helping our customers move money faster, security, and with complete trust in us. We don't want to be able to, we don't want to be in the business of managing infrastructure. And that's one of the reasons we started to, you know, migrate and move our journey to the cloud. AWS, obviously chosen first because of its, you know, first in the game, has more locations, and more data centers around the world where we operate. But we still have, you know, existing infrastructure, which is in some countries, which is still localized because AWS hasn't reached there, or we don't have a comparable provider there. We still manage those. And we have to comply by those laws. Our data privacy and our data localization tech stack is pretty good, I would say. We manage our data very well, we manage our customer data very well, but it comes with a lot of complexity. You know, we get a lot of requests from European Union, we get a lot of requests from Asia Pacific every pretty much on a weekly basis to explain, you know, how we are taking controls and putting measures in place to make sure that the data is secured and is in the right place. So it's a complex environment. We do have exposure to other clouds as well, like Google and Azure. And as much as we would love to be completely, you know, very, very hybrid kind of an organization, it's still at a stage where we are still very heavily focused on AWS yet, but at some point, you know, we would love to see a world which is not reliant on a single provider, but it's more a little bit more democratized, you know, as and when what I want to use, I should be able to use, and pay-per-use. And the concept started like that, but it's obviously it's now, again, there are like three big players in the market, and, you know, they're doing their own thing. Would love to see them come collaborate at some point. >> Yeah, wouldn't we all. I want to double-click on the whole multi-cloud strategy, but if I understand it correctly, and in a perfect world, everything on-premises would be in the cloud is, first of all, is that a correct statement? Is that nirvana for you or not necessarily? >> I would say it is nirvana for us, but I would also put a caveat, is it's very tricky because from a regulatory perspective, we are a regulated entity in many countries. The regulators would want to see some control if something happens with a relationship with AWS in one country, or with Google in another country, and it keeps happening, right? For example, Russia was a good example where we had to switch things off. We should be able to do that. But if let's say somewhere in Asia, this country decides that they don't want to partner with AWS, and majority of our stuff is on AWS, where do I go from there? So we have to have some level of confidence in our own infrastructure, so we do maintain some to be able to fail back into and move things it needs to be. So it's a tricky question. Yes, it's nirvana state that I don't have to manage infrastructure, but I think it's far less practical than it said. We will still own something that we call it our own where we have complete control, being a financial entity. >> And so do you try to, I'm sure you do, standardize between all the different on-premise, and in this case, the AWS cloud or maybe even other clouds. How do you do that? Do you work with, you know, different vendors at the various places of the stack to try to do that? Some of the vendors, you know, like a Snowflake is only in the cloud. You know, others, you know, whether it's whatever, analytics, or storage, or database, might be hybrid. What's your strategy with regard to creating as common an experience as possible between your on-prem and your clouds? >> You asked a question which I asked when I joined as well, right? Which question, this is one of the most important questions is how soon when I fail back, if I need to fail back? And how quickly can I, because not everything that is sitting on the cloud is comparable to on-prem or is backward compatible. And the reason I say backward compatible is, you know, there are, our on-prem cloud is obviously behind. We haven't taken enough time to kind of put it to a state where, because we started to migrate and now we have access to infrastructure on the cloud, most of the new things are being built there. But for critical application, I would say we have chronology that could be used to move back if need to be. So, you know, technologies like Couchbase, technologies like PostgreSQL, technologies like Db2, et cetera. We still have and maintain a fairly large portion of it on-prem where critical applications could potentially be serviced. We'll give you one example. We use Neo4j very heavily for our AML use cases. And that's an important one because if Neo4j on the cloud goes down, and it's happened in the past, again, even with three clusters, having all three clusters going down with a DR, we still need some accessibility of that because that's one of the biggest, you know, fraud and risk application it supports. So we do still maintain some comparable technology. Snowflake is an odd one. It's obviously there is none on-prem. But then, you know, Snowflake, I also feel it's more analytical based technology, not a transactional-based technology, at least in our ecosystem. So for me to replicate that, yes, it'll probably take time, but I can live with that. But my business will not stop because our transactional applications can potentially move over if need to. >> Yeah, and of course, you know, all these big market cap companies, so the Snowflake or Databricks, which is not public yet, but they've got big aspirations. And so, you know, we've seen things like Snowflake do a deal with Dell for on-prem object store. I think they do the same thing with Pure. And so over time, you see, Mongo, you know, extending its estate. And so over time all these things are coming together. I want to step out of this conversation for a second. I just ask you, given the current macroeconomic climate, what are the priorities? You know, obviously, people are, CIOs are tapping the breaks on spending, we've reported on that, but what is it? Is it security? Is it analytics? Is it modernization of the on-prem stack, which you were saying a little bit behind. Where are the priorities today given the economic headwinds? >> So the most important priority right now is growing the business, I would say. It's a different, I know this is more, this is not a very techy or a tech answer that, you know, you would expect, but it's growing the business. We want to acquire more customers and be able to service them as best needed. So the majority of our investment is going in the space where tech can support that initiative. During our earnings call, we released the new pillars of our organization where we will focus on, you know, omnichannel digital experience, and then one experience for customer, whether it's retail, whether it's digital. We want to open up our own experience stores, et cetera. So we are investing in technology where it's going to support those pillars. But the spend is in a way that we are obviously taking away from the things that do not support those. So it's, I would say it's flat for us. We are not like in heavily investing or aggressively increasing our tech budget, but it's more like, hey, switch this off because it doesn't make us money, but now switch this on because this is going to support what we can do with money, right? So that's kind of where we are heading towards. So it's not not driven by technology, but it's driven by business and how it supports our customers and our ability to compete in the market. >> You know, I think Harvir, that's consistent with what we heard in some other work that we've done, our ETR partner who does these types of surveys. We're hearing the same thing, is that, you know, we might not be spending on modernizing our on-prem stack. Yeah, we want to get to the cloud at some point and modernize that. But if it supports revenue, you know, we'll invest in that, and get the, you know, instant ROI. I want to ask you about, you know, this concept of supercloud, this abstracted layer of value on top of hyperscale infrastructure, and maybe on-prem. But we were talking about the integration, for instance, between Snowflake and Salesforce, where you got different data sources and you were explaining that you had great interest in being able to, you know, have a kind of, I'll say seamless, sorry, I know it's an overused word, but integration between the data sources and those two different platforms. Can you explain that and why that's attractive to you? >> Yeah, I'm a big supporter of action where the data is, right? Because the minute you start to move, things are already lost in translation. The time is lost, you can't get to it fast enough. So if, for example, for us, Snowflake, Salesforce, is our actionable platform where we action, we send marketing campaigns, we send customer communication via SMS, in app, as well as via email. Now, we would like to be able to interact with our customers pretty much on a, I would say near real time, but the concept of real time doesn't work well with me because I always feel that if you're observing something, it's not real time, it's already happened. But how soon can I react? That's the question. And given that I have to move that data all the way from our, let's say, engagement platforms like Adobe, and particles of the world into Snowflake first, and then do my modeling in some way, and be able to then put it back into Salesforce, it takes time. Yes, you know, I can do it in a few hours, but that few hours makes a lot of difference. Somebody sitting on my website, you know, couldn't find something, walked away, how soon do you think he will lose interest? Three hours, four hours, he'll probably gone, he will never come back. I think if I can react to that as fast as possible without too much data movement, I think that's a lot of good benefit that this kind of integration will bring. Yes, I can potentially take data directly into Salesforce, but I then now have two copies of data, which is, again, something that I'm not a big (indistinct) of. Let's keep the source of the data simple, clean, and a single source. I think this kind of integration will help a lot if the actions can be brought very close to where the data resides. >> Thank you for that. And so, you know, it's funny, we sometimes try to define real time as before you lose the customer, so that's kind of real time. But I want to come back to this idea of governed data sharing. You mentioned some other clouds, a little bit of Azure, a little bit of Google. In a world where, let's say you go more aggressively, and we know that for instance, if you want to use Google's AI tools, you got to use BigQuery. You know, today, anyway, they're not sort of so friendly with Snowflake, maybe different for the AWS, maybe Microsoft's going to be different as well. But in an ideal world, what I'm hearing is you want to keep the data in place. You don't want to move the data. Moving data is expensive, making copies is badness. It's expensive, and it's also, you know, changes the state, right? So you got governance issues. So this idea of supercloud is that you can leave the data in place and actually have a common experience across clouds. Let's just say, let's assume for a minute Google kind of wakes up, my words, not yours, and says, "Hey, maybe, you know what, partnering with a Snowflake or a Databricks is better for our business. It's better for the customers," how would that affect your business and the value that you can bring to your customers? >> Again, I would say that would be the nirvana state that, you know, we want to get to. Because I would say not everyone's perfect. They have great engineers and great products that they're developing, but that's where they compete as well, right? I would like to use the best of breed as much as possible. And I've been a person who has done this in the past as well. I've used, you know, tools to integrate. And the reason why this integration has worked is primarily because sometimes you do pick the best thing for that job. And Google's AI products are definitely doing really well, but, you know, that accessibility, if it's a problem, then I really can't depend on them, right? I would love to move some of that down there, but they have to make it possible for us. Azure is doing really, really good at investing, so I think they're a little bit more and more closer to getting to that state, and I know seeking our attention than Google at this point of time. But I think there will be a revelation moment because more and more people that I talk to like myself, they're also talking about the same thing. I'd like to be able to use Google's AdSense, I would like to be able to use Google's advertising platform, but you know what? I already have all this data, why do I need to move it? Can't they just go and access it? That question will keep haunting them (indistinct). >> You know, I think, obviously, Microsoft has always known, you know, understood ecosystems. I mean, AWS is nailing it, when you go to re:Invent, it's all about the ecosystem. And they think they realized they can make a lot more money, you know, together, than trying to have, and Google's got to figure that out. I think Google thinks, "All right, hey, we got to have the best tech." And that tech, they do have the great tech, and that's our competitive advantage. They got to wake up to the ecosystem and what's happening in the field and the go-to-market. I want to ask you about how you see data and cloud evolving in the future. You mentioned that things that are driving revenue are the priorities, and maybe you're already doing this today, but my question is, do you see a day when companies like yours are increasingly offering data and software services? You've been around for a long time as a company, you've got, you know, first party data, you've got proprietary knowledge, and maybe tooling that you've developed, and you're becoming more, you're already a technology company. Do you see someday pointing that at customers, or again, maybe you're doing it already, or is that not practical in your view? >> So data monetization has always been on the charts. The reason why it hasn't seen the light is regulatory pressure at this point of time. We are partnering up with certain agencies, again, you know, some pilots are happening to see the value of that and be able to offer that. But I think, you know, eventually, we'll get to a state where our, because we are trying to build accessible financial services, we will be in a state that we will be offering those to partners, which could then extended to their customers as well. So we are definitely exploring that. We are definitely exploring how to enrich our data with other data, and be able to complete a super set of data that can be used. Because frankly speaking, the data that we have is very interesting. We have trends of people migrating, we have trends of people migrating within the US, right? So if a new, let's say there's a new, like, I'll give you an example. Let's say New York City, I can tell you, at any given point of time, with my data, what is, you know, a dominant population in that area from migrant perspective. And if I see a change in that data, I can tell you where that is moving towards. I think it's going to be very interesting. We're a little bit, obviously, sometimes, you know, you're scared of sharing too much detail because there's too much data. So, but at the end of the day, I think at some point, we'll get to a state where we are confident that the data can be used for good. One simple example is, you know, pharmacies. They would love to get, you know, we've been talking to CVS and we are talking to Walgreens, and trying to figure out, if they would get access to this kind of data demographic information, what could they do be better? Because, you know, from a gene pool perspective, there are diseases and stuff that are very prevalent in one community versus the other. We could probably equip them with this information to be able to better, you know, let's say, staff their pharmacies or keep better inventory of products that could be used for the population in that area. Similarly, the likes of Walmarts and Krogers, they would like to have more, let's say, ethnic products in their aisles, right? How do you enable that? That data is primarily, I think we are the biggest source of that data. So we do take pride in it, but you know, with caution, we are obviously exploring that as well. >> My last question for you, Harvir, is I'm going to ask you to do a thought exercise. So in that vein, that whole monetization piece, imagine that now, Harvir, you are running a P&L that is going to monetize that data. And my question to you is a there's a business vector and a technology vector. So from a business standpoint, the more distribution channels you have, the better. So running on AWS cloud, partnering with Microsoft, partnering with Google, going to market with them, going to give you more revenue. Okay, so there's a motivation for multi-cloud or supercloud. That's indisputable. But from a technical standpoint, is there an advantage to running on multiple clouds or is that a disadvantage for you? >> It's, I would say it's a disadvantage because if my data is distributed, I have to combine it at some place. So the very first step that we had taken was obviously we brought in Snowflake. The reason, we wanted our analytical data and we want our historical data in the same place. So we are already there and ready to share. And we are actually participating in the data share, but in a private setting at the moment. So we are technically enabled to share, unless there is a significant, I would say, upside to moving that data to another cloud. I don't see any reason because I can enable anyone to come and get it from Snowflake. It's already enabled for us. >> Yeah, or if somehow, magically, several years down the road, some standard developed so you don't have to move the data. Maybe there's a new, Mogli is talking about a new data architecture, and, you know, that's probably years away, but, Harvir, you're an awesome guest. I love having you on, and really appreciate you participating in the program. >> I appreciate it. Thank you, and good luck (indistinct) >> Ah, thank you very much. This is Dave Vellante for John Furrier and the entire Cube community. Keep it right there for more great coverage from Supercloud 2. (uplifting music)

Published Date : Jan 6 2023

SUMMARY :

Harvir, it's good to see you again. a pleasure to talk to you. And the pressure to do so is only growing. and you you have what I would call But we still have, you know, you or not necessarily? that I don't have to Some of the vendors, you and it's happened in the past, And so, you know, we've and our ability to compete in the market. and get the, you know, instant ROI. Because the minute you start to move, and the value that you can that, you know, we want to get to. and cloud evolving in the future. But I think, you know, And my question to you So the very first step that we had taken and really appreciate you I appreciate it. Ah, thank you very much.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

AWSORGANIZATION

0.99+

WalmartsORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

MicrosoftORGANIZATION

0.99+

WalgreensORGANIZATION

0.99+

AsiaLOCATION

0.99+

Dave VellantePERSON

0.99+

HarvirPERSON

0.99+

Three hoursQUANTITY

0.99+

four hoursQUANTITY

0.99+

DellORGANIZATION

0.99+

New York CityLOCATION

0.99+

United NationsORGANIZATION

0.99+

KrogersORGANIZATION

0.99+

USLOCATION

0.99+

oneQUANTITY

0.99+

DatabricksORGANIZATION

0.99+

Western UnionORGANIZATION

0.99+

Harvir SinghPERSON

0.99+

10 yearsQUANTITY

0.99+

two copiesQUANTITY

0.99+

one countryQUANTITY

0.99+

183QUANTITY

0.99+

European UnionORGANIZATION

0.99+

MongoORGANIZATION

0.99+

three big playersQUANTITY

0.99+

first stepQUANTITY

0.99+

SnowflakeTITLE

0.98+

AdSenseTITLE

0.98+

more than 200 countriesQUANTITY

0.98+

todayDATE

0.98+

three clustersQUANTITY

0.98+

SnowflakeORGANIZATION

0.98+

MogliPERSON

0.98+

John FurrierPERSON

0.98+

supercloudORGANIZATION

0.98+

one exampleQUANTITY

0.97+

GDPRTITLE

0.97+

AdobeORGANIZATION

0.97+

SalesforceORGANIZATION

0.97+

200 countriesQUANTITY

0.97+

one experienceQUANTITY

0.96+

Harveer SinghPERSON

0.96+

one communityQUANTITY

0.96+

PureORGANIZATION

0.95+

One simple exampleQUANTITY

0.95+

two different platformsQUANTITY

0.95+

SalesforceTITLE

0.94+

firstQUANTITY

0.94+

CubeORGANIZATION

0.94+

BigQueryTITLE

0.94+

nirvanaLOCATION

0.93+

single sourceQUANTITY

0.93+

Asia PacificLOCATION

0.93+

first onesQUANTITY

0.92+

Matthew Scullion, Matillion & Harveer Singh, Western Union | Snowflake Summit 2022


 

>>Hey everyone. Welcome back to Las Vegas. This is the Cube's live coverage of day. One of snowflake summit 22 fourth annual. We're very happy to be here. A lot of people here, Lisa Martin with Dave Valante, David's always great to be at these events with you, but me. This one is shot out of the cannon from day one, data, data, data, data. That's what you heard of here. First, we have two guests joining us next, please. Welcome Matthew Scalian. Who's an alumni of the cube CEO and founder of Matillion and Jer staying chief data architect and global head of data engineering from Western union. Welcome gentlemen. Thank >>You. Great to be here. >>We're gonna unpack the Western union story in a second. I love that, but Matthew, I wanted to start with you, give the audience who might not be familiar with Matillion an overview, your vision, your differentiators, your joint value statement with snowflake, >>Of course. Well, first of all, thank you for having me on the cube. Again, Matillion S mission is to make the world's data useful, and we do that by providing a technology platform that allows our customers to load transform, synchronize, and orchestrate data on the snowflake data cloud. And on, on the cloud in general, we've been doing that for a number of years. We're co headquartered in the UK and the us, hence my dat accents. And we work with all sorts of companies, commercial scale, large end enterprises, particularly including of course, I'm delighted to say our friends at Western union. So that's why we're here today. >>And we're gonna talk about that in a second, but I wanna understand what's new with the data integration platform from Matillion perspective, lots of stuff coming out, give us an overview. >>Yeah, of course, it's been a really busy year and it's great to be here at snowflake summit to be able to share some of what we've been working on. You know, the Matillion platform is all about making our customers as productive as possible in terms of time to value insight on that analytics, data science, AI projects, like get you to value faster. And so the more technology we can put in the platform and the easier we can make it to use, the better we can achieve that goal. So this year we've, we've shipped a product that we call MDL 2.0, that's enterprise focused, exquisitely, easy to use batch data pipelines. So customers can load data even more simply into the snowflake data cloud, very excitingly we've also launched Matillion CDC. And so this is an industry first cloud native writer, head log based change data capture. >>I haven't come up with a shorter way of saying that, but, and surprise customers need this technology and it's been around for years, but mostly pre-cloud technology. That's been repurposed for the cloud. And so Matillion has rebuilt that concept for the cloud. And we launched that earlier this year. And of course we've continued to build out the core Matillion ETL platform that today over a thousand joint snowflake Matillion customers use, including Western union, of course we've been adding features there such as universal connectivity. And so a challenge that all data integration vendors have is having the right connectors for their source systems. Universal connectivity allows you to connect to any source system without writing code point and click. We shape that as well. So it's been a busy year, >>Was really simple. Sorry. I love that. He said that and it also sounded great with your accent. I didn't wanna >>Thank you. Excellent. Javier, talk about your role at Western union in, in what you've seen in terms of the evolution of the, the data stack. >>So in the last few years, well, a little bit of Western union, a 70 or 170 year old company, pretty much everybody knows what Western union is, right? Driving an interesting synergy from what Matthew says, when data moves money moves, that's what we do when he moves the da, he moves the data. We move the money. That's the synergy between, you know, us and the organization that support us from data move perspective. So what I've seen in the last few years is obviously a shift towards the cloud, but, you know, within the cloud itself, obviously there's a lot of players as well. And we as customers have always been wishing to have a short, smaller footprint of data so that the movement becomes a little lesser. You know, interestingly enough, in this conference, I've heard some very interesting stuff, which kind of helping me to bring that footprint down to a manageable number, to be more governed, to be more, you know, effective in terms of delivering more end results for my customers as well. >>So Matillion has been a great partner for us from our cloud adoption perspective. During the COVID times, we were a re we are a, you know, multi-channel organization. We have retail stores as well, our digital presence, but people just couldn't go to the retail stores. So we had to find ways to accelerate our adoption, make sure our systems are scaling and making sure that we are delivering the same experience to our customers. And that's where, you know, tools like Matillion came in and really, really partnered up with us to kind of bring it up to the level. >>So talk specifically about the stack evolution. Cause I have this sort of theory that everybody talks about injecting data and, and machine intelligence and AI and machine learning into apps. But the application development stack is like totally separate from the, the data analytics and the data pipeline stack. And the database is somewhere over here as well. How is that evolving? Are those worlds coming together? >>Some part of those words are coming together, but where I still see the difference is your heavy lifting will still happen on the data stack. You cannot have that heavy lifting on the app because if once the apps becomes heavy, you'll have trouble communicating with, with, with the organizations. You know, you need to be as lean as possible in the front end and make sure things are curated. Things are available on demand as soon as possible. And that's why you see all these API driven applications are doing really, really well because they're delivering those results back to the, the leaner applications much faster. So I'm a big proponent of, yes, it can be hybrid, but the majority of the heavy lifting still needs to happen down at the data layer, which is where I think snowflake plays a really good role >>In APIs are the connective tissue >>APIs connections. Yes. >>Also I think, you know, in terms of the, the data stack, there's another parallel that you can draw from applications, right? So technology is when they're new, we tend to do things in a granular way. We write a lot of code. We do a lot of sticking of things together with plasters and sticky tape. And it's the purview of high end engineers and people enthusiastic about that to get started. Then the business starts to see the value in this stuff, and we need to move a lot faster. And technology solutions come in and this is what the, the data cloud is all about, right? The technology getting out of the way and allowing people to focus on higher order problems of innovating around analytics, data applications, AI, machine learning, you know, that's also where Matillion sit as well as other companies in this modern enterprise data stack is technology vendors are coming in allowing organizations to move faster and have high levels of productivity. So I think that's a good parallel to application development. >>And's just follow up on that. When you think about data prep and you know, all the focus on data quality, you've got a data team, you know, in the data pipeline, a very specialized, maybe even hyper specialized data engineers, quality engineers, data, quality engineers, data analysts, data scientist, but they, and they serve a lot of different business lines. They don't necessarily have the business, they don't have the business context typically. So it's kind of this back and forth. Do you see that changing in your organization or, or the are the lines of business taking more responsibility for the data and, and addressing that problem? It's, >>It's like you die by thousand paper cuts or you just die. Right? That's the kind >>Of, right, >>Because if I say it's, it's good to be federated, it comes with its own flaws. But if I say, if it's good to be decentralized, then I'm the, the guy to choke, right? And in my role, I'm the guy to choke. So I've selectively tried to be a pseudo federated organization, where do I do have folks reporting into our organization, but they sit close to the line of business because the business understands data better. We are working with them hand in glove. We have dedicated teams that support them. And our problem is we are also regional. We are 200 countries. So the regional needs are very different than our us needs. Majority of the organizations that you probably end up talking to have like very us focused, 50 per more than 50% of our revenue is international. So we do, we are dealing with people who are international, their needs for data, their needs for quality and their needs for the, the delivery of those analytics and the data is completely different. And so we have to be a little bit more closer to the business than traditionally. Some, some organizations feel that they need >>To, is there need for the underlying infrastructure and the operational details that as diverse, or is that something that you bring standardization to the, >>So the best part about this, the cloud that happened to us is exactly that, because at one point of time, I had infrastructure in one country. I had another infrastructure sitting in another country, regional teams, making different different decisions of bringing in different tools. Now I can standardize. I will say, Matillion is our standard for doing ETL work. If this is the use case, but then it gets deployed across the geographies because the cloud helps us or the cloud platform helps us to manage it. Sitting down here. I have three centers around the world, you know, Costa Rica, India, and the us. I can manage 24 7 sitting here. No >>Problem. So the underlying our infrastructure is, is global, but the data needs are dealt with locally. Yep. >>One of the pav question, I was just thinking JVE is super well positioned funds for you, which is around that business domain knowledge versus technical expertise. Cause again, early in technology journeys tend, things tend to be very technical and therefore only high end engineers can do it, but high end engineers are scar. Right? Right. And, and also, I mean, we survey our hundreds of large enterprise customers and they tell us they spend two thirds of their time doing stuff they don't really want to do like reinventing the wheel, basic data movement and the low order staff. And so if you can make those people more productive and allow them to focus on higher value problems, but also bring pseudo technical people into it. Overall, the business can go a lot faster. And the way you do that is by making it easier. That's why Matillion is a low code NOCO platform, but Jer and Western union are doing this right. I >>Mean, I can't compete with AWS and Google to hire people. So I need to find people who are smart to figure the products that we have to make them work. I don't want them to spend time on infrastructure, Adam, I don't want them to spend time on trying to manage platforms. I want them to deliver the data, deliver the results to the business so that they can build and serve their customers better. So it's a little bit of a different approach, different mindset. I used to be in consulting for 17 years. I thought I knew it all, but it changed overnight when I own all of these systems. And I'm like, I need to be a little bit more smarter than this. I need to be more proactive and figure out what my business needs rather than what just from a technology needs. It's more what the business needs and how I can deliver that needs to them. So simple analogy, you know, I can build the best architecture in the world. It's gonna cost me an arm and leg, but I can't drive it because the pipeline is not there. So I can have a Ferrari, but I can't drive it. It's still capped at 80, 80 miles an hour. So rather than spend, rather than building one Ferrari, let me have 10 Toyotas or 10 Fs, which will go further along and do better for my cus my, for my customers. >>So how do you see this whole, we hearing about the data cloud. We hear about the marketplace, data products now, application development inside the data cloud. How do you see that affecting not so much the productivity of the data teams. I don't wanna necessarily say, but the product, the value that, that customers like you can get out >>Data. So data is moving closer to the business. That's the value I see, because you are injecting the business and you're injecting the application much more closer to the data because it, in the past, it was days and days of, you know, churn the data to actually clear results. Now the data has moved much closer. So I have a much faster turnaround time. The business can adapt and actually react much, much faster. It took us like 16 to 30 days to deliver, you know, data for marketing. Now I can turn it down in four hours. If I see something happening, I'll give you an example. The war in Ukraine happened. Let is shut down operations in Russia. Ukraine is cash swamp. There's no cash in Ukraine. We have cash. We roll out campaign, $0 money, transferred to Ukraine within four hours of the world going on. That's the impact that we have >>Massive impact. That's huge, especially with such a macro challenge going on, on the, in, in the world. Thank you so much for sharing the Matillion snowflake partnership story, how it's helping Western union really transform into a data company. We love hearing stories of organizations that are 170 years old that have always really been technology focused, but to see it come to life so quickly is pretty powerful. Guys. Thank you so much for your time. Thanks >>Guys. Thank you, having it. Thank >>You >>For Dave Velante and our guests. I'm Lisa Martin. You're watching the cubes live coverage of snowflake summit 22 live from Las Vegas. Stick around. We'll be back after a short break.

Published Date : Jun 14 2022

SUMMARY :

Who's an alumni of the cube give the audience who might not be familiar with Matillion an overview, your vision, And on, on the cloud in general, we've been doing that for a number of And we're gonna talk about that in a second, but I wanna understand what's new with the data integration platform from Matillion And so the more technology we can put in the platform and the easier we can make it to use, And so Matillion has rebuilt that concept for the cloud. He said that and it also sounded great with your accent. in what you've seen in terms of the evolution of the, the data stack. That's the synergy between, you know, us and the organization that support us from data move perspective. are delivering the same experience to our customers. So talk specifically about the stack evolution. but the majority of the heavy lifting still needs to happen down at the data layer, Then the business starts to see the value or the are the lines of business taking more responsibility for the data and, That's the kind And in my role, I'm the guy to choke. So the best part about this, the cloud that happened to us is exactly that, So the underlying our infrastructure is, is global, And the way you do that is by making it easier. the data, deliver the results to the business so that they can build and serve their customers but the product, the value that, that customers like you can get out it, in the past, it was days and days of, you know, churn the data to actually clear in, in the world. Thank For Dave Velante and our guests.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MatthewPERSON

0.99+

Dave ValantePERSON

0.99+

Lisa MartinPERSON

0.99+

Dave VelantePERSON

0.99+

Matthew ScalianPERSON

0.99+

JavierPERSON

0.99+

80QUANTITY

0.99+

DavidPERSON

0.99+

Las VegasLOCATION

0.99+

16QUANTITY

0.99+

MatillionORGANIZATION

0.99+

UkraineLOCATION

0.99+

$0QUANTITY

0.99+

AWSORGANIZATION

0.99+

Western unionORGANIZATION

0.99+

70QUANTITY

0.99+

UKLOCATION

0.99+

17 yearsQUANTITY

0.99+

IndiaLOCATION

0.99+

GoogleORGANIZATION

0.99+

two guestsQUANTITY

0.99+

Matthew ScullionPERSON

0.99+

AdamPERSON

0.99+

FirstQUANTITY

0.99+

RussiaLOCATION

0.99+

four hoursQUANTITY

0.99+

30 daysQUANTITY

0.99+

Costa RicaLOCATION

0.99+

FerrariORGANIZATION

0.99+

MatillionPERSON

0.99+

todayDATE

0.99+

OneQUANTITY

0.99+

Western UnionLOCATION

0.98+

JerORGANIZATION

0.98+

firstQUANTITY

0.98+

one countryQUANTITY

0.97+

two thirdsQUANTITY

0.97+

earlier this yearDATE

0.97+

oneQUANTITY

0.97+

JVEORGANIZATION

0.96+

200 countriesQUANTITY

0.96+

one pointQUANTITY

0.96+

10QUANTITY

0.96+

Harveer SinghPERSON

0.95+

three centersQUANTITY

0.95+

this yearDATE

0.95+

day oneQUANTITY

0.94+

Snowflake Summit 2022EVENT

0.92+

170 years oldQUANTITY

0.91+

170 year oldQUANTITY

0.91+

Western unionLOCATION

0.9+

50 perQUANTITY

0.9+

22QUANTITY

0.86+

80 miles an hourQUANTITY

0.85+

thousand paper cutsQUANTITY

0.84+

UkraiLOCATION

0.83+

fourth annualQUANTITY

0.83+

over a thousand joint snowflakeQUANTITY

0.82+

more than 50%QUANTITY

0.8+

CubeORGANIZATION

0.76+

Matillion CDCORGANIZATION

0.76+

lastDATE

0.75+

hundreds of large enterprise customersQUANTITY

0.74+

JerPERSON

0.73+

24 7QUANTITY

0.72+

MDL 2.0OTHER

0.71+

yearsDATE

0.71+

WesternORGANIZATION

0.65+

cubeORGANIZATION

0.64+

10COMMERCIAL_ITEM

0.62+

secondQUANTITY

0.61+

ToyotasORGANIZATION

0.58+

Phil Bullinger, Western Digital | CUBE Conversation, August 2020


 

>> Announcer: From theCUBE Studios in Palo Alto and Boston, connecting with thought leaders all around the world, this is a Cube conversation. >> Hey welcome back everybody, Jeff Frick here with theCUBE. We are in our Palo Alto studios, COVID is still going on, so all of the interviews continue to be remote, but we're excited to have a Cube alumni, he hasn't been on for a long time, and this guy has been in the weeds of the storage industry for a very very long time and we're happy to have him on and get an update because there continues to be a lot of exciting developments. He's Phil Bullinger, he is the SVP and general manager, data center business unit from Western Digital joining us, I think for Colorado, so Phil, great to see you, how's the weather in Colorado today? >> Hi Jeff, it's great to be here. Well, it's a hot, dry summer here, I'm sure like a lot of places. But yeah, enjoying the summer through these unusual times. >> It is unusual times, but fortunately there's great things like the internet and heavy duty compute and store out there so we can get together this way. So let's jump into it. You've been in the he business a long time, you've been at Western Digital, you were at EMC, you worked on Isilon, and you were at storage companies before that. And you've seen kind of this never-ending up and to the right slope that we see kind of ad nauseum in terms of the amount of storage demands. It's not going anywhere but up, and please increase complexity in terms of unstructure data, sources of data, speed of data, you know the kind of classic big V's of big data. So I wonder, before we jump into specifics, if you can kind of share your perspective 'cause you've been kind of sitting in the Catford seat, and Western Digital's a really unique company; you not only have solutions, but you also have media that feeds other people solutions. So you guys are really seeing and ultimately all this compute's got to put this data somewhere, and a whole lot of it's sitting on Western Digital. >> Yeah, it's a great intro there. Yeah, it's been interesting, through my career, I've seen a lot of advances in storage technology. Speeds and feeds like we often say, but the advancement through mechanical innovation, electrical innovation, chemistry, physics, just the relentless growth of data has been driven in many ways by the relentless acceleration and innovation of our ability to store that data, and that's been a very virtuous cycle through what, for me, has been 30 years in enterprise storage. There are some really interesting changes going on though I think. If you think about it, in a relatively short amount of time, data has gone from this artifact of our digital lives to the very engine that's driving the global economy. Our jobs, our relationships, our health, our security, they all kind of depend on data now, and for most companies, kind of irrespective of size, how you use data, how you store it, how you monetize it, how you use it to make better decisions to improve products and services, it becomes not just a matter of whether your company's going to thrive or not, but in many industries, it's almost an existential question; is your company going to be around in the future, and it depends on how well you're using data. So this drive to capitalize on the value of data is pretty significant. >> It's a really interesting topic, we've had a number of conversations around trying to get a book value of data, if you will, and I think there's a lot of conversations, whether it's accounting kind of way, or finance, or kind of good will of how do you value this data? But I think we see it intrinsically in a lot of the big companies that are really data based, like the Facebooks and the Amazons and the Netflixes and the Googles, and those types of companies where it's really easy to see, and if you see the valuation that they have, compared to their book value of assets, it's really baked into there. So it's fundamental to going forward, and then we have this thing called COVID hit, which I'm sure you've seen all the memes on social media. What drove your digital transformation, the CEO, the CMO, the board, or COVID-19? And it became this light switch moment where your opportunities to think about it are no more; you've got to jump in with both feet, and it's really interesting to your point that it's the ability to store this and think about it now differently as an asset driving business value versus a cost that IT has to accommodate to put this stuff somewhere, so it's a really different kind of a mind shift and really changes the investment equation for companies like Western Digital about how people should invest in higher performance and higher capacity and more unified and kind of democratizing the accessibility that data, to a much greater set of people with tools that can now start making much more business line and in-line decisions than just the data scientist kind of on Mahogany Row. >> Yeah, as you mentioned, Jeff, here at Western Digital, we have such a unique kind of perch in the industry to see all the dynamics in the OEM space and the hyperscale space and the channel, really across all the global economies about this growth of data. I have worked at several companies and have been familiar with what I would have called big data projects and fleets in the past. But at Western Digital, you have to move the decimal point quite a few digits to the right to get the perspective that we have on just the volume of data that the world has just relentless insatiably consuming. Just a couple examples, for our drive projects we're working on now, our capacity enterprise drive projects, you know, we used to do business case analysis and look at their lifecycle capacities and we measured them in exabytes, and not anymore, now we're talking about zettabyte, we're actually measuring capacity enterprise drive families in terms of how many zettabyte they're going to ship in their lifecycle. If we look at just the consumption of this data, the last 12 months of industry TAM for capacity enterprise compared to the 12 months prior to that, that annual growth rate was north of 60%. And so it's rare to see industries that are growing at that pace. And so the world is just consuming immense amounts of data, and as you mentioned, the COVID dynamics have been both an accelerant in some areas, as well as headwinds in others, but it's certainly accelerated digital transformation. I think a lot of companies we're talking about, digital transformation and hybrid models and COVID has really accelerated that, and it's certainly driving, continues to drive just this relentless need to store and access and take advantage of data. >> Yeah, well Phil, in advance of this interview, I pulled up the old chart with all the different bytes, kilobytes, megabytes, gigabytes, terabytes, petabytes, exabytes, and zettabytes, and just per the Wikipedia page, what is a zettabyte? It's as much information as there are grains of sand in all the world's beaches. For one zettabyte. You're talking about thinking in terms of those units, I mean, that is just mind boggling to think that that is the scale in which we're operating. >> It's really hard to get your head wrapped around a zettabyte of storage, and I think a lot of the industry thinks when we say zettabyte scale era, that it's just a buzz word, but I'm here to say it's a real thing. We're measuring projects in terms of zettabytes now. >> That's amazing. Well, let's jump into some of the technology. So I've been fortunate enough here at theCUBE to be there at a couple of major announcements along the way. We talked before we turned the cameras on, the helium announcement and having the hard drive sit in the fish bowl to get all types of interesting benefits from this less dense air that is helium versus oxygen. I was down at the Mammer and Hammer announcement, which was pretty interesting; big heavy technology moves there, to again, increase the capacity of the hard drive's base systems. You guys are doing a lot of stuff on RISC-V I know is an Open source project, so you guys have a lot of things happening, but now there's this new thing, this new thing called zonedd storage. So first off, before we get into it, why do we need zoned storage, and really what does it now bring to the table in terms of a capability? >> Yeah, great question, Jeff. So why now, right? Because I mentioned storage, I've been in storage for quite some time. In the last, let's just say in the last decade, we've seen the advent of the hyperscale model and certainly a whole nother explosion level of data and just the veracity with which they hyperscalers can create and consume and process and monetize data. And of course with that, has also come a lot of innovation, frankly, in the compute space around how to process that data and moving from what was just a general purpose CPU model to GPU's and DPU's and so we've seen a lot of innovation on that side, but frankly, in the storage side, we haven't seen much change at all in terms of how operating systems, applications, file systems, how they actually use the storage or communicate with the storage. And sure, we've seen advances in storage capacities; hard drives have gone from two to four, to eight, to 10 to 14, 16, and now our leading 18 and 20 terabyte hard drives. And similarly, on the SSD side, now we're dealing with the capacities of seven, and 15, and 30 terabytes. So things have gotten larger, as you expect. And some interfaces have improved, I think NVME, which we'll talk about, has been a nice advance in the industry; it's really now brought a very modern scalable low latency multi-threaded interface to a NAM flash to take advantage of the inherent performance of transistor based persistent storage. But really when you think about it, it hasn't changed a lot. But what has changed is workloads. One thing that definitely has evolved in the space of the last decade or so is this, the thing that's driving a lot of this explosion of data in the industry is around workloads that I would characterize as sequential in nature, they're serial, you can capture it in written. They also have a very consistent life cycle, so you would write them in a big chunk, you would read them maybe in smaller pieces, but the lifecycle of that data, we can treat more as a chunk of data, but the problem is applications, operating systems, vial systems continue to interface with storage using paradigms that are many decades old. The old 512 byte or even Forte, Sector size constructs were developed in the hard drive industry just as convenient paradigms to structure what is an unstructured sea of magnetic grains into something structured that can be used to store and access data. But the reality is when we talk about SSDs, structure really matters, and so what has changed in the industry is the workloads are driving very very fresh looks at how more intelligence can be applied to that application OS storage device interface to drive much greater efficiency. >> Right, so there's two things going on here that I want to drill down on. On one hand, you talked about kind of the introduction of NAND and flash, and treating it like you did, generically you did a regular hard drive. But you could get away and you could do some things because the interface wasn't taking full advantage of the speed that was capable in the NAND. But NVME has changed that, and now forced kind of getting rid of some of those inefficient processes that you could live with, so it's just kind of classic next level step up and capabilities. One is you get the better media, you just kind of plug it into the old way. Now actually you're starting to put in processes that take full advantage of the speed that that flash has. And I think obviously prices have come down dramatically since the first introduction, where before it was always kind of a clustered off or super high end, super low latency, super high value apps, it just continues to spread and proliferate throughout the data center. So what did NVME force you to think about in terms of maximizing the return on the NAND and flash? >> Yeah, NVME, which we've been involved in the standardization, I think it's been a very successful effort, but we have to remember NVME is about a decade old, or even more when the original work started around defining this interface, but it's been very successful. The NVME standard's body is very productive cross company effort, it's really driven a significant change, and what we see now is the rapid adoption of NVME in all of data center architectures, whether it's very large hyperscale to classic on prem enterprise to even smaller applications, it's just a very efficient interface mechanism for connecting SSDs into a server. So we continue to see evolution at NVME, which is great, and we'll talk about ZNS today as one of those evolutions. We're also very keenly interested in NVME protocol over fabrics, and so one of the things that Western Digital has been talking about a lot lately is incorporating NVME over fabrics as a mechanism for now connecting shared storage into multiple post architectures. We think this is a very attractive way to build shared storage architectures of the future that are scalable, that are composable, that really have a lot more agility with respect to rack level infrastructure and applying that infrastructure to applications. >> Right, now one thing that might strike some people as kind of counterintuitive is within the zoned storage in zoning off parts of the media, to think of the data also kind of in these big chunks, is it feels contrary to kind of atomization that we're seeing in the rest of the data center, right? So smaller units of compute, smaller units of store, so that you can assemble and disassemble them in different quantities as needed. So what was the special attribute that you had to think about and actually come back and provide a benefit in actually kind of re-chunking, if you will, in these zones versus trying to get as atomic as possible? >> Yeah, it's a great question, Jeff, and I think it's maybe not intuitive in terms of why zoned storage actually creates a more efficient storage paradigm when you're storing stuff essentially in larger blocks of data, but this is really where the intersection of structure and workload and sort of the nature of the data all come together. If you turn back the clock maybe four or five years when SMR hard drives host managers SMR hard drives first emerged on the scene. This was really taking advantage of the fact that the right head on a hard disk drive is larger than the read head, or the read head can be much smaller, and so the notion of overlapping or shingling the data on the drive, giving the read head a smaller target to read, but the writer a larger write pad to write the data could actually, what we found was it increases aerial density significantly. And so that was really the emergence of this notion of sequentially written larger blocks of data being actually much more efficiently stored when you think about physically how it's being stored. What's very new now and really gaining a lot of traction is the SSD corollary to SMR on the hard drive, on the SSD side, we had the ZNS specification, which is, very similarly where you'd divide up the name space of an SSD into fixed size zones, and those zones are written sequentially, but now those zones are intimately tied to the underlying physical architecture of the NAND itself; the dyes, the planes, the read pages, the erase pages. So that, in treating data as a block, you're actually eliminating a lot of the complexity and the work that an SSD has to do to emulate a legacy hard drive, and in doing so, you're increasing performance and endurance and the predictable performance of the device. >> I just love the way that you kind of twist the lens on the problem, and on one hand, by rule, just looking at my notes here, the zoned storage device is the ZSD's introduce a number of restrictions and limitations and rules that are outside the full capabilities of what you might do. But in doing so, an aggregate, the efficiency, and the performance of the system in the whole is much much better, even though when you first look at it, you think it's more of a limiter, but it's actually opens up. I wonder if there's any kind of performance stats you can share or any kind of empirical data just to give people kind of a feel for what that comes out as. >> So if you think about the potential of zoned storage in general and again, when I talk about zoned storage, there's two components; there's an HDD component of zoned storage that we refer to as SMR, and there's an SSD version of that that we call ZNS. So we think about SMR, the value proposition there is additional capacity. So effectively in the same drive architecture, with roughly the same bill of material used to build the drive, we can overlap or shingle the data on the drive. And generally for the customer, additional capacity. Today with our 18, 20 terabyte offerings that's on the order of just over 10%, but that delta is going to increase significantly going forward to 20% or more. And when you think about a hyperscale customer that has not hundreds or thousands of racks, but tens of thousands of racks. A 10 or 20% improvement in effective capacity is a tremendous TCO benefit, and the reason we do that is obvious. I mean, the economic paradigm that drives large at-scale data centers is total custom ownership, both acquisition costs and operating costs. And if you can put more storage in a square tile of data center space, you're going to generally use less power, you're going to run it more efficiently, you're actually, from an acquisition cost, you're getting a more efficient purchase of that capacity. And in doing that, our innovation, we benefit from it and our customers benefit from it. So the value proposition for zoned storage in capacity enterprise HDV is very clear, it's additional capacity. The exciting thing is, in the SSD side of things, or ZNS, it actually opens up even more value proposition for the customer. Because SSDs have had to emulate hard drives, there's been a lot of inefficiency and complexity inside an enterprise SSD dealing with things like garbage collection and right amplification reducing the endurance of the device. You have to over-provision, you have to insert as much as 20, 25, even 28% additional man bits inside the device just to allow for that extra space, that working space to deal with delete of data that are smaller than the block erase that the device supports. So you have to do a lot of reading and writing of data and cleaning up. It creates for a very complex environment. ZNS by mapping the zoned size with the physical structure of the SSD essentially eliminates garbage collection, it reduces over-provisioning by as much as 10x. And so if you were over provisioning by 20 or 25% on an enterprise SSD, and a ZNS SSD, that can be one or two percent. The other thing I have to keep in mind is enterprise SSD is typically incorporate D RAM and that D RAM is used to help manage all those dynamics that I just mentioned, but with a much simpler structure where the pointers to the data can be managed without all the D RAM. We can actually reduce the amount of D RAM in an enterprise SSD by as much as eight X. And if you think about the MILA material of an enterprise SSD, D RAM is number two on the list in terms of the most expensive bomb components. So ZNS and SSDs actually have a significant customer total cost of ownership impact. It's an exciting standard, and now that we have the standard ratified through the NVME working group, it can really accelerate the development of the software ecosystem around. >> Right, so let's shift gears and talk a little bit about less about the tech and more about the customers and the implementation of this. So you talked kind of generally, but are there certain types of workloads that you're seeing in the marketplace where this is a better fit or is it just really the big heavy lifts where they just need more and this is better? And then secondly, within these hyperscale companies, as well as just regular enterprises that are also seeing their data demands grow dramatically, are you seeing that this is a solution that they want to bring in for kind of the marginal kind of next data center, extension of their data center, or their next cloud region? Or are they doing lift and shift and ripping stuff out? Or do they enough data growth organically that there's plenty of new stuff that they can put in these new systems? >> Yeah, I love that. The large customers don't rip and shift; they ride their assets for a long lifecycle, 'cause with the relentless growth of data, you're primarily investing to handle what's coming in over the transom. But we're seeing solid adoption. And in SMRS you know we've been working on that for a number of years. We've got significant interest and investment, co-investment, our engineering, and our customer's engineering adapting the application environment's to take advantage of SMR. The great thing is now that we've got the NVME, the ZNS standard gratified now in the NVME working group, we've got a very similar, and all approved now, situation where we've got SMR standards that have been approved for some time, and the SATA and SCSI standards. Now we've got the same thing in the NVME standard, and the great thing is once a company goes through the lift, so to speak, to adapt an application, file system, operating system, ecosystem, to zoned storage, it pretty much works seamlessly between HDD and SSD, and so it's not an incremental investment when you're switching technologies. Obviously the early adopters of these technologies are going to be the large companies who design their own infrastructure, who have mega fleets of racks of infrastructure where these efficiencies really really make a difference in terms of how they can monetize that data, how they compete against the landscape of competitors they have. For companies that are totally reliant on kind of off the shelf standard applications, that adoption curve is going to be longer, of course, because there are some software changes that you need to adapt to enable zoned storage. One of the things Western Digital has done and taken the lead on is creating a landing page for the industry with zoned storage.io. It's a webpage that's actually an area where many companies can contribute Open source tools, code, validation environments, technical documentation. It's not a marketeering website, it's really a website built to land actual Open source content that companies can use and leverage and contribute to to accelerate the engineering work to adapt software stacks to zoned storage devices, and to share those things. >> Let me just follow up on that 'cause, again, you've been around for a while, and get your perspective on the power of Open source. And it used to be the best secrets, the best IP were closely guarded and held inside, and now really we're in an age where it's not necessarily. And the brilliant minds and use cases and people out there, just by definition, it's more groups of engineers, more engineers outside your building than inside your building, and how that's really changed kind of a strategy in terms of development when you can leverage Open source. >> Yeah, Open source clearly has accelerated innovation across the industry in so many ways, and it's the paradigm around which companies have built business models and innovated on top of it, I think it's always important as a company to understand what value ad you're bringing, and what value ad the customers want to pay for. What unmet needs in your customers are you trying to solve for, and what's the best mechanism to do that? And do you want to spend your RND recreating things, or leveraging what's available and innovating on top of it? It's all about ecosystem. I mean, the days where a single company could vertically integrate top to bottom a complete end solution, you know, those are fewer and far between. I think it's about collaboration and building ecosystems and operating within those. >> Yeah, it's such an interesting change, and one more thing, again, to get your perspective, you run the data center group, but there's this little thing happening out there that we see growing, IOT, in the industrial internet of things, and edge computing as we try to move more compute and store and power kind of outside the pristine world of the data center and out towards where this data is being collected and processed when you've got latency issues and all kinds of reasons to start to shift the balance of where the compute is and where the store and relies on the network. So when you look back from the storage perspective in your history in this industry and you start to see basically everything is now going to be connected, generating data, and a lot of it is even Opensource. I talked to somebody the other day doing kind of Opensource computer vision on surveillance video. So the amount of stuff coming off of these machines is growing in crazy ways. At the same time, it can't all be processed at the data center, it can't all be kind of shipped back and then have a decision and then ship that information back out to. So when you sit back and look at Edge from your kind of historical perspective, what goes through your mind, what gets you excited, what are some opportunities that you see that maybe the laymen is not paying close enough attention to? >> Yeah, it's really an exciting time in storage. I get asked that question from time to time, having been in storage for more than 30 years, you know, what was the most interesting time? And there's been a lot of them, but I wouldn't trade today's environment for any other in terms of just the velocity with which data is evolving and how it's being used and where it's being used. A TCO equation may describe what a data center looks like, but data locality will determine where it's located, and we're excited about the Edge opportunity. We see that as a pretty significant, meaningful part of the TAM as we look three to five years. Certainly 5G is driving much of that, I think just any time you speed up the speed of the connected fabric, you're going to increase storage and increase the processing the data. So the Edge opportunity is very interesting to us. We think a lot of it is driven by low latency work loads, so the concept of NVME is very appropriate for that, we think, in general SSDs deployed and Edge data centers defined as anywhere from a meter to a few kilometers from the source of the data. We think that's going to be a very strong paradigm. The workloads you mentioned, especially IOT, just machine-generated data in general, now I believe, has eclipsed human generated data, in terms of just the amount of data stored, and so we think that curve is just going to keep going in terms of machine generated data. Much of that data is so well suited for zoned storage because it's sequential, it's sequentially written, it's captured, and it has a very consistent and homogenous lifecycle associated with it. So we think what's going on with zoned storage in general and ZNS and SMR specifically are well suited for where a lot of the data growth is happening. And certainly we're going to see a lot of that at the Edge. >> Well, Phil, it's always great to talk to somebody who's been in the same industry for 30 years and is excited about today and the future. And as excited as they have been throughout their whole careers. So that really bodes well for you, bodes well for Western Digital, and we'll just keep hoping the smart people that you guys have over there, keep working on the software and the physics, and the mechanical engineering and keep moving this stuff along. It's really just amazing and just relentless. >> Yeah, it is relentless. What's exciting to me in particular, Jeff, is we've driven storage advancements largely through, as I said, a number of engineering disciplines, and those are still going to be important going forward, the chemistry, the physics, the electrical, the hardware capabilities. But I think as widely recognized in the industry, it's a diminishing curve. I mean, the amount of energy, the amount of engineering effort, investment, that cost and complexity of these products to get to that next capacity step is getting more difficult, not less. And so things like zoned storage, where we now bring intelligent data placement to this paradigm, is what I think makes this current juncture that we're at very exciting. >> Right, right, well, it's applied AI, right? Ultimately you're going to have more and more compute power driving the storage process and how that stuff is managed. As more cycles become available and they're cheaper, and ultimately compute gets cheaper and cheaper, as you said, you guys just keep finding new ways to move the curve in. And we didn't even get into the totally new material science, which is also coming down the pike at some point in time. >> Yeah, very exciting times. >> It's been great to catch up with you, I really enjoy the Western Digital story; I've been fortunate to sit in on a couple chapters, so again, congrats to you and we'll continue to watch and look forward to our next update. Hopefully it won't be another four years. >> Okay, thanks Jeff, I really appreciate the time. >> All right, thanks a lot. All right, he's Phil, I'm Jeff, you're watching theCUBE. Thanks for watching, we'll see you next time.

Published Date : Aug 25 2020

SUMMARY :

all around the world, this so all of the interviews Hi Jeff, it's great to be here. in terms of the amount of storage demands. be around in the future, that it's the ability to store this and the channel, really across and just per the Wikipedia and I think a lot of the and having the hard drive of data and just the veracity with which kind of the introduction and so one of the things of the data center, right? and so the notion of I just love the way that you kind of and the reason we do that is obvious. and the implementation of this. and the great thing is And the brilliant minds and use cases and it's the paradigm around which and all kinds of reasons to start to shift and increase the processing the data. and the mechanical engineering I mean, the amount of energy, driving the storage process I really enjoy the Western Digital story; really appreciate the time. we'll see you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

Western DigitalORGANIZATION

0.99+

Phil BullingerPERSON

0.99+

Western DigitalORGANIZATION

0.99+

ColoradoLOCATION

0.99+

oneQUANTITY

0.99+

20QUANTITY

0.99+

PhilPERSON

0.99+

August 2020DATE

0.99+

AmazonsORGANIZATION

0.99+

30 yearsQUANTITY

0.99+

both feetQUANTITY

0.99+

NetflixesORGANIZATION

0.99+

18QUANTITY

0.99+

threeQUANTITY

0.99+

two percentQUANTITY

0.99+

BostonLOCATION

0.99+

twoQUANTITY

0.99+

Palo AltoLOCATION

0.99+

FacebooksORGANIZATION

0.99+

15QUANTITY

0.99+

25%QUANTITY

0.99+

20%QUANTITY

0.99+

fourQUANTITY

0.99+

10QUANTITY

0.99+

GooglesORGANIZATION

0.99+

hundredsQUANTITY

0.99+

28%QUANTITY

0.99+

TodayDATE

0.99+

eightQUANTITY

0.99+

theCUBEORGANIZATION

0.99+

14QUANTITY

0.99+

five yearsQUANTITY

0.99+

30 terabytesQUANTITY

0.99+

EMCORGANIZATION

0.99+

COVID-19OTHER

0.99+

10xQUANTITY

0.99+

more than 30 yearsQUANTITY

0.99+

sevenQUANTITY

0.99+

CubeORGANIZATION

0.99+

two componentsQUANTITY

0.99+

OpensourceORGANIZATION

0.99+

two thingsQUANTITY

0.99+

four yearsQUANTITY

0.99+

IsilonORGANIZATION

0.99+

firstQUANTITY

0.98+

25QUANTITY

0.98+

20 terabyteQUANTITY

0.98+

bothQUANTITY

0.98+

18, 20 terabyteQUANTITY

0.98+

16QUANTITY

0.98+

over 10%QUANTITY

0.98+

COVIDOTHER

0.97+

tens of thousands of racksQUANTITY

0.97+

first introductionQUANTITY

0.97+

todayDATE

0.96+

TAMORGANIZATION

0.96+

theCUBE StudiosORGANIZATION

0.96+

NVMEORGANIZATION

0.95+

last decadeDATE

0.94+

EdgeORGANIZATION

0.94+

Mammer and HammerORGANIZATION

0.93+

OneQUANTITY

0.92+

COVIDORGANIZATION

0.92+

Janet George, Western Digital | WiDS 2019


 

>> Live from Stanford University. It's the Cube covering global Women in Data Science conference brought to you by Silicon Angle media. >> Welcome back to the key. We air live at Stanford University for the fourth annual Women in Data Science Conference. The Cube has had the pleasure of being here all four years on I'm welcoming Back to the Cube, one of our distinguished alumni Janet George, the fellow chief data officer, scientists, big data and cognitive computing at Western Digital. Janet, it's great to see you. Thank you. Thank you so much. So I mentioned yes. Fourth, Annie will women in data science. And it's been, I think I met you here a couple of years ago, and we look at the impact. It had a chance to speak with Margo Garrett's in a about an hour ago, one of the co founders of Woods saying, We're expecting twenty thousand people to be engaging today with the Livestream. There are wigs events in one hundred and fifty locations this year, fifty plus countries expecting about one hundred thousand people to engage the attention. The focus that they have on data science and the opportunities that it has is really palpable. Tell us a little bit about Western Digital's continued sponsorship and what makes this important to you? >> So Western distal has recently transformed itself as a company, and we are a data driven company, so we are very much data infrastructure company, and I think that this momentum off A is phenomenal. It's just it's a foundational shift in the way we do business, and this foundational shift is just gaining tremendous momentum. Businesses are realizing that they're going to be in two categories the have and have not. And in order to be in the half category, you have started to embrace a You've got to start to embrace data. You've got to start to embrace scale and you've got to be in the transformation process. You have to transform yourself to put yourself in a competitive position. And that's why Vest Initial is here, where the leaders in storage worldwide and we'd like to be at the heart of their data is. >> So how has Western Digital transform? Because if we look at the evolution of a I and I know you're give you're on a panel tan, you're also giving a breakout on deep learning. But some of the importance it's not just the technical expertise. There's other really important skills. Communication, collaboration, empathy. How has Western digital transformed to really, I guess, maybe transform the human capital to be able to really become broad enough to be ableto tow harness. Aye, aye, for good. >> So we're not just a company that focuses on business for a We're doing a number of initiatives One of the initiatives were doing is a I for good, and we're doing data for good. This is related to working with the U. N. We've been focusing on trying to figure out how climate change the data that impacts climate change, collecting data and providing infrastructure to store massive amounts of species data in the environment that we've never actually collected before. So climate change is a huge area for us. Education is a huge area for us. Diversity is a huge area for us. We're using all of these areas as launching pad for data for good and trying to use data to better mankind and use a eye to better mankind. >> One of the things that is going on at this year's with second annual data fun. And when you talk about data for good, I think this year's Predictive Analytics Challenge was to look at satellite imagery to train the model to evaluate which images air likely tohave oil palm plantations. And we know that there's a tremendous social impact that palm oil and oil palm plantations in that can can impact, such as I think in Borneo and eighty percent reduction in the Oregon ten population. So it's interesting that they're also taking this opportunity to look at data for good. And how can they look at predictive Analytics to understand how to reduce deforestation like you talked about climate and the impact in the potential that a I and data for good have is astronomical? >> That's right. We could not build predictive models. We didn't have the data to put predictive boats predictive models. Now we have the data to put put out massively predictive models that can help us understand what change would look like twenty five years from now and then take corrective action. So we know carbon emissions are causing very significant damage to our environment. And there's something we can do about it. Data is helping us do that. We have the infrastructure, economies of scale. We can build massive platforms that can store this data, and then we can. Alan, it's the state at scale. We have enough technology now to adapt to our ecosystem, to look at disappearing grillers, you know, to look at disappearing insects, to look at just equal system that be living, how, how the ecosystem is going to survive and be better in the next ten years. There's a >> tremendous amount of power that data for good has, when often times whether the Cube is that technology conferences or events like this. The word trust issues yes, a lot in some pretty significant ways. And we often hear that data is not just the life blood of an organization, whether it's in just industry or academia. To have that trust is essential without it. That's right. No, go. >> That's right. So the data we have to be able to be discriminated. That's where the trust comes into factor, right? Because you can create a very good eh? I'm odder, or you can create a bad air more so a lot depends on who is creating the modern. The authorship of the model the creator of the modern is pretty significant to what the model actually does. Now we're getting a lot of this new area ofthe eyes coming in, which is the adversarial neural networks. And these areas are really just springing up because it can be creators to stop and block bad that's being done in the world next. So, for example, if you have malicious attacks on your website or hear militias, data collection on that data is being used against you. These adversarial networks and had built the trust in the data and in the so that is a whole new effort that has started in the latest world, which is >> critical because you mentioned everybody. I think, regardless of what generation you're in that's on. The planet today is aware of cybersecurity issues, whether it's H vac systems with DDOS attacks or it's ah baby boomer, who was part of the fifty million Facebook users whose data was used without their knowledge. It's becoming, I won't say accepted, but very much commonplace, Yes, so training the A I to be used for good is one thing. But I'm curious in terms of the potential that individuals have. What are your thoughts on some of these practices or concepts that we're hearing about data scientists taking something like a Hippocratic oath to start owning accountability for the data that they're working with. I'm just curious. What's >> more, I have a strong opinion on this because I think that data scientists are hugely responsible for what they are creating. We need a diversity of data scientists to have multiple models that are completely divorce, and we have to be very responsible when we start to create. Creators are by default, have to be responsible for their creation. Now where we get into tricky areas off, then you are the human auto or the creator ofthe Anay I model. And now the marshal has self created because it a self learned who owns the patent, who owns the copyright to those when I becomes the creator and whether it's malicious or non malicious right. And that's also ownership for the data scientist. So the group of people that are responsible for creating the environment, creating the morals the question comes into how do we protect the authors, the uses, the producers and the new creators off the original piece of art? Because at the end of the day, when you think about algorithms and I, it's just art its creation and you can use the creation for good or bad. And as the creation recreates itself like a learning on its own with massive amounts of data after an original data scientist has created the model well, how we how to be a confident. So that's a very interesting area that we haven't even touched upon because now the laws have to change. Policies have to change, but we can't stop innovation. Innovation has to go, and at the same time we have to be responsible about what we innovate >> and where do you think we are? Is a society in terms of catching As you mentioned, we can't. We have to continue innovation. Where are we A society and society and starting to understand the different principles of practices that have to be implemented in order for proper management of data, too. Enable innovation to continue at the pace that it needs. >> June. I would say that UK and other countries that kind of better than us, US is still catching up. But we're having great conversations. This is very important, right? We're debating the issues. We're coming together as a community. We're having so many discussions with experts. I'm sitting in so many panels contributing as an Aye aye expert in what we're creating. What? We see its scale when we deploy an aye aye, modern in production. What have we seen as the longevity of that? A marker in a business setting in a non business setting. How does the I perform and were now able to see sustained performance of the model? So let's say you deploy and am are in production. You're able inform yourself watching the sustained performance of that a model and how it is behaving, how it is learning how it's growing, what is its track record. And this knowledge is to come back and be part of discussions and part of being informed so we can change the regulations and be prepared for where this is going. Otherwise will be surprised. And I think that we have started a lot of discussions. The community's air coming together. The experts are coming together. So this is very good news. >> Theologian is's there? The moment of Edward is building. These conversations are happening. >> Yes, and policy makers are actively participating. This is very good for us because we don't want innovators to innovate without the participation of policymakers. We want the policymakers hand in hand with the innovators to lead the charter. So we have the checks and balances in place, and we feel safe because safety is so important. We need psychological safety for anything we do even to have a conversation. We need psychological safety. So imagine having a >> I >> systems run our lives without having that psychological safety. That's bad news for all of us, right? And so we really need to focus on the trust. And we need to focus on our ability to trust the data or a right to help us trust the data or surface the issues that are causing the trust. >> Janet, what a pleasure to have you back on the Cube. I wish we had more time to keep talking, but it's I can't wait till we talk to you next year because what you guys are doing and also your pact, true passion for data science for trust and a I for good is palpable. So thank you so much for carving out some time to stop by the program. Thank you. It's my pleasure. We want to thank you for watching the Cuba and Lisa Martin live at Stanford for the fourth annual Women in Data Science conference. We back after a short break.

Published Date : Mar 4 2019

SUMMARY :

global Women in Data Science conference brought to you by Silicon Angle media. We air live at Stanford University for the fourth annual Women And in order to be in the half category, you have started to embrace a You've got to start Because if we look at the evolution of a initiatives One of the initiatives were doing is a I for good, and we're doing data for good. So it's interesting that they're also taking this opportunity to We didn't have the data to put predictive And we often hear that data is not just the life blood of an organization, So the data we have to be able to be discriminated. But I'm curious in terms of the creating the morals the question comes into how do we protect the We have to continue innovation. And this knowledge is to come back and be part of discussions and part of being informed so we The moment of Edward is building. We need psychological safety for anything we do even to have a conversation. And so we really need to focus on the trust. I can't wait till we talk to you next year because what you guys are doing and also your pact,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Janet GeorgePERSON

0.99+

JanetPERSON

0.99+

AlanPERSON

0.99+

BorneoLOCATION

0.99+

next yearDATE

0.99+

fifty millionQUANTITY

0.99+

Western DigitalORGANIZATION

0.99+

Lisa MartinPERSON

0.99+

OregonLOCATION

0.99+

twenty thousand peopleQUANTITY

0.99+

JuneDATE

0.99+

Silicon AngleORGANIZATION

0.99+

eighty percentQUANTITY

0.99+

two categoriesQUANTITY

0.99+

AnniePERSON

0.99+

Stanford UniversityORGANIZATION

0.99+

Western distalORGANIZATION

0.99+

fifty plus countriesQUANTITY

0.98+

Vest InitialORGANIZATION

0.98+

oneQUANTITY

0.98+

this yearDATE

0.98+

OneQUANTITY

0.97+

Women in Data ScienceEVENT

0.97+

second annualQUANTITY

0.96+

FacebookORGANIZATION

0.96+

todayDATE

0.96+

CubeORGANIZATION

0.95+

StanfordLOCATION

0.95+

Western digitalORGANIZATION

0.94+

Women in Data Science ConferenceEVENT

0.93+

about one hundred thousand peopleQUANTITY

0.92+

one hundred and fifty locationsQUANTITY

0.92+

FourthQUANTITY

0.91+

EdwardPERSON

0.9+

USORGANIZATION

0.89+

Women in Data Science conferenceEVENT

0.88+

ten populationQUANTITY

0.88+

couple of years agoDATE

0.85+

WiDS 2019EVENT

0.85+

one thingQUANTITY

0.85+

CubaLOCATION

0.85+

Margo GarrettPERSON

0.84+

about an hour agoDATE

0.82+

U. N.LOCATION

0.82+

twenty five yearsQUANTITY

0.81+

LivestreamORGANIZATION

0.77+

next ten yearsDATE

0.73+

fourth annualEVENT

0.69+

annualQUANTITY

0.65+

halfQUANTITY

0.62+

fourthEVENT

0.6+

WoodsORGANIZATION

0.59+

fourQUANTITY

0.58+

UKLOCATION

0.58+

wigsQUANTITY

0.56+

CubeCOMMERCIAL_ITEM

0.52+

AnayPERSON

0.31+

Christopher Bergey, Western Digital | Autotech Council 2018


 

>> Announcer: From Milpitas, California at the edge of Silicon Valley, it's The CUBE. Covering autonomous vehicles. Brought to you by Western Digital. >> Hey, welcome back everybody. Jeff Frick here with The Cube. We are at the Autotech Council Autonomous Vehicle event at Western Digital. Part of our Data Makes Possible Program with Western Digital where we're looking at all these cool applications and a lot of cutting edge technology that at the end of the day, it's data dependent and data's got to sit somewhere. But really what's interesting here is that the data, and more and more of the data is moving out to the edge and edge computing and nowhere is that more apparent than in autonomous vesicles so we're really excited to have maybe the best title at Western Digital, I don't know. Chris Bergey, VP of Product Marketing. That's not so special, but all the areas that he's involved with: mobile, compute, automotive, connected homes, smart cities, and if that wasn't enough, industrial IOT. Chris, you must be a busy guy. >> Hey, we're having a lot of fun here. This data world is an exciting place to be right now. >> so we're her at the Autonomous Vehicle event. We could talk about smart cities, which is pretty interesting, actually ties to it and internet of things and industrial internets, but what are some of the really unique challenges in autonomous vehicles that most people probably aren't thinking of? >> Well, I think that we all understand that really, autonomous vehicles are being made possible by just the immense amount of sensors that are being put into the car. Not much different than as our smartphones or our phones evolved from really not having a lot of sensors to today's smartphones have many, many sensors. Whether it's sensing your face, gyroscopes, GPS, all these kind of things. The car is having the exact thing happen but many, many more sensors. And, of course, those sensors just drive a tremendous amount of data and then it's really about trying to pull the intelligence out of that data and that's really what the whole artificial intelligence or autonomous is really trying to do is, okay, we've got all this data, how do I understand what's happening in the autonomous vehicle in a very short period of time? >> Right, and there's two really big factors that you've talked about and some of the other things that you've done. I did some homework and one of them is the metadata around the data, so there's the raw data itself that's coming off those sensors, but the metadata is a whole nother level, and a big level, and even more importantly is the context. What is the context of that data and without context, it's just data. It's not really intelligence or smarts or things you can do anything about so that baseline sensor data gets amplified significantly in terms of actually doing anything with that information. >> That's correct. I think one of the examples I give that's easier for people to understand is surveillance, right? We're very familiar with walking into a retail store where there's surveillance cameras and they're recording in the case that maybe there's a theft or something goes wrong, but there's so much data there that's not acutely being processed, right? How may people walked into the store? What was the average time a person came to the store? How many men? How many women? That's the context of the data and that's what's really would be very valuable if you were, say, an owner of the store or a regional manager. So that's really pulling the context out of the raw data. And in the car example, autonomous vehicles, hey, there's going to be something, my sensors are seeing something, and then, of course, you'd use multiple sensors. That's the sensor fusion between them of, "Hey, that's a person, that's a deer, oh, don't worry, "that's a car moving alongside of us and he's "staying in his lane." Those are the types of decisions we're making with this data and that's the context. >> Right, and even they had in the earlier presentation today the reflection of the car off the side of a bus, I mean, these are the nuance things that aren't necessarily obvious when you first start exploring. >> And we're dealing with human life, I mean, so obviously it needs to be right 99.999 plus percent. So that's the challenge, right? It's the corner cases and I think that's what we see with autonomous vehicles. It's really exciting to see the developments going on and, of course, there's been a couple challenges, but we just have so much learning to do to really get to that fifth nine or whatever it is from a probability point of view. And that's where we'll continue to work on those corner cases, but the technology is coming along so fast, it's just mind-boggling how quickly we are starting to attack these more difficult challenges. And we'll get there but it's going to take time like anything. >> The other really important thing, especially now where we're in the rise of Cloud, if you will. Amazon is going bananas. Google Cloud Platform, Microsoft Azure, so we're seeing this huge move of Cloud and enterprise IT. But in a car, right, there's this little thing called latency and this other thing called physics where you've got a real issue when you have to make a quick decision based on data and those sensors when something jumps out in front of the car. So really, the rise of edge computing and moving so much of that stored compute and intelligence into the vehicle and then deciding what goes back to the car to retrain the algorithm. So it's really a shift to back out to the edge, if you will, dependent because of this latency issue. >> Yeah, I mean, they're very complimentary, right? But there's a lot of decisions you can make locally and, obviously, there's a lot of advantages in doing that. Latency being one of them, but just cost of communications and again, what people don't necessarily understand is how big this data is. You see statistics thrown out there, one gigabit per second, two gigabits per second. I mean, that is just massive data. At the end of the day, actually, in some of the development, it's pretty interesting that we have the car developers actually FedExing the terabyte drives that they've captured data because it's the easiest way for them to actually transfer the data. I mean, people think, "Oh, internet connectivity, no problem." You try to ship 80 terabytes in a cost effective manner, FedEx ends up being the best shot right now. So it's pretty interesting. >> The old sneaker, that is pretty funny. But the quantities of this data are so big. I was teasing you on Twitter earlier today. I think we took it up to an xobyte, a zedobyte, a yodabyte, and then the crowd responded. No, it's a brontosaurousbyte is even bigger than a yodabyte. We were at Flink Forward earlier this week and really this whole idea of stream processing, it's really taking new approaches to data processing. You'll be able to take all that stuff in in real time, which probably state of the market now is financial trading and advertising markets. But to do that now in a car where if you make a mistake, there's really significant consequences. It's a really different challenge. >> It is and again, that's really this advent of the sensor data, right? The sensor data is going to swamp probably every other data set that's in the world, but a lot of it's not interesting because you don't know when that interesting event is going to happen. So what you actually find is that you try to put it's intelligence as close as you can to the data, end storage, and again, storage may be 30 seconds to if you had an accident, you want to be able to go back 30 seconds. It may be lifetimes. So just thinking about these data flows and what's the half life of the data relative to the value? But what we're actually finding with many of the machine learning is that data we thought was not valuable, data we thought, "Oh, we have the right amount of granularity," now with machine learning we're going back and saying, "Oh, why didn't we record at an even higher granularity?" We could have pulled out more of these trends or more of these corner cases. So I think that's one of the challenges enterprise are going through right now is that everyone's so scared of getting rid of any data, yet there's just tremendous data growth. And we're sitting right here in the middle of it at Western Digital. >> Well, thankfully for you guys, you're going to store all that data and it is really important, though, because it used to be, it's funny to me. It used to be a sample of things that happened in the past is how you would make your decisions. Now it's not a sample, it's all of what's happening now and hopefully you can make a decision while you still have time to have an impact. So it's a very different world but sampling is going away when, in theory, you don't know what you're going to need that data for and you have the ability to store it. >> Making real-time decisions but then also learning how to use that decision to make better decisions in the future. That's really where Silicon Valley's focused right now. >> All right, Chris, well you're a busy guy so we're going to let you get back to it because you also have to do IOT and industrial internet and mobile an compute. So thanks for taking ... >> And I try to eat in between there too. >> And you try to eat and hopefully see your kids Friday night, so hopefully you'll take >> Absolutely. your wife out to a movie tonight. >> All right, Chris, great to see you. Thanks for taking a few minutes. >> Chris: Thank you very much. >> All right, I'm Jeff Frick. You're watching The CUBE from Autotech Council Autonomous Vehicle event. Thanks for watching.

Published Date : Apr 14 2018

SUMMARY :

Brought to you by Western Digital. and more and more of the data is moving out to the edge Hey, we're having a lot of fun here. and internet of things and industrial internets, that are being put into the car. and a big level, and even more importantly is the context. So that's really pulling the context out of the raw data. necessarily obvious when you first start exploring. I mean, so obviously it needs to be right So it's really a shift to back out to the edge, captured data because it's the easiest way for them But to do that now in a car where if you make a mistake, of the sensor data, right? and hopefully you can make a decision while you still Making real-time decisions but then also learning how to so we're going to let you get back to it And I try to eat your wife out to a movie tonight. All right, Chris, great to see you. All right, I'm Jeff Frick.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jeff FrickPERSON

0.99+

ChrisPERSON

0.99+

Chris BergeyPERSON

0.99+

Western DigitalORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

30 secondsQUANTITY

0.99+

Christopher BergeyPERSON

0.99+

Silicon ValleyLOCATION

0.99+

oneQUANTITY

0.99+

Friday nightDATE

0.99+

FedExORGANIZATION

0.99+

80 terabytesQUANTITY

0.99+

Flink ForwardORGANIZATION

0.99+

todayDATE

0.99+

tonightDATE

0.98+

MicrosoftORGANIZATION

0.98+

Milpitas, CaliforniaLOCATION

0.97+

two gigabits per secondQUANTITY

0.95+

FedExingORGANIZATION

0.95+

Autotech Council Autonomous VehicleEVENT

0.95+

fifth nineQUANTITY

0.93+

one gigabit per secondQUANTITY

0.93+

TwitterORGANIZATION

0.93+

firstQUANTITY

0.92+

earlier this weekDATE

0.92+

earlier todayDATE

0.91+

99.999 plus percentQUANTITY

0.91+

two really big factorsQUANTITY

0.9+

Autotech CouncilEVENT

0.84+

Cloud PlatformTITLE

0.82+

Autonomous VehicleEVENT

0.79+

The CubeORGANIZATION

0.76+

zedobyteCOMMERCIAL_ITEM

0.69+

GoogleORGANIZATION

0.67+

coupleQUANTITY

0.63+

Autotech CouncilORGANIZATION

0.57+

decDATE

0.48+

2018DATE

0.47+

yodabyteCOMMERCIAL_ITEM

0.46+

AzureTITLE

0.46+

CUBETITLE

0.43+

yodabyteOTHER

0.3+

Oded Sagee, Western Digital | Autotech Council 2018


 

>> Announcer: From Milpitas, California at the edge of Silicon Valley, it's theCUBE, covering autonomous vehicles. Brought to you by Western Digital. >> Hey, welcome back everybody. Jeff Frick, here with theCUBE. We're in Milpitas, California, at Western Digital, at the Autotech Council Autonomous Vehicle Event. About 300 people, really deep into this space. It's a developing ecosystem. You know, we think about Tesla, that's kind of got a complete, closed system. But there's a whole ecosystem of other companies getting into the autonomous vehicle space, and as was mentioned in the keynote, there are, literally thousands of problems. A great opportunity for startups. So we're excited to have Oded Sagee, he's a senior director of product marketing from Western Digital. Oded, great to see you. >> Thank you very much, Jeff. >> So you were just on the panel and, really that was a big topic, is there are thousands of problems to solve and this ecosystem's trying to come together, but it's complicated, right? It's not just the big car manufacturers anymore, and the tier one providers, but there's this whole ecosystem that's now growing up to try to solve these problems. So what are you seeing from your point of view? >> Yes, correct. So, definitely in the past automotive was a tough market to play in, but it was simple from the amount of players and people you needed to talk to to design your product inside. With the disruption of connectivity, smart vehicles, even before autonomous, there are so many new systems in the car now that generate data or consume data. And so, for us, to kind of figure out what's the use case, right? How is this going to look in the future? Who's going to define it? Who's going to buy it? Who's going to pay for it? It has become more and more complex. Happily, storage is in the center of all this. >> Jeff: Right. >> So we get a seat at the table and everyone wants to talk to us, but yes, it's a very big ecosystem now. And trying to resolve that problem, it's going to take some time. >> So what are some of the unique characteristics, from a storage point of view, that you have to worry about? Obviously environmental jumps out. We had the guy on before talking about bumpy roads, you know, the huge impacts on vibration. And now you spent a lot of money for a Toughbook back in the day to put a laptop in a cop car, this is a whole other level of expense, investment, and data flow. >> Right. So, for us, I think with all this disruption happening of full autonomous, people are, very much focused on making that autonomous work, right? So, for them it's all about connectivity, it's all about the sensor, whether it's Lidar, or, you know, cameras. Just making that work, right? All the algorithms and the software. And so, for them storage, currently is an afterthought, right? They were saying, once we meet mass production we'll just go and buy some storage and everything's going to be fine. So while they're prototyping, right? They can use any storage that they want. But, if you think about a full autonomous vehicle out there driving, not two hours a day like we are driving today, right? 20 hours a day, from cold to hot, going through areas without connectivity. Suddenly, the storage requirements are very, very different. And this is what we're trying to drive and explain that, if we don't design the future storage solutions today, What's going to end up, is that people are going to pay much more for storage just to make a basic use case work. >> Right. >> But if we start working now, and I'm talking about five, seven years out, we can have affordable solutions to make those business models work. >> And is that resonating in the industry, or are they just too focused on, you know, better cameras? >> It definitely does, but as companies change, right? So let's just take the car makers for a second. They didn't necessarily have a CTO in place, right? To drive engineering and semi-conductor. So you got to find those figures, and you got to start working and educating them. It definitely resonates if you have the right person. Once you find him, yes, it's on the list of priority. So we need to push. But it is happening. Yes, it is resonating. >> And it's so different because you do have this edge case. You have so much data being collected out in the field, if you will, within that vehicle. Some, to go back to the cloud, but you've got latency is always an issue, right? For safety. So, a little different storage challenge. So are there significant design thoughts that you guys are bringing into play on why this is so different and what is it going to take to really have kind of an optimal solution for autonomous vehicles? >> Yes, definitely there are a couple of vectors I would say, or knobs we need to work on. One of them is temperature. So, again vehicles do tend to go between hot and cold. Unlike many other components that just need to make sure that they operate between hot and cold, we actually have a big challenge on keeping data being accurate between hot and cold. So if you program cold and read hot and vice versa, data gets corrupted. >> Oh, even within the structures within the media? >> Yes. >> Okay. >> And people don't know that. So, for us to figure out, what's the temperature range that the car, through its lifetime, is going to go through. And make sure that we meet the use case, that's a big one. What we call the endurance on the cycling of the storage, again, if you cannot rely on connectivity, cannot rely on cloud because of latency, you need to record a lot of data in the car. So, again, a car drives for seven years, 15 years, and you want to record constantly, how much do you need to record? We don't necessarily have the technology today to meet that use case and we need to work with the ecosystem, in figuring it out. So these are just two examples. >> And I would imagine clean power, as you're saying these things, but they can need others. You're not in daddy's data center anymore. This is a pretty harsh environment, I would imagine. >> Very harsh. >> Ugly power, inconsistent power, turning off the car before everything is spun down. There's all kinds of little, kind of environmental impacts in that whole realm that you would never think of in, kind of a typical data center, for instance. >> Correct. And even, you touched power, that's very interesting because even some people think, oh, there's not power limitation in a car. You can just enjoy how much power you want. Actually, it's very, very sensitive. The battery, if you think about an EV car now has so many components to run and so even the power consumption, right? Just the energy that you need to consume is becoming critical for each, and every component >> in the vehicle. >> Right. And it's everybody's AI comparison, right? Is if Kasparov had to fight the computer with the same amount of power, it wouldn't have been much of a match. So the power to run all this AI stuff is not insignificant, so it is going to be a huge drain on these electric vehicles. Pretty exciting times. So when you get up in the morning, what's the biggest thing, when you talk to people about autonomous vehicles, that they just don't get? That people should really be thinking about. >> Yeah, so it goes back to some of the things we've discussed. Definitely, again, we're seeing the use cases change. We are working again with the broad ecosystem to explain the fundamental challenges that we have, right? What is our design cycle? What are the challenges that we have? So we start with educating the ecosystem, so they know what we have. And from that we trigger a discussion because they realize, oh, okay, because I do have a use case that, probably, you don't have a solution for, how do we go together? And we're doing it across the board. It's not only happening in automotive. It's happening in surveillance. It's happening in the home space. A lot of people don't know, but the home space, if you think about it, again, set-top boxes used to be huge, sat outside in the room. People are moving to these sticks, right? And they're behind the TV and they have no ventilation and they're small and they record all the time. And they get to temperatures that we've never seen in the past. So we even need to educate the telcos of the world, the set-top box makers. Everything is changing. Automotive is definitely ahead in a lot of innovation and disruption, but everything is changing for us. >> Right, a lot of those are fond of just the bright shiny object that everybody can see, right? We can't necessarily see a lot of IOT that GE's putting in to connect their factories. Alright, Oded, well thanks for taking a few minutes out of your busy day and I really appreciate the insight. >> Thank you very much. >> All right, he's Oded, I'm Jeff, You're watching theCUBE from Western Digital at The Autonomous Vehicle Event for the Autotech Council. Thanks for watching. Catch you next time. (electronic music)

Published Date : Apr 14 2018

SUMMARY :

Brought to you by Western Digital. at the Autotech Council Autonomous Vehicle Event. So what are you seeing from your point of view? and people you needed to talk to So we get a seat at the table that you have to worry about? is that people are going to pay much more for storage just to make those business models work. So you got to find those figures, And it's so different because you do have this edge case. So if you program cold and read hot and vice versa, And make sure that we meet the use case, And I would imagine clean power, that you would never think of in, Just the energy that you need to consume So the power to run all this AI stuff but the home space, if you think about it, again, and I really appreciate the insight. at The Autonomous Vehicle Event for the Autotech Council.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

OdedPERSON

0.99+

Western DigitalORGANIZATION

0.99+

Pat GelsingerPERSON

0.99+

seven yearsQUANTITY

0.99+

15 yearsQUANTITY

0.99+

Oded SageePERSON

0.99+

KasparovPERSON

0.99+

Silicon ValleyLOCATION

0.99+

TeslaORGANIZATION

0.99+

DellORGANIZATION

0.99+

two examplesQUANTITY

0.99+

GEORGANIZATION

0.99+

FirstQUANTITY

0.99+

Milpitas, CaliforniaLOCATION

0.98+

Dell TechnologiesORGANIZATION

0.98+

two hours a dayQUANTITY

0.98+

20 hours a dayQUANTITY

0.98+

ThreeQUANTITY

0.97+

About 300 peopleQUANTITY

0.97+

theCUBEORGANIZATION

0.96+

VMwareORGANIZATION

0.96+

todayDATE

0.96+

fiveQUANTITY

0.93+

SKO V2COMMERCIAL_ITEM

0.9+

TwoQUANTITY

0.9+

Autotech CouncilORGANIZATION

0.89+

eachQUANTITY

0.88+

2018DATE

0.84+

Autonomous Vehicle EventEVENT

0.83+

The CubeTITLE

0.83+

One ofQUANTITY

0.82+

Autotech Council Autonomous Vehicle EventEVENT

0.82+

thousands of problemsQUANTITY

0.77+

ToughbookORGANIZATION

0.75+

this decadeDATE

0.69+

vectorsQUANTITY

0.62+

coupleQUANTITY

0.59+

a secondQUANTITY

0.56+

moneyQUANTITY

0.56+

lotQUANTITY

0.53+

Autotech CouncilEVENT

0.52+

EMCORGANIZATION

0.5+

oneQUANTITY

0.37+

Dave Tang, Western Digital & Martin Fink, Western Digital l | CUBEConversation Feb 2018


 

(inspirational music) >> Hey, welcome back everybody. Jeff Frick here with theCUBE. We are in our Palo Alto studio. The conference season hasn't really kicked off yet into full swing so we can do a lot more kind of intimate stuff here in the studio, for a CUBE Conversation. And we're really excited to have a many time CUBE alum on, and a new guest, both from Western Digital. So Dave Tang, Senior Vice President at Western Digital. Great to see you again, Dave. >> Great to be here, Jeff. >> Absolutely and Martin Fink, he is the Chief Technology Officer at Western Digital, a longtime HP alum. I'm sure people recognized you from that and our great machine keynotes we were talking about it. So great to finally meet you, Martin. >> Thank you, nice to be here. >> Absolutely, so you guys are here talking about and we've got an ongoing program actually with Western Digital about Data Makes Possible, right. With all the things that are going on in tech at the end of the day, right, there's data, it's got to be stored somewhere and then of course there's processes and things going on. We've been exploring media and entertainment, sports, healthcare, autonomous vehicles, you know. All the places that this continues to reach out and it's such a fun project because you guys are a rising tide, lifts all boats, kind of company and really enjoy watching this whole ecosystem grow. So I really want to thank you for that. But now there's some new things that we want to talk about that you guys are doing to continue really in that same theme, and that's the support of this RISC-V. So first off, for people who have no idea, what is RISC-V? Let's jump into that and then kind of what is the announcement and why it's important. >> Sure, so RISC-V is an, you know, the tagline is, it's an open source instruction set architecture. So what does that mean, just so people can kind of understand. So today the world is dominated by two instruction set architectures. For the most part the, we'll call the desktop enterprise world is dominated by the Intel instruction set architecture and that's what's in most PCs, what people talk about as x86. And then the embedded and mobile space tends to be dominated by Arm, or by Arm Holdings. And so both of those are great architectures but they're also proprietary, they're owned by their respective companies. So RISC-V is essentially a third entrant, we'll say, into this world, but the distinction is that it's completely open source. So everything about the instruction set is available to all and anybody can implement it. We can all share the implementations. We can share the code that makes up that instruction set architecture, and very importantly for us and part of our motivation is the freedom to innovate. So we now have the ability to modify the instruction set or change the implementation of the instruction set, to optimize it for our devices and our storage and our drives, etc. >> So is this the first kind of open source play in microprocessor architecture? >> No, there's been other attempts at this. OpenSpark kind of comes to mind, and things like that, but the ability to get a community of individuals to kind of rally around this in a meaningful way has really been a challenge. And so I'd say that right now, RISC-V presents probably the best sort of clean slate, let's take some thing new to the market out there. >> So open source, obviously we've seen you know, take over the software world, first in the operating system which everybody is familiar with Linux but then we see it time and time again in different applications, Hadoop. I mean, there's just a proliferation of open source projects. The benefits are tremendous. Pretty easy to ascertain at a typical software case, how is that going to be applied do you think within the microprocessor world? >> So it's a little bit different. When we're talking about open source hardware or open source chips and microprocessors, you're dealing with a physical device. So even though you can open source all of the designs and the code associated with that device, you still have to fabricate it. You still have to create a physical design and you still have to call up a fab and say, will you make this for me at these particular volumes? And so that's the difference. So there are some differences between open source software where it's, you know, you create the bits and then you distribute those bits through the Internet and all is good. Whereas here, you still have a physical need to fabricate something. >> Now, how much more flexibility can you do then for the output when you can actually impact the architecture as opposed to just creating a custom chip design, on top of somebody else's architecture? >> Well, let me give you probably a really simple, concrete example that kind of people can internalize of some of our motivation behind this, because that might sort of help get people through this. If you think of a very typical surveillance application, you have a camera pointed into a room or a hallway. The reality is we're basically grabbing a ton of video frames but very few of them change, right? So the typical surveillance application is it never changes and you really want, only know when stuff changes. Well, today, in very simple terms, all of those frames get routed up to some big server somewhere and that server spends a lot of time trying to figure out, okay have I got a frame that changed? Have I got a frame that changed, and so on. And then eventually it'll find maybe two or three or five frames that have got something interesting. So in the world what we're trying to do is to say, okay well why don't we take that, find no changes, and push that right down to the device? So we basically store all those frames, why don't we go figure out all the frames that mean nothing, and only ship up to that big bad server the frames that have something interesting and something you want to go analyze and do some work on? So that's a very typical application that's quite meaningful because we can do all of that work at the device. We can eliminate shipping a whole bunch of data to where it's just going to get discarded anyways, and we can allow the end customer to really focus on the data that matters, and get some intelligence. >> And that's critical as we get more and more immersed in a data-centric world, where we have realtime applications like Martin described as well as large data-centric applications like of course, big data analytics, but also training for AI systems or machine learning. These workloads are going to become more and more diverse and they're going to need more specialized architectures and more specialized processing. So big data is getting bigger and faster and these realtime fast data applications are getting faster and bigger. So we need ways to contend with that, that really go beyond what's available with general purpose architectures. >> So that's a great point because if we take this example of video frames, now if I can build a processor that is customized to only do that, that's the only thing it does. It can be very low power, very efficient, and do that one thing very very well, and the cost adder, if you want to call it that, to the device where we put it, is a tiny fraction, but the cost savings of the overall solution is significant. So this ability to customize the instruction set to only do what you need it to do for that very special purpose, that's gold. >> So I just wanted to, Dave, we've talked about a lot of interesting innovations that you guys have come up with over the years, with the helium launch. Which I don't know, a couple, two, three years ago, you were just at the MAMR event, really energy assisted recording. So this is really kind of foundational within the storage and the media itself and how you guys do better and take advantage of evolving land space. This is a kind of a different play for Western Digital, this isn't a direct kind of improvement in the way that storage media and architecture works but this is really more of, I'm going to ask you. What is the Western Digital play here? Why is this an important space for you guys in your core storage business? >> Well we're really broadening our focus to really develop and innovate around technologies that really help the world extract more value from data as a whole, right. So it's way beyond storage these days, right. We're looking for better ways to capture, preserve, access, and transform the data. And unless you transform it, you can't really extract the value out of it so as we see all these new applications for data and the vast possibilities for data, we really want to pave the path and help the industry innovate to bring all those applications to reality. >> It's interesting too because one of the great topics always in computing is you know, you got compute and store, which has to go to which, right. And nobody wants to move a lot of data, that's hard and may or may not be easy to get compute. Especially these IoT applications, remote devices, tough conditions and power, which we mentioned a little bit before we went on air. So the the landscape for the for the need for compute and store in networking is radically changing than either the desktop or what we're seeing a consolidation in clouds. So what's interesting here, where does the scale come, right? At the end of the day, scale always wins. And that's where we've seen historically where the general-purpose microprocessor architectures is dominated but used to be a slew of specialty purpose architectures but now there's an opportunity to bring scale to this. So how does that scale game continue to evolve? >> So it's a great point that scale does matter and we've seen that repeatedly and so it's a significant part of the reason why we decided to go early with a significant commitment was to tell the world that we were bringing scale to the equation. And so what we communicated to the marketplace is we ship on the order of a billion processor cores a year, most people don't realize that all of our devices from USB sticks to hard drives, all have processors on them. And so we said, hey we're going to basically go all-in and go big and that translates into a billion cores that we ship every year and we're going to go on a program to essentially migrate all of those cores to RISC-V. It'll take a few years to get there but we'll migrate all of those cores and so we basically were signaling to the market, hey scale is now here. Scale is here, you can make the investments, you can go forward, you can make that commitment to RISC-V because essentially we've got your back. >> So just to make sure we get that clear. So you guys have announced that you're going to slowly migrate over time your micro processors that power your devices to the tune of approximately a billion with a B, cores per year to this new architecture. >> That is correct. >> And has that started? >> So the design has started. So we have started to design and develop our first two cores but the actual manifestation into devices probably in the early stage of 2020. >> Okay, okay. But that's a pretty significant commitment and again, the ideas you explicitly said it's a signal to the ecosystem, this is worth your investment because there is some scale here. >> Martin: That's right. >> Yeah, pretty exciting. And how do you think it's going to open up the ability for you to do new things with your devices that you before either couldn't do or we're too expensive with dollars or power. >> Martin: So we're going to step and iterate through this and one key point here is a lot of people tend to want to start in this processor world at the very high end, right. I'm going to go take on a Xeon processor or something like that. It's not what we're doing. We're basically saying, we're going to go at the small end, the tiny end where power matters. Power matters a lot in our devices and where can we achieve the optimum combination of power and performance. So even in our small devices like a USB stick or a client SSD or something like that, if we can reduce power consumption and even just maintain performance that's a huge win for our customers, you know. If you think about your laptop and if I reduce the power consumption of that SSD in there so that you have longer battery life and you can get you know through the day better, that's a huge win, right. And I don't impact performance in the process, that's a huge win. So what we do, what we're doing right now is we're developing the cores based on the RISC-V architecture and then what we're going to do is once we've got that sort of design, sort of complete is we want to take all of the typical client workloads and profile them on that. Then we want to find out, okay where are the hot spots? What are the two or three things that are really consuming all the power and how do we go optimize, by either creating two or three instructions or by optimizing the micro architecture for an existing instruction. And then iterate through that a few times so that we really get a big win, even at the very low end of the spectrum and then we just iterate through that with time. >> We're in a unique position I think in that the technologies that we develop span everything from the actual media where the bits are stored, whether it's solid-state flash or rotating magnetic disk and the recording heads. We take those technologies and build them all the way up into devices and platforms and full-fledged data center systems. And if we can optimize and tune all the way from that core media level all the way up through into the system level, we can deliver significantly higher value, we believe, to the marketplace. So this is the start of that, that enables us to customize command sets and optimize the flow of data so that we can we can allow users to access it when and where they need it. >> So I think there's another actually really cool point, which goes back to the open source nature of this and we try to be very clear about this. We're not going to develop our cores for all applications. We want the world to develop all sorts of different cores. And so for many applications somebody else might come in and say, hey we've got a really cool core. So one of the companies we've partnered with and invested in for example, is Esperanto. They've actually decided to go at the high end and do a machine learning accelerator. Hey, maybe we'll use that for some machine learning applications in our system level performance. So we don't have to do it all but we've got a common architecture across the portfolio and that speaks to that sort of open source nature of the RISC-V architecture is we want the world to get going. We want our competitors to get on board, we want partners, we want software providers, we want everybody on board. >> It's such a different ecosystem with open-source and the way the contributions are made and the way contributions are valued and the way that people can go find niches that are underserved. It's this really interesting kind of bifurcation of the market really, you don't really want to be in the big general-purpose middle anymore. That's not a great place to be, there's all kinds of specialty places where you can build the competence and with software and you know with, thank goodness for Moore's law decreasing prices of the power of the compute and now the cloud, which is basically always available. Really a exciting time to develop a myriad of different applications. >> Right and you talked before about scale in terms of points of implementation that will drive adoption and drive this to critical mass but there's another aspect of scale relative to the architecture within a single system that's also important that I think RISC-V helps to break down some barriers. Because with general purpose computer architectures, they assume a certain ratio of memory and storage and processing and bandwidth for interconnect and if you exceed those ratios, you have to add a whole new processor. Even though you don't need to need the processing capability, you need it for scale. So that's another great benefit of these new architectures is that the diversity of data needs where some are going to be large data sets, some are going to be small data sets that need need high bandwidth. You can customize and blend that recipe as you need to, you're not at the mercy of these fixed ratios. >> Yeah and I think you know it's so much of kind of what is cloud computing. And the atomic nature of it, that you can apply the ratios, the amount that you need as you need, you can change it on the fly, you can tone it up, tone it down. And I think the other interesting thing that you touched on is some of these new, which are now relatively special-purpose but are going to be general-purpose very soon in terms of machine learning and AI and applying those to different places and applying them closer to the problem. It's a very very interesting evolution of the landscape but what I want to do is kind of close on you Martin, especially because again kind of back to the machine. Not the machine specifically but you have been in the business of looking way down the road for a long time. So you came out, I'd looked at your LinkedIn, you retired for three months, congratulations. (laughs) Hope you got some my golf in but you came back to Western Digital so why did you come back? And as you look down the road a ways, what do you see that it excites you, that got you off that three-month little tour around the golf course and I'm sorry I had to tease about that. But what do you see? What are you excited about that you came back and got involved in an open source microprocessor project? >> So the the short answer was that, I saw the opportunity at Western Digital to be where data lives. So I had spent my entire career, will call it at the compute or the server side of things and the interesting thing is I had a very close relationship with SanDisk, which was acquired by Western Digital. And so I had, we'll call it an insider view, of what was possible there and so what triggered was essentially what we're talking about here was given that about half the world's data lands on Western Digital devices, taking that from a real position of strength in the marketplace and say, what could we go do to make data more intelligent and rather than start kind of at that server end and so that I saw that potential there and it was just incredible, so that's that's what made me want to join. >> Exciting times. Dave good get. (laughs) >> We're delighted to have Martin with us. >> All right, well we look forward to watch it evolve. We've got another another whole set of events we're going to do again together with Western Digital that we're excited about. Again, covering Data Makes Possible but you know kind of uplifting into the application space as a lot of the cool things that people are doing in innovation. So Martin, great to finally meet you and thanks for stopping by. >> Thanks for the time. >> David as always and I think we'll see in a month or so. >> Right, always a pleasure Jeff, thanks. >> All right Martin Fink, Dave Tang. I'm Jeff Frick, you're watching theCUBE. Thanks for watching, we'll catch you next time. (inspirational music)

Published Date : Feb 1 2018

SUMMARY :

Great to see you again, Dave. So great to finally meet you, Martin. and that's the support of this RISC-V. So everything about the instruction set is available to all but the ability to get a community of individuals how is that going to be applied do you think and the code associated with that device, and something you want to go analyze and do some work on? and they're going to need more specialized architectures and the cost adder, if you want to call it that, and how you guys do better and the vast possibilities for data, So how does that scale game continue to evolve? and so it's a significant part of the reason why So just to make sure we get that clear. So the design has started. and again, the ideas you explicitly said that you before either couldn't do so that you have longer battery life and and optimize the flow of data and that speaks to that sort of open source nature and with software and you know with, is that the diversity of data needs where the amount that you need as you need, and the interesting thing is I had (laughs) So Martin, great to finally meet you David as always and I think Thanks for watching, we'll catch you next time.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
NutanixORGANIZATION

0.99+

Western DigitalORGANIZATION

0.99+

JohnPERSON

0.99+

DavidPERSON

0.99+

KristaPERSON

0.99+

Bernie HannonPERSON

0.99+

Jeff FrickPERSON

0.99+

BerniePERSON

0.99+

H3CORGANIZATION

0.99+

CitrixORGANIZATION

0.99+

September of 2015DATE

0.99+

Dave TangPERSON

0.99+

Krista SatterthwaitePERSON

0.99+

SanDiskORGANIZATION

0.99+

MartinPERSON

0.99+

James WhitePERSON

0.99+

SuePERSON

0.99+

CiscoORGANIZATION

0.99+

Carol DweckPERSON

0.99+

Martin FinkPERSON

0.99+

JeffPERSON

0.99+

HPEORGANIZATION

0.99+

twoQUANTITY

0.99+

Stu MinimanPERSON

0.99+

DavePERSON

0.99+

Dave allantePERSON

0.99+

John FurrierPERSON

0.99+

RaghuPERSON

0.99+

Raghu NandanPERSON

0.99+

Palo AltoLOCATION

0.99+

threeQUANTITY

0.99+

Lee CaswellPERSON

0.99+

HPORGANIZATION

0.99+

Antonio NeriPERSON

0.99+

five yearsQUANTITY

0.99+

three-monthQUANTITY

0.99+

four-yearQUANTITY

0.99+

one minuteQUANTITY

0.99+

GaryPERSON

0.99+

AntonioPERSON

0.99+

Feb 2018DATE

0.99+

2023DATE

0.99+

seven dollarsQUANTITY

0.99+

three monthsQUANTITY

0.99+

Arm HoldingsORGANIZATION

0.99+

firstQUANTITY

0.99+

John Rydning, IDC | Western Digital the Next Decade of Big Data 2017


 

>> Announcer: Live from San Jose, California, it's theCUBE covering innovating to fuel the next decade of big data. Brought to you by Western Digital. >> Hey, welcome back everybody. Jeff Frick, here with theCUBE. We are at the Western Digital Headquarters in San Jose, California. It's the Al-Mady Campus. A historic campus. It's had a lot of great innovation, especially in hard drives for years and years and years. This event's called Innovating to Fuel the Next Data Big Data. And we're excited to have a big brain on. We like to get smart people who's been watching this story for a while and will give us a little bit of historical perspective. It's John Rydning. He is the Research Vice President for Hard Drives for IEC. John, Welcome. >> Thank you, Jeff. >> Absolutely. So, what is your take on today's announcement? >> I think it's our very meaningful announcement, especially when you consider that the previous BIGIT Technology announcement for the industry was Helium, about four or five years ago. But, really, the last big technology announcement prior to that was back in 2005, 2006, when the industry announced making this transition to what they called at that time, "Perpendicular Magnetic Recording." And when that was announced it was kind of a similar problem at that time in the industry that we have today, where the industry was just having a difficult time putting more data on each disc inside that drive. And, so they kind of hit this technology wall. And they announced Perpendicular Magnetic Recording and it really put them on a new S curve in terms of their ability to pack more data on each disc and just kind of put it in some perspective. So, after they announce Perpendicular Magnetic Recording, the capacity per disc increased about 30% a year for about five years. And then over, really, a ten year period, increased about an average of about 20% a year. And, so today's announcement is I see a lot of parallels to that. You know, back when Perpendicular Magnetic Recording was announced, really they build. They increased the capacity per platter was growing very slowly. That's where we are today. And with this announcement of MAMR Technology the direction that Western Digital's choosing really could put the industry on a new S curve and putting in terms of putting more capacity, storage capacity on each one of those discs. >> It's interesting. Always reminds me kind of back to the OS in Microsoft in Intel battles. Right? Intel would come out with a new chip and then Microsoft would make a bigger OS and they go back and back and forth and back and forth. >> John: Yeah, that's very >> And we're seeing that here, right? Cuz the demands for the data are growing exponentially. I think one of the numbers that was thrown out earlier today that the data thrown off by people and the data thrown off by machines is so exponentially larger than the data thrown off by business, which has been kind of the big driver of IT spin. And it's really changing. >> It's a huge fundamental shift. It really is >> They had to do something. Right? >> Yeah, the demand for a storage capacity by these large data centers is just phenomenal and yet at the same time, they don't want to just keep building new data center buildings. And putting more and more racks. They want to put more storage density in that footprint inside that building. So, that's what's really pushing the demand for these higher capacity storage devices. They want to really increase the storage capacity per cubic meter. >> Right, right. >> Inside these data centers. >> It's also just fascinating that our expectation is that they're going to somehow pull it off, right? Our expectation that Moore's laws continue, things are going to get better, faster, cheaper, and bigger. But, back in the back room, somebody's actually got to figure out how to do it. And as you said, we hit these kind of seminal moments where >> Yeah, that's right. >> You do get on a new S curve, and without that it does flatten out over time. >> You know, what's interesting though, Jeff, is really about the time that Perpendicular Magnetic Recording was announced way back in 2005, 2006, the industry was really, already at that time, talking about these thermal assist technologies like MAMR that Western Digital announced today. And it's always been a little bit of a question for those folks that are either in the industry or watching the industry, like IDC. And maybe even even more importantly for some of the HDD industry customers. They're kind of wondering, so what's really going to be the next technology race horse that takes us to that next capacity point? And it's always been a bit of a horse race between HAMR and MAMR. And there's been this lack of clarity or kind of a huge question mark hanging over the industry about which one is it going to be. And Western Digital certainly put a stake in the ground today that they see MAMR as that next technology for the future. >> (mumbles words) Just read a quote today (rushes through name) key alumni just took a new job. And he's got a pin tweet at the top of his thing. And he says, "The smart man looks for ways "To solve the problem. "Or looks at new solutions. "The wise man really spends his time studying the problem." >> I like that. >> And it's really interesting here cuz it seems kind of obvious there. Heat's never necessarily a good thing with electronics and data centers as you mentioned trying to get efficiency up. There's pressure as these things have become huge, energy consumption machines. That said, they're relatively efficient, based on other means that we've been doing they compute and the demand for this compute continues to increase, increase, increase, increase. >> Absolutely >> So, as you kind of look forward, is there anything kind of? Any gems in the numbers that maybe those of us at a layman level are kind of a first read are missing that we should really be paying attention that give us a little bit of a clue of what the feature looks like? >> Well, there's a couple of major trends going on. One is that, at least for the hard drive industry, if you kind of look back the last ten years or so, a pretty significant percentage of the revenue that they've generated a pretty good percentage of the petabytes that they ship have really gone into the PC market. And that's fundamentally shifting. And, so now it's really the data centers, so that by the time you get to 2020, 2021, about 60 plus percent of the petabytes that the industry's shipping is going into data centers, where if you look back a few years ago, 60% was going into PCs. That's a big, big change for the industry. And it's really that kind of change that's pushing the need for these higher capacity hard drives. >> Jeff: Right. >> So, that's, I think, one of the biggest shifts has taking place. >> Well, the other thing that's interesting in that comment because we know scale drives innovation better than anything and clearly Intel microprocessors rode the PC boom to get out scale to drive the innovation. And, so if you're saying, now, that the biggest scale is happening in the data center Then, that's a tremendous force for innovation in there versus Flash, which is really piggy-backing on the growth of these jobs, because that's where it's getting it's scale. So, when you look at kind of the Flash hard drive comparison, right? Obviously, Flash is the shiny new toy getting a lot of buzz over the last couple years. Western Digital has a play across the portfolio, but the announcement earlier today said, you're still going to have like this TenX cost differentiation. >> Yeah, that's right. >> Even through, I think it was 20, 25. I don't want to say what the numbers were. Over a long period of time. You see that kind of continuing DC&E kind of conflict between those two? Or is there a pretty clear stratification between what's going to go into Flash systems, or what's going to hard drives? >> That's a great question, now. So, even in the very large HyperScale data centers and we definitely see where Flash and hard disk drives are very complimentary. They're really addressing different challenges, different problems, and so I think one of the charts that we saw today at the briefing really is something that we agree with strongly at IDC. Today, maybe, about 7% or 8% of all of the combined HDD SSD petabyte shipped for enterprise are SSD petabytes. And then, that grows to maybe ten. >> What was it? Like 7% you said? >> 6% to 7%. >> 6% to 7% okay. Yeah, so we still have 92, 93%, 94% of all petabytes that again are HDD SSD petabytes for enterprise. Those are still HDD petabytes. And even when you get out to 2020, 2021, again, still bought 90%. We agree with what Western Digital talked about today. About 90% of the combined HDD SSD petabytes that are shipping for enterprise continue to be HDD. So, we do see the two technologies very complementary. Talked about SSD is kind of getting their scale on PCs and that's true. They really are going to quickly continue to become a bigger slice of the storage devices attached to new PCs. But, in the data center you really need that bulk storage capacity, low cost capacity. And that's where we see that the two SSDs and HDDs are going to live together for a long time. >> Yeah, and as we said the conflict barrier, complimentary nature of the two different applications are very different. You need the big data to build the models, to run the algorithms, to do stuff. But, at the same time, you need the fast data that's coming in. You need the real time analytics to make modifications to the algorithms and learn from the algorithms >> That's right, yeah. It's the two of those things together that are one plus one makes three type of solution. Exactly, and especially to address latency. Everybody wants their data fast. When you type something into Google, you want your response right away. And that's where SSDs really come into play, but when you do deep searches, you're looking through a lot of data that has been collected over years and a lot of that's probably sitting on hard disc drives. >> Yeah. The last piece of the puzzle, I just want to you to address before we sign off, That was an interesting point is that not just necessarily the technology story, but the ecosystem story. And I thought that was really kind of, I thought, the most interesting part of the MAMR announcement was that it fits in the same form factor, there's no change to OS, there's no kind of change in the ecosystem components in which you plug this in. >> Yeah, that's right. It's just you take out the smaller drive, the 10, or the 12, or whatever, or 14 I guess is coming up. And plug in. They showed a picture of a 40 terabyte drive. >> Right. >> You know, that's the other part of the story that maybe doesn't get as much play as it should. You're playing in an ecosystem. You can't just come up with this completely, kind of independent, radical, new thing, unless it'S so radical that people are willing to swap out their existing infrastructure. >> I completely agree. It's can be very difficult for the customer to figure out how to adopt some of these new technologies and actually, the hard disk drive industry has thrown a couple of technologies at their customers over the past five, six years, that have been a little challenging for them to adopt. So, one was when the industry went from a native 512 by sectors to 4K sectors. Seems like a pretty small change that you're making inside the drive, but it actually presented some big challenges for some of the enterprise customers. And even the single magnetic recording technologies. So, it has a way to get more data on the disc, and Western Digital certainly talked about that today. But, for the customer trying to plug and play that into a system and SMR technology actually created some real challenges for them to figure out how to adopt that. So, I agree that what was shown today about the MAMR technology is definitely a plug and play. >> Alright, we'll give you the last word as people are driving away today from the headquarters. They got a bumper sticker as to why this is so important. What's it say on the bumper sticker about MAMR? It says that we continue to get more capacity at a lower cost. >> (chuckles) Isn't that just always the goal? >> I agree. >> (chuckles) Alright, well thank you for stopping by and sharing your insight. Really appreciate it. >> Thanks, Jeff. >> Alright. Jeff Frick here at Western Digital. You're watching theCUBE! Thanks for watching. (futuristic beat)

Published Date : Oct 12 2017

SUMMARY :

Brought to you by Western Digital. He is the Research Vice President So, what is your take on today's announcement? for the industry was Helium, about four or five years ago. Always reminds me kind of back to the OS that the data thrown off by people It's a huge fundamental shift. They had to do something. Yeah, the demand for a storage capacity But, back in the back room, and without that it does flatten out over time. as that next technology for the future. "To solve the problem. and the demand for this compute continues And it's really that kind of change that's pushing the need one of the biggest shifts has taking place. and clearly Intel microprocessors rode the PC boom You see that kind of continuing DC&E kind of conflict So, even in the very large HyperScale data centers of the storage devices attached to new PCs. You need the big data to build the models, It's the two of those things together is that not just necessarily the technology story, the 10, or the 12, or whatever, or 14 I guess is coming up. that's the other part of the story that maybe doesn't get And even the single magnetic recording technologies. What's it say on the bumper sticker about MAMR? and sharing your insight. Thanks for watching.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jeff FrickPERSON

0.99+

John RydningPERSON

0.99+

JeffPERSON

0.99+

Western DigitalORGANIZATION

0.99+

JohnPERSON

0.99+

2005DATE

0.99+

twoQUANTITY

0.99+

7%QUANTITY

0.99+

90%QUANTITY

0.99+

2006DATE

0.99+

60%QUANTITY

0.99+

2020DATE

0.99+

two technologiesQUANTITY

0.99+

94%QUANTITY

0.99+

92, 93%QUANTITY

0.99+

each discQUANTITY

0.99+

MicrosoftORGANIZATION

0.99+

2021DATE

0.99+

ten yearQUANTITY

0.99+

TodayDATE

0.99+

San Jose, CaliforniaLOCATION

0.99+

20QUANTITY

0.99+

todayDATE

0.99+

8%QUANTITY

0.99+

oneQUANTITY

0.99+

BIGITORGANIZATION

0.99+

two different applicationsQUANTITY

0.99+

40 terabyteQUANTITY

0.99+

25QUANTITY

0.99+

MAMRORGANIZATION

0.98+

tenQUANTITY

0.98+

IDCORGANIZATION

0.98+

12QUANTITY

0.98+

first readQUANTITY

0.98+

MAMR TechnologyORGANIZATION

0.98+

IntelORGANIZATION

0.97+

theCUBEORGANIZATION

0.97+

about 20% a yearQUANTITY

0.97+

about 30% a yearQUANTITY

0.97+

about five yearsQUANTITY

0.97+

OneQUANTITY

0.96+

GoogleORGANIZATION

0.96+

about 7%QUANTITY

0.96+

HAMRORGANIZATION

0.96+

14QUANTITY

0.95+

next decadeDATE

0.95+

five years agoDATE

0.95+

10QUANTITY

0.94+

6%QUANTITY

0.94+

512QUANTITY

0.93+

About 90%QUANTITY

0.91+

about 60 plus percentQUANTITY

0.91+

last couple yearsDATE

0.91+

earlier todayDATE

0.9+

singleQUANTITY

0.89+

six yearsQUANTITY

0.89+

few years agoDATE

0.88+

Janet George , Western Digital | Western Digital the Next Decade of Big Data 2017


 

>> Announcer: Live from San Jose, California, it's theCUBE, covering Innovating to Fuel the Next Decade of Big Data, brought to you by Western Digital. >> Hey welcome back everybody, Jeff Frick here with theCUBE. We're at Western Digital at their global headquarters in San Jose, California, it's the Almaden campus. This campus has a long history of innovation, and we're excited to be here, and probably have the smartest person in the building, if not the county, area code and zip code. I love to embarrass here, Janet George, she is the Fellow and Chief Data Scientist for Western Digital. We saw you at Women in Data Science, you were just at Grace Hopper, you're everywhere and get to get a chance to sit down again. >> Thank you Jeff, I appreciate it very much. >> So as a data scientist, today's announcement about MAMR, how does that make you feel, why is this exciting, how is this going to make you be more successful in your job and more importantly, the areas in which you study? >> So today's announcement is actually a breakthrough announcement, both in the field of machine learning and AI, because we've been on this data journey, and we have been very selectively storing data on our storage devices, and the selection is actually coming from the preconstructed queries that we do with business data, and now we no longer have to preconstruct these queries. We can store the data at scale in raw form. We don't even have to worry about the format or the schema of the data. We can look at the schema dynamically as the data grows within the storage and within the applications. >> Right, cause there's been two things, right. Before data was bad 'cause it was expensive to store >> Yes. >> Now suddenly we want to store it 'cause we know data is good, but even then, it still can be expensive, but you know, we've got this concept of data lakes and data swamps and data all kind of oceans, pick your favorite metaphor, but we want the data 'cause we're not really sure what we're going to do with it, and I think what's interesting that you said earlier today, is it was schema on write, then we evolved to schema on read, which was all the rage at Hadoop Summit a couple years ago, but you're talking about the whole next generation, which is an evolving dynamic schema >> Exactly. >> Based whatever happens to drive that query at the time. >> Exactly, exactly. So as we go through this journey, we are now getting independent of schema, we are decoupled from schema, and what we are finding out is we can capture data at its raw form, and we can do the learning at the raw form without human interference, in terms of transformation of the data and assigning a schema to that data. We got to understand the fidelity of the data, but we can train at scale from that data. So with massive amounts of training, the models already know to train itself from raw data. So now we are only talking about incremental learning, as the train model goes out into the field in production, and actually performs, now we are talking about how does the model learn, and this is where fast data plays a very big role. >> So that's interesting, 'cause you talked about that also earlier in your part of the presentation, kind of the fast data versus big data, which kind of maps the flash versus hard drive, and the two are not, it's not either or, but it's really both, because within the storage of the big data, you build the base foundations of the models, and then you can adapt, learn and grow, change with the fast data, with the streaming data on the front end, >> Exactly >> It's a whole new world. >> Exactly, so the fast data actually helps us after the training phase, right, and these are evolving architectures. This is part of your journey. As you come through the big data journey you experience this. But for fast data, what we are seeing is, these architectures like Lambda and Kappa are evolving, and especially the Lambda architecture is very interesting, because it allows for batch processing of historical data, and then it allows for what we call a high latency layer or a speed layer, where this data can then be promoted up the stack for serving purposes. And then Kappa architecture's where the data is being streamed near real time, bounded and unbounded streams of data. So this is again very important when we build machine learning and AI applications, because evolution is happening on the fly, learning is happening on the fly. Also, if you think about the learning, we are mimicking more and more on how humans learn. We don't really learn with very large chunks of data all at once, right? That's important for initially model training and model learning, but on a regular basis, we are learning with small chunks of data that are streamed to us near real time. >> Right, learning on the Delta. >> Learning on the Delta. >> So what is the bound versus the unbound? Unpack that a little bit. What does that mean? >> So what is bounded is basically saying, hey we are going to get certain amounts of data, so you're sizing the data for example. Unbounded is infinite streams of data coming to you. And so if your architecture can absorb infinite streams of data, like for example, the sensors constantly transmitting data to you, right? At that point you're not worried about whether you can store that data, you're simply worried about the fidelity of that data. But bounded would be saying, I'm going to send the data in chunks. You could also do bounded where you basically say, I'm going to pre-process the data a little bit just to see if the data's healthy, or if there is signal in the data. You don't want to find that out later as you're training, right? You're trying to figure that out up front. >> But it's funny, everything is ultimately bounded, it just depends on how you define the unit of time, right, 'cause you take it down to infinite zero, everything is frozen. But I love the example of the autonomous cars. We were at the event with, just talking about navigation just for autonomous cars. Goldman Sachs says it's going to be a seven billion dollar industry, and the great example that you used of the two systems working well together, 'cause is it the car centers or is it the map? >> Janet: That's right. >> And he says, well you know, you want to use the map, and the data from the map as much as you can to set the stage for the car driving down the road to give it some level of intelligence, but if today we happen to be paving lane number two on 101, and there's cones, now it's the real time data that's going to train the system. But the two have to work together, and the two are not autonomous and really can't work independent of each other. >> Yes. >> Pretty interesting. >> It makes perfect sense, right. And why it makes perfect sense is because first the autonomous cars have to learn to drive. Then the autonomous cars have to become an experienced driver. And the experience cannot be learned. It comes on the road. So one of the things I was watching was how insurance companies were doing testing on these cars, and they had a human, a human driving a car, and then an autonomous car. And the autonomous car, with the sensors, were predicting the behavior, every permutation and combination of how a bicycle would react to that car. It was almost predicting what the human on the bicycle would do, like jump in front of the car, and it got it right 80% of the cases. But a human driving a car, we're not sure how the bicycle is going to perform. We don't have peripheral vision, and we can't predict how the bicycle is going to perform, so we get it wrong. Now, we can't transmit that knowledge. If I'm a driver and I just encountered a bicycle, I can't transmit that knowledge to you. But a driverless car can learn, it can predict the behavior of the bicycle, and then it can transfer that information to a fleet of cars. So it's very powerful in where the learning can scale. >> Such a big part of the autonomous vehicle story that most people don't understand, that not only is the car driving down the road, but it's constantly measuring and modeling everything that's happening around it, including bikes, including pedestrians, including everything else, and whether it gets in a crash or not, it's still gathering that data and building the model and advancing the models, and I think that's, you know, people just don't talk about that enough. I want follow up on another topic. So we were both at Grace Hopper last week, which is a phenomenal experience, if you haven't been, go. Ill just leave it at that. But Dr. Fei-Fei Li gave one of the keynotes, and she made a really deep statement at the end of her keynote, and we were both talking about it before we turned the cameras on, which is, there's no question that AI is going to change the world, and it's changing the world today. The real question is, who are the people that are going to build the algorithms that train the AI? So you sit in your position here, with the power, both in the data and the tools and the compute that are available today, and this brand new world of AI and ML. How do you think about that? How does that make you feel about the opportunity to define the systems that drive the cars, et cetera. >> I think not just the diversity in data, but the diversity in the representation of that data are equally powerful. We need both. Because we cannot tackle diverse data, diverse experiences with only a single representation. We need multiple representation to be able to tackle that data. And this is how we will overcome bias of every sort. So it's not the question of who is going to build the AI models, it is a question of who is going to build the models, but not the question of will the AI models be built, because the AI models are already being built, but some of the models have biases into it from any kind of lack of representation. Like who's building the model, right? So I think it's very important. I think we have a powerful moment in history to change that, to make real impact. >> Because the trick is we all have bias. You can't do anything about it. We grew up in the world in which we grew up, we saw what we saw, we went to our schools, we had our family relationships et cetera. So everyone is locked into who they are. That's not the problem. The problem is the acceptance of bring in some other, (chuckles) and the combination will provide better outcomes, it's a proven scientific fact. >> I very much agree with that. I also think that having the freedom, having the choice to hear another person's conditioning, another person's experiences is very powerful, because that enriches our own experiences. Even if we are constrained, even if we are like that storage that has been structured and processed, we know that there's this other storage, and we can figure out how to get the freedom between the two point of views, right? And we have the freedom to choose. So that's very, very powerful, just having that freedom. >> So as we get ready to turn the calendar on 2017, which is hard to imagine it's true, it is. You look to 2018, what are some of your personal and professional priorities, what are you looking forward to, what are you working on, what's top of mind for Janet George? >> So right now I'm thinking about genetic algorithms, genetic machine learning algorithms. This has been around for a while, but I'll tell you where the power of genetic algorithms is, especially when you're creating powerful new technology memory cell. So when you start out trying to create a new technology memory cell, you have materials, material deformations, you have process, you have hundred permutation combination, and the genetic algorithms, we can quickly assign a cause function, and we can kill all the survival of the fittest, all that won't fit we can kill, arriving to the fastest, quickest new technology node, and then from there, we can scale that in mass production. So we can use these survival of the fittest mechanisms that evolution has used for a long period of time. So this is biology inspired. And using a cause function we can figure out how to get the best of every process, every technology, all the coupling effects, all the master effects of introducing a program voltage on a particular cell, reducing the program voltage on a particular cell, resetting and setting, and the neighboring effects, we can pull all that together, so 600, 700 permutation combination that we've been struggling on and not trying to figure out how to quickly narrow down to that perfect cell, which is the new technology node that we can then scale out into tens of millions of vehicles, right? >> Right, you're going to have to >> Getting to that spot. >> You're going to have to get me on the whiteboard on that one, Janet. That is amazing. Smart lady. >> Thank you. >> Thanks for taking a few minutes out of your time. Always great to catch up, and it was terrific to see you at Grace Hopper as well. >> Thank you, I really appreciate it, I appreciate it very much. >> All right, Janet George, I'm Jeff Frick. You are watching theCUBE. We're at Western Digital headquarters at Innovating to Fuel the Next Generation of Big Data. Thanks for watching.

Published Date : Oct 11 2017

SUMMARY :

the Next Decade of Big Data, in San Jose, California, it's the Almaden campus. the preconstructed queries that we do with business data, Right, cause there's been two things, right. of the data and assigning a schema to that data. and especially the Lambda architecture is very interesting, So what is the bound versus the unbound? the sensors constantly transmitting data to you, right? and the great example that you used and the data from the map as much as you can and it got it right 80% of the cases. and advancing the models, and I think that's, So it's not the question of who is going to Because the trick is we all have bias. having the choice to hear another person's conditioning, So as we get ready to turn the calendar on 2017, and the genetic algorithms, we can quickly assign You're going to have to get me on the whiteboard and it was terrific to see you at Grace Hopper as well. I appreciate it very much. at Innovating to Fuel the Next Generation of Big Data.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Janet GeorgePERSON

0.99+

JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

JanetPERSON

0.99+

Western DigitalORGANIZATION

0.99+

80%QUANTITY

0.99+

two thingsQUANTITY

0.99+

2018DATE

0.99+

last weekDATE

0.99+

2017DATE

0.99+

Goldman SachsORGANIZATION

0.99+

San Jose, CaliforniaLOCATION

0.99+

two systemsQUANTITY

0.99+

twoQUANTITY

0.99+

todayDATE

0.99+

bothQUANTITY

0.99+

seven billion dollarQUANTITY

0.99+

Fei-Fei LiPERSON

0.98+

AlmadenLOCATION

0.98+

two pointQUANTITY

0.97+

oneQUANTITY

0.97+

firstQUANTITY

0.95+

Grace HopperORGANIZATION

0.95+

theCUBEORGANIZATION

0.95+

hundred permutationQUANTITY

0.95+

MAMRORGANIZATION

0.94+

Women in Data ScienceORGANIZATION

0.91+

tens of millions of vehiclesQUANTITY

0.9+

one ofQUANTITY

0.89+

KappaORGANIZATION

0.89+

Dr.PERSON

0.88+

single representationQUANTITY

0.83+

a couple years agoDATE

0.83+

earlier todayDATE

0.82+

Next DecadeDATE

0.81+

LambdaTITLE

0.8+

101OTHER

0.8+

600, 700 permutationQUANTITY

0.77+

LambdaORGANIZATION

0.7+

of dataQUANTITY

0.67+

keynotesQUANTITY

0.64+

Hadoop SummitEVENT

0.62+

zeroQUANTITY

0.6+

numberOTHER

0.55+

DeltaOTHER

0.54+

twoOTHER

0.35+

Mark Grace, Western Digital | Western Digital the Next Decade of Big Data 2017


 

>> Announcer: Live from San Jose, California, it's theCUBE, covering Innovating to Fuel the Next Decade of Big Data, brought to you by Western Digital. >> Hey welcome back everybody, Jeff Frick here with theCUBE. We're at Western Digital's headquarters in San Jose, California at the Almaden campus. Lot of innovation's been going on here, especially in storage for decades, and we're excited to be at this special press and analyst event that Western Digital put on today to announce some exciting new products. It's called Innovating to Fuel the Next Decade of Data. I'm super happy to have a long-time industry veteran, he just told me, 35 years, I don't know if I can tell (Mark laughs) that on air or not. He's Mark Grace, he's the Senior Vice President of Devices for Western Digital, Mar, great to have you on. >> Thanks Jeff, glad to be here. >> Absolutely, so you've seen this movie over and over and over, I mean that's one of the cool things about being in the Valley, is this relentless pace of innovation. So how does today's announcement stack up as you kind of look at this versus kind of where we've come from? >> Oh I think this is maybe one of the, as big as it comes, Jeff, to be honest. I think we've plotted a course now that I think was relatively uncertain for the hard drive industry and the data center, and plotted a course that I think we can speak clearly to the market, and clearly to customers about the value proposition for rotating magnetic storage for decades to come. >> Which is pretty interesting, 'cause, you know, rotating drives have been taking a hit over the last couple of years, right, flash has been kind of the sexy new kid on the block, so this is something new, >> Mark: It is. >> And a new S curve I think as John said. >> I agree, we're jumping onto a, we're extending the S curve, let's call it that. I think there's actually plenty of other S curve opportunities for us, but in this case, I think the industry, and I would say our customer base, we have been less than clear with those guys about how we see the future of rotating storage in the cloud and enterprise space, and I think today's announcement clarifies that and gives some confidence about architectural decisions relative to rotating storage going forward for a long time. >> Well I think it's pretty interesting, 'cause compared to the other technology that was highlighted, the other option, the HAMR versus the MAMR, this was a much more elegant, simpler way to add this new S curve into an existing ecosystem. >> You know, elegant's probably a good word for it, and it's always the best solution I would say. HAMR's been a push for many years. I can't remember the first time I heard about HAMR. It's still something we're going to continue to explore and invest in, but it has numerous hurdles compared to the simplicity and elegance, as you say, of MAMR, not the least of which is we're going to operate at normal ambient temperatures versus apply tremendous heat to try and energize the recording and the technologies. So any time you introduce extraordinary heat you face all kinds of ancillary engineering challenges, and this simplifies those challenges down to one critical innovation, which is the oscillator. >> Pretty interesting, 'cause it seems pretty obvious that heat's never a good thing. So it's curious that in the quest for this next S curve that the HAMR path was pursued for as long as it was, it sounds like, because it sounds like that's a pretty tough thing to overcome. >> Yeah, I think it initially presented perhaps the most longevity perhaps in early exploration days. I would say that HAMR has certainly received the most press as far as trying to assert it as the extending recording technology for enterprise HDDs. I would say we've invested for almost as long in MAMR, but we've been extremely quiet about it. This is kind of our nature. When we're ready to talk about something, you can kind of be sure we're ready to go with it, and ready to think about productization. So we're quite confident in what we're doing. >> But I'm curious from your perspective, having been in the business a long time, you know, we who are not directly building these magical machines, just now have come to expect that Moore's Law will contain, has zero to do with semiconductor physics anymore, it's really an attitude and this relentless pace of innovation that now is expected and taken for granted. You're on the other side, and have to face real physics and mechanical limitations of the media and the science and everything else. So is that something that gets you up every day >> Mark: Keeps me awake every night! >> Obviously keeps you awake at night and up every day. You've been doing it for 35 years, so there must be some appeal. >> Yeah. (laughs) >> But you know, it's a unique challenge, 'cause at the same time not only has it got to be better and faster and bigger, it's got to be cheaper, and it has been. So when you look at that, how does that kind of motivate you, the teams here, to deliver on that promise? >> Yeah, I mean in this case, we are a little bit defensive, in the sense of the flash expectations that you mentioned, and I think as we digest our news today, we'll be level setting a little bit more in a more balanced way the expectations for contribution from rotating magnetic storage and solid state storage to what I think is a more accurate picture of its future going forward in the enterprise and hyperscale space. To your point about just relentless innovation, a few of us were talking the other day in advance of this announcement that this MAMR adventure feels like the early days of PMR, perpendicular, the current recording technology. It feels like we understand a certain amount of distance ahead of us, and that's about this four-terabit per inch kind of distance, but it feels like the early days where we could only see so far but the road actually goes much further, and we're going to find more and more ways to extend this technology, and keep that order of magnitude cost advantage going from a hard drive standpoint versus flash. >> I wonder how this period compares to that, just to continue, in terms of on the demand side, 'cause you know, back in the day, the demand and the applications for these magical compute machines weren't near, I would presume, as pervasive as now, or am I missing the boat? 'Cause now clearly there is no shortage of demand for storage and compute. >> Yeah, depending on where you're coming from, you could take two different views of that. The engine that drove the scale of the hard drive industry to date has, a big piece of it in the long history of the hard drive industry has been the PC space. So you see that industry converting to flash and solid state storage more aggressively, and we embrace that, you know we're invested in flash and we have great products in that space, and we see that happening. The opportunity in the hyperscale and cloud space is we're only at the tip of the iceberg, and therefore I think, as we think about this generation, we think about it differently than those opportunities in terms of breadth of applications, PCs, and all that kind of create the foundation for the hard drive, but what we see here is the virtuous cycle of more storage, more economical storage begets more value proposition, more opportunities to integrate more data, more data collection, more storage. And that virtuous cycle seems to me that we're just getting started. So long live data, that's what we say. (both laugh) >> The other piece that I find interesting is before the PCs were the driver of scale relative to an enterprise data center, but with the hyperscale guys and the proliferation of cloud and actually the growth of PCs is slowing down dramatically, that it's kind of flipped the bit. Now the data centers themselves have the scale to drive >> Absolutely. >> the scale innovation that before was before was really limited to either a PC or a phone or some more consumer device. >> Absolutely the case. When you take that cross-section of hard drive applications, that's a hundred percent the case, and in fact, we look at the utilization as a vertically integrated company we look at our media facilities for the disks, we look at our wafer facilities for heads, and those facilities as we look forward are going to be as busy as busier than they've ever been. I mean the amount of data is relative to the density as well as disks and heads and how many you can employ. So we see this in terms of fundamental technology and component construction, manufacturing, busier than it's ever been. We'll make fewer units. I mean there will be fewer units as they become bigger and denser for this application space, but the fundamental consumption of magnetic recording technology and components is all-time records. >> Right. And you haven't even talked about the software-defined piece that's dragging the utilization of that data across multiple applications. >> And it's just one of these that come in to help everybody there too, yeah. >> Jeff: You got another 35 years more years in you? (both laugh) >> I hope so. >> All right. >> But that would be the edge of it, I think. >> All right, we're going to take Mark Grace here, only 35 more years, Lord knows what he'll be working on. Well Mark, thanks for taking a few minutes and answering your prospective >> No that's fine, thanks a lot. >> Absolutely, Mark Grace, I'm Jeff Frick, you're watching theCUBE from Western Digital headquarters in San Jose, California. Thanks for watching. >> Mark: All right.

Published Date : Oct 11 2017

SUMMARY :

the Next Decade of Big Data, in San Jose, California at the Almaden campus. and over, I mean that's one of the cool things and clearly to customers about the value proposition in the cloud and enterprise space, the HAMR versus the MAMR, and it's always the best solution I would say. So it's curious that in the quest for this next S curve and ready to think about productization. and mechanical limitations of the media and the science Obviously keeps you awake at night and up every day. 'cause at the same time not only has it got to be in the sense of the flash expectations that you mentioned, and the applications for these magical compute machines PCs, and all that kind of create the foundation and actually the growth of PCs is slowing down dramatically, the scale innovation I mean the amount of data is relative to the density piece that's dragging the utilization of that data that come in to help everybody there too, yeah. and answering your prospective No that's fine, in San Jose, California.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jeff FrickPERSON

0.99+

JeffPERSON

0.99+

Western DigitalORGANIZATION

0.99+

JohnPERSON

0.99+

Mark GracePERSON

0.99+

35 yearsQUANTITY

0.99+

MarkPERSON

0.99+

San Jose, CaliforniaLOCATION

0.99+

AlmadenLOCATION

0.99+

MarPERSON

0.98+

bothQUANTITY

0.98+

35 more yearsQUANTITY

0.97+

todayDATE

0.97+

first timeQUANTITY

0.97+

theCUBEORGANIZATION

0.96+

oneQUANTITY

0.95+

hundred percentQUANTITY

0.94+

zeroQUANTITY

0.94+

four-terabit per inchQUANTITY

0.93+

two different viewsQUANTITY

0.93+

yearsDATE

0.92+

HAMRORGANIZATION

0.91+

MoorePERSON

0.9+

both laughQUANTITY

0.9+

35 years more yearsQUANTITY

0.88+

decadesQUANTITY

0.86+

2017DATE

0.75+

one critical innovationQUANTITY

0.71+

Innovating to Fuel the Next Decade ofEVENT

0.7+

MAMROTHER

0.7+

last coupleDATE

0.66+

Next DecadeTITLE

0.62+

Innovating to Fuel the Next Decade of Big DataTITLE

0.6+

SeniorPERSON

0.53+

PresidentPERSON

0.53+

HAMROTHER

0.45+

DataTITLE

0.44+

MAMRORGANIZATION

0.38+

Dave Tang, Western Digital | Western Digital the Next Decade of Big Data 2017


 

(upbeat techno music) >> Announcer: Live from San Jose, California it's theCUBE, covering Innovating to Fuel the Next Decade of Big Data, brought to you by Western Digital. >> Hey, welcome back everybody. Jeff Frick here at theCUBE. We're at the Western Digital Headquarters off Almaden down in San Jose, a really important place. Western Digital's been here for a while, their headquarters. A lot of innovation's been going on here forever. So we're excited to be here really for the next generation. The event's called Innovating to Fuel the Next Generation of big data, and we're joined by many time Cuber, Dave Tang. He is the SVP in corporate marketing from Western Digital. Dave, always great to see you. >> Yeah. Always great to be here, Jeff. Thanks. >> Absolutely. So you got to MC the announcement today. >> Yes. >> So for the people that weren't there, let's give them a quick overview on what the announcement was and then we can dive in a little deeper. >> Great, so what we were announcing was a major breakthrough in technology that's going to allow us to drive the increase in capacity in density to support big data for the next decade and beyond, right? So capacities and densities had been starting to level off in terms of hard drive technology capability. So what we announced was microwave-assisted magnetic recording technology that will allow us to keep growing that areal density up and reducing the cost per terabyte. >> You know, it's fascinating cause everyone loves to talk about Moore's Law and have these silly architectural debates, whether Moore's Law is alive or dead, but, as anyone who's lived here knows, Moore's Law is really an attitude much more it is than the specific physics of microprocessor density growth. And it's interesting to see. As we know the growth of data is growing in giant and the types of data, and not only regular big data, but now streaming data are bigger and bigger and bigger. I think you talked about stuff coming off of people and machines compared to business data is way bigger. >> Right. >> But you guys continue to push limits and break through, and even though we expect everything to be cheaper, faster, and better, you guys actually have to execute it-- >> Dave: Right. >> Back at the factory. >> Right, well it's interesting. There's this healthy tension, right, a push and pull in the environment. So you're right, it's not just Moore's Law that's enabling a technology push, but we have this virtuous cycle, right? We've realized what the value is of data and how to extract the possibilities and value of data, so that means that we want to store more of that data and access more of that data, which drives the need for innovation to be able to support all of that in a cost effective way. But then that triggers another wave of new applications, new ways to tap into the possibilities of data. So it just feeds on itself, and fortunately we have great technologists, great means of innovation, and a great attitude and spirit of innovation to help drive that. >> Yeah, so for people that want more, they can go to the press releases and get the data. We won't dive deep into the weeds here on the technology, but I thought you had Janet George speak, and she's chief data scientist. Phenomenal, phenomenal big brain. >> Dave: Yes. >> A smart lady. But she talked about, from her perspective we're still just barely even getting onto this data opportunity in terms of automation, and we see over and over at theCUBE events, innovation's really not that complicated. Give more people access to the data, give them more access to the tools, and let them try things easier and faster and feel quick, there's actually a ton of innovation that companies can unlock within their own four walls. But the data is such an important piece of it, and there's more and more and more of this. >> Dave: Right, right. >> What used to be digital exhaust now is, I think maybe you said, or maybe it was Dave, that there's a whole economy now built on data like we used to do with petroleum. I thought that was really insightful. >> Yeah, right. It's like a gold mine. So not only are the sources of data increasing, which is driving increased volume, but, as Janet was alluding to, we're starting to come up with the tools and the sophistication with machine learning and artificial intelligence to be able to put that data to new use as well as to find the pieces of data to interconnect, to drive these new capabilities and new insights. >> Yeah, but unlike petroleum it doesn't get used up. I mean that's the beauty of data. (laughing) >> Yeah, that's right. >> It's a digital process that can be used over and over and over again. >> And a self-renewing, renewing resource. And you're right, in that sense that it's being used over and over again that the longevity of that data, the use for life is growing exponentially along with the volume. >> Right, and Western Digital's in a unique position cause you have systems and you have big systems that could be used in data centers, but you also have the media that powers a whole bunch of other people's systems. So I thought one of the real important announcements today was, yes it's an interesting new breakthrough technology that uses energy assist to get more density on the drives, but it's done in such a way that the stuff's all backward compatible. It's plug and play. You've got production scheduled in a couple years I think with test out the customers-- >> Dave: That's right. >> Next year. So, you know, that is such an important piece beyond the technology. What's the commercial acceptance? What are the commercial barriers? And this sounds like a pretty interesting way to skin that cow. >> Right, often times the best answers aren't the most complex answers. They're the more elegant and simplistic answers. So it goes from the standpoint of a user being able to plug and play with older systems, older technologies. That's beautiful, and for us, to be able to, the ability to manufacture it in high volume reliably and cost effectively is equally as important. >> And you also talked, which I think was interesting, is kind of the relationship between hard drives and flash, because, obviously, flash is a, I want to say the sexy new toy, but it's not a sexy new toy anymore. >> Right. >> It's been around for a while, but, with that pressure on flash performance, you're still seeing the massive amounts of big data, which is growing faster than that. And there is a rule for the high density hard drives in that environment, and, based on the forecast you shared, which I'm presuming came from IDC or people that do numbers for a living, still a significant portion of a whole lot of data is not going to be on flash. >> Yeah, that's right. I think we have a tendency, especially in technology, to think either or, right? Something is going to take over from something else, but in this case it's definitely an and, right. And a lot of that is driven by this notion that there's fast data and big data, and, while our attention seems to shift over to maybe some fast data applications like autonomous vehicles and realtime applications, surveillance applications, there's still a need for big data because the algorithms that drive those realtime applications have to come from analysis of vast amounts of data. So big data is here to stay. It's not going away or shifting over. >> I think it's a really interesting kind of cross over, which Janet talked about too, where you need the algorithms to continue sharing the system that are feeding, continuing, and reacting to the real data, but then that just adds more vocabulary to their learning set so they can continue to evolve overtime. >> Yeah, what really helps us out in the market place is that because we have technologies and products across that full spectrum of flash and rotating magnetic recording, and we sell to customers who buy devices as well as platforms and systems, we see a lot of applications, a lot of uses of data, and we're able to then anticipate what those needs are going to be in the near future and in the distant future. >> Right, so we're getting towards the end of 2017, which I find hard to say, but as you look forward kind of to 2018 and this insatiable desire for more storage, cause this insatiable creation of more data, what are some of your priorities for 2018 and what are you kind of looking at as, like I said, I can't believe we're going to actually flip the calendar here-- >> Dave: Right, right. >> In just a few short months. (laughing) >> Well, I think for us, it's the realization that all these applications that are coming at us are more and more diverse, and their needs are very specialized. So it's not just the storage, although we're thought of as a storage company, it's not just about the storage of that data, but you have contrive complete environments to capture and preserve and access and transform that data, which means we have to go well beyond storage and think about how that data is accessed, technical interfaces to our memory products as well as storage products, and then where compute sits. Does it still sit in a centralized place or do you move compute to out closer to where the data sits. So, all this innovation and changing the way that we think about how we can mine that data is top of the mind for us for the next year and beyond. >> It's only job security for you, Dave. (laughing) >> Dave: Funny to think about. >> Alright. He's Dave Tang. Thanks for inviting us and again congratulations on the presentation. >> Always a pleasure. >> Alright, Dave Tang, I'm Jeff Frick. You're watching theCUBE from Western Digital headquarters in San Jose, California. Thanks for watching. (upbeat techno music)

Published Date : Oct 11 2017

SUMMARY :

brought to you by Western Digital. He is the SVP in corporate marketing Always great to be here, Jeff. So you got to MC the announcement today. So for the people that weren't there, and reducing the cost per terabyte. and machines compared to business data and how to extract the possibilities and get the data. Give more people access to the data, that there's a whole economy now the pieces of data to interconnect, I mean that's the beauty of data. It's a digital process that can be used that the longevity of that data, that the stuff's all backward compatible. What are the commercial barriers? the ability to manufacture it in high volume is kind of the relationship between hard drives and, based on the forecast you shared, So big data is here to stay. and reacting to the real data, in the near future and in the distant future. (laughing) So it's not just the storage, It's only job security for you, Dave. and again congratulations on the in San Jose, California.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JanetPERSON

0.99+

Janet GeorgePERSON

0.99+

Jeff FrickPERSON

0.99+

Dave TangPERSON

0.99+

DavePERSON

0.99+

Western DigitalORGANIZATION

0.99+

JeffPERSON

0.99+

2018DATE

0.99+

San Jose, CaliforniaLOCATION

0.99+

AlmadenLOCATION

0.99+

Next yearDATE

0.99+

San JoseLOCATION

0.99+

next yearDATE

0.99+

todayDATE

0.98+

next decadeDATE

0.97+

oneQUANTITY

0.95+

Moore's LawTITLE

0.91+

Western DigitalLOCATION

0.89+

CuberPERSON

0.88+

theCUBEORGANIZATION

0.88+

end of 2017DATE

0.88+

Innovating to Fuel the Next Generation ofEVENT

0.81+

IDCORGANIZATION

0.8+

2017DATE

0.8+

Moore'sTITLE

0.76+

MoorePERSON

0.72+

monthsQUANTITY

0.68+

couple yearsQUANTITY

0.65+

lot of applicationsQUANTITY

0.61+

four wallsQUANTITY

0.6+

Innovating to Fuel the Next Decade of Big DataEVENT

0.59+

Next DecadeDATE

0.58+

Mike Cordano, Western Digital | Western Digital the Next Decade of Big Data 2017


 

>> Announcer: Live from San Jose, California, it's The Cube. Covering Innovating to Fuel the Next Decade of Big Data. Brought to you by Western Digital. >> Hey, welcome back everybody. Jeff Frick here with The Cube. We're at the Western Digital headquarters in San Jose, the Great Oaks Campus, a really historic place in the history of Silicon Valley and computing. It's The Innovating to Fuel the Next Generation of Big Data event with Western Digital. We're really excited to be joined by our next guest, Mike Cordano. He's the president and chief operating officer of Western Digital. Mike, great to see you. >> Great to see you as well. Happy you guys could be here. It's an exciting day. >> Absolutely. First off, I think the whole merger thing is about done, right? That's got to feel good. >> Yeah, it's done, but there's legs to it, right? So we've combined these companies now, three of them, three large ones, so obviously Western Digital and Hitachi Global Storage, now we've added SanDisk into one Western Digital, so we're all together. Obviously more to do, as you expect in a large scale integration. There will be a year or two of bringing all those business processes and systems together, but I got to say, the teams are coming together great, showing up in our financial performance and our product execution, so things are really coming together. >> Yeah, not an easy task by any stretch of the imagination. >> No, not easy, but certainly a compliment to our team. I mean, we've got great people. You know, like anything, if you can harness the capabilities of your team, there's a lot you can accomplish, and it really is a compliment to the team. >> Excellent. Well, congratulations on that, and talking a bit about this event here today, you've even used "Big Data" in the title of the event, so you guys are obviously in a really unique place, Western Digital. You make systems, big systems. You also make the media that feeds a lot of other people's systems, but as the big data grows, the demand for data grows, it's got to live somewhere, so you're sitting right at the edge where this stuff's got to sit. >> Yeah, that's right, and it's central to our strategy, right? So if you think about it, there's three fundamental technologies that we think are just inherent in all of the evolution of compute and IT architecture. Obviously, there is compute, there is storage or memory, and then there's sort of movement, or interconnect. We obviously live in the storage or memory node, and we have a very broad set of capabilities, all the way from rotating magnetic media, which was our heritage, now including non-volatile memory and flash, and that's just foundational to everything that is going to come, and as you said, we're not going to stop there. It's not just a devices or component company, we're going to continue to innovate above that into platforms and systems, and why that becomes important to us, is there's a lot of technology innovation we can do that enhances the offering that we can bring to market when we control the entire technology stat. >> Right. Now, we've had some other guests on and people can get more information on the nitty-gritty details of the announcement today, the main announcement. Basically, in a nutshell, enabling you to get a lot more capacity in hard drives. But, I thought in your opening remarks this morning, there were some more high-level things I wanted to dig into with you, and specifically, you made an analogy of the data economy, and compared it to the petroleum economy. I've never... A lot of times, they talk about big data, but no one really talks about it, that I've heard, in those terms, because when you think about the petroleum economy, it's so much more than fuel and cars, and the second-order impacts, and the third-order impacts on society are tremendous, and you're basically saying, "We're going to "do this all over again, but now it's based on data." >> Yeah, that's right, and I think it puts it into a form that people can understand, right? I think it's well-proven what happened around petroleum, so the discovery of petroleum, and then the derivative industries, whether it be automobiles, whether it be plastics, you pick it, the entire economy revolved around, and, to some degree, still revolves around petroleum. The same thing will occur around data. You're seeing it with investments, you hear now things like machine learning, or artificial intelligence, that is all ways to transform and mine data to create value. >> Right. >> And we're going to see industries change rapidly. Autonomous cars, that's going to be enabled by data, and capabilities here, so pick your domain. There's going to be innovation across a lot of fronts, across a lot of traditional vertical industries, that is all going to be about data and driven by data. >> It's interesting what Janet, Doctor Janet George talked about too a little bit is the types of data, and the nozzles of the data is also evolving very quickly from data at rest to data in motion, to real-time analytics, to, like you said, the machine learning and the AI, which is based on modeling prior data, but then ingesting new data, and adjusting those models so even the types and the rate and the speed of the data is under dramatic change right now. >> Yeah, that's right, and I think one of the things that we're helping enable is you kind of get to this concept of what do you need to do to do what you describe? There has to be an infrastructure there that actually enables it. So, when you think about the scale of data we're dealing with, that's one thing that we're innovating around, then the issue is, how do you allow multiple applications to simultaneously access and update and transform that? Those are all problems that need to be solved in the infrastructure to enable things like AI, right? And so, where we come into play, is creating that infrastructure layer that actually makes that possible. The other thing I talked about briefly in the Q and A was, think about the problem of a future where the data set is just too large to actually move it in a substantive way to the compute. We actually have to invert that model over time architecturally, and bring the compute to the data, right? Because it becomes too complicated and too expensive to move from the storage layer up to compute and back, right? That is a complex operation. That's why those three pillars of technology are so important. >> And you've talked, and we're seeing in the Cloud right, because this continuing kind of atomization, atomic, not automatic, but making these more atomic. A smaller unit that the Cloud has really popularized, so you need a lot, you need a little, really, by having smaller bits and bytes, it makes that that much more easy. But another concept that you delved into a little was fast data versus big data, and clearly flash has been the bright, shiny object for the last couple years, and you guys play in that market as well, but it is two very different ways to think of the data, and I thought the other statistic that was shared is you know, the amount of data coming off of the machines and people dwarfs the business data, which has been the driver of IT spend for the last several decades. >> Yeah, no, that's right, and sort of that... You think about that, and the best analogy is a broader definition of IOT, right? Where you've got all of these censors, whether it be camera censors, because that's just a censor, creating an image or a video, or if it's more industrialized too, you've got all these sources of data, and they're going to proliferate at an exponential rate, and our ability to aggregate that in some sort of an organized way, and then act upon it, again, let's use the autonomous car as the example. You've got all these censors that are in constant motion. You've got to be able to aggregate the data, and make decisions on it at the edge, so that's not something... You can't deal with latency up to the Cloud and back, if it's an automobile, and it needs to make an instantaneous decision, so you've got to create that capability locally, and so when you think about the evolution of all this, it's really the integration of the Cloud, which, as Janet talked about, is the ability to tap into this historical or legacy data to help inform a decision, but then there's things happening out at the edge that are real time, and you have to have the capability to ingest the content, make a decision on it very quickly, and then act on it. >> Right. There's a great example. We went to the autonomous... Just navigation for the autonomous vehicles. It's own subset that I think Goldman-Sachs said it a seven billion dollar industry in the not-too-distant future, and the great example is this combination of the big data and the live data is, when they actually are working on the road. So you've got maps that tell you, and are updated, kind of what the road looks like, but on Tuesday, they were shifting the lane, and that particular lane now has cones in it, so the combination of the two is such a powerful thing. >> That's right. >> I want to dive into another topic we talked about, which is really architecting for the future. Unlike oil, data doesn't get consumed and is no longer available, right? It's a reusable asset, and you talked about classic stove-topping of data within an application center world where now you want that data available for multiple applications, so very different architecture to be able to use it across many fronts, some of which you don't even know yet. >> That's right. I think that's a key point. One of the things, when we talk to CEOs, or CIOs I should say, what they're realizing, to the extent you can enable a cost-effective mechanism for me to store and keep everything, I don't know how I'll derive value from it some time in the future, because as applications evolve, we're finding new insights into what can help drive decisions or innovation, or, to take it to health care, some sort of innovation that cures disease. That's one of the things that everybody wants to do. I want to build aggregate everything. If I could do that cost effectively enough, I'll find a way to get value out of it over time, and that's something where, when we're thinking about big data and what we talked about today, that's central to that idea, and enabling it. >> Right, and digital transformation, right, the hot buzz word, but we hear, time and time again, such a big piece of that is giving the democratization. Democratization of the data, so more people have access to it, democratization of the tools to manipulate that data, not just Mahogany Row super smart people, and then to have a culture that lets people actually try, experiment, fail fast, and there's a lot of innovation that would be unlocked right within your four walls, that probably are not being tapped into. >> Well, that's right, and that's something that innovation, and an innovation culture is something that we're working hard at, right? So if you think about Western Digital, you might think of us as, you know, legacy Western Digital as sort of a fast following, very operational-centric company. We're still good at those things, but over the last five years, we've really pushed this notion of innovation, and really sort of pressing in to becoming more influential in those feature architectures. That drives a culture that, if we think about the technical community, if we create the right sort of mix of opportunity, appetite for some risk, that allows the best creativity to come out of our technical... Innovating along these lines. >> Right, I'll give you the last word. I can't believe we're going to turn the calendar here on 2017, which is a little scary. As you look forward to 2018, what are some of your top priorities? What are you going to be working on as we come into the new calendar year? >> Yeah, so as we look into 2018 and beyond, we really want to drive this continued architectural shift. You'll see us be very active, and I think you talked about it, you'll see us getting increasingly active in this democratization. So we're going to have to figure out how we engage the broader open-source development world, whether it be hardware or software. We agree with that mantra, we will support that. Obviously we can do unique development, but with some hooks and keys that we can drive a broader ecosystem movement, so that's something that's central to us, and one last word would be, one of the things that Martin Fink has talked about which is really part of our plans as we go onto the new year, is really this inverting the model, where we want to continue to drive an architecture that brings compute to the storage and enables some things that just can't be done today. >> All right, well Mike Cordano, thanks for taking a few minutes, and congratulations on the terrific event. >> Thank you. Appreciate it. >> He's Mike Cordano, I'm Jeff Frick, you're watching The Cube, we're at Western Digital headquarters in San Jose, Great Oaks Campus, it's historic. Check it out. Thanks for watching.

Published Date : Oct 11 2017

SUMMARY :

Brought to you by Western Digital. It's The Innovating to Fuel the Next Generation of Big Data Great to see you as well. That's got to feel good. Obviously more to do, as you expect and it really is a compliment to the team. of the event, so you guys are obviously in a really unique that is going to come, and as you said, more information on the nitty-gritty details of the and mine data to create value. that is all going to be about data and driven by data. to real-time analytics, to, like you said, the machine architecturally, and bring the compute to the data, right? and people dwarfs the business data, which has been talked about, is the ability to tap into this historical now has cones in it, so the combination of the two to be able to use it across many fronts, some of which that's central to that idea, and enabling it. and then to have a culture that lets people actually and really sort of pressing in to becoming more influential the new calendar year? architecture that brings compute to the storage and enables and congratulations on the terrific event. Thank you. The Cube, we're at Western Digital headquarters in San Jose,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Mike CordanoPERSON

0.99+

Western DigitalORGANIZATION

0.99+

JanetPERSON

0.99+

Jeff FrickPERSON

0.99+

San JoseLOCATION

0.99+

Janet GeorgePERSON

0.99+

TuesdayDATE

0.99+

MikePERSON

0.99+

2018DATE

0.99+

threeQUANTITY

0.99+

2017DATE

0.99+

Goldman-SachsORGANIZATION

0.99+

Silicon ValleyLOCATION

0.99+

Martin FinkPERSON

0.99+

San Jose, CaliforniaLOCATION

0.99+

Hitachi Global StorageORGANIZATION

0.99+

OneQUANTITY

0.99+

seven billion dollarQUANTITY

0.99+

twoQUANTITY

0.99+

SanDiskORGANIZATION

0.99+

The CubeTITLE

0.99+

a yearQUANTITY

0.99+

todayDATE

0.98+

three pillarsQUANTITY

0.97+

three fundamental technologiesQUANTITY

0.97+

Great Oaks CampusLOCATION

0.97+

oneQUANTITY

0.97+

FirstQUANTITY

0.95+

two very different waysQUANTITY

0.94+

one thingQUANTITY

0.93+

this morningDATE

0.92+

Next DecadeDATE

0.92+

three largeQUANTITY

0.91+

one last wordQUANTITY

0.91+

DoctorPERSON

0.87+

The CubeORGANIZATION

0.84+

third-orderQUANTITY

0.84+

last couple yearsDATE

0.83+

Mahogany RowORGANIZATION

0.82+

second-orderQUANTITY

0.81+

CloudTITLE

0.76+

last five yearsDATE

0.7+

last several decadesDATE

0.69+

four wallsQUANTITY

0.64+

Big DataEVENT

0.49+

Brendan Collins, Western Digital | Western Digital the Next Decade of Big Data 2017


 

>> Male voiceover: Live from San Jose California, it's the Cube, covering Innovating to Fuel the Next Decade of Big Data. Brought to you by Western Digital. >> Hey, welcome back everybody, Jeff Frick here with the Cube. We're at the Western Digital World Headquarters It's the Almaden Campus in San Jose. If you know anything about the tech world, you know there's a lot of innovation that's been happening on this campus for years and years and years. Big announcement today called Innovating to Fuel the Next Generation of Big Data. Lot of exciting announcements and here to join us to tell us all about it is Brendan Collins. He's the Vice President of Product Marketing Devices for Western Digital. Brendan, great to see you. >> Thank you, glad to be here. >> Absolutely so, really exciting announcement. You know, I've talked to Kim Stevenson at Intel, we had an interview talking about Moore's law. And one thing she really reinforced is that Moore's law is really more of an attitude than it is specifically physics, and whether you want to argue the physics is one thing, but the attitude for innovation, to continue to deliver a lot more for less, just continues, continues, and continues, and you guys announced a huge step in that direction today. >> Yeah, we have a challenge that storage is growing at a rate of about 40 percent per year. And budgets from the data centers are not growing, right? So the challenge is for us to develop new technologies that allow us to stay on the technology curve, and cut costs and do that efficiently. >> Then this is a big one, so let's jump in. So actually it was years ago I was actually at the event when you guys introduced the Helium drives, and that was a big deal there, and you've continued to kind of move that innovation but then you can see a plateau. And the density of this data, so you guys had to come up with something new. >> Yeah, what we've seen is that our PMR technology that we use currently is slowly running out of steam, right? So in order to come down the cost curve, we needed to boost areal density. And luckily we were able to come up with a new breakthrough in MAMR technology that will allow us to do that for the next decade. >> It's interesting in the talk, you talked about you guys could see this kind of coming and you actually put a lot of bets on the table, you didn't just bet on MAMR, you bet on HAMR, and you continued along a number of multiple tracks, and you've been at this for a while. What was kind of the innovation that finally gave you a breakthrough moment that got us to where we are today? >> Well, there were multiple technologies that we could have invested in, and we decided to continue on the two major ones which were HAMR and MAMR but we made a decision to invest in a process called, a head fabrication process called damascene that allowed us to extend the life of PMR for the last five to six years, and it's been in all the products we've been shipping since 2013. >> And you talked the areal density, so that's basically the amount of information we can put on the square inch of surface area And you've really, you attacked it on two vectors. One is how many tracks, just think of a record, how many tracks can you get on an album, in terms of the number of lines, and then how much density then you can have on each of those tracks. >> That's right, that's right. And you're now seeing major improvements on both of those factors. >> Well if you look at, we've had three enabling technologies in our products for the past three to four years, right. One is helium, one is micro actuation, and the other is the damascene process. Damascene and micro actuation actually push track density which enables higher capacity. But the newer technology that we're talking about, MAMR, addresses both factors. So we push the track density even tighter together, But we also boost the linear density at the same time, and we do that without adding cost. >> Right. The other thing you talked about, and I think it's a really important piece, right it's not only the technology breakthrough, but it's also how does that fit within the existing ecosystem of your customers, and obviously big giant data centers and big giant cloud providers, we actually have a show going on at a big cloud show right now, and this technology was innovative in that you've got a breakthrough on density, but not so crazy that you introduced a whole bunch of new factors into the ecosystem that would then have to be incorporated into all these systems, because you guys not only make your own systems, but you make the media that feeds a whole host of ecosystems, and that was a pretty important piece. >> If you look at some previous technologies we've introduced whether it be even 4K sectors in the industry, or shingled magnetic reporting, both of those require whole side modifications. Any time you have whole side modifications, it generally slows down the adoption, right? With HAMR, one of the challenges that we had was because of the concerns with thermals on the media, we needed a process called wear leveling, and that required whole software changes. In contrast, when we go to MAMR, everything is seamless, everything is transparent, and it's great. >> Right. I thought it was much simpler than that. I thought just heat is bad, HAMR is heat, and MAMR is microwave, and you know, heat and efficiency and data centers and all those, kind of again, system-level concerns; heat's never a good thing in electronics. >> Well, and in the case of MAMR versus HAMR, there's like an order of magnitude difference in the temperature on the disk, which is the key concern. >> And then of course as you mentioned in the key note, this is real, you've got sample units going on, correct me if I'm wrong, as early as next year >> That's right. >> you're hoping you'd be in scale production in 2020. Where some of these other competing technologies, there's really still no forecasted ship date on the horizon. >> Yeah, you can generate samples, you can build lower quantities of these HAMR drives, but you still have that big concern out there in front of you, how do I address the reliability, how do I address the complexity of all these new materials, and then if I got all of that to work, how do I do it commercially because of the cost additives. >> Right; so I just want to get your perspective before we let you go, you're busy, there's a high demand for your time, as you kind of think back and look at these increasing demands for storage, this increasing demand for computers, and I think one of the data points given is, you know, the data required for humans and machines and IOT is growing way way way way faster than business focused data which has been the driver of a lot of this stuff, if you just kind of sit back and take a look, you know, what are some of your thoughts because I'm sure not that long ago you could have never imagined that there would be the demand for the types of capacities that we're talking about now and we both know that when we sit down five years from now, ten years ago, you know, ten years from now, we're going to look back at today and think, you know, that was zero. >> Yeah, way back in the day there were just PCs and servers and there was traditional IT with rate, today with autonomous cars and IOT and AI and machine learning, it's just going to continue, so that exponential growth that you saw, there's no sign of that slowing down, which is good news for us. >> Yeah, good job security for you for sure. >> You bet! >> Alright Brendan, well, again, thanks for taking a few minutes to sit down and congratulations on the great event and the launch of these new products. >> Thank you, thank you. >> He's Brendan Collins, I'm Jeff Frick, you're watching the Cube from the Western Digital Headquarters in San Jose California. Thanks for watching.

Published Date : Oct 11 2017

SUMMARY :

Brought to you by Western Digital. and here to join us to tell us all about it and you guys announced a huge step in that direction today. and cut costs and do that efficiently. and that was a big deal there, that we use currently and you actually put a lot of bets on the table, and it's been in all the products and then how much density then you can have And you're now seeing major improvements and the other is the damascene process. but not so crazy that you introduced and that required whole software changes. and you know, heat and efficiency and data centers Well, and in the case of MAMR versus HAMR, Where some of these other competing technologies, and then if I got all of that to work, and we both know that when we sit down five years from now, so that exponential growth that you saw, for you for sure. and the launch of these new products. Western Digital Headquarters in San Jose California.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Brendan CollinsPERSON

0.99+

Kim StevensonPERSON

0.99+

Jeff FrickPERSON

0.99+

Western DigitalORGANIZATION

0.99+

BrendanPERSON

0.99+

2020DATE

0.99+

IntelORGANIZATION

0.99+

San Jose CaliforniaLOCATION

0.99+

MoorePERSON

0.99+

two vectorsQUANTITY

0.99+

San JoseLOCATION

0.99+

bothQUANTITY

0.99+

both factorsQUANTITY

0.99+

2013DATE

0.99+

next yearDATE

0.99+

eachQUANTITY

0.98+

OneQUANTITY

0.98+

next decadeDATE

0.98+

oneQUANTITY

0.98+

todayDATE

0.97+

ten years agoDATE

0.97+

Next DecadeDATE

0.95+

two major onesQUANTITY

0.95+

three enabling technologiesQUANTITY

0.94+

zeroQUANTITY

0.94+

Almaden CampusLOCATION

0.92+

MAMRORGANIZATION

0.92+

one thingQUANTITY

0.91+

five yearsQUANTITY

0.89+

years agoDATE

0.89+

Vice PresidentPERSON

0.88+

four yearsQUANTITY

0.85+

six yearsQUANTITY

0.84+

about 40 percent per yearQUANTITY

0.8+

CubeCOMMERCIAL_ITEM

0.76+

yearsQUANTITY

0.76+

2017DATE

0.75+

NextDATE

0.74+

HAMRORGANIZATION

0.73+

Marketing DevicesPERSON

0.72+

MAMROTHER

0.71+

InnovatingTITLE

0.7+

bunch of new factorsQUANTITY

0.68+

WesternLOCATION

0.67+

ten yearsDATE

0.61+

threeQUANTITY

0.61+

heliumOTHER

0.6+

Innovating to FuelTITLE

0.59+

HeliumCOMMERCIAL_ITEM

0.58+

DigitalORGANIZATION

0.58+

fiveQUANTITY

0.55+

Big DataEVENT

0.53+

DecadeTITLE

0.5+

Big DataTITLE

0.49+

pastDATE

0.45+

4KOTHER

0.44+

lastQUANTITY

0.4+

Dave Tang, Western Digital – When IoT Met AI: The Intelligence of Things - #theCUBE


 

>> Presenter: From the Fairmont Hotel, in the heart of Silicon Valley, it's theCUBE. Covering When IoT Met AI The Intelligence of Things. Brought to you by Western Digital. >> Hey welcome back everybody, Jeff Frick here with theCUBE. We're in downtown San Jose at the Fairmont Hotel, at an event called When IoT Met AI The Intelligence of Things. You've heard about the internet of things, and on the intelligence of things, it's IoT, it's AI, it's AR, all this stuff is really coming to play, it's very interesting space, still a lot of start-up activity, still a lot of big companies making plays in this space. So we're excited to be here, and really joined by our host, big thanks to Western Digital for hosting this event with WDLabs' Dave Tang. Got newly promoted since last we spoke. The SVP of corporate marketing and communications, for Western Digital, Dave great to see you as usual. >> Well, great to be here, thanks. >> So I don't think the need for more storage is going down anytime soon, that's kind of my takeaway. >> No, no, yeah. If this wall of data just keeps growing. >> Yeah, I think the term we had yesterday at the Ag event that we were at, also sponsored by you, is really the flood of data using an agricultural term. But it's pretty fascinating, as more, and more, and more data is not only coming off the sensors, but coming off the people, and used in so many more ways. >> That's right, yeah we see it as a virtual cycle, you create more data, you find more uses for that data to harness the power and unleash the promise of that data, and then you create even more data. So, when that virtual cycle of creating more, and finding more uses of it, and yeah one of the things that we find interesting, that's related to this event with IoT and AI, is this notion that data is falling into two general categories. There's big data, and there's fast data. So, big data I think everyone is quite familiar with by this time, these large aggregated likes of data that you can extract information out of. Look for insights and connections between data, predict the future, and create more prescriptive recommendations, right? >> Right. >> And through all of that you can gain algorithms that help to make predictions, or can help machines run based on that data. So we've gone through this phase where we focused a lot on how we harness big data, but now we're taking these algorithms that we've gleaned from that, and we're able to put them in real time applications, and that's sort of been the birth of fast data, it's been really-- >> Right, the streaming data. We cover Spark Summit, we cover Flink, and New, a new kind of open source project that came out of Berlin. That some people would say the next generation of Spark, and the other thing, you know, good for you guys, is that it used to be, not only was it old data, but it was a sampling of old data. Now on this new data, and the data stream that's all of the data. And I would actually challenge, I wonder if that separation as you describe, will stay, because I got to tell you, the last little drive I bought, just last week, was an SSD drive, you know, one terabyte. I needed some storage, and I had a choice between spinning disc and not, and I went with the flat. I mean, 'cause what's fascinating to me, is the second order benefits that we keep hearing time, and time, and time again, once people become a data-driven enterprise, are way more than just that kind of top-level thing that they thought. >> Exactly, and that's sort of that virtual cycle, you got to taste, and you learn how to use it, and then you want more. >> Jeff: Right, right. >> And that's the great thing about the breadth of technologies and products that Western Digital has, is from the solid state products, the higher performance flash products that we have, to the higher capacity helium-filled drive technologies, as well as devices going on up into systems, we cover this whole spectrum of fast data and big data. >> Right, right. >> I'll give an example. So credit card fraud detection is an interesting area. Billions of dollars potentially being lost there. Well to learn how to predict when transactions are fraudulent, you have to study massive amounts of data. Billions of transactions, so that's the big data side of it, and then as soon as you do that, you can take those algorithms and run them in real time. So as transactions come in for authorization, those algorithms can determine, before they're approved, that one's fraudulent, and that one's not. Save a lot of time and processing for fraud claims. So that's a great example of once you learn something from big data, you apply it to the real-time realm, and it's quite dire right? And then that spawned you to collect even more data, because you want to find new applications and new uses. >> Right, and too kind of this wave of computing back and forth from the shared services computer, then the desktop computer, now it's back to the cloud, and then now it's-- >> Dave: Out with the edge. >> IoT, it's all about the edge. >> Yeah, right. >> And at the end of the day, it's going to be application-specific. What needs to be processed locally, what needs to be processed back at the computer, and then all the different platforms. We were again at a navigation for autonomous vehicles show, who knew there was such a thing that small? And even the attributes of the storage required in the ecosystem of a car, right? And the environmental conditions-- >> That's right. >> Is the word I'm looking for. Completely different, new opportunity, kind of new class of hardware required to operate in that environment, and again that still combines cloud and Edge, sensors and maps. So just the, I don't think that the man's going down David. >> Yeah, absolutely >> I think you're in a good spot. (Jeff laughing) >> You're absolutely right, and even though we try to simplify into fast data, and big data, and Core and Edge, what we're finding is that applications are increasingly specialized, and have specialized needs in terms of the type of data. Is it large amounts of data, is it streaming? You know, what are the performance characteristics, and how is it being transformed, what's the compute aspect of it? And what we're finding, is that the days of general-purpose compute and storage, and memory platforms, are fading, and we're getting into environments with increasingly specialized architectures, across all those elements. Compute, memory and storage. So that's what's really exciting to be in our spot in the industry, is that we're looking at creating the future by developing new technologies that continue to fuel that growth even further, and fuel the uses of data even further. >> And fascinating just the ongoing case of Moore's law, which I know is not, you know you're not making microprocessors, but I think it's so powerful. Moore's law really is a philosophy, as opposed to an architectural spec. Just this relentless pace of innovation, and you guys just continue to push the envelope. So what are your kind of priorities? I can't believe we're halfway through 2017 already, but for kind of the balance of the year kind of, what are some of your top-of-mind things? I know it's exciting times, you're going through the merger, you know, the company is in a great space. What are your kind of top priorities for the next several months? >> Well, so, I think as a company that has gone through serial acquisitions and integrations, of course we're continuing to drive the transformation of the overall business. >> But the fun stuff right? It's not to increase your staff (Jeff laughing). >> Right, yeah, that is the hardware. >> Stitching together the European systems. >> But yeah, the fun stuff includes pushing the limits even further with solid state technologies, with our 3D NAND technologies. You know, we're leading the industry in 64 layer 3D NAND, and just yesterday we announced a 96 layer 3D NAND. So pushing those limits even further, so that we can provide higher capacities in smaller footprints, lower power, in mobile devices and out on the Edge, to drive all these exciting opportunities in IoT an AI. >> It's crazy, it's crazy. >> Yeah it is, yeah. >> You know, terabyte SD cards, terabyte Micro SD cards, I mean the amount of power that you guys pack into these smaller and smaller packages, it's magical. I mean it's absolutely magic. >> Yeah, and the same goes on the other end of the spectrum, with high-capacity devices. Our helium-filled drives are getting higher and higher capacity, 10, 12, 14 terabyte high-capacity devices for that big data core, that all the data has to end up with at some point. So we're trying to keep a balance of pushing the limits on both ends. >> Alright, well Dave, thanks for taking a few minutes out of your busy day, and congratulations on all your success. >> Great, good to be here. >> Alright, he's Dave Tang from Western Digital, he's changing your world, my world, and everyone else's. We're here in San Jose, you're watching theCUBE, thanks for watching.

Published Date : Jul 3 2017

SUMMARY :

in the heart of Silicon Valley, it's theCUBE. and on the intelligence of things, is going down anytime soon, that's kind of my takeaway. If this wall of data just keeps growing. is not only coming off the sensors, and then you create even more data. and that's sort of been the birth of fast data, and the other thing, you know, good for you guys, and then you want more. And that's the great thing about the breadth and then as soon as you do that, And at the end of the day, and again that still combines cloud and Edge, I think you're in a good spot. is that the days of general-purpose compute and storage, but for kind of the balance of the year kind of, of the overall business. But the fun stuff right? in mobile devices and out on the Edge, I mean the amount of power that you guys pack that all the data has to end up with at some point. and congratulations on all your success. and everyone else's.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Jeff FrickPERSON

0.99+

Dave TangPERSON

0.99+

JeffPERSON

0.99+

San JoseLOCATION

0.99+

Western DigitalORGANIZATION

0.99+

DavePERSON

0.99+

12QUANTITY

0.99+

10QUANTITY

0.99+

BerlinLOCATION

0.99+

Silicon ValleyLOCATION

0.99+

DavidPERSON

0.99+

yesterdayDATE

0.99+

last weekDATE

0.99+

2017DATE

0.99+

second orderQUANTITY

0.99+

both endsQUANTITY

0.98+

Billions of dollarsQUANTITY

0.98+

one terabyteQUANTITY

0.97+

FlinkORGANIZATION

0.96+

The Intelligence of ThingsTITLE

0.95+

14 terabyteQUANTITY

0.95+

AgEVENT

0.94+

oneQUANTITY

0.94+

two general categoriesQUANTITY

0.91+

EuropeanOTHER

0.9+

theCUBEORGANIZATION

0.87+

#theCUBEORGANIZATION

0.87+

Billions of transactionsQUANTITY

0.87+

Fairmont HotelLOCATION

0.87+

WDLabs'ORGANIZATION

0.81+

64QUANTITY

0.81+

Spark SummitEVENT

0.71+

96 layerQUANTITY

0.67+

MoorePERSON

0.66+

yonePERSON

0.66+

next several monthsDATE

0.64+

CoreORGANIZATION

0.59+

EdgeTITLE

0.58+

terabyteORGANIZATION

0.55+

layer 3DOTHER

0.55+

SparkTITLE

0.46+

theCUBETITLE

0.42+

When IoTTITLE

0.36+

3DQUANTITY

0.26+

Janet George, Western Digital –When IoT Met AI: The Intelligence of Things - #theCUBE


 

(upbeat electronic music) >> Narrator: From the Fairmont Hotel in the heart of Silicon Valley, it's theCUBE. Covering when IoT met AI, The Intelligence of Things. Brought to you by Western Digital. >> Welcome back here everybody, Jeff Frick here with theCUBE. We are at downtown San Jose at the Fairmont Hotel. When IoT met AI it happened right here, you saw it first. The Intelligence of Things, a really interesting event put on by readwrite and Western Digital and we are really excited to welcome back a many time CUBE alumni and always a fan favorite, she's Janet George. She's Fellow & Chief Data Officer of Western Digital. Janet, great to see you. >> Thank you, thank you. >> So, as I asked you when you sat down, you're always working on cool things. You're always kind of at the cutting edge. So, what have you been playing with lately? >> Lately I have been working on neural networks and TensorFlow. So really trying to study and understand the behaviors and patterns of neural networks, how they work and then unleashing our data at it. So trying to figure out how it's training through our data, how many nets there are, and then trying to figure out what results it's coming with. What are the predictions? Looking at how the predictions are, whether the predictions are accurate or less accurate and then validating the predictions to make it more accurate, and so on and so forth. >> So it's interesting. It's a different tool, so you're learning the tool itself. >> Yes. >> And you're learning the underlying technology behind the tool. >> Yes. >> And then testing it actually against some of the other tools that you guys have, I mean obviously you guys have been doing- >> That's right. >> Mean time between failure analysis for a long long time. >> That's right, that's right. >> So, first off, kind of experience with the tool, how is it different? >> So with machine learning, fundamentally we have to go into feature extraction. So you have to figure out all the features and then you use the features for predictions. With neural networks you can throw all the raw data at it. It's in fact data-agnostic. So you don't have to spend enormous amounts of time trying to detect the features. Like for example, If you throw hundreds of cat images at the neural network, the neural network will figure out image features of the cat; the nose, the eyes, the ears and so on and so forth. And once it trains itself through a series of iterations, you can throw a lot of deranged cats at the neural network and it's still going to figure out what the features of a real cat is. >> Right. >> And it will predict the cat correctly. >> Right. So then, how does that apply to, you know, the more specific use case in terms of your failure analysis? >> Yeah. So we have failures and we have multiple failures. Some failures through through the human eye, it's very obvious, right? But humans get tired, and over a period of time we can't endure looking at hundreds and millions of failures, right? And some failures are interconnected. So there is a relationship between these failure patterns or there is a correlation between two failures, right? It could be an edge failure. It could a radial failure, eye pattern type failure. It could be a radial failure. So these failures, for us as humans, we can't escape. >> Right. >> And we used to be able to take these failures and train them at scale and then predict. Now with neural networks, we don't have to take and do all that. We don't have to extract these labels and try to show them what these failures look like. Training is almost like throwing a lot of data at the neural networks. >> So it almost sounds like kind of the promise of the data lake if you will. >> Yes. >> If you have heard about, from the Hadoop Summit- >> Yes, yes, yes. >> For ever and ever and ever. Right? You dump it all in and insights will flow. But we found, often, that that's not true. You need hypothesis. >> Yes, yes. >> You need to structure and get it going. But what you're describing though, sounds much more along kind of that vision. >> Yes, very much so. Now, the only caveat is you need some labels, right? If there is no label on the failure data, it's very difficult for the neural networks to figure out what the failure is. >> Jeff: Right. >> So you have to give it some labels to understand what patterns it should learn. >> Right. >> Right, and that is where the domain experts come in. So we train it with labeled data. So if you are training with a cat, you know the features of a cat, right? In the industrial world, cat is really what's in the heads of people. The domain knowledge is not so authoritative. Like the sky or the animals or the cat. >> Jeff: Right. >> The domain knowledge is much more embedded in the brains of the people who are working. And so we have to extract that domain knowledge into labels. And then you're able to scale the domain. >> Jeff: Right. >> Through the neural network. >> So okay so then how does it then compare with the other tools that you've used in the past? In terms of, obviously the process is very different, but in terms of just pure performance? What are you finding? >> So we are finding very good performance and actually we are finding very good accuracy. Right? So once it's trained, and it's doing very well on the failure patterns, it's getting it right 90% of the time, right? >> Really? >> Yes, but in a machine learning program, what happens is sometimes the model is over-fitted or it's under-fitted or there is bias in the model and you got to remove the bias in the model or you got to figure out, well, is the model false-positive or false-negative? You got to optimize for something, right? >> Right, right. >> Because we are really dealing with mathematical approximation, we are not dealing with preciseness, we are not dealing with exactness. >> Right, right. >> In neural networks, actually, it's pretty good, because it's actually always dealing with accuracy. It's not dealing with precision, right? So it's accurate most of the time. >> Interesting, because that's often what's common about the kind of difference between computer science and statistics, right? >> Yes. >> Computers is binary. Statistics always has a kind of a confidence interval. But what you're describing, it sounds like the confidence is tightening up to such a degree that it's almost reaching binary. >> Yeah, yeah, exactly. And see, brute force is good when your traditional computing programing paradigm is very brute force type paradigm, right? The traditional paradigm is very good when the problems are simpler. But when the problems are of scale, like you're talking 70 petabytes of data or you're talking 70 billion roles, right? Find all these patterns in that, right? >> Jeff: Right. >> I mean you just, the scale at which that operates and at the scale at which traditional machine learning even works is quite different from how neural networks work. >> Jeff: Okay. >> Right? Traditional machine learning you still have to do some feature extraction. You still have to say "Oh I can't." Otherwise you are going to have dimensionality issues, right? It's too broad to get the prediction anywhere close. >> Right. >> Right? And so you want to reduce the dimensionality to get a better prediction. But here you don't have to worry about dimensionality. You just have to make sure the labels are right. >> Right, right. So as you dig deeper into this tool and expose all these new capabilities, what do you look forward to? What can you do that you couldn't do before? >> It's interesting because it's grossly underestimating the human brain, right? The human brain is supremely powerful in all aspects, right? And there is a great deal of difficulty in trying to code the human brain, right? But with neural networks and because of the various propagation layers and the ability to move through these networks we are coming closer and closer, right? So one example: When you think about driving, recently, Google driverless car got into an accident, right? And where it got into an accident was the driverless car was merging into a lane and there was a bus and it collided with the bus. So where did A.I. go wrong? Now if you train an A.I., birds can fly, and then you say penguin is a bird, it is going to assume penguin can fly. >> Jeff: Right, right. >> We as humans know penguin is a bird but it can't fly like other birds, right? >> Jeff: Right. >> It's that anomaly thing, right? Naturally when are driving and a bus shows up, even if it's yield, the bus goes. >> Jeff: Right, right. >> We yield to the bus because it's bigger and we know that. >> A.I. doesn't know that. It was taught that yield is yield. >> Right, right. >> So it collided with the bus. But the beauty is now large fleets of cars can learn very quickly based on what it just got from that one car. >> Right, right. >> So now there are pros and cons. So think about you driving down Highway 85 and there is a collision, it's Sunday morning, you don't know about the collision. You're coming down on the hill, right? Blind corner and boom that's how these crashes happen and so many people died, right? If you were driving a driverless car, you would have knowledge from the fleet and from everywhere else. >> Right. >> So you know ahead of time. We don't talk to each other when we are in cars. We don't have universal knowledge, right? >> Car-to-car communication. >> Car-to-car communications and A.I. has that so directly it can save accidents. It can save people from dying, right? But people still feel, it's a psychology thing, people still feel very unsafe in a driverless car, right? So we have to get over- >> Well they will get over that. They feel plenty safe in a driverless airplane, right? >> That's right. Or in a driveless light rail. >> Jeff: Right. >> Or, you know, when somebody else is driving they're fine with the driver who's driving. You just sit in the driver's car. >> But there's that one pesky autonomous car problem, when the pedestrian won't go. >> Yeah. >> And the car is stopped it's like a friendly battle-lock. >> That's right, that's right. >> Well good stuff Janet and always great to see you. I'm sure we will see you very shortly 'cause you are at all the great big data conferences. >> Thank you. >> Thanks for taking a few minutes out of your day. >> Thank you. >> Alright she is Janet George, she is the smartest lady at Western Digital, perhaps in Silicon Valley. We're not sure but we feel pretty confident. I am Jeff Frick and you're watching theCUBE from When IoT meets AI: The Intelligence of Things. We will be right back after this short break. Thanks for watching. (upbeat electronic music)

Published Date : Jul 2 2017

SUMMARY :

Brought to you by Western Digital. We are at downtown San Jose at the Fairmont Hotel. So, what have you been playing with lately? Looking at how the predictions are, So it's interesting. behind the tool. So you have to figure out all the features So then, how does that apply to, you know, So these failures, for us as humans, we can't escape. at the neural networks. the promise of the data lake if you will. But we found, often, that that's not true. But what you're describing though, sounds much more Now, the only caveat is you need some labels, right? So you have to give it some labels to understand So if you are training with a cat, in the brains of the people who are working. So we are finding very good performance we are not dealing with preciseness, So it's accurate most of the time. But what you're describing, it sounds like the confidence the problems are simpler. and at the scale at which traditional machine learning Traditional machine learning you still have to But here you don't have to worry about dimensionality. So as you dig deeper into this tool and because of the various propagation layers even if it's yield, the bus goes. It was taught that yield is yield. So it collided with the bus. So think about you driving down Highway 85 So you know ahead of time. So we have to get over- Well they will get over that. That's right. You just sit in the driver's car. But there's that one pesky autonomous car problem, I'm sure we will see you very shortly 'cause you are Alright she is Janet George, she is the smartest lady

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JeffPERSON

0.99+

Jeff FrickPERSON

0.99+

Janet GeorgePERSON

0.99+

JanetPERSON

0.99+

Western DigitalORGANIZATION

0.99+

90%QUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

one carQUANTITY

0.99+

Highway 85LOCATION

0.99+

Sunday morningDATE

0.99+

two failuresQUANTITY

0.99+

70 billion rolesQUANTITY

0.99+

GoogleORGANIZATION

0.99+

CUBEORGANIZATION

0.98+

one exampleQUANTITY

0.96+

The Intelligence of ThingsTITLE

0.94+

hundreds of cat imagesQUANTITY

0.93+

firstQUANTITY

0.92+

theCUBEORGANIZATION

0.84+

San JoseLOCATION

0.8+

one pesky autonomous carQUANTITY

0.77+

70 petabytes of dataQUANTITY

0.77+

hundreds andQUANTITY

0.76+

IoTORGANIZATION

0.74+

millions of failuresQUANTITY

0.66+

Fairmont HotelLOCATION

0.66+

ollisionPERSON

0.65+

meetsTITLE

0.64+

#theCUBEORGANIZATION

0.57+

Hadoop SummitEVENT

0.51+

ofTITLE

0.47+

Western Digital Taking the Cloud to the Edge, Panel 2 | DataMakesPossible


 

>> They are disruptive technologies. And if you think about the disruption that's happening in business, with IoT, with OT, and with big data, you can't get anything more disruptive to the whole of the business chain as this particular area. It's an area that I focused on myself, asking the question, should everything go to the cloud? Is that the new future? Is 90% of the computing going to go to the cloud with just little mobile devices right on the edge? Felt wrong when I did the math on it, I did some examples of real-world environments, wind farms, et cetera, it clearly was not the right answer, things need to be near the edge. And I think one of the areas to me that solidified it was when you looked at an area like video. Huge amounts of data, real important decisions being made on the content of that video, for example, recognizing a face, a white hat or a black hat. If you look at the technology, sending that data somewhere to do that recognition just does not make sense. Where is it going? It's going actually into the camera itself, right next to the data, because that's where you have the raw data, that's where you have the maximum granularity of data, that's where you need to do the processing of which faces are which, right close to the edge itself, and then you can send the other data back up to the cloud, for example, to improve those algorithms within that camera, to do all that sort of work on the batch basis over time, that's what I was looking at, and looking at the cost justification for doing that sort of work. So today, we've got a set people here on the panel, and we want to talk about coming down one level to where IoT and IT are going to have to connect together. So on the panel I've got, I'm going to get these names really wrong, Sanjeev Kumar? >> Yes, that's right. >> From FogHorn, could you introduce yourself and what you're doing where the data is meeting the people and the machines? >> Sure, sure, so my name is Sanjeev Kumar, I actually run engineering for a company called FogHorn Systems, we are actually bringing analytics and machine learning to the edge, and, so our goal and motto is to take computing to where the data is, than the other way around. So it's a two-year-old company that started, was incubated in the hive, and we are in the process of getting our second release of the product out shortly. >> Excellent, so let me start at the other end, Rohan, can you talk about your company and what contribution you're focusing on? >> Sure, I'm head product marketing for Maana, Maana is a startup, about three years old, what we're doing is we're offering an enterprise platform for large enterprises, we're helping the likes of Shell and Maersk and Chevron digitally transform, and that simply means putting the focus on subject matter experts, putting the focus on the people, and data's definitely an important part of it, but allowing them to bring their expertise into the decision flows, so that ultimately the key decisions that are driving the revenue for these behemoths, are made at a higher quality and faster. >> Excellent. Well, two software companies, we have a practitioner here who is actually doing fog computing, doing it for real, has been doing it for some time, so could you like, Janet George from Western Digital, can you introduce yourself, and say something from the trenches, of what's really going on? >> Okay, very good, thank you. I actually build infrastructure for the edge to deal with fog computing, and so for Western Digital, we're very lucky, because we are the largest storage manufacture, and we have what we call Internet of Things, and Internet of Test Equipment, and I process petabytes of data that comes out of the Internet of Things, which is basically our factories, and then I take these petabytes of data, I process them both on the cloud and then on the edge, but primarily, to be able to consume that data. And the way we consume that data is by building very high-profile models through artificial intelligence and machine learning, and I'll talk a lot more about that, but at the end of the day, it's all about consuming the data that you collect from anywhere, Internet of Things, computer equipment, data that's being produced through products, you have to figure out a way to compute that, and the cloud has many advantages and many trade-offs, and so we're going to talk about the trade-offs, that's where the gap for computing comes into play. >> Excellent, thanks very much. And last but not least, we have Val, and I can never pronounce your surname. >> Bercovici. >> Thank you. (chuckling) You are in the midst of a transition yourself, so talk about where you have been and where you're going. >> For the better part of this century, I've been with NetApp, working at various functions, obviously enterprise storage, and around 2008, my developer instinct kind of fired up, and this thing called cloud became very interesting to me. So I became a self-anointed cloud czar at NetApp, and I ended up initiating a lot of our projects which we know today as the NetApp Data Fabric, that culminated about 18 months ago, in acquisition of SolidFire, and I'm now the acting CTO of SolidFire, but I plan to retire from the storage industry at the end of our fiscal year, at the end of April, and I'm spending a lot of time with particularly the Cloud Native Compute Foundation, that is, the opensource home of Google's Kubernetes Technology and about seven other related projects, we keep adding some almost every month, and I'm starting to lose track, and spending a lot of time on the data gravity challenge. It's a challenge in the cloud, it's a particularly new and interesting challenge at the edge, and I look forward to talking about that. >> Okay, and data gravity is absolutely key, isn't it, it's extremely expensive and extremely heavy to move around. >> And the best analogy is workloads are like electricity, they move fairly easily and lightly, data's like water, it's really hard to move, particularly large bodies around. >> Great. I want to start with one question though, just in the problem, the core problem, particularly in established industries, of how do we get change to work? In an IT shop, we have enough problems dealing with operations and development. In the industrial world, we have the IT and the OT, who look at each other with less than pleasure, and mainly disdain. How do we solve the people problem in trying to put together solutions? You must be right in the middle of it, would you like to start with that question? >> Absolutely, so we are 26 years old, probably more than that, but we have very old and new mix of manufacturing equipment, it's a storage industry, and in our storage industry, we are used to doing things a certain way. We have existing data, we have historical data, we have trend data, you can't get rid of what you already have. The goal is to make connectors such that you can move from where you're at to where you're going, and so you have to be able to take care of the shift that is happening in the market, so at the end of the day, if you look at five years from now, it's all going to be machine learning and AI, right? Agent technology's already here, it's proven, we can see, Siri is out here, we can see Alexa, we can see these agent technologies out there, so machine learning is a getting a lot of momentum, deep learning and neural networks, things like that. So we got to be able to look at that data and tap into our data, near realistically, very different, and the way to do that is really making these connections happen, tapping into old versus new. Like for example, if you look at storage, you have file storage, you have block storage, and then you have object storage, right? We've not really tapped into the field of object storage, and the reason is because if you are going to process one trillion objects like Amazon is doing right now with S3, you can't do it with the file system level storage or with the blog system level storage, you have to go to objects. Think Internet of Things. How many trillions of objects are going to come out of these Internet of Things? So one, you have to be positioned from an infrastructure standpoint. Two, you have to be positioned from a use case prototyping perspective, and three, you got to be able to scale that very rapidly, very quickly, and that's how change happens, change does not happen because you ask somebody to change their behavior, change happens when you show value, and people are so eager to get that value out of what you've shown them in real life, that they are so quick to adapt. >> That's an excellent-- >> If I could comment on that as well, which is, we just got through training a bunch of OT guys on our software, and two analogies that actually work very well, one is sort of, the operational people are very familiar with circuit diagrams, and so, and sort of, flow of things through essentially black boxes, you can think of these as something that has a bunch of inputs and has a bunch of outputs. So that's one thing that worked very well. The second thing that works very well is the PLC model, and there are direct analogies between PLC's and analytics, which people on the floor can actually relate to. So if you have software that's basically based on data streams and time, as a first-class citizen, the PLC model again works very well in terms of explaining the new software to the OT people. >> Excellent, okay, would you want to come in on that as well? >> Sure, I think a couple of points to add to what Janet said, I couldn't agree more in terms of the result, I think Maana did a few projects, a few pilots to convince customers of their value, and we typically focus very heavily on operationalizing the output, so we are very focused on making sure that there is some measurable value that comes out of it, and it's not until the end user started seeing that value that they were willing and open to adopt the newer methodologies. A second point to that is, a lot of the more recent techniques available to solve certain challenges, there are deep learning neural nets there's all sorts of sophisticated AI and machine learning algorithms that are out there, a lot of these are very sophisticated in their ability to deliver results, but not necessarily in the transparency of how you got that, and I think that's another thing that Maana's learning, is yes, we have this arsenal of fantastic algorithms to throw at problems, but we try to start with the simplest approach first, we don't unnecessarily try to brute force, because I think an enterprise, they are more than willing to have that transparency in how they're solving something, so if they're able to see how they were able to get to us, how the software was able to get to a certain conclusion, then they are a lot happier with that approach. >> Could you maybe just give one example, a real-world example, make it a little bit real? >> Right, absolutely, so we did a project for a very large organization for collections, they have a lot of outstanding capital locked up and customers not paying, it's a standard problem, you're going to find it in pretty much any industry, and so for that outstanding invoice, what we did was we went ahead and we worked with the subject matter experts, we looked at all the historical accounts receivable data, we took data from a lot of other sources, and we were able to come up with models to predict when certain customers are likely to pay, and when they should be contacted. Ultimately, what we wanted to give the collection agent were a list of customers to call. It was fairly straightforward, of course, the solution was not very, very easy, but at least on a holistic level, it made a lot of sense to us. When we went to the collection agents, many of them actually refused to use that approach, and this is part of change management in some sense, they were so used to doing things their way, they were so used to trying to target the customers with the largest outstanding invoice, or the ones that hadn't paid for the longest amount of time, that it actually took us a while, because initially, what the feedback we got was that your approach is not working, we're not seeing the results. And when we dug into it, it was because it wasn't being used, so that would be one example. >> So again, proof points that you will actually get results from this. >> Absolutely, and the transparency, I think we actually sent some of our engineers to work with the collections agents to help them understand what approach is it that we're taking, and we showed them that this is not magic, we're actually, instead of looking at the final dollar value, we're looking, we're calculating time value lost, so we are coming up with a metric that allows us to incorporate not just the outstanding amount, or the time that they haven't paid for, but a lot of other factors as well. >> Excellent, Val. >> When you asked that question, I immediately went to more of a nontechnical business side of my brain to answer it, so my experience over the years has been particularly during major industry transitions, I'm old enough to remember the mainframe to client server transition, and now client server to virtualization and cloud, and really, sales reps have that well-earned reputation of being coin-operated, though it's remarkable how much you can adjust compensation plans for pretty much anyone, in a capitalist environment, and the IT/OT divide, if you will, is pretty easy to solve from a business perspective when you take someone with an IT supporting the business mentality, and you compensate them on new revenue streams, new business, all of a sudden, the world perspective changes sometimes overnight, or certainly when that contract is signed. That's probably the number one thing you can do from a people perspective, is incent them and motivate them to focus on these new things, the technology is, particularly nowadays is evolving to support them for these new initiatives, but nothing motivates like the right compensation plan. >> Excellent, a great series of different viewpoints. So the second question I have again coming down a bit to this level, is how do we architect a solution? We heard you got to architect it, and you've got less, like this, it seems to me that that's pretty difficult to do ahead of where you're going, that in general, you take smaller steps, one step at a time, you solve one problem, you go on to the next. Am I right in that? If I am, how would you suggest the people go about this decision-making of putting architectures together, and if you think I'm wrong and you have a great new way of doing it, I'd love to hear about it. >> I can take a shorter route. So we have a number of customers that are trying to adopt, are going through a phased way of adopting our technology and products, and so it begins with first gathering of the data, and replaying it back, to build the first level of confidence, in the sense that the product is actually doing what you're expecting it to do. So that's more from monitoring administration standpoint. The second stage is you should begin to capture analytical logic into the project, where it can start doing prediction for you, so you go into, so from operational, you go into a predictive maintenance, predictive maintenance, predictive models standpoint. The third part is prescriptive, where you actually help create a machine learning model, now, it's still in flux in terms of where the model gets created, whether it's on the cloud, in a central fashion, or some sort of a, the right place, the right context in a multi-level hierarchical fog layer, and then, you sort of operationalize that as close to the data again as possible, so you go through this operational to predictive to prescriptive adoption of the technology, and that's how people actually build confidence in terms of adopting something new into, let's say, a manufacturing environment, or things that are pretty expensive, so I give you another example where you have the case of capacitors being built on a assembly line, manufacturing, and so how do you, can you look at data across different stations and manufacturing on a assembly line? And can you predict on the second station that it's going to fail on the eighth one? By that, what you're doing is you are actually reducing the scrap that's coming off of the assembly line. So, that's the kind of usage that you're going to in the second and third stage. >> Host: Excellent. Janet, do you want to go on? >> Yeah, I agree and I have a slightly different point of view also. I think architecture's very difficult, it's like Thomas Edison, he spent a lot of time creating negative knowledge to get to that positive knowledge, and so that's kind of the way it is in the trenches, we spend a lot of time trying to think through, the keyword that comes to mind is abstraction layers, because where we came from, everything was tightly coupled, and tightly coupled, computer and storage are tightly coupled, structured and unstructured data are tightly coupled, they're tightly coupled with the database, schema is tightly coupled, so now we are going into this world of everything being decoupled. In that, multiple things, multiple operating systems should be able to use your storage. Multiple models should be able to use your data. You cannot structure your data in any kind of way that is customized to one particular model. Many models have to run on that data on the fly, retrain itself, and then run again, so when you think about that, you think about what suits best to stay in the cloud, maybe large amounts of training data, schema that's already processed can stay on the cloud. Schema that is very dynamic, schema that is on the fly, that you need to read, and data that's coming at you from the Internet of Things that's changing, I call it heteroscedastic data, which is very statistical in nature, and highly variable in nature, you don't have time to sit there and create rows and columns and structure this data and put it into some sort of a structured set, you need to have a data lake, you need to have a stack on top of that data lake that can then adapt, create metadata, process that data and make it available for your models, so, and then over time, like I totally believe that now we're running into near realtime compute bottleneck, processing all this pattern processing for the different models and training sets, so we need a stack that we can quickly replace with GPUs, which is where the future is going, with pattern processing and machine learning, so your architecture has to be extremely flexible, high layers of abstraction, ability to train and grow and iterate. >> Excellent. Do you want to go next? >> So I'll be a broken record, back to data gravity, I think in an edge context, you really got to look at the cost of processing data is orders of magnitude less than moving it or even storing it, and so I think that the real urgency, I don't know, there's 90% that think of data at the edge is kind of wasted, you can filter through it and find that signal through the noise, so processing data to make sure that you're dealing with really good data at the edge first, figuring out what's worth retaining for future steps, I love the manufacturing example, I have lots of customer examples ourselves where, for quality control in a high-moving assembly line, you want to take thousands of not millions of images and compare frame and frame exactly according to the schematics where the device is compared to where it should be, or where the components, and the device compared to where they should be, processing all of that data locally and making sure you extract the maximum value before you move data to a central data lake to correlate it against other anomalies or other similarities, that's really key, so really focus on that cost of moving and storing data, yeah. >> Yes, do you want the last word? >> Sure, Maana takes an interesting approach, I'm going to up-level a little bit. Whenever we are faced with a customer or a particular problem for a customer, we try to go over the question-answer approach, so we start with taking a very specific business question, we don't look at what data sources are available, we don't ask them whether they have a data lake, or we literally get their business leaders, their subject matter experts, we literally lock them up in a room and we say, "You have to define "a very specific problem statement "from which we start working backwards," each problem statement can be then broken down into questions, and what we believe is any question can be answered by a series of models, you talked about models, we go beyond just data models, we believe anything in the real world, in the case of, let's say, manufacturing, since we're talking about it, any smallest component of a machine should be represented in the form of a concept, relationships between people operating that machinery should be represented in the form of models, and even physics equations that are going into predicting behavior should be able to represent in the form of a model, so ultimately, what that allows us is that granularity, that abstraction that you were talking about, that it shouldn't matter what the data source is, any model should be able to plug into any data source, or any more sophisticated bigger model, I'll give you an example of that, we started solving a problem of predictive maintenance for a very large customer, and while we were solving that predictive maintenance problem, we came up with a number of models to go ahead and solve that problem. We soon realized that within that enterprise, there are several related problems, for example, replacement of part inventory management, so now that you figured out which machine is going to fail at roughly what instance of time from now, we can also figure out what parts are likely to fail, so now you don't have to go ahead and order a ton of replacement parts, because you know what parts are going to likely fail, and then you can take that a step further by figuring out which equipment engineer has the skillset to go ahead and solve that particular issue. Now, all of that, in today's world, is somewhat happening in some companies, but it is actually a series of point solutions that are not talking to each other, that's where our pattern technology graph is coming into play where each and every model is actually a note on the graph including computational models, so once you build 10 models to solve that first problem, you can reuse some of them to solve the second and third, so it's a time-to-value advantage. >> Well, you've been a fantastic panel, I think these guys would like to get to a drink at the bar, and there's an opportunity to talk to you people, I think this conversation could go on for a long, long time, there's so much to learn and so much to share in this particular information. So with that, over to you! >> I'll just wrap it up real quick, thanks everyone, give the panel a hand, great job. Thanks for coming out, we have drinks for the next hour or two here, so feel free to network and mingle, great questions to ask them privately one-on-one, or just have a great conversation, and thanks for coming, we really appreciate it, for our Big Data SV Event livestreamed out, it'll be on demand on YouTube.com/siliconangle, all the video, if you want to go back, look at the presentations, go to YouTube.com/siliconangle, and of course, siliconangle.com, and Wikibond.com for the research and content coverage, so thanks for coming, one more time, big round of applause for the panel, enjoy your evening, thanks so much.

Published Date : Mar 16 2017

SUMMARY :

Is 90% of the computing going to go to the cloud of getting our second release of the product out shortly. and that simply means putting the focus so could you like, Janet George from Western Digital, consuming the data that you collect from anywhere, and I can never pronounce your surname. so talk about where you have been the acting CTO of SolidFire, but I plan to retire Okay, and data gravity is absolutely key, isn't it, And the best analogy is workloads are like electricity, would you like to start with that question? and the reason is because if you are going to process in terms of explaining the new software to the OT people. but not necessarily in the transparency of how you got that, and we were able to come up with models to predict So again, proof points that you will actually Absolutely, and the transparency, and the IT/OT divide, if you will, and if you think I'm wrong and you have a great new way and then, you sort of operationalize that Janet, do you want to go on? the keyword that comes to mind is abstraction layers, Do you want to go next? and the device compared to where they should be, and then you can take that a step further and there's an opportunity to talk to you people, all the video, if you want to go back,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Janet GeorgePERSON

0.99+

JanetPERSON

0.99+

Western DigitalORGANIZATION

0.99+

Cloud Native Compute FoundationORGANIZATION

0.99+

90%QUANTITY

0.99+

10 modelsQUANTITY

0.99+

AmazonORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

ShellORGANIZATION

0.99+

SiriTITLE

0.99+

Sanjeev KumarPERSON

0.99+

secondQUANTITY

0.99+

RohanPERSON

0.99+

one questionQUANTITY

0.99+

MaanaORGANIZATION

0.99+

FogHorn SystemsORGANIZATION

0.99+

two analogiesQUANTITY

0.99+

thousandsQUANTITY

0.99+

BercoviciPERSON

0.99+

Thomas EdisonPERSON

0.99+

second questionQUANTITY

0.99+

second stationQUANTITY

0.99+

SolidFireORGANIZATION

0.99+

FogHornORGANIZATION

0.99+

thirdQUANTITY

0.99+

firstQUANTITY

0.99+

third partQUANTITY

0.99+

todayDATE

0.99+

second thingQUANTITY

0.98+

TwoQUANTITY

0.98+

two-year-oldQUANTITY

0.98+

one problemQUANTITY

0.98+

end of AprilDATE

0.98+

AlexaTITLE

0.98+

first problemQUANTITY

0.98+

one exampleQUANTITY

0.97+

threeQUANTITY

0.97+

second releaseQUANTITY

0.97+

third stageQUANTITY

0.97+

second pointQUANTITY

0.96+

one trillion objectsQUANTITY

0.96+

second stageQUANTITY

0.96+

oneQUANTITY

0.96+

one levelQUANTITY

0.95+

two software companiesQUANTITY

0.95+

NetApp Data FabricORGANIZATION

0.95+

eachQUANTITY

0.95+

millions of imagesQUANTITY

0.95+

2008DATE

0.95+

first levelQUANTITY

0.95+

bothQUANTITY

0.95+

eighth oneQUANTITY

0.94+

S3TITLE

0.93+

trillions of objectsQUANTITY

0.92+

Wikibond.comORGANIZATION

0.92+

each problem statementQUANTITY

0.92+

one thingQUANTITY

0.92+

one stepQUANTITY

0.91+

Big Data SV EventEVENT

0.91+

siliconangle.comOTHER

0.91+

NetAppORGANIZATION

0.91+

MaerskORGANIZATION

0.9+

about three years oldQUANTITY

0.89+

five yearsQUANTITY

0.89+

MaanaPERSON

0.89+

about 18 months agoDATE

0.88+

26 years oldQUANTITY

0.82+

one particular modelQUANTITY

0.82+

Kubernetes TechnologyORGANIZATION

0.82+

ValPERSON

0.82+

everyQUANTITY

0.81+

ChevronORGANIZATION

0.79+

Panel 2QUANTITY

0.77+

seven other related projectsQUANTITY

0.7+

next hourDATE

0.69+

toTITLE

0.66+

petabytesQUANTITY

0.64+

timeQUANTITY

0.64+

twoQUANTITY

0.63+

series of modelsQUANTITY

0.52+

Western Digital Taking the Cloud to the Edge - #DataMakesPossible - Panel 1


 

>> Why don't I spend just a couple minutes talking about what we mean by digital enactment, turning data in models and models into action. And then we'll jump directly into, I'll introduce the panelists after that, and we'll jump directly into the questions. So Wikibon SiliconAngle has been on a mission for quite sometime now to really understand what is the nature of digital transformation, or digital disruption. And historically, when we've talked about digital, people talk about a variety of different characteristics of it, so we'll talk about new types of channels and activity on the web, and a many number of other things. But to really make sense of this, we kind of felt that we had to go to a set of basic principles, and utilize those basic principles to build our observations up. And so what we started with is a simple observation that, if it's not digital, or if it's not data, it ain't digital. By that we mean fundamentally the idea of digital business is how are we going to use data as an asset to differentially drive our business forward? And if we borrowed from Drucker, Drucker used to like to talk about the idea that business exists to create sustained customers, and so we would say that digital business is about applying data assets to differentially create sustained customers. Now to do that successfully, we have to be able to, as businesses, be able to establish a set of strategic business capabilities that will allow us to differentially use data assets. And we think that there are a couple of core strategic business capabilities required. One is human beings and most businesses operate in the analog world, so it's how do we take that analog data and turn it into digital data that we can then process. So that's the first one, the notion of an IOT as a transducer of information so that we can generate these very rich data streams. Secondly we have to be able to do something with those data streams, and that's the basis of big data. So we utilize big data to create models, to create insights, and increasingly through a more declarative style, actually create new types of software systems that will be crucial to driving the business forward. That's the second capability. The third capability is one that we're still coming to understand, and that is we have to take the output of those models, the output of those insights, and then turn them back into some event that has a consequential moment in the real world, or what we call systems of an action. And so the three core business capabilities that have to be built are this capture data through IOT, big data to process it, systems of an action also through IOT, through actuators, to actually that have a consequential action in the real world. So that's the basis of what we're talking about. We're going to take Flavio's vision that he just laid out, and then we, in this panel, are going to talk about some of the business capabilities necessary to make that happen, and then after this, David Foyer will lead a panel on specifically some of the lower level technologies that are going to make it work. Make sense guys? >> Sounds good (mumbles). >> Okay, so let me introduce the panelists. Over, down there on the end, Ted Connell. Ted is from Intel, I don't know if we can get the slide up that has their names and their titles. Ted, why don't you very quickly introduce yourself. >> Yeah, thank you very much. I run Solution Architecture for the manufacturing and industrial vertical, where we put together end to end ecosystem solutions that solve our clients business problems. So we're not selling silicone or semiconductors, we're solving our clients problems, which as Flavio said, requires ecosystem solutions of software, system integrators, and other partners to come together to put together end solutions. >> Excellent, next to Ted is Steve Madden of Equinix. >> Yeah, Steve Madden. Equinix is the largest interconnection, global interconnection company and a lot of the ecosystems that you'll be hearing about, come together inside our locations. And one of the things I do in there is work with our big customers on industry vertical level solutions, IOT being one of them. >> Phu Hoang, from Data Torrent. >> Hi, my name's Phu Hoang, I'm co-founder and chief strategy of a company called Data Torrent, and at Data Torrent, our mission is really to build out solutions to allow enterprises to process big data in a streaming fashion. So that whole theme around ingestion, transformation, analytics, and taking action in sub second on massive data is what we're focusing on. >> And you're familiar with Flavio. Flavio, will you take a second to introduce yourself. >> Yes, thank you, I am leading a company that is trying to manifest the vision highlighted here, building a platform. Not so much the applications, we are hosting the applications (mumbles) the data management and so forth. And trying to apply the industrial vertical first. Big enough to keep us busy for quite a while. >> So in case you didn't know this, we have an interesting panel, we have use case, application, technol infrastructure, and platform. So what' we'll try to do is over the next, say, 10 minutes or so, we're going to spend a little bit of time, again, talking about some of these business capabilities. Let me start off by asking each of you a question, and I will take, if anybody is really burning to ask a question, raise your hand, I'll do my best to see you and I'll share the microphone for just long enough for you to ask it. Okay, so first question, digital business is data. That means we have to think about data differently. Ted, at Intel, what is Intel doing when they think about data as an asset? >> So, Intel has been working on what is now being called Fog, and big data analytics for over a generation. The modern xeon server we're selling, the wire in the electronics if you will, is 10 silicon atoms wide. So to control that process, we've had to do what is called Industry 4.0 20 years ago. So all of our production equipment has been connected for 20 years, we're running... One of our factories will produce a petabyte of data a day, and we're running big data analytics, including machine learning on the stuff currently. If you look at an Intel factory, we have 2,000 fit clients on the factory floor supported by 600 servers in our data center at the factory, just to control the process and run predictive yield analytics. >> Peter: So that's your itch? >> Our competitive advantage at Intel is the factory. We are a manufacturer, we're a world class manufacturer. Our front end factories have zero people in it, not that we don't like people, but we had to fully automate the factory because as I speak, tens of thousands of water molecules are leaving my mouth, and if one of those water molecules lands on a silicon, it ain't going to work. So we had to get people physically out of the factory, and so we were forced by Moore's Law, and the product we build, to build out what became Fog, when they came up with the term seven years ago, we just came to that conclusion because of cost, latency, and security, it made sense to, you know, look, you got data, you got compute, there's a network between. It doesn't matter where you do the compute, bring the compute to the data, the data to the compute. You're doing a compute function, it doesn't matter where you do it. So Fog is not complicated, it's just a distributed data center. >> So when you think about some of the technologies necessary to make this work, it's not just batch, we're going to be doing a lot of stuff in real time, continuously. So Phu, talk a little bit about the system software, the infrastructure software that has to be put in place to ensure that this works for them. >> I think that's great. A little bit about our background, the company was founded by a bunch of ex-Yahoos that had been out for 12, 15 years from the early days. So we sort of grew up in that period where we had to learn about big data, learn about making all the mistakes of big data, and really seeing that nowadays, it's not good enough to get insight, you have to get insight in a timely fashion enough to actually do something about it. And for a lot of enterprise, especially with human being carrying around mobile phones and moving around all over the place, and sensors sending thousands, if not millions of events per second, the need for the business to understand what's going on and react, have insight and react sub second, is crucial. And what that means is the stuff that used to be batch, offline, you know, can kind of go down, now has to be continuous, 24 by seven. You can't lose data, you got to be able to recover and come back to where you were as if nothing has happened with no human intervention. There's a lot of theme around no human intervention, because this stuff is so fast, you can't involve human beings in it, then you're not reacting fast enough. >> Can I real quickly add one thing first? >> Peter: Sure. >> We think of data at Intel in half life terms. >> Yeah, that's exactly right. >> The data has valuable right now. If you wait a second, literally a second, the data has a little bit of value. You wait two second, it's historical data you can run regressions, and tell you why you screwed up, but you ain't going to fix anything. >> Exactly. >> If you want to do anything with your data, you got to do it now. >> So that, ultimately, we need to develop experience, a creed experience about what we're doing. And the stuff we're doing in applications will eventually find itself into platforms. So Flavio, talk to us a little bit about the types of things that are going to end up in the platform to ensure that these use cases are made available to, certainly, businesses that perhaps aren't as sophisticated as Intel. >> Yes, so in many ways, we are learning from what is going on in the Cloud, and has to come through this continuum, all the way into the machines. This break between what's going inside the machine, and old 1980 microprocessor and the server, and the Cloud server with virtualization on the other side cannot leave. So it has to be a continuum of computing so you can move the same function, the same container, all the way through first. Second, you really have to take the real time very, very seriously, particularly at the edge, but even in the back so that when you have these end to end continuum, you can decide where you do what. And I think that one of the models that was in that picture with a concentric circle is really telling what we need to learn first. Bring the data back and learn, and that can take time. But then you can have models that are lightweight, that can be brought down to the front, and impact the reaction to the data there. And we heard from a car company, a big car company, how powerful this was when they learned that the angle of a screwdriver, and a few other parameters, can determine the success of screwing something into a body of a car, that could go well, or could go very, very bad and be very costly. So all the learning, massive data, can come down to a simple model that can save a lot of money and improve efficiency. But that has to be hosted along this continuum. >> So from a continuum, it means we still have to have machines somewhere to do something. >> Touching the ground, touching the physical world requires machines, actuators. >> Peter: Absolutely, so Steve, what is Equinix doing to simplify the thinking through of some of these infrastructure issues? >> Yeah, I mean, the biggest thing that people find when they start looking at millions of devices, millions of data capture points, transferring those data real time and streaming it, is one thing hasn't changed and that's physics. So where those things are, where they need to go, where the data needs to move to and how fast, starts with having to figure out your own topology of how you're moving that data. As much as it's easy to say we're just going to buy a platform and choose a device, and we'll clink them together, there's still a lot of other things that need to be solved, physics being the first one. The second one, primarily, is volumes. So how much bandwidth and (mumbles) you're going to require. How much of that data are you going to back haul to centralized data center before you send it up to a Cloud? How much of it are you going to leave at the edge? Where do you place that becomes a bigger deal. And the third one is pretty much every industry has to deal with regulations. Regulations control what you can and can't do in terms of IT delivery, where you can place stuff, where you cannot place stuff, data that can leave the country, data that can't. So all these things mean that you need to have a thought through process of where you're placing certain functions, and what you're defining as your itch between the digital and physical world. And Equinix is an interconnection company that's sitting there as a neutral party across all the networks, all the clouds, all the enterprises, all the providers to help people figure that out. >> So before I ask the audience a question, now that I'm down here so I can see you so be prepared, I'm going to ask some of you a question. When you think about the strategic business capabilities necessary to succeed, what is the first thing that the business has to do? So why don't I just take Ted, and just go right on down the line. >> Yeah, so I think this is really, really important. I work with many, many clients around the world who are doing five, 10, 15 POCs, pilots, and the internet things, and they haven't thought through a codified strategy. So they're doing five things that will never fit together, that you will never scale, and the learnings you're using, you really can't do that much with. So coming up with what is my architecture, what is my stack going to look like, how am I going to push data, what is my data... You know, because when you connect to these things, I can't tell you how much data you're going to get. You're going to be overwhelmed by the data, and that's why we all go to the edge, and I got to process this data real time. And oh, by the way, if I only have one source of data, like I'm connecting to production equipment, you're not going to learn anything. 98% of that data's useless, you got to contextualize the data with either an inspection step, or some kind of contextualization that tells you if this then that. You need the then that, without that, your data is basically worthless. So now you're pulling multiple sources of data together in real time to make an understanding. And so understanding what that architecture looks like, spend the time upfront. Look, most of us are engineers, you know five percent additional work upfront saves you 95% on the backend, that's true here. So think through the architecture, talk to some of us who have been working in this area for a long time. We'll share our architecture, we have reference architecture that we're working with companies. How do you go from industry 2.0 or industry 3.0, to industry 4.0? And there is a logical path to do it, but ultimately, where we're going to end up is a software defined universe. I mean, what's a cloud? It's a software defined data center. Now we're doing software defined networks, software defined storages, ultimately we're going to be doing software defined systems because it's cheaper. You get better capital utilization, better asset utilization, so we will go there, so what does that mean for you infrastructure, and what are you going to do from an architectural perspective, and then take all of your POCs and pilots, and force them to do that specifically around security. People are doing POCs with security that they don't even have any protocols, they're violating all their industry standards doing POCs, and that's going to get thrown out. It's wasted time, wasted effort, don't do it. >> Steve, a couple sentences? >> Yeah, essentially it's not going to be any prizes for me saying think interconnection first. A lot of our customers, if we look at what they've done with us, everyone from GE to real time facial recognition at the edge, it all comes down to how are you wired, topology wise, first. You can't use the internet for risk reasons, you can't necessarily pay for multiple (mumbles) bandwidth costs, et cetera. So low latency, 80% lower latency, seven times of bandwidth at half the cost is a scalable infrastructure to move (mumbles) around the planet. If you don't have that, the rest of the stuff (mumbles) breakdown. >> Peter: Phu? >> Well I would say that analytics is hard, analytics in real time is even harder. And I think with us talking to our customers, I feel for them, they're confused. There's like a million solutions out there, everybody's trying to claim to do the same thing. I think it's both sides, consumers have to get more educated, they have to be more intelligent about their POCs, but as an industry, we also have to get better at thinking about how do we help our customer succeed. It's not about let me give you some open source, and then let me spend the next 10 months charging you professional services to help you. We ought to think about software tools and enterprise tools to really help the customer be able to think about their total cost (mumbles) and time to value to handle this thing, because it's not easy. >> Peter: Flavio. >> Yeah, we're facing an interesting situation where the customers are ready, the needs are there, the marketing is going to be huge, but the plot, the solution, is not trivial. It is maturing and we are all trying to understand how to do it. And this is the confusion that you see in many of these half baked solution (mumbles). Everything is coming together, and you have to go up the stalk and down the stalk with full confidence, that's not easy. So we all have to really work together. Give ourselves time, be feeling that we are in a competitive world, preparing for addressing together a huge market. And trying to mature these solutions that then will be replicated more and more, but we have to be patient with each other, and with the technologies that are maturing and they're not fully there and understood. But the market is amazing. >> Peter: So we have a Twitter question. >> Man: It's being live streamed, the audience is really engaged online as well, digital. So we have a question from Twitter from Lauren Cooney saying, "Would like to know what industries would "be most impacted with digitization "over the next five years." >> Which one won't be? (men laughing) All of them, what we've seen, the business model is the data. I mean, our CEOs calling data the new gold. I mean, it's the new oil. So I don't know of anything, unless you're doing something that is just physical therapy, but that even data, you can do data on that. So yeah, everything, yeah, I don't know of anything that won't be. >> I think the real question is how is it going to move through industries. Obviously it's going to start with some of the digital native, it's all ready deep into that, deep into media, we're moving through the media right now. Intel's clearly a digital company, and you've been working, you've been on this path for quite some time. >> Let me give you a stat. Intel has a 105,000 people, and 144,000 servers. So we're about 1.5 server to people, that's what kind of computation we're (mumbles). >> Peter: We can help you work on that. >> If you do like the networking started by (mumbles) the internet, then content delivery, and media, hard media, et cetera, is gone. Financial services and trading exchanges pretty much show what digital market's going to be in the future. Cloud showed up, and now, I think he's right, it's effecting every industry. Manufacturing, industrial, health professional services are the top three right now. But people who shop to ask for help went from every industry on every country, for that matter. >> Our customers are, you know, the top players in almost every vertical. You start out as a small company thinking that you're going to attack one vertical, but as you start to talk about the capability, everybody (mumbles) wait, you're solving my problem. >> Peter: (mumbles) are followers, is what you mean. >> Yeah, because what business would say, hey, I don't want to know what's going on with my business, and I don't want to take any action. >> Add to that it's an ecosystem of ecosystems. No one, by themselves, is going to solve anything. They have to partner and connect with other people to solve the solution. >> So I'll close the panel by making these kind of summary comments, the business capabilities that we think are going to be most important are, first off, when we talk about the internet of things, we like to talk about the internet of things and people. That the people equation doesn't go away. So we're building on mobile, we're building on other things, but if there's a strategic capability that's going to be required, it's going to be how is this going to impact folks who actually create value in the business. The second one, I'll turn it around, is that IT organizations have gone through a number of different range wars, if you will, over the past 20 years. I lived through IT versus telecom, for example. The IT, OT conflict, or potential conflict, is non trivial. There's going to be some serious work that has to be done, so I would add to the conversation that we've heard thus far, the answers that we've heard thus far, is the degree to which people are going to be essential to making this work, and how we diffuse this knowledge into our employees, and into our IT and professional communities is going to be crucial, especially with developers because Flavio, if we are, right now, trying to figure stuff out, it really matures when we think about the developer world. Okay, so I want to close the first panel and get ready for the second panel. So thank you very much, and thank you very much to our panelists. (audience applauding) And if we could bring David Foyer and the second panel up, we'll get going on panel two. Oh, we're going to get together for a picture. (exciting rhythmic music)

Published Date : Mar 16 2017

SUMMARY :

Now to do that successfully, we have to be able to, Okay, so let me introduce the panelists. I run Solution Architecture for the manufacturing And one of the things I do in there is work with our and at Data Torrent, our mission is really to build Flavio, will you take a second to introduce yourself. Not so much the applications, I'll do my best to see you and I'll share the microphone in our data center at the factory, just to control and the product we build, to build out what became Fog, the infrastructure software that has to be put in and come back to where you were as if nothing has happened the data has a little bit of value. you got to do it now. And the stuff we're doing in applications will eventually and impact the reaction to the data there. So from a continuum, it means we still have to have Touching the ground, touching the physical world all the providers to help people figure that out. the business has to do? and what are you going to do from an architectural perspective, at the edge, it all comes down to how are you wired, and time to value to handle this thing, the marketing is going to be huge, saying, "Would like to know what industries would I mean, our CEOs calling data the new gold. Obviously it's going to start with some of the digital native, Let me give you a stat. in the future. but as you start to talk about the capability, and I don't want to take any action. They have to partner and connect with other people is the degree to which people are going to be

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
StevePERSON

0.99+

PeterPERSON

0.99+

Lauren CooneyPERSON

0.99+

Data TorrentORGANIZATION

0.99+

80%QUANTITY

0.99+

Steve MaddenPERSON

0.99+

EquinixORGANIZATION

0.99+

fiveQUANTITY

0.99+

firstQUANTITY

0.99+

David FoyerPERSON

0.99+

600 serversQUANTITY

0.99+

95%QUANTITY

0.99+

five percentQUANTITY

0.99+

second panelQUANTITY

0.99+

second panelQUANTITY

0.99+

98%QUANTITY

0.99+

20 yearsQUANTITY

0.99+

10 minutesQUANTITY

0.99+

Ted ConnellPERSON

0.99+

thousandsQUANTITY

0.99+

TedPERSON

0.99+

first panelQUANTITY

0.99+

144,000 serversQUANTITY

0.99+

105,000 peopleQUANTITY

0.99+

IntelORGANIZATION

0.99+

third oneQUANTITY

0.99+

two secondQUANTITY

0.99+

five thingsQUANTITY

0.99+

millions of devicesQUANTITY

0.99+

first questionQUANTITY

0.99+

one sourceQUANTITY

0.99+

SecondQUANTITY

0.99+

OneQUANTITY

0.99+

both sidesQUANTITY

0.99+

seven years agoDATE

0.99+

10QUANTITY

0.99+

first oneQUANTITY

0.99+

second capabilityQUANTITY

0.99+

third capabilityQUANTITY

0.98+

20 years agoDATE

0.98+

seven timesQUANTITY

0.98+

three coreQUANTITY

0.98+

panel twoQUANTITY

0.98+

second oneQUANTITY

0.98+

oneQUANTITY

0.97+

Phu HoangORGANIZATION

0.97+

GEORGANIZATION

0.97+

TwitterORGANIZATION

0.97+

eachQUANTITY

0.97+

24QUANTITY

0.97+

YahoosORGANIZATION

0.96+

sevenQUANTITY

0.96+

12, 15 yearsQUANTITY

0.96+

15 POCsQUANTITY

0.96+

FlavioPERSON

0.96+

a secondQUANTITY

0.96+

zero peopleQUANTITY

0.95+

one verticalQUANTITY

0.94+

DruckerORGANIZATION

0.93+

Panel 1QUANTITY

0.93+

MooreORGANIZATION

0.93+

millions of events per secondQUANTITY

0.92+

Western Digital Taking the Cloud to the Edge - #DataMakesPossible - Presentation by Flavio Bonomi


 

>> It's a pleasure to be here with you and to tell you about something I've been dreaming about and working for for many years and now is coming to the surface quite powerfully and quite usefully in many areas. I apologize, sometimes this flickers for some reason but I hope it doesn't disturb the story. I'd like to give you a little touch of history since I was there at the beginning of this journey and give you a brief introduction to what we mean for Fog Computing. And then go quickly to three powerful application spaces for this technology, together with industrial internet and one is industrial automation. That's the focus of our activity as Nebbiolo Technologies. The other one is one of my favorite ones and we'll get there is the automotive that caught fire here in Silicone Valley in the last years, the autonomous car, the connected vehicle and so on. And this is related to also to intelligent transportation and Smart Cities. And then a little touch on what Fog Computing means for Smart grid energy but many, many other sectors will find the same usefulness, the same architecture dimensions of Fog Computing applicable. So this is the story that comes back hopefully, here, the day in 2010 when Fog Computing, the word started here, oh God, is this jumping around? I think it's the connector, this is the age of the connector, this is the age of the Dongles. This is not an Apple Dongle and so we are having troubles. And this is not yet one of the last machines that are out. Let's hope for, I never had this problem, okay. Alright, this date 2010 at the Aquarium Research Center in Monterey where I gave a talk about robots going down deep in the bottom of those big valleys under the ocean and when I finished, the lady, Ginny in the middle approached me and told me, look, why don't you call what you're talking about fog computing? Because it's cloud computing brought too close to the ground and I protested for about 15 minutes. And on the drive home, I thought that's really a good name for what we are doing, what we have been doing in the last years and I started trying it out and using it and more and more I found good response and so seven years later, I'm still here talking about the same thing. What's happening is Fog, the edge of the metric zone was very important but it was always very important in IT, is still very important in IT in mobile, in content distribution but when IOT came to the surface, it became even more relevant to understand the need of resources, virtualized real time capable, secure, trusted with storage computing and networking coming together at the edge. At the edge of the IT network, now they are calling this mobile edge, they realize we are realizing that mobile can benefit from local resources at the edge, powerful real time capable resources but also and more importantly for what we are doing in this space of operational technologies, this is the space, the other and the other side of the boundary between information technologies and operational technologies and here is where we are living with Fog Computing these days so, apologize, I apologize for this behavior that is, maybe I have another Dongle, Apple Dongle. Maybe I could look at that, maybe Morris can help me out here, anyway, so what is Fog Computing? Fog Computing is really the platform that brings modern, Cloud inspired Computing storage here is important here for our friends at Western Digital and networking functions closer to the data producing sources. In our case, machines, things, but not just bringing Cloud down, it's also bringing functions up from the machine world, the real time, the safety functions, the trusting and reliability functions required in that area and this is a unified solution at the edge that really brings together communication, device management, data harvesting, analysis and control. So this is kind of new except for our friends in Wall Street. The real time part was not as sensitive. Now we are realizing how important it is and how important the position of resources is in the future of solutions in this space and so it's not boxes. It's a distributed layer of resources, well managed at the edge of the network and really has a lot of potential across multiple industries. Here we see the progress also in the awareness of this topic with the open fog control room that is now a very active and even the Vcs. Peter Levine here is talking about the importance of the edge. What is really happening is the the convergence. I think we should probably stop and use a different Dongle. Is this the one, no, no, this is not the right Dongle. The world of Dongles, sorry. Oh boy. Oh you have the computer with the, okay, is the right Dongle with the right computer, okay. Here we are, okay. Alright, we're getting back there. This is the new Apple. Okay, we are here, this looks better, thank you. Alright, so this is to be understood. This is the convergence of IT functionality, the modern IT functionality with the OT requirements and this is fundamentally the powerful angle that Fog Computing brings to IOT and machine world so all the nice things that happened in the Cloud come down but meet the requirements of resources, the needs and the timing of the Edge. And so when you look at what is brought into particularly the world of operations, you see these kind of functions that are not usually there. In fact, when you meet this operational world, you find microprocessors, you find Windows machines, industrial Pcs and so on, not so much Linux, not so much the modern approaches to computing. These are the type of dimensions that you'll see have a particular impact on the pain points seen in the wold of applications. So now we go to the Use cases in, use cases in the internet of things. I think it's on your side, I'm sorry. Because it's the second machine. Okay, well, maybe here's the solution. So we have seen this picture of IOT multiple times. A lot of verticals, we are concentrating on this tree, one is the industrial, the second one is the autonomous vehicle in intelligent transportation, the third one, just touched upon is the Smart Grid. This is the area of activity for Nebbiolo Technologies. Those kind of body shops and industrial floors with large robots with a lot of activity around those robots with cells protecting the activities within each working space, this is the world PLCs, industrial Pcs controlling robots, very fragmented. Here we are really finding even more critical this boundary between operational and informational technologies. This is a fire wall, also a mental fire wall between the two worlds and best practice is very different in one place than the other particularly also in the way we handle data, security, and many other areas. In this space, which is also a little more characterized here with this kind of machines that you see in this ISA 99 or ISA 95 type of picture, you see the boundary between the two spaces, once more when we come back. And alright, so the key message here, very tough to go across, it's very complex, the interaction between the two worlds. And there is where deeply we find a number of pain points at the security level, at the Hardware architecture level, at the data analytics and storage level, at the networking, software technologies and control architecture. There's a lot happening there that is old, 1980's time frame, very stable but in need of new approaches. And this is where Fog Computing has a very strong impact And we'll see, sorry, this is a disaster here. Alright, what do we do, alright. Maybe I should go around with this computer and show it to you. Okay, now it's there for a moment. Now, this is, maybe you have to remember one picture of all this talk, look at this, what is this? This is a graphical image of a body shop of a an important car company, you see the dots represent computers within boxes, industrial Pcs, PLCs, controllers for welding machines, tools and so on. That is, if you sum up the numbers, it's thousands of computers, each one of them is updated through a UPC, USB stick, sorry and is not managed remotely. It's not secure because there's a trust that the whole area is enclosed and protected through a fire wall on the other side but it's very stable but very rigid. So this is the world that we are finding with dedicated, isolated, not secure computing, this is Edge Computing. But it's not what we hope to be seeing soon as Fog Computing in action there so this is the situation. Very delicate, very powerful and very motivating. And now comes IOT and this is not the solution. It's helping, IOT tries to connect this big region, the operational region to the back end to the Clouds, to the power of computing that is there, very important, predicting maintenance, many other things can be done from there but it's still not solving the problem. Because now you have to put little machines, gateways into that region, one more machine to manage, one more machine to secure and now you're taking the data out. You are not solving a lot of the pain points. There's some important benefits, this is very, very good. But it's not the story, the story is sold once you really go one step deeper, in fact, from connectivity between information technologies and informational technologies to really Convergence and you see it here where you're starting to replace those machines supporting each cell with a fog node, with a powerful convergent point of computing, real time computing that can allow control, analytics and storage and networking in the same nodes so now these nodes are starting to replace all the objects controlling a cell. And offer more functions to the cell itself. And now, you can imagine where this goes, to a convergent architecture, much more compact, much more homogeneous, much more like Cloud. Much more like Cloud brought down to the Edge. When this comes back, okay, almost there. So this is okay, this is now the image that you can image leads to this final picture that is now even not, okay, do you see it, okay. Now you're seeing the operational space with the fabric of computing storage and networking that is modern, that is virtualized, that supports an application store, now you have containers there. You can imagine virtual machines and dockers living the operational space. At the same time, you have it continuing from the Cloud to the network, the modern network, moving to the Edge into the operational space. This is where we are going and this is where the world wants us to go and the picture representing this transition and this application of Fog Computing in this area is the following, the triangle, the pyramid is now showing a layer of modern computing that allows communications analysis control application hosting and orchestration in a new way. This is cataclysmic, really is a powerful shift, still not fully understood but with immense consequences. And now you can do control, tight, close to the machines, a little slower through the Fog and a little slower through the Cloud, this is where we are going. And there's many, many used cases, I don't dwell on those. But we are proceeding with some of our partners exactly in this direction. Now the exciting topics if I can have five more minutes making up the time wasted. What's going on here, the connected vehicle, the autonomous vehicle, the electrification of automobile are all converging and I think it's very clear that the para dime of Fog Computing is fundamental here. And in fact, imagine the equivalent of a manufacturing cell with a converging capabilities into the Fog and compare it with what's going on with the autonomous vehicle. This is a picture we used a Sysco seven years ago. But this is now, a car is a set of little control loops, ECUs, little dispersed, totally connected computers. Very difficult to program, same as the manufacturing cell. And now where are we going, we are going towards a Fog node on wheels, data center on wheels but better a Fog node on wheels with much better networking between, with a convergence of the intelligence, the control, the analytics, the communications in the middle and a modern network deterministic internet called TSN is going to replace all these CAN boxes and all these flakey things of the past. Same movement in industrial and in the automobile and then you look at what's going on in the intelligent transportation, you can imagine Fog Computing at the edge, controlling the junctions, the traffic lights, the interactions with cars, cars to cars and you see it here, this is the image, again where you have the operational space of transportation connected to the Clouds in a seamless way which these nodes of computing storage and networking at the junctions inside the cars talking to each other, so this is the beautiful movement coming to us and it requires the distribution of resources with real time capabilities, here you see it. And now, the Smart Grid, again, it cannot continue to go the same way with a utility data center controlling everything one way, it has to have and this is from Duke and a standardization body, you can see that there's a need of intelligence in the middle, Fog nodes, distributed computing that are allowing local decisions. Energy coming from a microcell into the grid and out, a car that wants to sell it's energy or buy energy doesn't need to go slowly to a utility data center to make decisions so again, same architecture, same technologies needed, very, very, very powerful. And we could go on and on and on, so what are we doing? We won't advertise here but the name has to be remembered. The name comes from a grape that grows in the Fog in Northern Italy, it's in Piedmont, my home town is behind that 13th century castle you see there. Out there is Northern Italy close to Switzerland. That vineyard is from my cousin, it's a good Nebbiolo, starting to be sold in California too. So this is the name Nebbia Fog comes to, Nebbiolo Technologies, we are building a platform for this space with all the features that we feel are required and we are applying it to industrial automation. And our funders are not so much from here, are from Germany, Austria, KUKA Robotics, TTTech, GiTV from Japan and a few bullets to complete my presentation. Fog Computing is really happening. There's a deep need for this converged infrastructure for IOT including Fog or Edge as someone calls it. But we need to continue to learn, demonstrate, validate through pilots and POCs and we need to continue to converge with each other and with the integrators because these solutions are big and they are not from a little start up. They are from integrators, customers, big customers at the other end, an ecosystem of creative companies. No body has all the pieces, no Sisco, no GE and so on. In fact, they are all trying to create the ecosystem. And so let's play, let's enjoy the Cloud, the Fog and the machines and try to solve some of the big problems of this world. >> Okay, Flavio, well done. >> Sorry for that. Sorry for the hiccups. >> Now we do that on purpose to see how you'd react and you're a pro, thank you so much for the great presentation. >> Alright. >> Alright, now we're going to get into panel one, looking at the data models and putting data to work.

Published Date : Mar 16 2017

SUMMARY :

the interactions with cars, cars to cars and you see it Sorry for the hiccups. Now we do that on purpose to see how you'd looking at the data models and putting data to work.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
KUKA RoboticsORGANIZATION

0.99+

CaliforniaLOCATION

0.99+

2010DATE

0.99+

Flavio BonomiPERSON

0.99+

TTTechORGANIZATION

0.99+

Nebbiolo TechnologiesORGANIZATION

0.99+

MontereyLOCATION

0.99+

SwitzerlandLOCATION

0.99+

GiTVORGANIZATION

0.99+

PiedmontLOCATION

0.99+

Wall StreetLOCATION

0.99+

AppleORGANIZATION

0.99+

SiscoORGANIZATION

0.99+

Silicone ValleyLOCATION

0.99+

GinnyPERSON

0.99+

Western DigitalORGANIZATION

0.99+

GEORGANIZATION

0.99+

five more minutesQUANTITY

0.99+

two worldsQUANTITY

0.99+

Peter LevinePERSON

0.99+

Northern ItalyLOCATION

0.99+

DukeORGANIZATION

0.99+

second machineQUANTITY

0.99+

oneQUANTITY

0.99+

LinuxTITLE

0.99+

13th centuryDATE

0.99+

JapanLOCATION

0.99+

each cellQUANTITY

0.99+

seven years agoDATE

0.98+

one placeQUANTITY

0.98+

seven years laterDATE

0.98+

Nebbia FogORGANIZATION

0.98+

thousands of computersQUANTITY

0.98+

WindowsTITLE

0.98+

one pictureQUANTITY

0.98+

each oneQUANTITY

0.97+

two spacesQUANTITY

0.97+

about 15 minutesQUANTITY

0.95+

CloudTITLE

0.95+

GermanyLOCATION

0.95+

third oneQUANTITY

0.94+

FlavioPERSON

0.93+

MorrisPERSON

0.93+

last yearsDATE

0.92+

one more machineQUANTITY

0.91+

SyscoORGANIZATION

0.88+

one more machineQUANTITY

0.86+

1980'sDATE

0.82+

each workingQUANTITY

0.81+

one wayQUANTITY

0.81+

NebbioloORGANIZATION

0.81+

second oneQUANTITY

0.77+

one stepQUANTITY

0.72+

AustriaLOCATION

0.72+

Aquarium Research CenterLOCATION

0.67+

ISA 95OTHER

0.63+

threeQUANTITY

0.63+

EdgeTITLE

0.62+

ISA 99OTHER

0.57+

casesQUANTITY

0.51+

DongleCOMMERCIAL_ITEM

0.51+

DongleORGANIZATION

0.45+

Janet George, Western Digital | Women in Data Science 2017


 

>> Male Voiceover: Live from Stanford University, it's The Cube covering the Women in Data Science Conference 2017. >> Hi, welcome back to The Cube, I'm Lisa Martin and we are live at Stanford University at the second annual Women in Data Science Technical Conference. It's a one day event here, incredibly inspiring morning we've had. We're joined by Janet George, who is the chief data scientist at Western Digital. Janet, welcome to the show. >> Thank you very much. >> You're a speaker at-- >> Very happy to be here. >> We're very happy to have you. You're a speaker at this event and we want to talk about what you're going to be talking about. Industrialized data science. What is that? >> Industrialized data science is mostly about how data science is applied in the industry. It's less about more research work, but it's more about practical application of industry use cases in which we actually apply machine learning and artificial intelligence. >> What are some of the use cases at Western Digital for that application? >> One of the use case that we use is, we are in the business of creating new technology nodes and for creating new technology nodes we actually create a lot of data. And with that data, we actually look at, can we understand pattern recognition at very large scale? We're talking millions of wafers. Can we understand memory holes? The shape, the type, the curvature, circularity, radius, can we detect these patterns at scale? And then how can we detect if the memory hole is warped or deformed and how can we have machine learning do that for us? We also look at things like correlations during the manufacturing process. Strong correlations, weak correlations, and we try to figure out interactions between different correlations. >> Fantastic. So if we look at big data, it's probably applicable across every industry. How has it helped to transform Western Digital, that's been an institution here in Silicon Valley for a while? >> We in Western Digital we move mountains of data. That's just part of our job, right? And so we are the leaders in storage technology, people store data in Western Digital products, and so data's inherently very familiar to us. We actually deal with data on a regular basis. And now we've started confronting our data with data science. And we started confronting our data with machine learning because we are very aware that artificial intelligence, machine learning can bring a different value to that data. We can look at the insides, we can develop intelligence about how we build our storage products. What we do with our storage. Failure analysis is a huge area for us. So we're really tapping into our data to figure out how can we make artificial intelligence and machine learning ingrained in the way we do work. >> So from a cultural perspective, you've really done a lot to evolve the culture of Western Digital to apply the learnings, to improve the values that you deliver to all of your customers. >> Yes, believe it or not, we've become a data-driven company. That's amazing, because we've invested in our own data, and we've said "Hey, if we are going to store the world's data, we need to lead, from a data perspective" and so we've sort of embraced machine learning and artificial intelligence. We've embraced new algorithms, technologies that's out there we can tap into to look at our data. >> So from a machine learning, human perspective, in storage manufacturing, is there still a dependence on human insight where storage manufacturing devices are concerned, or are you seeing the machine learning really, in this case, take more of a lead? >> No, I think humans play a huge role, right? Because these are domain experts. We're talking about Ph.D.'s in material science and device physics areas so what I see is the augmentation between machine learning and humans, and the domain experts. Domain experts will not be able to scale. When the scale of wafer production becomes very large. So let's talk about 3 million wafers. How is a machine going to physically look at all the failure patterns on those wafers? We're not going to be able to scale just having domain expertise. But taking our core domain expertise and using that as training data to build intelligence models that can inform the domain expert and be smart and come up with all the ideas, that's where we want to be. >> Excellent. So you talked a little bit about the manufacturing process. Who are some of the other constituents that you collaborate with as chief data scientist at Western Digital that are demanding access to data, marketing, etcetera, what are some of those key collaborators for your group? >> Many of our marketing department, as well as our customer service department, we also have collaborations going on with universities, but one of the things we found out was when a drive fails, and it goes to our customer, it's much better for us to figure out the failure. So we've started modeling out all the customer returns that we've received, and look at that and see "How can we predict the life cycle of our storage?" And get to those return possibilities or potential issues before it lands in the hands of customers. >> That's excellent. >> So that's one area we've been focusing quite a bit on, to look at the whole life cycle of failures. >> You also talked about collaborating with universities. Share a little bit about that in terms of, is there a program for internships for example? How are you helping to shape the next generation of computer scientists? >> We are very strongly embedded in universities. We usually have a very good internship program. Six to eight weeks, to 12 weeks in the summer, the interns come in. Ours is a little different where we treat our interns as real value add. They come in, and they're given a hypothesis, or problem domain that they need to go after. And within six to eight weeks, and they have access to tremendous amounts of data, so they get to play with all this industry data that they would never get to play with. They can quickly bring their academic background, or their academic learning to that data. We also take really hard research-ended problems or further out problems and we collaborate with universities on that, especially Stanford University, we've been doing great collaborations with them. I'm super encouraged with Feliz's work on computer vision, and we've been looking into things around deep neural networks. This is an area of great passion for me. I think the cognitive computing space is just started to open up and we have a lot to learn from neural networks and how they work and where the value can be added. >> Looking at, just want to explore the internship topic for a second. And we're at the second annual Women in Data Science Conference. There's a lot of young minds here, not just here in person, but in many cities across the globe. What are you seeing with some of the interns that come in? Are they confident enough to say "I'm getting access to real world data I wouldn't have access to in school", are they confident to play around with that, test out a hypothesis and fail? Or do they fear, "I need to get this right right away, this is my career at stake?" >> It's an interesting dichotomy because they have a really short time frame. That's an issue because of the time frame, and they have to quickly discover. Failing fast and learning fast is part of data science and I really think that we have to get to that point where we're really comfortable with failure, and the learning we get from the failure. Remember the light bulb was invented with 99% negative knowledge, so we have to get to that negative knowledge and treat that as learning. So we encourage a culture, we encourage a style of different learning cycles so we say, "What did we learn in the first learning cycle?" "What discoveries, what hypothesis did we figure out in the first learning cycle, which will then prepare our second learning cycle?" And we don't see it as a one-stop, rather more iterative form of work. Also with the internships, I think sometimes it's really essential to have critical thinking. And so the interns get that environment to learn critical thinking in the industry space. >> Tell us about, from a skills perspective, these are, you can share with us, presumably young people studying computer science, maybe engineering topics, what are some of the traditional data science skills that you think are still absolutely there? Maybe it's a hybrid of a hacker and someone who's got, great statistician background. What about the creative side and the ability to communicate? What's your ideal data scientist today? What are the embodiments of those? >> So this is a fantastic question, because I've been thinking about this a lot. I think the ideal data scientist is at the intersection of three circles. The first circle is really somebody who's very comfortable with data, mathematics, statistics, machine learning, that sort of thing. The second circle is in the intersection of implementation, engineering, computer science, electrical engineering, those backgrounds where they've had discipline. They understand that they can take complex math or complex algorithms and then actually implement them to get business value out of them. And the third circle is around business acumen, program management, critical thinking, really going deeper, asking the questions, explaining the results, very complex charts. The ability to visualize that data and understand the trends in that data. So it's the intersection of these very diverse disciplines, and somebody who has deep critical thinking and never gives up. (laughs) >> That's a great one, that never gives up. But looking at it, in that way, have you seen this, we're really here at a revolution, right? Have you seen that data science traditionalist role evolve into these three, the intersection of these three elements? >> Yeah, traditionally, if you did a lot of computer science, or you did a lot of math, you'd be considered a great data scientist. But if you don't have that business acumen, how do you look at the critical problems? How do you communicate what you found? How do you communicate that what you found actually matters in the scheme of things? Sometimes people talk about anomalies, and I always say "is the anomaly structured enough that I need to care about?" Is it systematic? Why should I care about this anomaly? Why is it different from an alert? If you have modeled all the behaviors, and you understand that this is a different anomaly than I've normally seen, and you must care about it. So you need to have business acumen to ask the right business questions and understand why that matters. >> So your background in computer science, your bachelor's Ph.D.? >> Bachelor's and master's in computer science, mathematics, and statistics, so I've got a combination of all of those and then my business experience comes from being in the field. >> Lisa: I was going to ask you that, how did you get that business acumen? Sounds like it was by in-field training, basically on-the-job? >> It was in the industry, it was on-the-job, I put myself in positions where I've had great opportunities and tackled great business problems that I had to go out and solve, very unique set of business problems that I had to dig deep into figuring out what the solutions were, and so then gained the experience from that. >> So going back to Western Digital, how you're leveraging data science to really evolve the company. You talked about the cultural evolution there, which we both were mentioning off-camera, is quite a feat because it's very challenging. Data from many angles, security, usage, is a board level, boardroom conversation. I'd love to understand, and you also talked about collaboration, so talk to us a little bit about how, and some of the ways, tangible ways, that data science and your team have helped evolve Western Digital. Improving products, improving services, improving revenue. >> I think of it as when an algorithm or a machine learning model is smart, it cannot be a threat. There's a difference between being smart and being a threat. It's smart when it actually provides value. It's a threat when it takes away or does something you would be wanting to do, and here I see that initially there's a lot of fear in the industry, and I think the fear is related to "oh, here's a new technology," and we've seen technologies come in and disrupt in a major way. And machine learning will make a lot of disruptions in the industry for sure. But I think that will cause a shift, or a change. Look at our phone industry, and how much the phone industry has gone through. We never complain that the smart phone is smarter than us. (laughs) We love the fact that the smartphone can show us maps and it can send us in the right, of course, it sends us in the wrong direction sometimes, most of the time it's pretty good. We've grown to rely on our cell phones. We've grown to rely on the smartness. I look at when technology becomes your partner, when technology becomes your ally, and when it actually becomes useful to you, there is a shift in culture. We start by saying "how do we earn the value of the humans?" How can machine learning, how can the algorithms we built, actually show you the difference? How can it come up with things you didn't see? How can it discover new things for you that will create a wow factor for you? And when it does create a wow factor for you, you will want more of it, so it's more, to me, it's most an intent-based progress, in terms of a culture change. You can't push any new technology on people. People will be reluctant to adapt. The only way you can, that people adopt to new technologies is when they the value of the technology instantly and then they become believers. It's a very grassroots-level change, if you will. >> For the foreseeable future, that from a fear perspective and maybe job security, that at least in the storage and manufacturing industry, people aren't going to be replaced by machines. You think it's going to maybe live together for a very long, long time? >> I totally agree. I think that it's going to augment the humans for a long, long time. I think that we will get over our fear, we worry that the humans, I think humans are incredibly powerful. We give way too little credit to ourselves. I think we have huge creative capacity. Machines do have processing capacity, they have very large scale processing capacity, and humans and machines can augment each other. I do believe that the time when we had computers and we relied on our computers for data processing. We're going to rely on computers for machine learning. We're going to get smarter, so we don't have to do all the automation and the daily grind of stuff. If you can predict, and that prediction can help you, and you can feed that prediction model some learning mechanism by reinforced learning or reading or ranking. Look at spam industry. We just taught the Spam-a-Guccis to become so good at catching spam, and we don't worry about the fact that they do the cleansing of that level of data for us and so we'll get to that stage first, and then we'll get better and better and better. I think humans have a natural tendency to step up, they always do. We've always, through many generations, we have always stepped up higher than where we were before, so this is going to make us step up further. We're going to demand more, we're going to invent more, we're going to create more. But it's not going to be, I don't see it as a real threat. The places where I see it as a threat is when the data has bias, or the data is manipulated, which exists even without machine learning. >> I love though, that the analogy that you're making is as technology is evolving, it's kind of a natural catalyst >> Janet: It is a natural catalyst. >> For us humans to evolve and learn and progress and that's a great cycle that you're-- >> Yeah, imagine how we did farming ten years ago, twenty years ago. Imagine how we drive our cars today than we did many years ago. Imagine the role of maps in our lives. Imagine the role of autonomous cars. This is a natural progression of the human race, that's how I see it, and you can see the younger, young people now are so natural for them, technology is so natural for them. They can tweet, and swipe, and that's the natural progression of the human race. I don't think we can stop that, I think we have to embrace that it's a gift. >> That's a great message, embracing it. It is a gift. Well, we wish you the best of luck this year at Western Digital, and thank you for inspiring us and probably many that are here and those that are watching the livestream. Janet George, thanks so much for being on The Cube. >> Thank you. >> Thank you for watching The Cube. We are again live from the second annual Women in Data Science conference at Stanford, I'm Lisa Martin, don't go away. We'll be right back. (upbeat electronic music)

Published Date : Feb 3 2017

SUMMARY :

it's The Cube covering the Women in I'm Lisa Martin and we are going to be talking about. data science is applied in the industry. One of the use case How has it helped to in the way we do work. apply the learnings, to to look at our data. that can inform the a little bit about the the things we found out quite a bit on, to look at the helping to shape the next started to open up and we but in many cities across the globe. That's an issue because of the time frame, the ability to communicate? So it's the intersection of the intersection of I always say "is the So your background in computer science, comes from being in the field. problems that I had to You talked about the how can the algorithms we built, that at least in the I do believe that the time of the human race, Well, we wish you the We are again live from the second annual

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JanetPERSON

0.99+

Lisa MartinPERSON

0.99+

Janet GeorgePERSON

0.99+

Western DigitalORGANIZATION

0.99+

SixQUANTITY

0.99+

LisaPERSON

0.99+

99%QUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

12 weeksQUANTITY

0.99+

Stanford UniversityORGANIZATION

0.99+

third circleQUANTITY

0.99+

first circleQUANTITY

0.99+

twenty years agoDATE

0.99+

second circleQUANTITY

0.99+

eight weeksQUANTITY

0.99+

sixQUANTITY

0.99+

The CubeTITLE

0.99+

ten years agoDATE

0.99+

OneQUANTITY

0.98+

threeQUANTITY

0.98+

eight weeksQUANTITY

0.98+

three circlesQUANTITY

0.98+

Women in Data Science Technical ConferenceEVENT

0.98+

this yearDATE

0.98+

FelizPERSON

0.97+

StanfordLOCATION

0.97+

three elementsQUANTITY

0.97+

oneQUANTITY

0.97+

Women in Data Science Conference 2017EVENT

0.97+

bothQUANTITY

0.96+

Women in Data Science ConferenceEVENT

0.96+

many years agoDATE

0.96+

second learning cycleQUANTITY

0.96+

Women in Data ScienceEVENT

0.96+

one day eventQUANTITY

0.96+

first learning cycleQUANTITY

0.94+

first learning cycleQUANTITY

0.93+

todayDATE

0.91+

one areaQUANTITY

0.91+

Women in Data Science conferenceEVENT

0.89+

secondQUANTITY

0.88+

millions of wafersQUANTITY

0.87+

firstQUANTITY

0.86+

one-stopQUANTITY

0.86+

about 3 million wafersQUANTITY

0.84+

-a-GuccisORGANIZATION

0.81+

The CubeORGANIZATION

0.77+

UniversityORGANIZATION

0.6+

second annualQUANTITY

0.56+

2017DATE

0.51+

CubePERSON

0.36+

Ken Byrnes, Dell Technologies & David Trigg, Dell Technologies | MWC Barcelona 2023


 

>> Narrator: TheCUBE's live coverage is made possible by funding from Dell Technologies. Creating technologies that drive human progress. >> All right, welcome back to the Fira in Barcelona. This is Dave Vellante with Dave Nicholson. Day 4 of coverage MWC 23. We've been talking all week about the disaggregation of the telco networks, how telcos need to increase revenue how they're not going to let the over the top providers do it again. They want to charge Netflix, right? And Netflix is punching back. There maybe are better ways to do revenue acceleration. We're going to talk to that topic with Dave Trigg who's the Global Vice President of Telecom systems business at Dell Technologies. And Ken Burns, who's a global telecom partner, sales lead. Guys, good to see you. >> Good to see you. Great to be here. >> Dave, you heard my, you're welcome. You heard my intro. It's got to be better ways to, for the telcos to make money. How can they accelerate revenue beyond taxing Netflix? >> Yeah, well, well first of all, sort of the promise of 5G, and a lot of people talk about 5G as the enterprise G. Right? So the promise of 5G is to really help drive revenue enterprise use cases. And so, it's sort of the promise of the next generation of technology, but it's not easy to figure out how we monetize that. And so we think Dell has a pretty significant role to play. It's a CEO conversation for every telco and how they accelerate. And so it's an area we're investing heavily into three different areas for telcos. One is the IT space. Dell's done that forever. 90% of the companies leaning in on that. The other places network, network's more about cost takeout. And the third area where we're investing in is working with what we call their line of businesses, but it's really their business units, right? How can we sit down with them and really understand what services do they take to market? Where do they go? So, we're making significant investments. So one way they can do it is working with Dell and and we're making big investments 'cause in most Geos we have a fairly significant sales force. We've brought in an industry leader to help us put it together. And we're getting very focused on this space and, you know, looking forward to talking more about it. >> So Ken, you know, the space inside and out, we just had at AT&T on... >> Dave Trigg: Yep. >> And they were saying we have to be hypersensitive because of our platinum brand to the use of personal information. >> Ken: Yeah. >> So we're not going to go there yet. We're not going to go directly monetize, but yet I'm thinking well, Netflix knows what I'm watching and they're making recommendations and they're, and and that's how they make money. And so the, the telcos are, are shy about doing that for right reasons, but they want to make better offers. They want to put, put forth better bundles. You know, they don't, they don't want to spend all their time trying to figure that out and not being able to change when they need to change. So, so what is the answer? If they're not going to go toward that direct monetization of data? >> Ken: Yeah. >> How do they get there? >> So I, I joined Dell in- at the end of June and brought on, as David said, to, to build and lead this what we call the line of business strategy, right? And ultimately what it is is tying together Dell technology solutions and the best of breed of what the telecoms bring to bear to solve the business outcomes of our joint customers. And there's a few jewels inside of Dell. One of it is that we have 35,000 sellers out there all touching enterprise business customers. And we have a really good understanding of what those customer needs are and you know what their outcomes needs to be. The other jewel is we have a really good understanding of how to solve those business outcomes. Dell is an open company. We work with thousands of integrators, and we have a really good insight in terms of how to solve those business outcomes, right? And so in my conversations with the telecom companies when you talk about, you know combining the best assets of Dell with their capabilities and we're all talking to the same customers, right? And if we're giving them the same story on these solutions solving business outcomes it's a beautiful thing. It's a time to market. >> What's an example of a, of a, of a situation where you'll partner with telcos that's going to drive revenue for, for both of you and value for the customer? >> Yeah, great question. So we've been laser focused on four key areas, cyber, well, let me start off with connected laptops, cyber, private mobility, and edge. Right? Now, the last two are a little bit squishy, but I'll I'll get to that in a bit, right? Because ultimately I feel like with this 5G market, we could actually make the market. And the way that we've been positioning this is almost, almost on a journey for IOT. When we talk about laptops, right? Dell is the, is the number one company in the world to sell business laptops. Well, if we start selling connected laptops the telcos are starting to say, well, you know what? If all of those laptops get connected to my network, that's a ton of 5G activations, right? We have the used cases on why having a connected workforce makes sense, right? So we're sharing that with the telcos to not simply sell a laptop, but to sell the company on why it makes sense to have that connected workforce. >> Dave Vellante: Why does it make sense? It could change the end customer. >> Ken: Yeah. So, you know, I'm probably not the best to answer that one right? But, but ultimately, you know Dell is selling millions and millions of laptops out there. And, and again, the Verizon's, the AT&T's, the T-mobile's, they're seeing the opportunity that, you know, connecting those laptops, give those the 5G activations right? But Dave, you know, the way that we've been positioning this is it's not simply a laptop could be really a Trojan horse into this IOT journey. Because ultimately, if you sell a thousand laptops to an enterprise company and you're connecting a thousand of their employees, you're connecting people, right? And we can give the analytics around that, what they're using it for, you know, making sure that the security, the bios, all of that is up to date. So now that you're connecting their people you could open up the conversation to why don't we we connect your place and, you know, allowing the telecom companies to come in and educate customers and the Dell sales force on why a private 5G mobility network makes sense to connecting places. That's a great opportunity. When you connect the place, the next part of that journey is connecting things in that place. Robotics, sensors, et cetera, right? And, and so really, so we're on the journey of people, places, things. >> So they got the cyber angle angle in there, Dave. That, that's clear benefit. If you, you know, if you got all these bespoke laptops and they're all at different levels you're going to get, you know, you're going to get hacked anyway. >> Ken: That's right. >> You're going to get hacked worse. >> Yeah. I'm curious, as you go to market, do you see significant differences? You don't have to name any names, but I imagine that there are behemoths that could be laggards because essentially they feel like they're the toll booth and all they have to do is collect, keep collecting the tolls. Whereas some of the smaller, more nimble, more agile entities that you might deal with might be more receptive to this message. That seems to be the sort of way the circle of life are. Are you seeing that? Are you seeing the big ones? Are you seeing the, you know, the aircraft carriers realizing that we got to turn into the wind guys and if we don't start turning into the wind now we're going to be in trouble. >> So this conference has been absolutely fantastic allowing us to speak with, you know, probably 30 plus telecom operators around this strategy, right? And all of the big guys, they've invested hundreds of billions of dollars in their 5G network and they haven't really seen the ROI. So when we're coming into them with a story about how Dell can help monetize their 5G network I got to tell you they're pretty excited >> Dave Nicholson: So they're receptive? >> Oh my God. They are very receptive >> So that's the big question, right? I mean is, who's, is anybody ever going to make any money off of 5G? And Ken, you were saying that private mobility and edge are a little fuzzy but I think from a strategy standpoint I mean that is a potential gold mine. >> Yeah, but it, for, for lot of the telcos and most telcos it's a pretty significant shift in mentality, right? Cause they are used to selling sim cards to some degree and how many sim cards are they selling and how many, what other used cases? And really to get to the point where they understand the use case, 'cause to get into the enterprise to really get into what can they do to help power a enterprise business more wholly. They've got to understand the use case. They got to understand the more complete solution. You know, Dell's been doing that for years. And that's where we can bring our Salesforce, our capabilities, our understanding of the customer. 'cause even your original question around AT&T and trying to understand the data, that's just really a how do you get better understanding of your customer, right? >> Right. Absolutely. >> And, and combined we're better together 'cause we bring a more complete picture of understanding our customers and then how can we help them understand what the edge is. Cause nobody's ever bought an Edge, right? They're buying an Edge to get a business outcome. You know, back in the day, nobody ever bought a data lake, right? Like, you know, they're buying an outcome. They want to use, use that data lake or they want to use the edge to deliver something. They want to use 5G. And 5G has very real capabilities. It's got intrinsic security, which, you know a lot of the wifi doesn't. It's got guaranteed on time, you know, for areas where you can't lose connectivity: autonomous vehicles, et cetera. So it's got very real capabilities that helps deliver that outcome. But you got to be able to translate that into the en- enterprise language to help them solve a problem. And that's where we think we need the help of the telcos. I think the telcos we can help them as well and, and really go drive that outcome. >> So Dell's bringing its go to market expertise and its technology. The telcos obviously have the the connectivity piece and what they do. There's no overlap in terms of the... >> Yeah. >> The, the equipment and the software that you're selling. I mean, they're going to, they're going to take your equipment and create new networks. Beautiful. And, and it's interesting you, like, you think about how Dell has transformed prior to EMC, Dell was, you know, PC maker with a subpar enterprise business, right? Kind of a wannabe enterprise business. Sorry Dell, it's the truth. And then EMC was largely, you know, a company sold storage boxes, but you owned VMware and then brought those two together. Now all of a sudden you had Dell powerhouse leader and Michael Dell, you had VMware incredibly strategic and important and it got EMC with amazing go to market. All of a sudden this Dell, Dell technologies became incredibly attractive to CIOs, C-level executives, board level. And you've come out of that transition VMware's now a separate company, right? And now, but now you have these relationships and you got the shops to be able to go into these edge locations at companies And actually go partner with the telcos. And you got a very compelling value proposition. >> Well, it's been interesting as in, in this show, again most telcos think of Dell as a server provider, you know? Important, but not overly strategic in their journey. But as we've started to invest in this business we've started to invest in things like automation. We've brought together things in our Infra Blocks and then we help them develop revenue. We're not only helping 'em take costs out of their network we're not helping 'em take risk out of deploying that network. We're helping them accelerate the deployment of that network. And then we're helping 'em drive revenue. We are having, you know, they're starting to see us in a new light. Not done yet, but, you know, you can start to see, one, how they're looking at Dell and two, and then how we can go to market. And you know, a big part of that is helping 'em drive and generate revenue. >> Yeah. Well, as, as a, as a former EMC person myself, >> Yeah? >> I will assert that that strategic DNA was injected into Dell by the acquisition of, of EMC. And I'm sticking... >> I won't say that. Okay I'll believe you on that. >> I'm sticking with the story. And it makes sense when you think about moving up market, that's the natural thing. What's, what's what's nearly impossible is to say, we sell semi-trucks but we want to get into the personal pickup truck market. That's that, that doesn't work. Going the other way works. >> Dave Trigg: Yeah. >> Now, now back to the conversation that you had with, with, with AT&T. I'm not buying this whole, no offense to AT&T, but I'm not buying this whole story that, you know, oh we're concerned about our branded customer data. That sounds like someone who's a little bit too comfortable with their existing revenue stream. If I'm out there, I want to be out partnering with folks who are truly aggressive about, about coming up with the next cool thing. You guys are talking about being connected in a laptop. Someone would say, well I got wifi. No, no, no. I'm thinking I want to sim in my laptop cause I don't want to screw around with wifi. Okay, fine. If I know I'm going to be somewhere with excellent wifi connectivity, great. But most of the time it's not excellent. >> That's right. >> So the idea that I could maybe hit F2 and have it switch over to my sim and know that anywhere that I've got coverage, I have high speed connections. Just the convenience of that. >> Ken: Absolutely. >> I'd pay extra for that as an end user consumer. >> Absolutely. >> And I pay for the service. >> Like I tell you, if it interests AT&T I think it's more not, they ask, they're comfortable. They don't know how to monetize that data. Now, of course, AT&T has a media >> Dave Nicholson: Business necessity is the mother of invention. If they don't see the necessity then they're not going to think about it. >> It's a mentality shift. Yes, but, but when you start talking about private mobility and edge, there's there's no concern about personal information there. You're going in with basically a business transformation. Hey, your, your business is, is not, not digital. It's not automated. Now we're going to automate that and digitize that. It's like the, the Dell booth with the beer guys. >> Right. >> You saw that, right? >> I mean that's, I mean that's a simple application. Yeah, a perfect example of how you network and use this technology. >> I mean, how many non-digital businesses are that that need to go digital? >> Dave Nicholson: Like, hundred percent of them. >> Everyone. >> Dave Nicholson: Pretty much. >> Yeah. And this, and this jewel that we have inside of Dell our global industries group, right, where we're investing really heavily in terms of what is the manufacturing industry looking for retail, finance, et cetera. So we have a CTO that came in, that it would be the CTO of manufacturing that gives us a really good opportunity to go to at AT&T or to Verizon or any telco out there, right? To, to say, these are the outcomes. There's Dell technology already in place. How do we connect it to your network? How do we leverage your assets, your manager professional services to provide a richer experience? So it's, there's, you said before Dave, there's really no overlap between Dell and, and our telecom partners. >> You guys making some serious investments here. I mean I, I've been, I was been critical over the years of, hey, you can't just take an X86 block, put a name on it that says edge something and throw it over the fence because that's what you were doing. >> Dave Trigg: And we would agree. >> Yeah. Right. But, of course, but that's all you had at the time. And so you put some... >> We may not have agreed then, but we would agree. >> You bought, brought some people in, you know, like Ken, who really know the business. You brought people into the technical side and you can really see it happening. It's not going to happen overnight. You know, I mean, you know if I were an investor in Dell, I'd be like, okay when are you going to start making money at this business? I'd be like, be patient. You know, it's going to take some time but look at the TAM. >> Yep. >> You know, you guys do a good, good TAM. Tennis is a pro at this stuff. >> We've been at, we've been at this two, three years and we're just now coming with some real material products. You've seen our server line really start to get more purpose-built, really start to get in there as we've started to put out some software that allows for quicker automation, quicker deployments. We have some telcos that are using it to deploy at 10,000 locations. They're literally turning up thousands of locations a week. And so yeah, we're starting to put out some real capability. Got a long way to go. A lot of exciting things on the roadmap. But to your point, it doesn't, you know the ship doesn't turn overnight, you know. >> It could be a really meaningful portion of Dell's business. I'm, I'm excited for the day that Tom Sweet starts reporting on it. Here's our telco business. Yeah. The telco business. But that's not going to happen overnight. But you know, Dell's pretty good at things like ROI. And so you guys do a lot of planning a lot of TAM analysis, a lot of technical analysis, bringing the ecosystem together. That's what this business needs. I, I just don't, it's, it feels unstoppable. You know, you're at this show everybody recognizes the need to open up. Some telcos are moving faster than others. The ones that move faster are going to disrupt. They're going to probably make some mistakes, you know but they're going to get there first. >> Well we've, we've seen the disruptors are making some mistakes and are kind of re- they're already at the phase where they're reevaluating, you know, their approach. Which is great. You know, you, you learn and adjust. You know, you run into a wall, you, you make a turn. And the interesting thing, one of the biggest learnings I've taken out of the show is talking to a bunch of the telcos that are a little bit more of the laggards. They're like, Nope, we, we don't believe in open. We don't think we can do it. We don't have the skillset. They're maybe in a geo that it's hard to find the skillset. As they've been talking to us, and we've been talking about, there's almost a glimmer of hope. They're not convinced yet, but they're like, well wait, maybe we can do this. Maybe open, you know, does give us choice. Maybe it can help us accelerate revenue. So it's been interesting to see a little bit of the, just a little bit, but a little bit of that shift. >> We all remember at 2010, 2011, you talked to banks and financial services companies about, the heck, the Cloud is happening, the Cloud's going to take over the world. We're never going to go into the Cloud. Now they're the biggest, you know Capital One's launching Cloud businesses, Western Union, I mean, they're all in the cloud, right? I mean, it's the same thing's going to happen here. Might, it might take a different pattern. Maybe it takes a little longer, but it's, it's it's a fate are completely >> I was in high school then, so I don't remember all that. >> Sorry, Dave. >> Wow, that was a low blow, like you know? >> But, but the, but the one thing that is for sure there's money to be made convincing people to get off of the backs of the dinosaurs they're riding. >> Dave Vellante: That's right. >> And also, the other thing that's a certainty is that it's not easy. And because it's not easy, there's opportunity there. So I know, I know it's, it, it, it, it, it all sounds great to talk about the the wonderful vision of the future, but I know how hard the the road is that you have to go down to get people, especially if you're comfortable with the revenue stream, if you're comfortable running the plumbing. If you're so comfortable that you can get up on stage and say, I want more money from you to pump your con- your content across my network. I love the Netflix retort, right Dave? >> Yeah, totally Dave. And, but the, the other thing is, telco's a great business. It's, they got monopolies that print money. So... >> Dave Nicholson: It's rational. It's rational. I understand. >> There's less of an incentive to move but what's going to be the incentive is guys like Dish Network coming in saying, we're going to, we're going to disrupt, we're going to build new apps. >> That's right. >> Yeah. >> Well and it's, you know, revenue acceleration, the board level, the CEO level know that they have to, you know, do things different. But to your point, it's just hard, and there's so much gravity there. There's hundreds of years literally of gravity of how they've operated their business. To your point, a lot of them, you know, lot- most of 'em were regulated and most Geos around the world at one point, right? They were government owned or government regulated entities. It's, it's a big ship to turn and it's really hard. We're not claiming we can help them turn the ship overnight but we think we can help evolve them. We think we can go along with the journey and we do think we are better together. >> IT the network and the line of business. Love the strategy. Guys, thanks so much for coming in theCUBE. >> Thank you so much. >> Thank you. >> All right, for Dave, Nicholson, Dave Vellante here, John Furrier is in our Palo Alto studio banging out all the news, keep it right there. TheCUBE's coverage of MWC 23. We'll be right back.

Published Date : Mar 2 2023

SUMMARY :

that drive human progress. of the telco networks, how Great to be here. for the telcos to make money. 90% of the companies leaning in on that. So Ken, you know, the space of our platinum brand to the If they're not going to go toward that of how to solve those business outcomes. the telcos are starting to the end customer. allowing the telecom companies to come in and they're all at different levels and all they have to do is collect, I got to tell you they're pretty excited So that's the big question, right? And really to get Right. a lot of the wifi doesn't. the connectivity piece and what they do. And then EMC was largely, you know, And you know, a big part a former EMC person myself, into Dell by the acquisition I'll believe you on that. And it makes sense when you think about But most of the time it's not excellent. So the idea that I could I'd pay extra for that They don't know how to monetize that data. then they're not going to think about it. Yes, but, but when you start talking Yeah, a perfect example of how you network Dave Nicholson: Like, a really good opportunity to over the years of, hey, you And so you put some... then, but we would agree. You know, it's going to take some time You know, you guys do a good, good TAM. the ship doesn't turn overnight, you know. everybody recognizes the need to open up. of the telcos that are a little the Cloud's going to take over the world. I was in high school then, there's money to be made the road is that you have that print money. I understand. There's less of an incentive to move of them, you know, lot- the line of business. banging out all the news,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

DavePERSON

0.99+

Ken BurnsPERSON

0.99+

Dave NicholsonPERSON

0.99+

Dave TriggPERSON

0.99+

KenPERSON

0.99+

VerizonORGANIZATION

0.99+

Dave VellantePERSON

0.99+

DellORGANIZATION

0.99+

AT&TORGANIZATION

0.99+

NicholsonPERSON

0.99+

Ken ByrnesPERSON

0.99+

AT&T.ORGANIZATION

0.99+

David TriggPERSON

0.99+

millionsQUANTITY

0.99+

NetflixORGANIZATION

0.99+

Tom SweetPERSON

0.99+

Capital OneORGANIZATION

0.99+

90%QUANTITY

0.99+

Dell TechnologiesORGANIZATION

0.99+

EMCORGANIZATION

0.99+

2010DATE

0.99+

10,000 locationsQUANTITY

0.99+

Dish NetworkORGANIZATION

0.99+

VMwareORGANIZATION

0.99+

Michael DellPERSON

0.99+

Palo AltoLOCATION

0.99+

bothQUANTITY

0.99+

thousandsQUANTITY

0.99+

35,000 sellersQUANTITY

0.99+

twoQUANTITY

0.99+

BarcelonaLOCATION

0.99+

hundred percentQUANTITY

0.99+

John FurrierPERSON

0.99+

30 plus telecom operatorsQUANTITY

0.99+

three yearsQUANTITY

0.99+

Ed Walsh & Thomas Hazel | A New Database Architecture for Supercloud


 

(bright music) >> Hi, everybody, this is Dave Vellante, welcome back to Supercloud 2. Last August, at the first Supercloud event, we invited the broader community to help further define Supercloud, we assessed its viability, and identified the critical elements and deployment models of the concept. The objectives here at Supercloud too are, first of all, to continue to tighten and test the concept, the second is, we want to get real world input from practitioners on the problems that they're facing and the viability of Supercloud in terms of applying it to their business. So on the program, we got companies like Walmart, Sachs, Western Union, Ionis Pharmaceuticals, NASDAQ, and others. And the third thing that we want to do is we want to drill into the intersection of cloud and data to project what the future looks like in the context of Supercloud. So in this segment, we want to explore the concept of data architectures and what's going to be required for Supercloud. And I'm pleased to welcome one of our Supercloud sponsors, ChaosSearch, Ed Walsh is the CEO of the company, with Thomas Hazel, who's the Founder, CTO, and Chief Scientist. Guys, good to see you again, thanks for coming into our Marlborough studio. >> Always great. >> Great to be here. >> Okay, so there's a little debate, I'm going to put you right in the spot. (Ed chuckling) A little debate going on in the community started by Bob Muglia, a former CEO of Snowflake, and he was at Microsoft for a long time, and he looked at the Supercloud definition, said, "I think you need to tighten it up a little bit." So, here's what he came up with. He said, "A Supercloud is a platform that provides a programmatically consistent set of services hosted on heterogeneous cloud providers." So he's calling it a platform, not an architecture, which was kind of interesting. And so presumably the platform owner is going to be responsible for the architecture, but Dr. Nelu Mihai, who's a computer scientist behind the Cloud of Clouds Project, he chimed in and responded with the following. He said, "Cloud is a programming paradigm supporting the entire lifecycle of applications with data and logic natively distributed. Supercloud is an open architecture that integrates heterogeneous clouds in an agnostic manner." So, Ed, words matter. Is this an architecture or is it a platform? >> Put us on the spot. So, I'm sure you have concepts, I would say it's an architectural or design principle. Listen, I look at Supercloud as a mega trend, just like cloud, just like data analytics. And some companies are using the principle, design principles, to literally get dramatically ahead of everyone else. I mean, things you couldn't possibly do if you didn't use cloud principles, right? So I think it's a Supercloud effect, you're able to do things you're not able to. So I think it's more a design principle, but if you do it right, you get dramatic effect as far as customer value. >> So the conversation that we were having with Muglia, and Tristan Handy of dbt Labs, was, I'll set it up as the following, and, Thomas, would love to get your thoughts, if you have a CRM, think about applications today, it's all about forms and codifying business processes, you type a bunch of stuff into Salesforce, and all the salespeople do it, and this machine generates a forecast. What if you have this new type of data app that pulls data from the transaction system, the e-commerce, the supply chain, the partner ecosystem, et cetera, and then, without humans, actually comes up with a plan. That's their vision. And Muglia was saying, in order to do that, you need to rethink data architectures and database architectures specifically, you need to get down to the level of how the data is stored on the disc. What are your thoughts on that? Well, first of all, I'm going to cop out, I think it's actually both. I do think it's a design principle, I think it's not open technology, but open APIs, open access, and you can build a platform on that design principle architecture. Now, I'm a database person, I love solving the database problems. >> I'm waited for you to launch into this. >> Yeah, so I mean, you know, Snowflake is a database, right? It's a distributed database. And we wanted to crack those codes, because, multi-region, multi-cloud, customers wanted access to their data, and their data is in a variety of forms, all these services that you're talked about. And so what I saw as a core principle was cloud object storage, everyone streams their data to cloud object storage. From there we said, well, how about we rethink database architecture, rethink file format, so that we can take each one of these services and bring them together, whether distributively or centrally, such that customers can access and get answers, whether it's operational data, whether it's business data, AKA search, or SQL, complex distributed joins. But we had to rethink the architecture. I like to say we're not a first generation, or a second, we're a third generation distributed database on pure, pure cloud storage, no caching, no SSDs. Why? Because all that availability, the cost of time, is a struggle, and cloud object storage, we think, is the answer. >> So when you're saying no caching, so when I think about how companies are solving some, you know, pretty hairy problems, take MySQL Heatwave, everybody thought Oracle was going to just forget about MySQL, well, they come out with Heatwave. And the way they solve problems, and you see their benchmarks against Amazon, "Oh, we crush everybody," is they put it all in memory. So you said no caching? You're not getting performance through caching? How is that true, and how are you getting performance? >> Well, so five, six years ago, right? When you realize that cloud object storage is going to be everywhere, and it's going to be a core foundational, if you will, fabric, what would you do? Well, a lot of times the second generation say, "We'll take it out of cloud storage, put in SSDs or something, and put into cache." And that adds a lot of time, adds a lot of costs. But I said, what if, what if we could actually make the first read hot, the first read distributed joins and searching? And so what we went out to do was said, we can't cache, because that's adds time, that adds cost. We have to make cloud object storage high performance, like it feels like a caching SSD. That's where our patents are, that's where our technology is, and we've spent many years working towards this. So, to me, if you can crack that code, a lot of these issues we're talking about, multi-region, multicloud, different services, everybody wants to send their data to the data lake, but then they move it out, we said, "Keep it right there." >> You nailed it, the data gravity. So, Bob's right, the data's coming in, and you need to get the data from everywhere, but you need an environment that you can deal with all that different schema, all the different type of technology, but also at scale. Bob's right, you cannot use memory or SSDs to cache that, that doesn't scale, it doesn't scale cost effectively. But if you could, and what you did, is you made object storage, S3 first, but object storage, the only persistence by doing that. And then we get performance, we should talk about it, it's literally, you know, hundreds of terabytes of queries, and it's done in seconds, it's done without memory caching. We have concepts of caching, but the only caching, the only persistence, is actually when we're doing caching, we're just keeping another side-eye track of things on the S3 itself. So we're using, actually, the object storage to be a database, which is kind of where Bob was saying, we agree, but that's what you started at, people thought you were crazy. >> And maybe make it live. Don't think of it as archival or temporary space, make it live, real time streaming, operational data. What we do is make it smart, we see the data coming in, we uniquely index it such that you can get your use cases, that are search, observability, security, or backend operational. But we don't have to have this, I dunno, static, fixed, siloed type of architecture technologies that were traditionally built prior to Supercloud thinking. >> And you don't have to move everything, essentially, you can do it wherever the data lands, whatever cloud across the globe, you're able to bring it together, you get the cost effectiveness, because the only persistence is the cheapest storage persistent layer you can buy. But the key thing is you cracked the code. >> We had to crack the code, right? That was the key thing. >> That's where the plans are. >> And then once you do that, then everything else gets easier to scale, your architecture, across regions, across cloud. >> Now, it's a general purpose database, as Bob was saying, but we use that database to solve a particular issue, which is around operational data, right? So, we agree with Bob's. >> Interesting. So this brings me to this concept of data, Jimata Gan is one of our speakers, you know, we talk about data fabric, which is a NetApp, originally NetApp concept, Gartner's kind of co-opted it. But so, the basic concept is, data lives everywhere, whether it's an S3 bucket, or a SQL database, or a data lake, it's just a node on the data mesh. So in your view, how does this fit in with Supercloud? Ed, you've said that you've built, essentially, an enabler for that, for the data mesh, I think you're an enabler for the Supercloud-like principles. This is a big, chewy opportunity, and it requires, you know, a team approach. There's got to be an ecosystem, there's not going to be one Supercloud to rule them all, so where does the ecosystem fit into the discussion, and where do you fit into the ecosystem? >> Right, so we agree completely, there's not one Supercloud in effect, but we use Supercloud principles to build our platform, and then, you know, the ecosystem's going to be built on leveraging what everyone else's secret powers are, right? So our power, our superpower, based upon what we built is, we deal with, if you're having any scale, or cost effective scale issues, with data, machine generated data, like business observability or security data, we are your force multiplier, we will take that in singularly, just let it, simply put it in your object storage wherever it sits, and we give you uniformity access to that using OpenAPI access, SQL, or you know, Elasticsearch API. So, that's what we do, that's our superpower. So I'll play it into data mesh, that's a perfect, we are a node on a data mesh, but I'll play it in the soup about how, the ecosystem, we see it kind of playing, and we talked about it in just in the last couple days, how we see this kind of possibly. Short term, our superpowers, we deal with this data that's coming at these environments, people, customers, building out observability or security environments, or vendors that are selling their own Supercloud, I do observability, the Datadogs of the world, dot dot dot, the Splunks of the world, dot dot dot, and security. So what we do is we fit in naturally. What we do is a cost effective scale, just land it anywhere in the world, we deal with ingest, and it's a cost effective, an order of magnitude, or two or three order magnitudes more cost effective. Allows them, their customers are asking them to do the impossible, "Give me fast monitoring alerting. I want it snappy, but I want it to keep two years of data, (laughs) and I want it cost effective." It doesn't work. They're good at the fast monitoring alerting, we're good at the long-term retention. And yet there's some gray area between those two, but one to one is actually cheaper, so we would partner. So the first ecosystem plays, who wants to have the ability to, really, all the data's in those same environments, the security observability players, they can literally, just through API, drag our data into their point to grab. We can make it seamless for customers. Right now, we make it helpful to customers. Your Datadog, we make a button, easy go from Datadog to us for logs, save you money. Same thing with Grafana. But you can also look at ecosystem, those same vendors, it used to be a year ago it was, you know, its all about how can you grow, like it's growth at all costs, now it's about cogs. So literally we can go an environment, you supply what your customer wants, but we can help with cogs. And one-on one in a partnership is better than you trying to build on your own. >> Thomas, you were saying you make the first read fast, so you think about Snowflake. Everybody wants to talk about Snowflake and Databricks. So, Snowflake, great, but you got to get the data in there. All right, so that's, can you help with that problem? >> I mean we want simple in, right? And if you have to have structure in, you're not simple. So the idea that you have a simple in, data lake, schema read type philosophy, but schema right type performance. And so what I wanted to do, what we have done, is have that simple lake, and stream that data real time, and those access points of Search or SQL, to go after whatever business case you need, security observability, warehouse integration. But the key thing is, how do I make that click, click, click answer, and do it quickly? And so what we want to do is, that first read has to be fast. Why? 'Cause then you're going to do all this siloing, layers, complexity. If your first read's not fast, you're at a disadvantage, particularly in cost. And nobody says I want less data, but everyone has to, whether they say we're going to shorten the window, we're going to use AI to choose, but in a security moment, when you don't have that answer, you're in trouble. And that's why we are this service, this Supercloud service, if you will, providing access, well-known search, well-known SQL type access, that if you just have one access point, you're at a disadvantage. >> We actually talked about Snowflake and BigQuery, and a different platform, Data Bricks. That's kind of where we see the phase two of ecosystem. One is easy, the low-hanging fruit is observability and security firms. But the next one is, what we do, our super power is dealing with this messy data that schema is changing like night and day. Pipelines are tough, and it's changing all the time, but you want these things fast, and it's big data around the world. That's the next point, just use us alongside, or inside, one of their platforms, and now we get the best of both worlds. Our superpower is keeping this messy data as a streaming, okay, not a batch thing, allow you to do that. So, that's the second one. And then to be honest, the third one, which plays you to Supercloud, it also plays perfectly in the data mesh, is if you really go to the ultimate thing, what we have done is made object storage, S3, GCS, and blob storage, we made it a database. Put, get, complex query with big joins. You know, so back to your original thing, and Muglia teed it up perfectly, we've done that. Now imagine if that's an ecosystem, who would want that? If it's, again, it's uniform available across all the regions, across all the clouds, and it's right next to where you are building a service, or a client's trying, that's where the ecosystem, I think people are going to use Superclouds for their superpowers. We're really good at this, allows that short term. I think the Snowflakes and the Data Bricks are the medium term, you know? And then I think eventually gets to, hey, listen if you can make object storage fast, you can just go after it with simple SQL queries, or elastic. Who would want that? I think that's where people are going to leverage it. It's not going to be one Supercloud, and we leverage the super clouds. >> Our viewpoint is smart object storage can be programmable, and so we agree with Bob, but we're not saying do it here, do it here. This core, fundamental layer across regions, across clouds, that everyone has? Simple in. Right now, it's hard to get data in for access for analysis. So we said, simply, we'll automate the entire process, give you API access across regions, across clouds. And again, how do you do a distributed join that's fast? How do you do a distributed join that doesn't cost you an arm or a leg? And how do you do it at scale? And that's where we've been focused. >> So prior, the cloud object store was a niche. >> Yeah. >> S3 obviously changed that. How standard is, essentially, object store across the different cloud platforms? Is that a problem for you? Is that an easy thing to solve? >> Well, let's talk about it. I mean we've fundamentally, yeah we've extracted it, but fundamentally, cloud object storage, put, get, and list. That's why it's so scalable, 'cause it doesn't have all these other components. That complexity is where we have moved up, and provide direct analytical API access. So because of its simplicity, and costs, and security, and reliability, it can scale naturally. I mean, really, distributed object storage is easy, it's put-get anywhere, now what we've done is we put a layer of intelligence, you know, call it smart object storage, where access is simple. So whether it's multi-region, do a query across, or multicloud, do a query across, or hunting, searching. >> We've had clients doing Amazon and Google, we have some Azure, but we see Amazon and Google more, and it's a consistent service across all of them. Just literally put your data in the bucket of choice, or folder of choice, click a couple buttons, literally click that to say "that's hot," and after that, it's hot, you can see it. But we're not moving data, the data gravity issue, that's the other. That it's already natively flowing to these pools of object storage across different regions and clouds. We don't move it, we index it right there, we're spinning up stateless compute, back to the Supercloud concept. But now that allows us to do all these other things, right? >> And it's no longer just cheap and deep object storage. Right? >> Yeah, we make it the same, like you have an analytic platform regardless of where you're at, you don't have to worry about that. Yeah, we deal with that, we deal with a stateless compute coming up -- >> And make it programmable. Be able to say, "I want this bucket to provide these answers." Right, that's really the hope, the vision. And the complexity to build the entire stack, and then connect them together, we said, the fabric is cloud storage, we just provide the intelligence on top. >> Let's bring it back to the customers, and one of the things we're exploring in Supercloud too is, you know, is Supercloud a solution looking for a problem? Is a multicloud really a problem? I mean, you hear, you know, a lot of the vendor marketing says, "Oh, it's a disaster, because it's all different across the clouds." And I talked to a lot of customers even as part of Supercloud too, they're like, "Well, I solved that problem by just going mono cloud." Well, but then you're not able to take advantage of a lot of the capabilities and the primitives that, you know, like Google's data, or you like Microsoft's simplicity, their RPA, whatever it is. So what are customers telling you, what are their near term problems that they're trying to solve today, and how are they thinking about the future? >> Listen, it's a real problem. I think it started, I think this is a a mega trend, just like cloud. Just, cloud data, and I always add, analytics, are the mega trends. If you're looking at those, if you're not considering using the Supercloud principles, in other words, leveraging what I have, abstracting it out, and getting the most out of that, and then build value on top, I think you're not going to be able to keep up, In fact, no way you're going to keep up with this data volume. It's a geometric challenge, and you're trying to do linear things. So clients aren't necessarily asking, hey, for Supercloud, but they're really saying, I need to have a better mechanism to simplify this and get value across it, and how do you abstract that out to do that? And that's where they're obviously, our conversations are more amazed what we're able to do, and what they're able to do with our platform, because if you think of what we've done, the S3, or GCS, or object storage, is they can't imagine the ingest, they can't imagine how easy, time to glass, one minute, no matter where it lands in the world, querying this in seconds for hundreds of terabytes squared. People are amazed, but that's kind of, so they're not asking for that, but they are amazed. And then when you start talking on it, if you're an enterprise person, you're building a big cloud data platform, or doing data or analytics, if you're not trying to leverage the public clouds, and somehow leverage all of them, and then build on top, then I think you're missing it. So they might not be asking for it, but they're doing it. >> And they're looking for a lens, you mentioned all these different services, how do I bring those together quickly? You know, our viewpoint, our service, is I have all these streams of data, create a lens where they want to go after it via search, go after via SQL, bring them together instantly, no e-tailing out, no define this table, put into this database. We said, let's have a service that creates a lens across all these streams, and then make those connections. I want to take my CRM with my Google AdWords, and maybe my Salesforce, how do I do analysis? Maybe I want to hunt first, maybe I want to join, maybe I want to add another stream to it. And so our viewpoint is, it's so natural to get into these lake platforms and then provide lenses to get that access. >> And they don't want it separate, they don't want something different here, and different there. They want it basically -- >> So this is our industry, right? If something new comes out, remember virtualization came out, "Oh my God, this is so great, it's going to solve all these problems." And all of a sudden it just got to be this big, more complex thing. Same thing with cloud, you know? It started out with S3, and then EC2, and now hundreds and hundreds of different services. So, it's a complex matter for a lot of people, and this creates problems for customers, especially when you got divisions that are using different clouds, and you're saying that the solution, or a solution for the part of the problem, is to really allow the data to stay in place on S3, use that standard, super simple, but then give it what, Ed, you've called superpower a couple of times, to make it fast, make it inexpensive, and allow you to do that across clouds. >> Yeah, yeah. >> I'll give you guys the last word on that. >> No, listen, I think, we think Supercloud allows you to do a lot more. And for us, data, everyone says more data, more problems, more budget issue, everyone knows more data is better, and we show you how to do it cost effectively at scale. And we couldn't have done it without the design principles of we're leveraging the Supercloud to get capabilities, and because we use super, just the object storage, we're able to get these capabilities of ingest, scale, cost effectiveness, and then we built on top of this. In the end, a database is a data platform that allows you to go after everything distributed, and to get one platform for analytics, no matter where it lands, that's where we think the Supercloud concepts are perfect, that's where our clients are seeing it, and we're kind of excited about it. >> Yeah a third generation database, Supercloud database, however we want to phrase it, and make it simple, but provide the value, and make it instant. >> Guys, thanks so much for coming into the studio today, I really thank you for your support of theCUBE, and theCUBE community, it allows us to provide events like this and free content. I really appreciate it. >> Oh, thank you. >> Thank you. >> All right, this is Dave Vellante for John Furrier in theCUBE community, thanks for being with us today. You're watching Supercloud 2, keep it right there for more thought provoking discussions around the future of cloud and data. (bright music)

Published Date : Feb 17 2023

SUMMARY :

And the third thing that we want to do I'm going to put you right but if you do it right, So the conversation that we were having I like to say we're not a and you see their So, to me, if you can crack that code, and you need to get the you can get your use cases, But the key thing is you cracked the code. We had to crack the code, right? And then once you do that, So, we agree with Bob's. and where do you fit into the ecosystem? and we give you uniformity access to that so you think about Snowflake. So the idea that you have are the medium term, you know? and so we agree with Bob, So prior, the cloud that an easy thing to solve? you know, call it smart object storage, and after that, it's hot, you can see it. And it's no longer just you don't have to worry about And the complexity to and one of the things we're and how do you abstract it's so natural to get and different there. and allow you to do that across clouds. I'll give you guys and we show you how to do it but provide the value, I really thank you for around the future of cloud and data.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
WalmartORGANIZATION

0.99+

Dave VellantePERSON

0.99+

NASDAQORGANIZATION

0.99+

Bob MugliaPERSON

0.99+

ThomasPERSON

0.99+

Thomas HazelPERSON

0.99+

Ionis PharmaceuticalsORGANIZATION

0.99+

Western UnionORGANIZATION

0.99+

Ed WalshPERSON

0.99+

BobPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Nelu MihaiPERSON

0.99+

SachsORGANIZATION

0.99+

Tristan HandyPERSON

0.99+

twoQUANTITY

0.99+

AmazonORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

two yearsQUANTITY

0.99+

Supercloud 2TITLE

0.99+

firstQUANTITY

0.99+

Last AugustDATE

0.99+

threeQUANTITY

0.99+

OracleORGANIZATION

0.99+

SnowflakeORGANIZATION

0.99+

bothQUANTITY

0.99+

dbt LabsORGANIZATION

0.99+

John FurrierPERSON

0.99+

EdPERSON

0.99+

GartnerORGANIZATION

0.99+

Jimata GanPERSON

0.99+

third oneQUANTITY

0.99+

one minuteQUANTITY

0.99+

secondQUANTITY

0.99+

first generationQUANTITY

0.99+

third generationQUANTITY

0.99+

GrafanaORGANIZATION

0.99+

second generationQUANTITY

0.99+

second oneQUANTITY

0.99+

hundreds of terabytesQUANTITY

0.98+

SQLTITLE

0.98+

fiveDATE

0.98+

oneQUANTITY

0.98+

DatabricksORGANIZATION

0.98+

a year agoDATE

0.98+

ChaosSearchORGANIZATION

0.98+

MugliaPERSON

0.98+

MySQLTITLE

0.98+

both worldsQUANTITY

0.98+

third thingQUANTITY

0.97+

MarlboroughLOCATION

0.97+

theCUBEORGANIZATION

0.97+

todayDATE

0.97+

SupercloudORGANIZATION

0.97+

ElasticsearchTITLE

0.96+

NetAppTITLE

0.96+

DatadogORGANIZATION

0.96+

OneQUANTITY

0.96+

EC2TITLE

0.96+

each oneQUANTITY

0.96+

S3TITLE

0.96+

one platformQUANTITY

0.95+

Supercloud 2EVENT

0.95+

first readQUANTITY

0.95+

six years agoDATE

0.95+

Welcome to Supercloud2


 

(bright upbeat melody) >> Hello everyone, welcome back to Supercloud2. I'm John Furrier, my co-host Dave Vellante, here at theCUBE in Palo Alto, California, for our live stage performance all day for Supercloud2. Unpacking this next generation movement in cloud computing. Dave, Supercloud1 was in August. We had great response and acceleration of that momentum. We had some haters too. We had some folks out there throwing shade on this. But at the same time, a lot of leaders came out of the woodwork, a lot of practitioners. And this Supercloud2 event I think will expose and illustrate some of the examples of what's happening in the industry and more importantly, kind of where it's going. >> Well it's great to be back in our studios in Palo Alto, John. Seems like just yesterday was August 9th, where the community was really refining the definition of Super Cloud. We were identifying the essential characteristics, with some of the leading technologists in Silicon Valley. We were digging into the deployment models. Whereas this Supercloud, Supercloud2 is really taking a practitioner view. We're going to hear from Walmart today. They've built a Supercloud. They called it the Walmart Cloud native platform. We're going to hear from other data practitioners, like Saks. We're going to hear from Western Union. They've got 200 locations around the world, how they're dealing with data sovereignty. And of course we've got some local technologists and practitioners coming in, analysts, consultants, theCUBE community. I'm really excited to be here. >> And we've got some great keynotes from executives at VMware. We're going to expose some of the things that they're working on around cross cloud services, which leads into multicloud. I think the practitioner angle highlights my favorite part of this program, 'cause you're starting to see the builders, a term coined by Andy Jassy, early days of AWS. That builder movement has been continuing to go. And you're seeing the enterprise, global enterprises adopt this builder mentality with Cloud Native. This is going to power the next generation global economy. And I think the role of the cloud computing vendors like AWS, Azure, Google, Alibaba are going to be the source engine of innovation. And what gets built on top of and with the clouds will be a big significant market value for all businesses and their business models. So I think the market wants the supercloud, the business models are pointing to Supercloud. The technology needs supercloud. And society, from an economic standpoint and from a use case standpoint, needs supercloud. You're seeing it today. Everyone's talking about chat GPT. This is an example of what will come out of this next generation and it's just getting started. So to me, you're either on the supercloud side of the camp or you're on the old school, hugging onto the old school mentality of wait a minute, that's cloud computing. So I think if you're not on the super cloud wave, you're going to be driftwood. And that's a term coined by Pat Gelsinger. And this is really the reality. Are you on the super cloud side? Or are you on the old huggin' the old model? And that's going to be a determinant. And you're going to see who's going to be the players on that, Dave. This is going to be a real big year. >> Everybody's heard the phrase follow the money. Well, my philosophy is follow the data. And that's a big part of what Supercloud2 is, because the data is where the money is across the clouds. And people want more simplicity, or greater simplicity across the clouds. So it's really, there's two forces here. You've got the ecosystem that's saying, hey the hyperscalers, they've done a great job but there's problems that they're not solving. So we're going to lean in and solve those problems. At the same time, you have the practitioners saying we have multicloud, we have to deal with this, help us. It's got to be simpler. Because we want to share data across clouds. We want to build data products, we want to monetize and drive revenue and cut costs. >> This is the key thing. The builder movement is hitting a wall, and that wall will be broken down because the business models of the companies themselves are demanding that the value from the data with security has to be embedded. So I think you're going to see a big year this next year or so where the builders will accelerate through this next generation, supercloud wave, will be a builder's wave for business. And I think that's going to be the nuance here. And all the people that are on the side of Supercloud are all pro-business, pro-technology. The ones that aren't are like, wait a minute I used to do things differently. They're stuck. And so I think this is going to be a question of are we stuck? Are builders accelerating? Will the business models develop around it? That's digital transformation. At the end of the day, the market's speaking, Dave. The market wants more. Chat GPT, you're seeing AI starting to flourish, powered by data. It's unstoppable, supercloud's unstoppable. >> One of our headliners today is Zhamak Dehghani, the creator of Data Mesh. We've got some news around her. She's going to be live in studio. Super excited about that. Kit Colbert in Supercloud, the first Supercloud in last August, laid out an initial architecture for Supercloud. He's going to advance that today, tell us what's changed, and really dig into and really talk about the meat on the bone, if you will. And we've got some other technologists that are coming in saying, Hey, is it a platform? Is it an architecture? What's the right model here? So we're going to debate that a little bit today. >> And before we close, I'll just say look at the guests, look at the talk tracks. You're seeing a diversity of startups doing cloud networking, you're seeing big practitioners building their own thing, being builders for business value and business model advantages. And you got companies like VMware, who have been on the wave of virtualization. So the, everyone who's involved in super cloud, they're seeing it, they're on the front lines. They're seeing the trend. They are riding that wave. And they have, they're bringing data to the table. So to me, you look at who's involved and you judge it that way. To me, that's the way I look at this. And because we're making it open, Supercloud is going to continue to be debated. But more importantly, the results are going to come in. The market supports it, the business needs it, tech's there, and will it happen? So I think the builders movement, Dave, is going to be big to watch. And then ultimately how that business transformation kicks in, and I think those are the two variables that I would watch on Supercloud. >> Our mission has always been around free content, giving back to the community. So I really want to thank our sponsors today. We've had a great partnership with VMware, who's not only contributed some financial support, but also great content. Alkira, ChaosSearch, prosimo, all phenomenal, allowing us to achieve our mission of serving our audiences and really trying to give more than we take from. >> Free content, that's our mission. Dave, great to kick it off. Kickin' off Supercloud2 all day, we've got some great programs here. We've got VMware coming up next. We have Victoria Viering, who's been on before. He's got a great vision for cross cloud service. We're getting also a keynote with Kit Colbert, who's going to lay out the fragmentation and the benefits that that solves, from solvent fragmentation and silos, breaking down the silos and bringing multicloud future to the table via Super Cloud. So stay with us. We'll be right back after this short break. (bright upbeat music) (music fades)

Published Date : Feb 17 2023

SUMMARY :

and illustrate some of the examples We're going to hear from Walmart today. And that's going to be a determinant. At the same time, you And so I think this is going to the meat on the bone, if you will. Dave, is going to be big to watch. giving back to the community. and the benefits that that solves,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

DavePERSON

0.99+

Pat GelsingerPERSON

0.99+

AlibabaORGANIZATION

0.99+

Kit ColbertPERSON

0.99+

Zhamak DehghaniPERSON

0.99+

WalmartORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Andy JassyPERSON

0.99+

GoogleORGANIZATION

0.99+

Silicon ValleyLOCATION

0.99+

AugustDATE

0.99+

Victoria VieringPERSON

0.99+

August 9thDATE

0.99+

John FurrierPERSON

0.99+

200 locationsQUANTITY

0.99+

VMwareORGANIZATION

0.99+

SupercloudORGANIZATION

0.99+

Palo Alto, CaliforniaLOCATION

0.99+

Supercloud2EVENT

0.99+

two forcesQUANTITY

0.99+

last AugustDATE

0.99+

yesterdayDATE

0.99+

firstQUANTITY

0.99+

two variablesQUANTITY

0.99+

todayDATE

0.98+

OneQUANTITY

0.98+

supercloudORGANIZATION

0.98+

AzureORGANIZATION

0.97+

ChaosSearchORGANIZATION

0.95+

super cloud waveEVENT

0.94+

Supercloud1EVENT

0.94+

Super CloudTITLE

0.93+

AlkiraPERSON

0.83+

Palo Alto, JohnLOCATION

0.83+

this next yearDATE

0.81+

Data MeshORGANIZATION

0.8+

supercloud waveEVENT

0.79+

wave ofEVENT

0.79+

Western UnionLOCATION

0.78+

SaksORGANIZATION

0.76+

GPTORGANIZATION

0.73+

Supercloud2ORGANIZATION

0.72+

Cloud NativeTITLE

0.69+

SupercloudTITLE

0.67+

Supercloud2COMMERCIAL_ITEM

0.66+

multicloudORGANIZATION

0.57+

SupercloudCOMMERCIAL_ITEM

0.53+

Supercloud2TITLE

0.53+

theCUBEORGANIZATION

0.51+

super cloudTITLE

0.51+

CloudTITLE

0.41+

Supercloud Applications & Developer Impact | Supercloud2


 

(gentle music) >> Okay, welcome back to Supercloud 2, live here in Palo Alto, California for our live stage performance. Supercloud 2 is our second Supercloud event. We're going to get these out as fast as we can every couple months. It's our second one, you'll see two and three this year. I'm John Furrier, my co-host, Dave Vellante. A panel here to break down the Supercloud momentum, the wave, and the developer impact that we bringing back Vittorio Viarengo, who's a VP for Cross-Cloud Services at VMware. Sarbjeet Johal, industry influencer and Analyst at StackPayne, his company, Cube alumni and Influencer. Sarbjeet, great to see you. Vittorio, thanks for coming back. >> Nice to be here. >> My pleasure. >> Vittorio, you just gave a keynote where we unpacked the cross-cloud services, what VMware is doing, how you guys see it, not just from VMware's perspective, but VMware looking out broadly at the industry and developers came up and you were like, "Developers, developer, developers", kind of a goof on the Steve Ballmer famous meme that everyone's seen. This is a huge star, sorry, I mean a big piece of it. The developers are the canary in the coal mines. They're the ones who are being asked to code the digital transformation, which is fully business transformation and with the market the way it is right now in terms of the accelerated technology, every enterprise grade business model's changing. The technology is evolving, the builders are kind of, they want go faster. I'm saying they're stuck in a way, but that's my opinion, but there's a lot of growth. >> Yeah. >> The impact, they got to get released up and let it go. Those developers need to accelerate faster. It's been a big part of productivity, and the conversations we've had. So developer impact is huge in Supercloud. What's your, what do you guys think about this? We'll start with you, Sarbjeet. >> Yeah, actually, developers are the masons of the digital empires I call 'em, right? They lay every brick and build all these big empires. On the left side of the SDLC, or the, you know, when you look at the system operations, developer is number one cost from economic side of things, and from technology side of things, they are tech hungry people. They are developers for that reason because developer nights are long, hours are long, they forget about when to eat, you know, like, I've been a developer, I still code. So you want to keep them happy, you want to hug your developers. We always say that, right? Vittorio said that right earlier. The key is to, in this context, in the Supercloud context, is that developers don't mind mucking around with platforms or APIs or new languages, but they hate the infrastructure part. That's a fact. They don't want to muck around with servers. It's friction for them, it is like they don't want to muck around even with the VMs. So they want the programmability to the nth degree. They want to automate everything, so that's how they think and cloud is the programmable infrastructure, industrialization of infrastructure in many ways. So they are happy with where we are going, and we need more abstraction layers for some developers. By the way, I have this sort of thinking frame for last year or so, not all developers are same, right? So if you are a developer at an ISV, you behave differently. If you are a developer at a typical enterprise, you behave differently or you are forced to behave differently because you're not writing software.- >> Well, developers, developers have changed, I mean, Vittorio, you and I were talking earlier on the keynote, and this is kind of the key point is what is a developer these days? If everything is software enabled, I mean, even hardware interviews we do with Nvidia, and Amazon and other people building silicon, they all say the same thing, "It's software on a chip." So you're seeing the role of software up and down the stack and the role of the stack is changing. The old days of full stack developer, what does that even mean? I mean, the cloud is a half a stack kind of right there. So, you know, developers are certainly more agile, but cloud native, I mean VMware is epitome of operations, IT operations, and the Tan Zoo initiative, you guys started, you went after the developers to look at them, and ask them questions, "What do you need?", "How do you transform the Ops from virtualization?" Again, back to your point, so this hardware abstraction, what is software, what is cloud native? It's kind of messy equation these days. How do you guys grokel with that? >> I would argue that developers don't want the Supercloud. I dropped that up there, so, >> Dave: Why not? >> Because developers, they, once they get comfortable in AWS or Google, because they're doing some AI stuff, which is, you know, very trendy right now, or they are in IBM, any of the IPA scaler, professional developers, system developers, they love that stuff, right? Yeah, they don't, the infrastructure gets in the way, but they're just, the problem is, and I think the Supercloud should be driven by the operators because as we discussed, the operators have been left behind because they're busy with day-to-day jobs, and in most cases IT is centralized, developers are in the business units. >> John: Yeah. >> Right? So they get the mandate from the top, say, "Our bank, they're competing against". They gave teenagers or like young people the ability to do all these new things online, and Venmo and all this integration, where are we? "Oh yeah, we can do it", and then build it, and then deploy it, "Okay, we caught up." but now the operators are back in the private cloud trying to keep the backend system running and so I think the Supercloud is needed for the primarily, initially, for the operators to get in front of the developers, fit in the workflow, but lay the foundation so it is secure.- >> So, so I love this thinking because I love the rift, because the rift points to what is the target audience for the value proposition and if you're a developer, Supercloud enables you so you shouldn't have to deal with Supercloud. >> Exactly. >> What you're saying is get the operating environment or operating system done properly, whether it's architecture, building the platform, this comes back to architecture platform conversations. What is the future platform? Is it a vendor supplied or is it customer created platform? >> Dave: So developers want best to breed, is what you just said. >> Vittorio: Yeah. >> Right and operators, they, 'cause developers don't want to deal with governance, they don't want to deal with security, >> No. >> They don't want to deal with spinning up infrastructure. That's the role of the operator, but that's where Supercloud enables, to John's point, the developer, so to your question, is it a platform where the platform vendor is responsible for the architecture, or there is it an architectural standard that spans multiple clouds that has to emerge? Based on what you just presented earlier, Vittorio, you are the determinant of the architecture. It's got to be open, but you guys determine that, whereas the nirvana is, "Oh no, it's all open, and it just kind of works." >> Yeah, so first of all, let's all level set on one thing. You cannot tell developers what to do. >> Dave: Right, great >> At least great developers, right? Cannot tell them what to do. >> Dave: So that's what, that's the way I want to sort of, >> You can tell 'em what's possible. >> There's a bottle on that >> If you tell 'em what's possible, they'll test it, they'll look at it, but if you try to jam it down their throat, >> Yeah. >> Dave: You can't tell 'em how to do it, just like your point >> Let me answer your answer the question. >> Yeah, yeah. >> So I think we need to build an architect, help them build an architecture, but it cannot be proprietary, has to be built on what works in the cloud and so what works in the cloud today is Kubernetes, is you know, number of different open source project that you need to enable and then provide, use this, but when I first got exposed to Kubernetes, I said, "Hallelujah!" We had a runtime that works the same everywhere only to realize there are 12 different distributions. So that's where we come in, right? And other vendors come in to say, "Hey, no, we can make them all look the same. So you still use Kubernetes, but we give you a place to build, to set those operation policy once so that you don't create friction for the developers because that's the last thing you want to do." >> Yeah, actually, coming back to the same point, not all developers are same, right? So if you're ISV developer, you want to go to the lowest sort of level of the infrastructure and you want to shave off the milliseconds from to get that performance, right? If you're working at AWS, you are doing that. If you're working at scale at Facebook, you're doing that. At Twitter, you're doing that, but when you go to DMV and Kansas City, you're not doing that, right? So your developers are different in nature. They are given certain parameters to work with, certain sort of constraints on the budget side. They are educated at a different level as well. Like they don't go to that end of the degree of sort of automation, if you will. So you cannot have the broad stroking of developers. We are talking about a citizen developer these days. That's a extreme low, >> You mean Low-Code. >> Yeah, Low-Code, No-code, yeah, on the extreme side. On one side, that's citizen developers. On the left side is the professional developers, when you say developers, your mind goes to the professional developers, like the hardcore developers, they love the flexibility, you know, >> John: Well app, developers too, I mean. >> App developers, yeah. >> You're right a lot of, >> Sarbjeet: Infrastructure platform developers, app developers, yes. >> But there are a lot of customers, its a spectrum, you're saying. >> Yes, it's a spectrum >> There's a lot of customers don't want deal with that muck. >> Yeah. >> You know, like you said, AWS, Twitter, the sophisticated developers do, but there's a whole suite of developers out there >> Yeah >> That just want tools that are abstracted. >> Within a company, within a company. Like how I see the Supercloud is there shouldn't be anything which blocks the developers, like their view of the world, of the future. Like if you're blocked as a developer, like something comes in front of you, you are not developer anymore, believe me, (John laughing) so you'll go somewhere else >> John: First of all, I'm, >> You'll leave the company by the way. >> Dave: Yeah, you got to quit >> Yeah, you will quit, you will go where the action is, where there's no sort of blockage there. So like if you put in front of them like a huge amount of a distraction, they don't like it, so they don't, >> Well, the idea of a developer, >> Coming back to that >> Let's get into 'cause you mentioned platform. Get year in the term platform engineering now. >> Yeah. >> Platform developer. You know, I remember back in, and I think there's still a term used today, but when I graduated my computer science degree, we were called "Software engineers," right? Do people use that term "Software engineering", or is it "Software development", or they the same, are they different? >> Well, >> I think there's a, >> So, who's engineering what? Are they engineering or are they developing? Or both? Well, I think it the, you made a great point. There is a factor of, I had the, I was blessed to work with Adam Bosworth, that is the guy that created some of the abstraction layer, like Visual Basic and Microsoft Access and he had so, he made his whole career thinking about this layer, and he always talk about the professional developers, the developers that, you know, give him a user manual, maybe just go at the APIs, he'll build anything, right, from system engine, go down there, and then through obstruction, you get the more the procedural logic type of engineers, the people that used to be able to write procedural logic and visual basic and so on and so forth. I think those developers right now are a little cut out of the picture. There's some No-code, Low-Code environment that are maybe gain some traction, I caught up with Adam Bosworth two weeks ago in New York and I asked him "What's happening to this higher level developers?" and you know what he is told me, and he is always a little bit out there, so I'm going to use his thought process here. He says, "ChapGPT", I mean, they will get to a point where this high level procedural logic will be written by, >> John: Computers. >> Computers, and so we may not need as many at the high level, but we still need the engineers down there. The point is the operation needs to get in front of them >> But, wait, wait, you seen the ChatGPT meme, I dunno if it's a Dilbert thing where it's like, "Time to tic" >> Yeah, yeah, yeah, I did that >> "Time to develop the code >> Five minutes, time to decode", you know, to debug the codes like five hours. So you know, the whole equation >> Well, this ChatGPT is a hot wave, everyone's been talking about it because I think it illustrates something that's NextGen, feels NextGen, and it's just getting started so it's going to get better. I mean people are throwing stones at it, but I think it's amazing. It's the equivalent of me seeing the browser for the first time, you know, like, "Wow, this is really compelling." This is game-changing, it's not just keyword chat bots. It's like this is real, this is next level, and I think the Supercloud wave that people are getting behind points to that and I think the question of Ops and Dev comes up because I think if you limit the infrastructure opportunity for a developer, I think they're going to be handicapped. I mean that's a general, my opinion, the thesis is you give more aperture to developers, more choice, more capabilities, more good things could happen, policy, and that's why you're seeing the convergence of networking people, virtualization talent, operational talent, get into the conversation because I think it's an infrastructure engineering opportunity. I think this is a seminal moment in a new stack that's emerging from an infrastructure, software virtualization, low-code, no-code layer that will be completely programmable by things like the next Chat GPT or something different, but yet still the mechanics and the plumbing will still need engineering. >> Sarbjeet: Oh yeah. >> So there's still going to be more stuff coming on. >> Yeah, we have, with the cloud, we have made the infrastructure programmable and you give the programmability to the programmer, they will be very creative with that and so we are being very creative with our infrastructure now and on top of that, we are being very creative with the silicone now, right? So we talk about that. That's part of it, by the way. So you write the code to the particle's silicone now, and on the flip side, the silicone is built for certain use cases for AI Inference and all that. >> You saw this at CES? >> Yeah, I saw at CES, the scenario is this, the Bosch, I spoke to Bosch, I spoke to John Deere, I spoke to AWS guys, >> Yeah. >> They were showcasing their technology there and I was spoke to Azure guys as well. So the Bosch is a good example. So they are building, they are right now using AWS. I have that interview on camera, I will put it some sometime later on there online. So they're using AWS on the back end now, but Bosch is the number one, number one or number two depending on what day it is of the year, supplier of the componentry to the auto industry, and they are creating a platform for our auto industry, so is Qualcomm actually by the way, with the Snapdragon. So they told me that customers, their customers, BMW, Audi, all the manufacturers, they demand the diversity of the backend. Like they don't want all, they, all of them don't want to go to AWS. So they want the choice on the backend. So whatever they cook in the middle has to work, they have to sprinkle the data for the data sovereign side because they have Chinese car makers as well, and for, you know, for other reasons, competitive reasons and like use. >> People don't go to, aw, people don't go to AWS either for political reasons or like competitive reasons or specific use cases, but for the most part, generally, I haven't met anyone who hasn't gone first choice with either, but that's me personally. >> No, but they're building. >> Point is the developer wants choice at the back end is what I'm hearing, but then finish that thought. >> Their developers want the choice, they want the choice on the back end, number one, because the customers are asking for, in this case, the customers are asking for it, right? But the customers requirements actually drive, their economics drives that decision making, right? So in the middle they have to, they're forced to cook up some solution which is vendor neutral on the backend or multicloud in nature. So >> Yeah, >> Every >> I mean I think that's nirvana. I don't think, I personally don't see that happening right now. I mean, I don't see the parody with clouds. So I think that's a challenge. I mean, >> Yeah, true. >> I mean the fact of the matter is if the development teams get fragmented, we had this chat with Kit Colbert last time, I think he's going to come on and I think he's going to talk about his keynote in a few, in an hour or so, development teams is this, the cloud is heterogenous, which is great. It's complex, which is challenging. You need skilled engineering to manage these clouds. So if you're a CIO and you go all in on AWS, it's hard. Then to then go out and say, "I want to be completely multi-vendor neutral" that's a tall order on many levels and this is the multicloud challenge, right? So, the question is, what's the strategy for me, the CIO or CISO, what do I do? I mean, to me, I would go all in on one and start getting hedges and start playing and then look at some >> Crystal clear. Crystal clear to me. >> Go ahead. >> If you're a CIO today, you have to build a platform engineering team, no question. 'Cause if we agree that we cannot tell the great developers what to do, we have to create a platform engineering team that using pieces of the Supercloud can build, and let's make this very pragmatic and give examples. First you need to be able to lay down the run time, okay? So you need a way to deploy multiple different Kubernetes environment in depending on the cloud. Okay, now we got that. The second part >> That's like table stakes. >> That are table stake, right? But now what is the advantage of having a Supercloud service to do that is that now you can put a policy in one place and it gets distributed everywhere consistently. So for example, you want to say, "If anybody in this organization across all these different buildings, all these developers don't even know, build a PCI compliant microservice, They can only talk to PCI compliant microservice." Now, I sleep tight. The developers still do that. Of course they're going to get their hands slapped if they don't encrypt some messages and say, "Oh, that should have been encrypted." So number one. The second thing I want to be able to say, "This service that this developer built over there better satisfy this SLA." So if the SLA is not satisfied, boom, I automatically spin up multiple instances to certify the SLA. Developers unencumbered, they don't even know. So this for me is like, CIO build a platform engineering team using one of the many Supercloud services that allow you to do that and lay down. >> And part of that is that the vendor behavior is such, 'cause the incentive is that they don't necessarily always work together. (John chuckling) I'll give you an example, we're going to hear today from Western Union. They're AWS shop, but they want to go to Google, they want to use some of Google's AI tools 'cause they're good and maybe they're even arguably better, but they're also a Snowflake customer and what you'll hear from them is Amazon and Snowflake are working together so that SageMaker can be integrated with Snowflake but Google said, "No, you want to use our AI tools, you got to use BigQuery." >> Yeah. >> Okay. So they say, "Ah, forget it." So if you have a platform engineering team, you can maybe solve some of that vendor friction and get competitive advantage. >> I think that the future proximity concept that I talk about is like, when you're doing one thing, you want to do another thing. Where do you go to get that thing, right? So that is very important. Like your question, John, is that your point is that AWS is ahead of the pack, which is true, right? They have the >> breadth of >> Infrastructure by a lot >> infrastructure service, right? They breadth of services, right? So, how do you, When do you bring in other cloud providers, right? So I believe that you should standardize on one cloud provider, like that's your primary, and for others, bring them in on as needed basis, in the subsection or sub portfolio of your applications or your platforms, what ever you can. >> So yeah, the Google AI example >> Yeah, I mean, >> Or the Microsoft collaboration software example. I mean there's always or the M and A. >> Yeah, but- >> You're going to get to run Windows, you can run Windows on Amazon, so. >> By the way, Supercloud doesn't mean that you cannot do that. So the perfect example is say that you're using Azure because you have a SQL server intensive workload. >> Yep >> And you're using Google for ML, great. If you are using some differentiated feature of this cloud, you'll have to go somewhere and configure this widget, but what you can abstract with the Supercloud is the lifecycle manage of the service that runs on top, right? So how does the service get deployed, right? How do you monitor performance? How do you lifecycle it? How you secure it that you can abstract and that's the value and eventually value will win. So the customers will find what is the values, obstructing in making it uniform or going deeper? >> How about identity? Like take identity for instance, you know, that's an opportunity to abstract. Whether I use Microsoft Identity or Okta, and I can abstract that. >> Yeah, and then we have APIs and standards that we can use so eventually I think where there is enough pain, the right open source will emerge to solve that problem. >> Dave: Yeah, I can use abstract things like object store, right? That's pretty simple. >> But back to the engineering question though, is that developers, developers, developers, one thing about developers psychology is if something's not right, they say, "Go get fixing. I'm not touching it until you fix it." They're very sticky about, if something's not working, they're not going to do it again, right? So you got to get it right for developers. I mean, they'll maybe tolerate something new, but is the "juice worth the squeeze" as they say, right? So you can't go to direct say, "Hey, it's, what's a work in progress? We're going to get our infrastructure together and the world's going to be great for you, but just hang tight." They're going to be like, "Get your shit together then talk to me." So I think that to me is the question. It's an Ops question, but where's that value for the developer in Supercloud where the capabilities are there, there's less friction, it's simpler, it solves the complexity problem. I don't need these high skilled labor to manage Amazon. I got services exposed. >> That's what we talked about earlier. It's like the Walmart example. They basically, they took away from the developer the need to spin up infrastructure and worry about all the governance. I mean, it's not completely there yet. So the developer could focus on what he or she wanted to do. >> But there's a big, like in our industry, there's a big sort of flaw or the contention between developers and operators. Developers want to be on the cutting edge, right? And operators want to be on the stability, you know, like we want governance. >> Yeah, totally. >> Right, so they want to control, developers are like these little bratty kids, right? And they want Legos, like they want toys, right? Some of them want toys by way. They want Legos, they want to build there and they want make a mess out of it. So you got to make sure. My number one advice in this context is that do it up your application portfolio and, or your platform portfolio if you are an ISV, right? So if you are ISV you most probably, you're building a platform these days, do it up in a way that you can say this portion of our applications and our platform will adhere to what you are saying, standardization, you know, like Kubernetes, like slam dunk, you know, it works across clouds and in your data center hybrid, you know, whole nine yards, but there is some subset on the next door systems of innovation. Everybody has, it doesn't matter if you're DMV of Kansas or you are, you know, metaverse, right? Or Meta company, right, which is Facebook, they have it, they are building something new. For that, give them some freedom to choose different things like play with non-standard things. So that is the mantra for moving forward, for any enterprise. >> Do you think developers are happy with the infrastructure now or are they wanting people to get their act together? I mean, what's your reaction, or you think. >> Developers are happy as long as they can do their stuff, which is running code. They want to write code and innovate. So to me, when Ballmer said, "Developer, develop, Developer, what he meant was, all you other people get your act together so these developers can do their thing, and to me the Supercloud is the way for IT to get there and let developer be creative and go fast. Why not, without getting in trouble. >> Okay, let's wrap up this segment with a super clip. Okay, we're going to do a sound bite that we're going to make into a short video for each of you >> All right >> On you guys summarizing why Supercloud's important, why this next wave is relevant for the practitioners, for the industry and we'll turn this into an Instagram reel, YouTube short. So we'll call it a "Super clip. >> Alright, >> Sarbjeet, you want, you want some time to think about it? You want to go first? Vittorio, you want. >> I just didn't mind. (all laughing) >> No, okay, okay. >> I'll do it again. >> Go back. No, we got a fresh one. We'll going to already got that one in the can. >> I'll go. >> Sarbjeet, you go first. >> I'll go >> What's your super clip? >> In software systems, abstraction is your friend. I always say that. Abstraction is your friend, even if you're super professional developer, abstraction is your friend. We saw from the MFC library from C++ days till today. Abstract, use abstraction. Do not try to reinvent what's already being invented. Leverage cloud, leverage the platform side of the cloud. Not just infrastructure service, but platform as a service side of the cloud as well, and Supercloud is a meta platform built on top of these infrastructure services from three or four or five cloud providers. So use that and embrace the programmability, embrace the abstraction layer. That's the key actually, and developers who are true developers or professional developers as you said, they know that. >> Awesome. Great super clip. Vittorio, another shot at the plate here for super clip. Go. >> Multicloud is awesome. There's a reason why multicloud happened, is because gave our developers the ability to innovate fast and ever before. So if you are embarking on a digital transformation journey, which I call a survival journey, if you're not innovating and transforming, you're not going to be around in business three, five years from now. You have to adopt the Supercloud so the developer can be developer and keep building great, innovating digital experiences for your customers and IT can get in front of it and not get in trouble together. >> Building those super apps with Supercloud. That was a great super clip. Vittorio, thank you for sharing. >> Thanks guys. >> Sarbjeet, thanks for coming on talking about the developer impact Supercloud 2. On our next segment, coming up right now, we're going to hear from Walmart enterprise architect, how they are building and they are continuing to innovate, to build their own Supercloud. Really informative, instructive from a practitioner doing it in real time. Be right back with Walmart here in Palo Alto. Thanks for watching. (gentle music)

Published Date : Feb 17 2023

SUMMARY :

the Supercloud momentum, and developers came up and you were like, and the conversations we've had. and cloud is the and the role of the stack is changing. I dropped that up there, so, developers are in the business units. the ability to do all because the rift points to What is the future platform? is what you just said. the developer, so to your question, You cannot tell developers what to do. Cannot tell them what to do. You can tell 'em your answer the question. but we give you a place to build, and you want to shave off the milliseconds they love the flexibility, you know, platform developers, you're saying. don't want deal with that muck. that are abstracted. Like how I see the Supercloud is So like if you put in front of them you mentioned platform. and I think there's the developers that, you The point is the operation to decode", you know, the browser for the first time, you know, going to be more stuff coming on. and on the flip side, the middle has to work, but for the most part, generally, Point is the developer So in the middle they have to, the parody with clouds. I mean the fact of the matter Crystal clear to me. in depending on the cloud. So if the SLA is not satisfied, boom, 'cause the incentive is that So if you have a platform AWS is ahead of the pack, So I believe that you should standardize or the M and A. you can run Windows on Amazon, so. So the perfect example is abstract and that's the value Like take identity for instance, you know, the right open source will Dave: Yeah, I can use abstract things and the world's going to be great for you, the need to spin up infrastructure on the stability, you know, So that is the mantra for moving forward, Do you think developers are happy and to me the Supercloud is for each of you for the industry you want some time to think about it? I just didn't mind. got that one in the can. platform side of the cloud. Vittorio, another shot at the the ability to innovate thank you for sharing. the developer impact Supercloud 2.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

DavePERSON

0.99+

BMWORGANIZATION

0.99+

WalmartORGANIZATION

0.99+

JohnPERSON

0.99+

SarbjeetPERSON

0.99+

John FurrierPERSON

0.99+

BoschORGANIZATION

0.99+

VittorioPERSON

0.99+

NvidiaORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

AudiORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Steve BallmerPERSON

0.99+

QualcommORGANIZATION

0.99+

Adam BosworthPERSON

0.99+

Palo AltoLOCATION

0.99+

FacebookORGANIZATION

0.99+

New YorkLOCATION

0.99+

Vittorio ViarengoPERSON

0.99+

Kit ColbertPERSON

0.99+

BallmerPERSON

0.99+

fourQUANTITY

0.99+

Sarbjeet JohalPERSON

0.99+

five hoursQUANTITY

0.99+

VMwareORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

Palo Alto, CaliforniaLOCATION

0.99+

MicrosoftORGANIZATION

0.99+

Five minutesQUANTITY

0.99+

NextGenORGANIZATION

0.99+

StackPayneORGANIZATION

0.99+

Visual BasicTITLE

0.99+

second partQUANTITY

0.99+

12 different distributionsQUANTITY

0.99+

CESEVENT

0.99+

FirstQUANTITY

0.99+

TwitterORGANIZATION

0.99+

Kansas CityLOCATION

0.99+

second oneQUANTITY

0.99+

threeQUANTITY

0.99+

bothQUANTITY

0.99+

KansasLOCATION

0.98+

first timeQUANTITY

0.98+

WindowsTITLE

0.98+

last yearDATE

0.98+

Closing Remarks | Supercloud2


 

>> Welcome back everyone to the closing remarks here before we kick off our ecosystem portion of the program. We're live in Palo Alto for theCUBE special presentation of Supercloud 2. It's the second edition, the first one was in August. I'm John Furrier with Dave Vellante. Here to wrap up with our special guest analyst George Gilbert, investor and industry legend former colleague of ours, analyst at Wikibon. George great to see you. Dave, you know, wrapping up this day what in a phenomenal program. We had a contribution from industry vendors, industry experts, practitioners and customers building and redefining their company's business model. Rolling out technology for Supercloud and multicloud and ultimately changing how they do data. And data was the theme today. So very, very great program. Before we jump into our favorite parts let's give a shout out to the folks who make this possible. Free contents our mission. We'll always stay true to that mission. We want to thank VMware, alkira, ChaosSearch, prosimo for being sponsors of this great program. We will have Supercloud 3 coming up in a month or so, or two months. We'll see. Or sooner, we don't know. But it'll be more about security, but a lot more momentum. Okay, so that's... >> And don't forget too that this program not going to end now. We've got a whole ecosystem speaks track so stay tuned for that. >> John: Yeah, we got another 20 interviews. Feels like it. >> Well, you're going to hear from Saks, Veronika Durgin. You're going to hear from Western Union, Harveer Singh. You're going to hear from Ionis Pharmaceuticals, Nick Taylor. Brian Gracely chimes in on Supecloud. So he's the man behind the cloud cast. >> Yeah, and you know, the practitioners again, pay attention to also to the cloud networking interviews. Lot of change going on there that's going to be disruptive and actually change the landscape as well. Again, as Supercloud progresses to be the next big thing. If you're not on this next wave, you'll drift what, as Pat Gelsinger says. >> Yep. >> To kick off the closing segments, George, Dave, this is a wave that's been identified. Again, people debate the word all you want Supercloud. It is a gateway to multicloud eventually it is the standard for new applications, new ways to do data. There's new computer science being generated and customer requirements being addressed. So it's the confluence of, you know, tectonic plates shifting in the industry, new computer science seeing things like AI and machine learning and data at the center of it and new infrastructure all kind of coming together. So, to me, that's my takeaway so far. That is the big story and it's going to change society and ultimately the business models of these companies. >> Well, we've had 10, you know, you think about it we came out of the financial crisis. We've had 10, 12 years despite the Covid of tech success, right? And just now CIOs are starting to hit the brakes. And so my point is you've had all this innovation building up for a decade and you've got this massive ecosystem that is running on the cloud and the ecosystem is saying, hey, we can have even more value by tapping best of of breed across clouds. And you've got customers saying, hey, we need help. We want to do more and we want to point our business and our intellectual property, our software tooling at our customers and monetize our data. So you have all these forces coming together and it's sort of entering a new era. >> George, I want to go to you for a second because you are big contributor to this event. Your interview with Bob Moglia with Dave was I thought a watershed moment for me to hear that the data apps, how databases are being rethought because we've been seeing a diversity of databases with Amazon Web services, you know, promoting no one database rules of the world. Now it's not one database kind of architecture that's puling these new apps. What's your takeaway from this event? >> So if you keep your eye on this North Star where instead of building apps that are based on code you're building apps that are defined by data coming off of things that are linked to the real world like people, places, things and activities. Then the idea is, and the example we use is, you know, Uber but it could be, you know, amazon.com is defined by stuff coming off data in the Amazon ecosystem or marketplace. And then the question is, and everyone was talking at different angles on this, which was, where's the data live? How much do you hide from the developer? You know, and when can you offer that? You know, and you started with Walmart which was describing apps, traditional apps that are just code. And frankly that's easier to make that cross cloud and you know, essentially location independent. As soon as you have data you need data management technology that a customer does not have the sophistication to build. And then the argument was like, so how much can you hide from the developer who's building data apps? Tristan's version was you take the modern data stack and you start adding these APIs that define business concepts like bookings, billings and revenue, you know, or in the Uber example like drivers and riders, you know, and ETA's and prices. But those things execute still on the data warehouse or data lakehouse. Then Bob Muglia was saying you're not really hiding enough from the developer because you still got to say how to do all that. And his vision is not only do you hide where the data is but you hide how to sort of get at all that code by just saying what you want. You define how a car and how a driver and how a rider works. And then those things automatically figure out underneath the cover. >> So huge challenges, right? There's governance, there's security, they could be big blockers to, you know, the Supercloud but the industry's going to be attacking that problem. >> Well, what's your take? What's your favorite segment? Zhamak Dehghani came on, she's starting in that company, exclusive news. That was big notable moment for theCUBE. She launched her company. She pioneered the data mesh concept. And I think what George is saying and what data mesh points to is something that we've been saying for a long time. That data is now going to flip the script on how apps behave. And the Uber example I think is illustrated 'cause people can relate to Uber. But imagine that for every business whether it's a manufacturing business or retail or oil and gas or FinTech, they can look at their business like a game almost gamify it with data, riders, cars you know, moving data around the value of data. This is something that Adam Selipsky teased out at AWS, Dave. So what's your takeaway from this Supercloud? Where are we in your mind? Well big thing is data products and decentralizing your data architecture, but putting data in the hands of domain experts who can actually monetize the data. And I think that's, to me that's really exciting. Because look, data products financial industry has always been doing building data products. Mortgage backed securities is a data product. But why should the financial industry have all the fun? I mean virtually every organization can tap its ecosystem build data products, take its internal IP and processes and software and point it to the world and actually begin to make money out of it. >> Okay, so let's go around the horn. I'll start, I'll get you guys some time to think. Next question, what did you learn today? I learned that I think it's an infrastructure game and talking to Kit Colbert at VMware, I think it's all about infrastructure refactoring and I think the data's going to be an ingredient that's going to be operating system like. I think you're going to see the infrastructure influencing operations that will enable Superclouds to be real. And developers won't even know what a Supercloud is because they'll be using it. It's the operations focus is going to be very critical. Just like DevOps movements started Cloud native I think you're going to see a data native movement and I think infrastructure is critical as people go to the next level. That's my big takeaway today. And I'll say the data conversation is at the center. I think security, data are going to be always active horizontally scalable concepts, but every company's going to reset their infrastructure, how it looks and if it's not set up for data and or things that there need to be agile on, it's going to be a non-starter. So I think that's the cloud NextGen, distributed computing. >> I mean, what came into focus for me was I think the hyperscaler is going to continue to do their thing, you know, and be very, very successful and they're each coming at it from different approaches. We talk about this all the time in theCUBE. Amazon the best infrastructure, you know, Google's got its you know, data and AI thing and it's playing catch up and Microsoft's got this massive estate. Okay, cool. Check. The next wave of innovation which is coming from data, I've always said follow the data. That's where the where the money's going to be is going to come from other places. People want to be able to, organizations want to be able to share data across clouds across their organization, outside of their ecosystem and make money with that data sharing. They don't want to FTP it anymore. I got it. You take it. They want to work with live data in real time and I think the edge, we didn't talk much about the edge today is going to even take that to a new level real time inferencing at the edge, AI and and being able to do new things with data that we haven't even seen. But playing around with ChatGPT, it's blowing our mind. And I think you're right, it's like when we first saw the browser, holy crap, this is going to change the world. >> Yeah. And the ChatGPT by the way is going to create a wave of machine learning and data refactoring for sure. But also Howie Liu had an interesting comment, he was asked by a VC how much to replicate that and he said it's in the hundreds of millions, not billions. Now if you asked that same question how much does it cost to replicate AWS? The CapEx alone is unstoppable, they're already done. So, you know, the hyperscalers are going to continue to boom. I think they're going to drive the infrastructure. I think Amazon's going to be really strong at silicon and physics and squeeze every ounce atom out of every physical thing and then get latency as your bottleneck and the rest is all going to be... >> That never blew me away, a hundred million to create kind of an open AI, you know, competitor. Look at companies like Lacework. >> John: Some people have that much cash on the balance sheet. >> These are security companies that have raised a billion dollars, right? To compete. You know, so... >> If you're not shifting left what do you do with data, shift up? >> But, you know. >> What did you learn, George? >> I'm listening to you and I think you're helping me crystallize something which is the software infrastructure to enable the data apps is wide open. The way Zhamak described it is like if you want a data product like a sales and operation plan, that is built on other data products, like a sales plan which has a forecast in it, it has a production plan, it has a procurement plan and then a sales and operation plan is actually a composition of all those and they call each other. Now in her current platform, you need to expose to the developer a certain amount of mechanics on how to move all that data, when to move it. Like what happens if something fails. Now Muglia is saying I can hide that completely. So all you have to say is what you want and the underlying machinery takes care of everything. The problem is Muglia stuff is still a few years off. And Tristan is saying, I can give you much of that today but it's got to run in the data warehouse. So this trade offs all different ways. But again, I agree with you that the Cloud platform vendors or the ecosystem participants who can run across Cloud platforms and private infrastructure will be the next platform. And then the cloud platform is sort of where you run the big honking centralized stuff where someone else manages the operations. >> Sounds like middleware to me, Dave >> And key is, I'll just end with this. The key is being able to get to the data, whether it's in a data warehouse or a data lake or a S3 bucket or an object store, Oracle database, whatever. It's got to be inclusive that is critical to execute on the vision that you just talked about 'cause that data's in different systems and you're not going to put it all into some new system. >> So creating middleware in the cloud that sounds what it sounds like to me. >> It's like, you discovered PaaS >> It's a super PaaS. >> But it's platform services 'cause PaaS connotes like a tightly integrated platform. >> Well this is the real thing that's going on. We're going to see how this evolves. George, great to have you on, Dave. Thanks for the summary. I enjoyed this segment a lot today. This ends our stage performance live here in Palo Alto. As you know, we're live stage performance and syndicate out virtually. Our afternoon program's going to kick in now you're going to hear some great interviews. We got ChaosSearch. Defining the network Supercloud from prosimo. Future of Cloud Network, alkira. We got Saks, a retail company here, Veronika Durgin. We got Dave with Western Union. So a lot of customers, a pharmaceutical company Warner Brothers, Discovery, media company. And then you know, what is really needed for Supercloud, good panels. So stay with us for the afternoon program. That's part two of Supercloud 2. This is a wrap up for our stage live performance. I'm John Furrier with Dave Vellante and George Gilbert here wrapping up. Thanks for watching and enjoy the program. (bright music)

Published Date : Jan 17 2023

SUMMARY :

to the closing remarks here program not going to end now. John: Yeah, we got You're going to hear from Yeah, and you know, It is a gateway to multicloud starting to hit the brakes. go to you for a second the sophistication to build. but the industry's going to And I think that's, to me and talking to Kit Colbert at VMware, to do their thing, you know, I think Amazon's going to be really strong kind of an open AI, you know, competitor. on the balance sheet. that have raised a billion dollars, right? I'm listening to you and I think It's got to be inclusive that is critical So creating middleware in the cloud But it's platform services George, great to have you on, Dave.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
TristanPERSON

0.99+

Dave VellantePERSON

0.99+

George GilbertPERSON

0.99+

Dave VellantePERSON

0.99+

DavePERSON

0.99+

Adam SelipskyPERSON

0.99+

Pat GelsingerPERSON

0.99+

Bob MogliaPERSON

0.99+

Veronika DurginPERSON

0.99+

JohnPERSON

0.99+

Bob MugliaPERSON

0.99+

GeorgePERSON

0.99+

AmazonORGANIZATION

0.99+

Western UnionORGANIZATION

0.99+

Nick TaylorPERSON

0.99+

Palo AltoLOCATION

0.99+

10QUANTITY

0.99+

John FurrierPERSON

0.99+

UberORGANIZATION

0.99+

Brian GracelyPERSON

0.99+

Howie LiuPERSON

0.99+

Zhamak DehghaniPERSON

0.99+

hundreds of millionsQUANTITY

0.99+

AWSORGANIZATION

0.99+

Ionis PharmaceuticalsORGANIZATION

0.99+

AugustDATE

0.99+

Warner BrothersORGANIZATION

0.99+

Kit ColbertPERSON

0.99+

MicrosoftORGANIZATION

0.99+

WalmartORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

billionsQUANTITY

0.99+

ZhamakPERSON

0.99+

MugliaPERSON

0.99+

20 interviewsQUANTITY

0.99+

DiscoveryORGANIZATION

0.99+

second editionQUANTITY

0.99+

ChaosSearchORGANIZATION

0.99+

todayDATE

0.99+

two monthsQUANTITY

0.99+

Supercloud 2TITLE

0.98+

VMwareORGANIZATION

0.98+

SaksORGANIZATION

0.98+

PaaSTITLE

0.98+

amazon.comORGANIZATION

0.98+

first oneQUANTITY

0.98+

LaceworkORGANIZATION

0.98+

Harveer SinghPERSON

0.98+

OracleORGANIZATION

0.97+

alkiraPERSON

0.96+

firstQUANTITY

0.96+

SupercloudORGANIZATION

0.95+

Supercloud2TITLE

0.94+

WikibonORGANIZATION

0.94+

SupecloudORGANIZATION

0.94+

eachQUANTITY

0.93+

hundred millionQUANTITY

0.92+

multicloudORGANIZATION

0.92+

every ounce atomQUANTITY

0.91+

Amazon WebORGANIZATION

0.88+

Supercloud 3TITLE

0.87+

Breaking Analysis: Supercloud2 Explores Cloud Practitioner Realities & the Future of Data Apps


 

>> Narrator: From theCUBE Studios in Palo Alto and Boston bringing you data-driven insights from theCUBE and ETR. This is breaking analysis with Dave Vellante >> Enterprise tech practitioners, like most of us they want to make their lives easier so they can focus on delivering more value to their businesses. And to do so, they want to tap best of breed services in the public cloud, but at the same time connect their on-prem intellectual property to emerging applications which drive top line revenue and bottom line profits. But creating a consistent experience across clouds and on-prem estates has been an elusive capability for most organizations, forcing trade-offs and injecting friction into the system. The need to create seamless experiences is clear and the technology industry is starting to respond with platforms, architectures, and visions of what we've called the Supercloud. Hello and welcome to this week's Wikibon Cube Insights powered by ETR. In this breaking analysis we give you a preview of Supercloud 2, the second event of its kind that we've had on the topic. Yes, folks that's right Supercloud 2 is here. As of this recording, it's just about four days away 33 guests, 21 sessions, combining live discussions and fireside chats from theCUBE's Palo Alto Studio with prerecorded conversations on the future of cloud and data. You can register for free at supercloud.world. And we are super excited about the Supercloud 2 lineup of guests whereas Supercloud 22 in August, was all about refining the definition of Supercloud testing its technical feasibility and understanding various deployment models. Supercloud 2 features practitioners, technologists and analysts discussing what customers need with real-world examples of Supercloud and will expose thinking around a new breed of cross-cloud apps, data apps, if you will that change the way machines and humans interact with each other. Now the example we'd use if you think about applications today, say a CRM system, sales reps, what are they doing? They're entering data into opportunities they're choosing products they're importing contacts, et cetera. And sure the machine can then take all that data and spit out a forecast by rep, by region, by product, et cetera. But today's applications are largely about filling in forms and or codifying processes. In the future, the Supercloud community sees a new breed of applications emerging where data resides on different clouds, in different data storages, databases, Lakehouse, et cetera. And the machine uses AI to inspect the e-commerce system the inventory data, supply chain information and other systems, and puts together a plan without any human intervention whatsoever. Think about a system that orchestrates people, places and things like an Uber for business. So at Supercloud 2, you'll hear about this vision along with some of today's challenges facing practitioners. Zhamak Dehghani, the founder of Data Mesh is a headliner. Kit Colbert also is headlining. He laid out at the first Supercloud an initial architecture for what that's going to look like. That was last August. And he's going to present his most current thinking on the topic. Veronika Durgin of Sachs will be featured and talk about data sharing across clouds and you know what she needs in the future. One of the main highlights of Supercloud 2 is a dive into Walmart's Supercloud. Other featured practitioners include Western Union Ionis Pharmaceuticals, Warner Media. We've got deep, deep technology dives with folks like Bob Muglia, David Flynn Tristan Handy of DBT Labs, Nir Zuk, the founder of Palo Alto Networks focused on security. Thomas Hazel, who's going to talk about a new type of database for Supercloud. It's several analysts including Keith Townsend Maribel Lopez, George Gilbert, Sanjeev Mohan and so many more guests, we don't have time to list them all. They're all up on supercloud.world with a full agenda, so you can check that out. Now let's take a look at some of the things that we're exploring in more detail starting with the Walmart Cloud native platform, they call it WCNP. We definitely see this as a Supercloud and we dig into it with Jack Greenfield. He's the head of architecture at Walmart. Here's a quote from Jack. "WCNP is an implementation of Kubernetes for the Walmart ecosystem. We've taken Kubernetes off the shelf as open source." By the way, they do the same thing with OpenStack. "And we have integrated it with a number of foundational services that provide other aspects of our computational environment. Kubernetes off the shelf doesn't do everything." And so what Walmart chose to do, they took a do-it-yourself approach to build a Supercloud for a variety of reasons that Jack will explain, along with Walmart's so-called triplet architecture connecting on-prem, Azure and GCP. No surprise, there's no Amazon at Walmart for obvious reasons. And what they do is they create a common experience for devs across clouds. Jack is going to talk about how Walmart is evolving its Supercloud in the future. You don't want to miss that. Now, next, let's take a look at how Veronica Durgin of SAKS thinks about data sharing across clouds. Data sharing we think is a potential killer use case for Supercloud. In fact, let's hear it in Veronica's own words. Please play the clip. >> How do we talk to each other? And more importantly, how do we data share? You know, I work with data, you know this is what I do. So if you know I want to get data from a company that's using, say Google, how do we share it in a smooth way where it doesn't have to be this crazy I don't know, SFTP file moving? So that's where I think Supercloud comes to me in my mind, is like practical applications. How do we create that mesh, that network that we can easily share data with each other? >> Now data mesh is a possible architectural approach that will enable more facile data sharing and the monetization of data products. You'll hear Zhamak Dehghani live in studio talking about what standards are missing to make this vision a reality across the Supercloud. Now one of the other things that we're really excited about is digging deeper into the right approach for Supercloud adoption. And we're going to share a preview of a debate that's going on right now in the community. Bob Muglia, former CEO of Snowflake and Microsoft Exec was kind enough to spend some time looking at the community's supercloud definition and he felt that it needed to be simplified. So in near real time he came up with the following definition that we're showing here. I'll read it. "A Supercloud is a platform that provides programmatically consistent services hosted on heterogeneous cloud providers." So not only did Bob simplify the initial definition he's stressed that the Supercloud is a platform versus an architecture implying that the platform provider eg Snowflake, VMware, Databricks, Cohesity, et cetera is responsible for determining the architecture. Now interestingly in the shared Google doc that the working group uses to collaborate on the supercloud de definition, Dr. Nelu Mihai who is actually building a Supercloud responded as follows to Bob's assertion "We need to avoid creating many Supercloud platforms with their own architectures. If we do that, then we create other proprietary clouds on top of existing ones. We need to define an architecture of how Supercloud interfaces with all other clouds. What is the information model? What is the execution model and how users will interact with Supercloud?" What does this seemingly nuanced point tell us and why does it matter? Well, history suggests that de facto standards will emerge more quickly to resolve real world practitioner problems and catch on more quickly than consensus-based architectures and standards-based architectures. But in the long run, the ladder may serve customers better. So we'll be exploring this topic in more detail in Supercloud 2, and of course we'd love to hear what you think platform, architecture, both? Now one of the real technical gurus that we'll have in studio at Supercloud two is David Flynn. He's one of the people behind the the movement that enabled enterprise flash adoption, that craze. And he did that with Fusion IO and he is now working on a system to enable read write data access to any user in any application in any data center or on any cloud anywhere. So think of this company as a Supercloud enabler. Allow me to share an excerpt from a conversation David Flore and I had with David Flynn last year. He as well gave a lot of thought to the Supercloud definition and was really helpful with an opinionated point of view. He said something to us that was, we thought relevant. "What is the operating system for a decentralized cloud? The main two functions of an operating system or an operating environment are one the process scheduler and two, the file system. The strongest argument for supercloud is made when you go down to the platform layer and talk about it as an operating environment on which you can run all forms of applications." So a couple of implications here that will be exploring with David Flynn in studio. First we're inferring from his comment that he's in the platform camp where the platform owner is responsible for the architecture and there are obviously trade-offs there and benefits but we'll have to clarify that with him. And second, he's basically saying, you kill the concept the further you move up the stack. So the weak, the further you move the stack the weaker the supercloud argument becomes because it's just becoming SaaS. Now this is something we're going to explore to better understand is thinking on this, but also whether the existing notion of SaaS is changing and whether or not a new breed of Supercloud apps will emerge. Which brings us to this really interesting fellow that George Gilbert and I RIFed with ahead of Supercloud two. Tristan Handy, he's the founder and CEO of DBT Labs and he has a highly opinionated and technical mind. Here's what he said, "One of the things that we still don't know how to API-ify is concepts that live inside of your data warehouse inside of your data lake. These are core concepts that the business should be able to create applications around very easily. In fact, that's not the case because it involves a lot of data engineering pipeline and other work to make these available. So if you really want to make it easy to create these data experiences for users you need to have an ability to describe these metrics and then to turn them into APIs to make them accessible to application developers who have literally no idea how they're calculated behind the scenes and they don't need to." A lot of implications to this statement that will explore at Supercloud two versus Jamma Dani's data mesh comes into play here with her critique of hyper specialized data pipeline experts with little or no domain knowledge. Also the need for simplified self-service infrastructure which Kit Colbert is likely going to touch upon. Veronica Durgin of SAKS and her ideal state for data shearing along with Harveer Singh of Western Union. They got to deal with 200 locations around the world in data privacy issues, data sovereignty how do you share data safely? Same with Nick Taylor of Ionis Pharmaceutical. And not to blow your mind but Thomas Hazel and Bob Muglia deposit that to make data apps a reality across the Supercloud you have to rethink everything. You can't just let in memory databases and caching architectures take care of everything in a brute force manner. Rather you have to get down to really detailed levels even things like how data is laid out on disk, ie flash and think about rewriting applications for the Supercloud and the MLAI era. All of this and more at Supercloud two which wouldn't be complete without some data. So we pinged our friends from ETR Eric Bradley and Darren Bramberm to see if they had any data on Supercloud that we could tap. And so we're going to be analyzing a number of the players as well at Supercloud two. Now, many of you are familiar with this graphic here we show some of the players involved in delivering or enabling Supercloud-like capabilities. On the Y axis is spending momentum and on the horizontal accesses market presence or pervasiveness in the data. So netscore versus what they call overlap or end in the data. And the table insert shows how the dots are plotted now not to steal ETR's thunder but the first point is you really can't have supercloud without the hyperscale cloud platforms which is shown on this graphic. But the exciting aspect of Supercloud is the opportunity to build value on top of that hyperscale infrastructure. Snowflake here continues to show strong spending velocity as those Databricks, Hashi, Rubrik. VMware Tanzu, which we all put under the magnifying glass after the Broadcom announcements, is also showing momentum. Unfortunately due to a scheduling conflict we weren't able to get Red Hat on the program but they're clearly a player here. And we've put Cohesity and Veeam on the chart as well because backup is a likely use case across clouds and on-premises. And now one other call out that we drill down on at Supercloud two is CloudFlare, which actually uses the term supercloud maybe in a different way. They look at Supercloud really as you know, serverless on steroids. And so the data brains at ETR will have more to say on this topic at Supercloud two along with many others. Okay, so why should you attend Supercloud two? What's in it for me kind of thing? So first of all, if you're a practitioner and you want to understand what the possibilities are for doing cross-cloud services for monetizing data how your peers are doing data sharing, how some of your peers are actually building out a Supercloud you're going to get real world input from practitioners. If you're a technologist, you're trying to figure out various ways to solve problems around data, data sharing, cross-cloud service deployment there's going to be a number of deep technology experts that are going to share how they're doing it. We're also going to drill down with Walmart into a practical example of Supercloud with some other examples of how practitioners are dealing with cross-cloud complexity. Some of them, by the way, are kind of thrown up their hands and saying, Hey, we're going mono cloud. And we'll talk about the potential implications and dangers and risks of doing that. And also some of the benefits. You know, there's a question, right? Is Supercloud the same wine new bottle or is it truly something different that can drive substantive business value? So look, go to Supercloud.world it's January 17th at 9:00 AM Pacific. You can register for free and participate directly in the program. Okay, that's a wrap. I want to give a shout out to the Supercloud supporters. VMware has been a great partner as our anchor sponsor Chaos Search Proximo, and Alura as well. For contributing to the effort I want to thank Alex Myerson who's on production and manages the podcast. Ken Schiffman is his supporting cast as well. Kristen Martin and Cheryl Knight to help get the word out on social media and at our newsletters. And Rob Ho is our editor-in-chief over at Silicon Angle. Thank you all. Remember, these episodes are all available as podcast. Wherever you listen we really appreciate the support that you've given. We just saw some stats from from Buzz Sprout, we hit the top 25% we're almost at 400,000 downloads last year. So really appreciate your participation. All you got to do is search Breaking Analysis podcast and you'll find those I publish each week on wikibon.com and siliconangle.com. Or if you want to get ahold of me you can email me directly at David.Vellante@siliconangle.com or dm me DVellante or comment on our LinkedIn post. I want you to check out etr.ai. They've got the best survey data in the enterprise tech business. This is Dave Vellante for theCUBE Insights, powered by ETR. Thanks for watching. We'll see you next week at Supercloud two or next time on breaking analysis. (light music)

Published Date : Jan 14 2023

SUMMARY :

with Dave Vellante of the things that we're So if you know I want to get data and on the horizontal

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Bob MugliaPERSON

0.99+

Alex MyersonPERSON

0.99+

Cheryl KnightPERSON

0.99+

David FlynnPERSON

0.99+

VeronicaPERSON

0.99+

JackPERSON

0.99+

Nelu MihaiPERSON

0.99+

Zhamak DehghaniPERSON

0.99+

Thomas HazelPERSON

0.99+

Nick TaylorPERSON

0.99+

Dave VellantePERSON

0.99+

Jack GreenfieldPERSON

0.99+

Kristen MartinPERSON

0.99+

Ken SchiffmanPERSON

0.99+

Veronica DurginPERSON

0.99+

WalmartORGANIZATION

0.99+

Rob HoPERSON

0.99+

Warner MediaORGANIZATION

0.99+

Tristan HandyPERSON

0.99+

Veronika DurginPERSON

0.99+

George GilbertPERSON

0.99+

Ionis PharmaceuticalORGANIZATION

0.99+

George GilbertPERSON

0.99+

Bob MugliaPERSON

0.99+

David FlorePERSON

0.99+

DBT LabsORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

BobPERSON

0.99+

Palo AltoLOCATION

0.99+

21 sessionsQUANTITY

0.99+

Darren BrambermPERSON

0.99+

33 guestsQUANTITY

0.99+

Nir ZukPERSON

0.99+

BostonLOCATION

0.99+

AmazonORGANIZATION

0.99+

Harveer SinghPERSON

0.99+

Kit ColbertPERSON

0.99+

DatabricksORGANIZATION

0.99+

Sanjeev MohanPERSON

0.99+

Supercloud 2TITLE

0.99+

SnowflakeORGANIZATION

0.99+

last yearDATE

0.99+

Western UnionORGANIZATION

0.99+

CohesityORGANIZATION

0.99+

SupercloudORGANIZATION

0.99+

200 locationsQUANTITY

0.99+

AugustDATE

0.99+

Keith TownsendPERSON

0.99+

Data MeshORGANIZATION

0.99+

Palo Alto NetworksORGANIZATION

0.99+

David.Vellante@siliconangle.comOTHER

0.99+

next weekDATE

0.99+

bothQUANTITY

0.99+

oneQUANTITY

0.99+

secondQUANTITY

0.99+

first pointQUANTITY

0.99+

OneQUANTITY

0.99+

FirstQUANTITY

0.99+

VMwareORGANIZATION

0.98+

Silicon AngleORGANIZATION

0.98+

ETRORGANIZATION

0.98+

Eric BradleyPERSON

0.98+

twoQUANTITY

0.98+

todayDATE

0.98+

SachsORGANIZATION

0.98+

SAKSORGANIZATION

0.98+

SupercloudEVENT

0.98+

last AugustDATE

0.98+

each weekQUANTITY

0.98+