Image Title

Search Results for Jupiter:

Erkang Zheng, JupiterOne | AWS re:Invent 2022 - Global Startup Program


 

well hello everybody John Wallace here on thecube he's continuing our segments here on the AWS Global startup showcase we are at day three of Reinventing irking Zhang is joining us now he is the CEO co-founder of Jupiter one um first off before we get going talking about you know security and big world for you guys I know what's your take on the show what's been going on out here at re invent yeah yeah ring event has been one of my favorite shows there's a lot of people here there's a lot of topics of course it's not just cyber security a lot of cloud infrastructure and just technology in general so you get a lot you know if you go walk the floor you see a lot of vendors you look at us go into sessions you can learn a lot but you're the Hot Topic right everybody's focused on Cyber yeah big time and with good reason right because as we know the Bad actors are getting even smarter and even faster and even more Nimble so just paint the landscape for me here in general right now as you see uh security Cloud Security in particular and and kind of where we are in that battle well we are clearly not winning so I think that in itself is a bit of a uh interesting problem right so as a it's not just Cloud security if you think about cyber security in general as an industry it has it has not been around for that long right but if you just look at the history of it uh we haven't done that while so uh pick another industry say medicine which has been around forever and if you look at the history of Medicine well I would argue you has done tremendously well because people live longer right when you get sick you get access to health care and yeah exactly you have Solutions and and you can see the trend even though there are problems in healthcare of course right but the trend is is good it's going well but not in cyber security more breaches more attacks more attackers we don't know what the hell we're doing with that many solutions and you know that's been one of my struggles as a former CSO and security practitioner for many years you know why is it that we're not getting better all right so I'm going to ask you the question yeah okay why aren't we getting better you know how come we can't stay ahead of the curve on this thing that for some reason it's like whack-a-mole times a hundred every time we think we solve one problem we have a hundred more that show up over here exactly and we have to address that and and our attention keeps floating around yeah I think you said it right so because we're taking this guacamole approach and we're looking for the painkiller of the day and you know we're looking for uh the Band-Aids right so and then we ended up well I I think to be fair to be fair to your industry the industry moves so quickly technology in general moves so quickly and security has been playing catch-up over time we're still playing catch-up so when you're playing catch-up you you can almost only uh look at you know what's the painkiller of what's the band name of the day so I can stop the bleeding right but I do think that we're we're to a point or we have enough painkillers and Band-Aids and and we need to start looking at how can we do better fundamentally with the basics and do the basics well because a lot of times the basics that get you into trouble so fundamentally the foundation I if I hear you right what you're saying is um you know quick changing industry right things are moving rapidly but we're not blocking and tackling we're not doing the X's and O's and so forget changing and we we got to get back to the basis and do those things right exactly you can only seem so simple it seems so simple but it's so hard right so you can you can think about you know uh even in case of building a starter building a company and and in order at one point right so we're blocking uh blocking tackling and then when we grow to a certain size we have to scale we have to figure out how to scale the business this is the same problem that happens in security as an industry we've been blocking happening for so long you know we're the industry is so young but we're to a point that we got to figure out how to scale this scale this in a fundamentally different way and I'll give you some example right so so what when we say the basics now it's easy to to think that say users should have MFA enabled is one of the basics right or another Basics will be you have endpoint protection on your devices you know maybe it's Cloud strike or Sentinel one or carbon black or whatever but the question being how do you know it is working 100 of the time right how do you know that how do you know right you find out too exactly that's right and how do you know that you have 100 coverage on your endpoints those Solutions are not going to tell you because they don't know what they don't know right if it's not enabled if it's not you know what what's the negative that you are not seeing so that's one of the things that you know that's in the basic state that you're now covering so the fundamentals it really goes to these five questions that I think that nobody has a really good answer for until now so the five questions goes what do I have right is it important what's important out of all the things I have you have a lot right you could have millions of things what important now for those that are important does it have a problem and if it has a problem who can fix it because the reality is in most cases security teams are not the ones fixing the problems they're they're the ones identical they're very good at recognizing but not so good exactly identifying the owner who can fix it right right could be could be business owner could be Engineers so the the asset ownership identification right so so these four questions and and then over time you know whether it's over a week or a month or a quarter or a year am I getting better right and then you just keep asking these questions in different areas in different domains with a different lens right so maybe that's endpoints maybe that's Cloud maybe that's you know users maybe that's a product and applications right but it really boils down to these five questions that's the foundation for any good security program if you can do that well I think we cover a lot of bases and we're going to be in much better shape than we have been all right so where do you come in man Jupiter one in terms of what you're providing because obviously you've identified this kind of pyramid yes this hierarchy of addressing needs and I assume obviously knowing you as I do and knowing the company as I do you've got Solutions that's exactly right right and and we precisely answer those five questions right for uh any organization uh from a asset perspective right because all the the answers to all those these five questions are based in assets it starts with knowing what I have right right so the the overall challenge of cyber security being broke broken I I believe is fundamentally that people do not understand and cannot uh probably deal with the complexity that we have within our own environments so again like you know using uh medicine as an example right so in order to come up with the right medicine for either it's a vaccine for covid-19 or whether it is a treatment for cancer or whatever that case may be you have to start with the foundations of understanding both the pathogen and to the human body like DNA sequencing right without those you cannot effectively produce the right medicine in modern uh you know Medicine sure right so that is the same thing that's happening in cyber security you know we spend a lot of times you know putting band days in patches right and then we spend a lot of time doing attacker research from the outside but we don't fundamentally understand in a complete way what's the complexity within our own environment in terms of digital assets and that's that's almost like the DNA of your own work what is that kind of mind-blowing in a way that if again hearing you what you're talking about is saying that the first step is to identify what you have that's right so it seems just so basic that that I should know what I what's under my hood I should know what is valuable and what is not I should prioritize what I really need to protect and what maybe can go on the second shelf yeah it has been a tough problem since the beginning of I.T not just the beginning of cyber security right so in the history of I.T we have this thing called cmdb configuration management database it is supposed to capture the configurations of it assets now over time that has become a lot more complex and and there's a lot more than just it asset that we have to understand from a security and attack service perspective right so we have to understand I.T environments we have to understand Cloud environments and applications and users and access and data and as and all of those things then then we have to take a different approach of sort of a modern cmdb right so what is the way that we can understand all of those complexity within all of those assets but not just independently within those silos but rather in a connected way so we can not only understand the attack surface but only but also understand the attack path that connect the dots from one thing to another right because everything in the organization is actually connected if if there's any one thing that sits on an island right so if you say you have a a a a server or a device or a user that is on an island that is not connected to the rest of the organization then why have it right and it doesn't matter so it's the understanding of that connect connected tissue this entire map where this you know DNA sequencing equivalent of a digital organization is what Jupiter one provides right so that visibility of the fundamental you know very granular uh level of assets and resources to answer those five questions and how does that how do I get better at that then I mean I have you to help me but but internally within our organization um I mean I don't want to be rude but I mean do I have do I have the skill for that do I have um do I have the the internal horsepower for that or or is there some need to close that Gap and how do I do it you know I'll tell you two things right so so one you mentioned the worst skills right so let me start there so because this one is very interesting we also have a huge skills shortage in cyber security we will we've all heard that for years and and and and for a long time but if you dig deeper into it why is that why is that and you know we have a lot of you know talented people right so why do we still have a skills shortage now what's interesting is if you think about what we're asking security people to do is mind-boggling so if you if you get a security analyst to say hey I want to understand how to protect something or or how to deal with an incident and what you're asking the person to do is not only to understand the security concept and be a domain expert in security you're also asking the person to and understand at the same time AWS or other clouds or endpoints or code or applications so that you can properly do the analysis and the in the response it's it's impossible it's like you know if you have you have to have a person who's an expert in everything know everything about everything that's right it's impossible so so so that's that's one thing that we have to to resolve is how do we use technology like Jupiter one to provide an abstraction so that there's Automation in place to help the security teams be better at their jobs without having to be an expert in deep technology right just add the abstract level of understanding because you know we can we can model the data and and provide the analysis and visual visualization out of the box for them so they can focus on just the security practices so that's one and the second thing is we have to change the mindset like take vulnerability management as an example right so the mindset for vulnerability management has been how do I manage findings now we have to change it to the concept of more proactive and how to manage assets so let's think about uh you know say log4j right that that happened and uh you know when it happened everybody scrambles and said hey which which devices or which you know uh systems have log4j and you know it doesn't matter what's the impact we can fix it right going back to those questions that that I mentioned before right and then um and then they try to look for a solution at a time say well where's that silver bullet that can give me the answers now what what what we struggle with though is that you know I want to maybe ask the question where were you six months ago where were you six months ago where you could have done the due diligence and put something in place that help you understand all of these assets and connections so you can go to one place and just ask for that question when something like that you know hit the fan so so if we do not fundamentally change the mindset to say I have to look at things not from a reactive findings perspective but really starting from an asset-centric you know day one perspective to look at that and have this Foundation have this map build we can't get there right so it's like you know if I need direction I go to Google Maps right but the the reason that it works is because somebody has done the work of creating the map right right if you haven't if you don't have the map and you just at you know when the time you say I gotta go somewhere and you expect the map to magically happen to show you the direction it's not going to work right right I imagine there are a lot of people out there right now are listening to thinking oh boy you know and that's what Jupiter one's all about they're there to answer your oh boy thanks for the time of course I appreciate the insights as well it's nice to know that uh at least somebody is reminding us to keep the front door locked too that's just the back door the side doors keep that front door and that garage locked up too definitely um all right we'll continue our coverage here at AWS re invent 22 this is part of the AWS Global startup showcase and you're watching the cube the leader in high-tech coverage foreign

Published Date : Dec 1 2022

SUMMARY :

all right so I'm going to ask you the

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
five questionsQUANTITY

0.99+

John WallacePERSON

0.99+

five questionsQUANTITY

0.99+

AWSORGANIZATION

0.99+

six months agoDATE

0.99+

four questionsQUANTITY

0.99+

first stepQUANTITY

0.99+

two thingsQUANTITY

0.99+

over a weekQUANTITY

0.99+

Google MapsTITLE

0.98+

Erkang ZhengPERSON

0.98+

second thingQUANTITY

0.98+

oneQUANTITY

0.98+

covid-19OTHER

0.98+

one placeQUANTITY

0.98+

ZhangPERSON

0.98+

bothQUANTITY

0.97+

one problemQUANTITY

0.97+

one thingQUANTITY

0.97+

JupiterLOCATION

0.96+

second shelfQUANTITY

0.95+

millions of thingsQUANTITY

0.93+

a quarterQUANTITY

0.92+

one pointQUANTITY

0.9+

100 coverageQUANTITY

0.86+

Band-AidsORGANIZATION

0.85+

Global startup showcaseEVENT

0.85+

a yearQUANTITY

0.85+

a lot of peopleQUANTITY

0.85+

day threeQUANTITY

0.84+

yearsQUANTITY

0.82+

a lot of peopleQUANTITY

0.82+

firstQUANTITY

0.81+

re inventEVENT

0.8+

a monthQUANTITY

0.8+

a hundred moreQUANTITY

0.79+

one of the thingsQUANTITY

0.77+

favoriteQUANTITY

0.76+

my strugglesQUANTITY

0.76+

cmdbTITLE

0.74+

100 ofQUANTITY

0.74+

a hundredQUANTITY

0.73+

lot of timesQUANTITY

0.72+

one ofQUANTITY

0.71+

lot of basesQUANTITY

0.71+

timeQUANTITY

0.71+

one thingQUANTITY

0.69+

re:Invent 2022 - Global Startup ProgramTITLE

0.67+

ringEVENT

0.64+

lot ofQUANTITY

0.64+

Reinventing irkingEVENT

0.62+

dayQUANTITY

0.61+

lotQUANTITY

0.61+

re invent 22EVENT

0.58+

lot of timesQUANTITY

0.57+

strikeOTHER

0.49+

every timeQUANTITY

0.43+

JupiterTITLE

0.43+

JupiterOneORGANIZATION

0.31+

Lucas Snyder, Indiana University and Karl Oversteyns, Purdue University | SuperComputing 22


 

(upbeat music) >> Hello, beautiful humans and welcome back to Supercomputing. We're here in Dallas, Texas giving you live coverage with theCUBE. I'm joined by David Nicholson. Thank you for being my left arm today. >> Thank you Savannah. >> It's a nice little moral. Very excited about this segment. We've talked a lot about how the fusion between academia and the private sector is a big theme at this show. You can see multiple universities all over the show floor as well as many of the biggest companies on earth. We were very curious to learn a little bit more about this from people actually in the trenches. And we are lucky to be joined today by two Purdue students. We have Lucas and Karl. Thank you both so much for being here. >> One Purdue, one IU, I think. >> Savannah: Oh. >> Yeah, yeah, yeah. >> I'm sorry. Well then wait, let's give Indiana University their fair do. That's where Lucas is. And Karl is at Purdue. Sorry folks. I apparently need to go back to school to learn how to read. (chuckles) In the meantime, I know you're in the middle of a competition. Thank you so much for taking the time out. Karl, why don't you tell us what's going on? What is this competition? What brought you all here? And then let's dive into some deeper stuff. >> Yeah, this competition. So we're a joint team between Purdue and IU. We've overcome our rivalries, age old rivalries to computer at the competition. It's a multi-part competition where we're going head to head against other teams from all across the world, benchmarking our super computing cluster that we designed. >> Was there a moment of rift at all when you came together? Or was everyone peaceful? >> We came together actually pretty nicely. Our two advisors they were very encouraging and so we overcame that, no hostility basically. >> I love that. So what are you working on and how long have you guys been collaborating on it? You can go ahead and start Lucas. >> So we've been prepping for this since the summer and some of us even before that. >> Savannah: Wow. >> And so currently we're working on the application phase of the competition. So everybody has different specialties and basically the competition gives you a set of rules and you have to accomplish what they tell you to do in the allotted timeframe and run things very quickly. >> And so we saw, when we came and first met you, we saw that there are lights and sirens and a monitor looking at the power consumption involved. So part of this is how much power is being consumed. >> Karl: That's right. >> Explain exactly what are the what are the rules that you have to live within? >> So, yeah, so the main constraint is the time as we mentioned and the power consumption. So for the benchmarking phase, which was one, two days ago there was a hard camp of 3000 watts to be consumed. You can't go over that otherwise you would be penalized for that. You have to rerun, start from scratch basically. Now there's a dynamic one for the application section where it's it modulates at random times. So we don't know when it's going to go down when it's going to go back up. So we have to adapt to that in real time. >> David: Oh, interesting. >> Dealing with a little bit of real world complexity I guess probably is simulation is here. I think that's pretty fascinating. I want to know, because I am going to just confess when I was your age last week, I did not understand the power of supercomputing and high performance computing. Lucas, let's start with you. How did you know this was the path you wanted to go down in your academic career? >> David: Yeah, what's your background? >> Yeah, give us some. >> So my background is intelligence systems engineering which is kind of a fusion. It's between, I'm doing bioengineering and then also more classical computer engineering. So my background is biology actually. But I decided to go down this path kind of on a whim. My professor suggested it and I've kind of fallen in love with it. I did my summer internship doing HPC and I haven't looked back. >> When did you think you wanted to go into this field? I mean, in high school, did you have a special teacher that sparked it? What was it? >> Lucas: That's funny that you say that. >> What was in your background? >> Yes, I mean, in high school towards the end I just knew that, I saw this program at IU and it's pretty new and I just thought this would be a great opportunity for me and I'm loving it so far. >> Do you have family in tech or is this a different path for you? >> Yeah, this is a different path for me, but my family is so encouraging and they're very happy for me. They text me all the time. So I couldn't be happier. >> Savannah: Just felt that in my heart. >> I know. I was going to say for the parents out there get the tissue out. >> Yeah, yeah, yeah. (chuckles) >> These guys they don't understand. But, so Karl, what's your story? What's your background? >> My background, I'm a major in unmanned Aerial systems. So this is a drones commercial applications not immediately connected as you might imagine although there's actually more overlap than one might think. So a lot of unmanned systems today a lot of it's remote sensing, which means that there's a lot of image processing that takes place. Mapping of a field, what have you, or some sort of object, like a silo. So a lot of it actually leverages high performance computing in order to map, to visualize much replacing, either manual mapping that used to be done by humans in the field or helicopters. So a lot of cost reduction there and efficiency increases. >> And when did you get this spark that said I want to go to Purdue? You mentioned off camera that you're from Belgium. >> Karl: That's right. >> Did you, did you come from Belgium to Purdue or you were already in the States? >> No, so I have family that lives in the States but I grew up in Belgium. >> David: Okay. >> I knew I wanted to study in the States. >> But at what age did you think that science and technology was something you'd be interested in? >> Well, I've always loved computers from a young age. I've been breaking computers since before I can remember. (chuckles) Much to my parents dismay. But yeah, so I've always had a knack for technology and that's sort of has always been a hobby of mine. >> And then I want to ask you this question and then Lucas and then Savannah will get some time. >> Savannah: It cool, will just sit here and look pretty. >> Dream job. >> Karl: Dream job. >> Okay. So your undergrad both you. >> Savannah: Offering one of my questions. Kind of, It's adjacent though. >> Okay. You're undergrad now? Is there grad school in your future do you feel that's necessary? Is that something you want to pursue? >> I think so. Entrepreneurship is something that's been in the back of my head for a while as well. So may be or something. >> So when I say dream job, understand could be for yourself. >> Savannah: So just piggyback. >> Dream thing after academia or stay in academia. What's do you think at this point? >> That's a tough question. You're asking. >> You'll be able to review this video in 10 years. >> Oh boy. >> This is give us your five year plan and then we'll have you back on theCUBE and see 2027. >> What's the dream? There's people out here watching this. I'm like, go, hey, interesting. >> So as I mentioned entrepreneurship I'm thinking I'll start a company at some point. >> David: Okay. >> Yeah. In what? I don't know yet. We'll see. >> David: Lucas, any thoughts? >> So after graduation, I am planning to go to grad school. IU has a great accelerated master's degree program so I'll stay an extra year and get my master's. Dream job is, boy, that's impossible to answer but I remember telling my dad earlier this year that I was so interested in what NASA was doing. They're sending a probe to one of the moons of Jupiter. >> That's awesome. From a parent's perspective the dream often is let's get the kids off the payroll. So I'm sure that your families are happy to hear that you have. >> I think these two will be right in that department. >> I think they're going to be okay. >> Yeah, I love that. I was curious, I want to piggyback on that because I think when NASA's doing amazing we have them on the show. Who doesn't love space. >> Yeah. >> I'm also an entrepreneur though so I very much empathize with that. I was going to ask to your dream job, but also what companies here do you find the most impressive? I'll rephrase. Because I was going to say, who would you want to work with? >> David: Anything you think is interesting? >> But yeah. Have you even had a chance to walk the floor? I know you've been busy competing >> Karl: Very little. >> Yeah, I was going to say very little. Unfortunately I haven't been able to roam around very much. But I look around and I see names that I'm like I can't even, it's crazy to see them. Like, these are people who are so impressive in the space. These are people who are extremely smart. I'm surrounded by geniuses everywhere I look, I feel like, so. >> Savannah: That that includes us. >> Yeah. >> He wasn't talking about us. Yeah. (laughs) >> I mean it's hard to say any of these companies I would feel very very lucky to be a part of, I think. >> Well there's a reason why both of you were invited to the party, so keep that in mind. Yeah. But so not a lot of time because of. >> Yeah. Tomorrow's our day. >> Here to get work. >> Oh yes. Tomorrow gets play and go talk to everybody. >> Yes. >> And let them recruit you because I'm sure that's what a lot of these companies are going to be doing. >> Yeah. Hopefully it's plan. >> Have you had a second at all to look around Karl. >> A Little bit more I've been going to the bathroom once in a while. (laughs) >> That's allowed I mean, I can imagine that's a vital part of the journey. >> I've ruin my gaze a little bit to what's around all kinds of stuff. Higher education seems to be very important in terms of their presence here. I find that very, very impressive. Purdue has a big stand IU as well, but also others all from Europe as well and Asia. I think higher education has a lot of potential in this field. >> David: Absolutely. >> And it really is that union between academia and the private sector. We've seen a lot of it. But also one of the things that's cool about HPC is it's really not ageist. It hasn't been around for that long. So, I mean, well, at this scale it's obviously this show's been going on since 1988 before you guys were even probably a thought. But I think it's interesting. It's so fun to get to meet you both. Thank you for sharing about what you're doing and what your dreams are. Lucas and Karl. >> David: Thanks for taking the time. >> I hope you win and we're going to get you off the show here as quickly as possible so you can get back to your teams and back to competing. David, great questions as always, thanks for being here. And thank you all for tuning in to theCUBE Live from Dallas, Texas, where we are at Supercomputing. My name's Savannah Peterson and I hope you're having a beautiful day. (gentle upbeat music)

Published Date : Nov 16 2022

SUMMARY :

Thank you for being my left arm today. Thank you both so much for being here. I apparently need to go back from all across the world, and so we overcame that, So what are you working on since the summer and some and you have to accomplish and a monitor looking at the So for the benchmarking phase, How did you know this was the path But I decided to go down I saw this program at They text me all the time. I was going to say for Yeah, yeah, yeah. But, so Karl, what's your story? So a lot of unmanned systems today And when did you get that lives in the States I can remember. ask you this question Savannah: It cool, will of my questions. Is that something you want to pursue? I think so. So when I say dream job, understand What's do you think at this point? That's a tough question. You'll be able to review and then we'll have you back What's the dream? So as I mentioned entrepreneurship I don't know yet. planning to go to grad school. to hear that you have. I think these two will I was curious, I want to piggyback on that I was going to ask to your dream job, Have you even had I can't even, it's crazy to see them. Yeah. I mean it's hard to why both of you were invited go talk to everybody. And let them recruit you Have you had a second I've been going to the I mean, I can imagine that's I find that very, very impressive. It's so fun to get to meet you both. going to get you off the show

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SavannahPERSON

0.99+

DavidPERSON

0.99+

David NicholsonPERSON

0.99+

BelgiumLOCATION

0.99+

KarlPERSON

0.99+

NASAORGANIZATION

0.99+

3000 wattsQUANTITY

0.99+

LucasPERSON

0.99+

IUORGANIZATION

0.99+

EuropeLOCATION

0.99+

Karl OversteynsPERSON

0.99+

Savannah PetersonPERSON

0.99+

five yearQUANTITY

0.99+

AsiaLOCATION

0.99+

Lucas SnyderPERSON

0.99+

Dallas, TexasLOCATION

0.99+

PurdueORGANIZATION

0.99+

two advisorsQUANTITY

0.99+

TomorrowDATE

0.99+

twoQUANTITY

0.99+

PurdueLOCATION

0.99+

1988DATE

0.99+

last weekDATE

0.99+

JupiterLOCATION

0.99+

bothQUANTITY

0.99+

Purdue UniversityORGANIZATION

0.99+

10 yearsQUANTITY

0.99+

OneQUANTITY

0.99+

todayDATE

0.99+

two days agoDATE

0.98+

oneQUANTITY

0.98+

Indiana UniversityORGANIZATION

0.98+

Indiana UniversityORGANIZATION

0.97+

earlier this yearDATE

0.93+

earthLOCATION

0.93+

firstQUANTITY

0.92+

SupercomputingORGANIZATION

0.9+

2027TITLE

0.86+

HPCORGANIZATION

0.8+

theCUBEORGANIZATION

0.8+

StatesLOCATION

0.56+

secondQUANTITY

0.48+

22QUANTITY

0.38+

Harry Glaser, Modlbit, Damon Bryan, Hyperfinity & Stefan Williams, Snowflake | Snowflake Summit 2022


 

>>Thanks. Hey, everyone, welcome back to the cubes. Continuing coverage of snowflakes. Summit 22 live from Caesars Forum in Las Vegas. Lisa Martin here. I have three guests here with me. We're gonna be talking about Snowflake Ventures and the snowflakes start up Challenge. That's in its second year. I've got Harry Glaser with me. Co founder and CEO of Model Bit Start Up Challenge finalist Damon Bryan joins us as well. The CTO and co founder of Hyper Affinity. Also a startup Challenge Finalists. And Stephane Williams to my left here, VP of Corporate development and snowflake Ventures. Guys, great to have you all on this little mini panel this morning. >>Thank you. >>Thank you. >>Let's go ahead, Harry, and we'll start with you. Talk to the audience about model. But what do you guys do? And then we'll kind of unpack the snowflake. The Snowflakes challenge >>Model bit is the easiest way for data scientists to deploy machine learning models directly into Snowflake. We make use of the latest snowflake functionality called Snow Park for python that allows those models to run adjacent to the data so that machine learning models can be much more efficient and much more powerful than they were before. >>Awesome. Damon. Give us an overview of hyper affinity. >>Yes, so hyper affinity were Decision Intelligence platform. So we helped. Specifically retailers and brands make intelligent decisions through the use of their own customer, data their product data and put data science in a I into the heart of the decision makers across their business. >>Nice Step seven. Tell us about the startup challenge. We talked a little bit about it yesterday with CMO Denise Pearson, but I know it's in its second year. Give us the idea of the impetus for it, what it's all about and what these companies embody. >>Yeah, so we This is the second year that we've done it. Um, we it was really out of, um Well, it starts with snowflake Ventures when we started to invest in companies, and we quickly realised that there's there's a massive opportunity for companies to be building on top of the Lego blocks, uh, of snowflake. And so, um, open up the competition. Last year it was the inaugural competition overlay analytics one, Um, and since then, you've seen a number of different functionalities and features as part of snowflakes snow part. Being one of them native applications is a really exciting one going forward. Um, the companies can really use to accelerate their ability to kind of deliver best in class applications using best in class technology to deliver real customer outcomes and value. Um, so we've we've seen tremendous traction across the globe, 250 applicants across 50. I think 70 countries was mentioned today, so truly global in nature. And it's really exciting to see how some of the start ups are taking snowflake to to to new and interesting use cases and new personas and new industries. >>So you had 200 over 250 software companies applied for this. How did you did you narrow it down to three? >>We did. Yeah, >>you do that. >>So, behind the scenes, we had a sub judging panel, the ones you didn't see up on stage, which I was luckily part of. We had kind of very distinct evaluation criteria that we were evaluating every company across. Um and we kind of took in tranches, right? We we took the first big garden, and we kind of try to get that down to a top 50 and top 50. Then we really went into the details and we kind of across, um, myself in ventures with some of my venture partners. Um, some of the market teams, some of the product and engineering team, all kind of came together and evaluated all of these different companies to get to the top 10, which was our semifinalists and then the semi finalists, or had a chance to present in front of the group. So we get. We got to meet over Zoom along the way where they did a pitch, a five minute pitch followed by a Q and A in a similar former, I guess, to what we just went through the startup challenge live, um, to get to the top three. And then here we are today, just coming out of the competition with with With folks here on the table. >>Wow, Harry talked to us about How did you just still down what model bit is doing into five minutes over Zoom and then five minutes this morning in person? >>I think it was really fun to have that pressure test where, you know, we've only been doing this for a short time. In fact model. It's only been a company for four or five months now, and to have this process where we pitch and pitch again and pitch again and pitch again really helped us nail the one sentence value proposition, which we hadn't done previously. So in that way, very grateful to step on in the team for giving us that opportunity. >>That helps tremendously. I can imagine being a 4 to 5 months young start up and really trying to figure out I've worked with those young start ups before. Messaging is challenging the narrative. Who are we? What do we do? How are we changing or chasing the market? What are our customers saying we are? That's challenging. So this was a good opportunity for you, Damon. Would you say the same as well for hyper affinity? >>Yeah, definitely conquer. It's really helped us to shape our our value proposition early and how we speak about that. It's quite complicated stuff, data science when you're trying to get across what you do, especially in retail, that we work in. So part of what our platform does is to help them make sense of data science and Ai and implement that into commercial decisions. So you have to be really kind of snappy with how you position things. And it's really helped us to do that. We're a little bit further down the line than than these guys we've been going for three years. So we've had the benefit of working with a lot of retailers to this point to actually identify what their problems are and shape our product and our proposition towards. >>Are you primarily working with the retail industry? >>Yes, Retail and CPG? Our primary use case. We have seen any kind of consumer related industries. >>Got it. Massive changes right in retail and CPG the last couple of years, the rise of consumer expectations. It's not going to go back down, right? We're impatient. We want brands to know who we are. I want you to deliver relevant content to me that if I if I bought a tent, go back on your website, don't show me more tense. Show me things that go with that. We have this expectation. You >>just explain the whole business. But >>it's so challenging because the brothers brands have to respond to that. How do you what is the value for retailers working with hyper affinity and snowflake together. What's that powerhouse? >>Yeah, exactly. So you're exactly right. The retail landscape is changing massively. There's inflation everywhere. The pandemic really impacted what consumers really value out of shopping with retailers. And those decisions are even harder for retailers to make. So that's kind of what our platform does. It helps them to make those decisions quickly, get the power of data science or democratise it into the hands of those decision makers. Um, so our platform helps to do that. And Snowflake really underpins that. You know, the scalability of snowflake means that we can scale the data and the capability that platform in tangent with that and snowflake have been innovating a lot of things like Snow Park and then the new announcements, announcements, uni store and a native APP framework really helping us to make developments to our product as quick as snowflakes are doing it. So it's really beneficial. >>You get kind of that tailwind from snowflakes acceleration. It sounds like >>exactly that. Yeah. So as soon as we hear about new things were like, Can we use it? You know, and Snow Park in particular was music to our ears, and we actually part of private preview for that. So we've been using that while and again some of the new developments will be. I'm on the phone to my guys saying, Can we use this? Get it, get it implemented pretty quickly. So yeah, >>fantastic. Sounds like a great aligned partnership there, Harry. Talk to us a little bit about model bit and how it's enabling customers. Maybe you've got a favourite customer example at model bit plus snowflake, the power that delivers to the end user customer? >>Absolutely. I mean, as I said, it allows you to deploy the M L model directly into snowflake. But sometimes you need to use the exact same machine learning model in multiple endpoints simultaneously. For example, one of our customers uses model bit to train and deploy a lead scoring model. So you know when somebody comes into your website and they fill out the form like they want to talk to a sales person, is this gonna be a really good customer? Do we think or maybe not so great? Maybe they won't pay quite as much, and that lead scoring model actually runs on the website using model bit so that you can deploy display a custom experience to that customer we know right away. If this is an A, B, C or D lead, and therefore do we show them a salesperson contact form? Do we just put them in the marketing funnel? Based on that lead score simultaneously, the business needs to know in the back office the score of the lead so that they can do things like routed to the appropriate salesperson or update their sales forecasts for the end of the quarter. That same model also runs in the in the snowflake warehouse so that those back office systems can be powered directly off of snowflake. The fact that they're able to train and deploy one model into two production environment simultaneously and manage all that is something they can only do with bottled it. >>Lead scoring has been traditionally challenging for businesses in every industry, but it's so incredibly important, especially as consumers get pickier and pickier with. I don't want I don't want to be measured. I want to opt out. What sounds like what model but is enabling is especially alignment between sales and marketing within companies, which is That's also a big challenge at many companies face for >>us. It starts with the data scientist, right? The fact that sales and marketing may not be aligned might be an issue with the source of truth. And do we have a source of truth at this company? And so the idea that we can empower these data scientists who are creating this value in the company by giving them best in class tools and resources That's our dream. That's our mission. >>Talk to me a little bit, Harry. You said you're only 4 to 5 months old. What were the gaps in the market that you and your co founders saw and said, Guys, we've got to solve this. And Snowflake is the right partner to help us do it. >>Absolutely. We This is actually our second start up, and we started previously a data Analytics company that was somewhat successful, and it got caught up in this big wave of migration of cloud tools. So all of data tools moved and are moving from on premise tools to cloud based tools. This is really a migration. That snowflake catalyst Snowflake, of course, is the ultimate in cloud based data platforms, moving customers from on premise data warehouses to modern cloud based data clouds that dragged and pulled the rest of the industry along with it. Data Science is one of the last pieces of the data industry that really hasn't moved to the cloud yet. We were almost surprised when we got done with our last start up. We were thinking about what to do next. The data scientists were still using Jupiter notebooks locally on their laptops, and we thought, This is a big market opportunity and we're We're almost surprised it hasn't been captured yet, and we're going to get in there. >>The other thing. I think it's really interesting on your business that we haven't talked about is just the the flow of data, right? So that the data scientist is usually taking data out of a of a of a day like something like Smoke like a data platform and the security kind of breaks down because then it's one. It's two, it's three, it's five, it's 20. Its, you know, big companies just gets really big. And so I think the really interesting thing with what you guys are doing is enabling the data to stay where it's at, not copping out keeping that security, that that highly governed environment that big companies want but allowing the data science community to really unlock that value from the data, which is really, really >>cool. Wonderful for small startups like Model Bit. Because you talk to a big company, you want them to become a customer. You want them to use your data science technology. They want to see your fed ramp certification. They want to talk to your C. So we're two guys in Silicon Valley with a dream. But if we can tell them the data is staying in snowflake and you have that conversation with Snowflake all the time and you trust them were just built on top. That is an easy and very smooth way to have that conversation with the customer. >>Would you both say that there's credibility like you got street cred, especially being so so early in this stage? Harry, with the partnership with With Snowflake Damon, we'll start with you. >>Yeah, absolutely. We've been using Snowflake from day one. We leave from when we started our company, and it was a little bit of an unknown, I guess maybe 23 years ago, especially in retail. A lot of retailers using all the legacy kind of enterprise software, are really starting to adopt the cloud now with what they're doing and obviously snowflake really innovating in that area. So what we're finding is we use Snowflake to host our platform and our infrastructure. We're finding a lot of retailers doing that as well, which makes it great for when they wanted to use products like ours because of the whole data share thing. It just becomes really easy. And it really simplifies it'll and data transformation and data sharing. >>Stephane, talk about the startup challenge, the innovation that you guys have seen, and only the second year I can. I can just hear it from the two of you. And I know that the winner is back in India, but tremendous amount of of potential, like to me the last 2.5 days, the flywheel that is snowflake is getting faster and faster and more and more powerful. What are some of the things that excite you about working on the start up challenge and some of the vision going forward that it's driving. >>I think the incredible thing about Snowflake is that we really focus as a company on the data infrastructure and and we're hyper focused on enabling and incubating and encouraging partners to kind of stand on top of a best of breed platform, um, unlocked value across the different, either personas within I T organisations or industries like hypothermia is doing. And so it's it's it's really incredible to see kind of domain knowledge and subject matter expertise, able to kind of plug into best of breed underlying data infrastructure and really divide, drive, drive real meaningful outcomes for for for our customers in the community. Um, it's just been incredible to see. I mean, we just saw three today. Um, there was 250 incredible applications that past the initial. Like, do they check all the boxes and then actually, wow, they just take you to these completely different areas. You never thought that the technology would go and solve. And yet here we are talking about, you know, really interesting use cases that have partners are taking us to two >>150. Did that surprise you? And what was it last year. >>I think it was actually close to close to 2 to 40 to 50 as well, and I think it was above to 50 this year. I think that's the number that is in my head from last year, but I think it's actually above that. But the momentum is, Yeah, it's there and and again, we're gonna be back next year with the full competition, too. So >>awesome. Harry, what is what are some of the things that are next for model bed as it progresses through its early stages? >>You know, one thing I've learned and I think probably everyone at this table has internalised this lesson. Product market fit really is everything for a start up. And so for us, it's We're fortunate to have a set of early design partners who will become our customers, who we work with every day to build features, get their feedback, make sure they love the product, and the most exciting thing that happened to me here this week was one of our early design partner. Customers wanted us to completely rethink how we integrate with gets so that they can use their CI CD workflows their continuous integration that they have in their own get platform, which is advanced. They've built it over many years, and so can they back, all of model, but with their get. And it was it was one of those conversations. I know this is getting a little bit in the weeds, but it was one of those conversations that, as a founder, makes your head explode. If we can have a critical mass of those conversations and get to that product market fit, then the flywheel starts. Then the investment money comes. Then you're hiring a big team and you're off to the races. >>Awesome. Sounds like there's a lot of potential and momentum there. Damon. Last question for you is what's next for hyper affinity. Obviously you've got we talked about the street cred. >>Yeah, what's >>next for the business? >>Well, so yeah, we we've got a lot of exciting times coming up, so we're about to really fully launch our products. So we've been trading for three years with consultancy in retail analytics and data science and actually using our product before it was fully ready to launch. So we have the kind of main launch of our product and we actually starting to onboard some clients now as we speak. Um, I think the climate with regards to trying to find data, science, resources, you know, a problem across the globe. So it really helps companies like ours that allow, you know, allow retailers or whoever is to democratise the use of data science. And perhaps, you know, really help them in this current climate where they're struggling to get world class resource to enable them to do that >>right so critical stuff and take us home with your overall summary of snowflake summit. Fourth annual, nearly 10,000 people here. Huge increase from the last time we were all in person. What's your bumper sticker takeaway from Summit 22 the Startup Challenge? >>Uh, that's a big closing statement for me. It's been just the energy. It's been incredible energy, incredible excitement. I feel the the products that have been unveiled just unlock a tonne, more value and a tonne, more interesting things for companies like the model bit I profanity and all the other startups here. And to go and think about so there's there's just this incredible energy, incredible excitement, both internally, our product and engineering teams, the partners that we have spoke. I've spoken here with the event, the portfolio companies that we've invested in. And so there's there's there's just this. Yeah, incredible momentum and excitement around what we're able to do with data in today's world, powered by underlying platform, like snowflakes. >>Right? And we've heard that energy, I think, through l 30 plus guests we've had on the show since Tuesday and certainly from the two of you as well. Congratulations on being finalist. We wish you the best of luck. You have to come back next year and talk about some of the great things. More great >>things hopefully will be exhibited next year. >>Yeah, that's a good thing to look for. Guys really appreciate your time and your insights. Congratulations on another successful start up challenge. >>Thank you so much >>for Harry, Damon and Stefan. I'm Lisa Martin. You're watching the cubes. Continuing coverage of snowflakes. Summit 22 live from Vegas. Stick around. We'll be right back with a volonte and our final guest of the day. Mhm, mhm

Published Date : Jun 16 2022

SUMMARY :

Guys, great to have you all on this little mini panel this morning. But what do you guys do? Model bit is the easiest way for data scientists to deploy machine learning models directly into Snowflake. Give us an overview of hyper affinity. So we helped. Give us the idea of the impetus for it, what it's all about and what these companies And it's really exciting to see how some of the start ups are taking snowflake to So you had 200 over 250 software companies applied We did. So, behind the scenes, we had a sub judging panel, I think it was really fun to have that pressure test where, you know, I can imagine being a 4 to 5 months young start up of snappy with how you position things. Yes, Retail and CPG? I want you to deliver relevant content to me that just explain the whole business. it's so challenging because the brothers brands have to respond to that. You know, the scalability of snowflake means that we can scale the You get kind of that tailwind from snowflakes acceleration. I'm on the phone to my guys saying, Can we use this? bit plus snowflake, the power that delivers to the end user customer? the business needs to know in the back office the score of the lead so that they can do things like routed to the appropriate I want to opt out. And so the idea that And Snowflake is the right partner to help us do it. dragged and pulled the rest of the industry along with it. So that the data scientist is usually taking data out of a of a of a day like something But if we can tell them the data is staying in snowflake and you have that conversation with Snowflake all the time Would you both say that there's credibility like you got street cred, especially being so so are really starting to adopt the cloud now with what they're doing and obviously snowflake really innovating in that area. And I know that the winner is back in India, but tremendous amount of of and really divide, drive, drive real meaningful outcomes for for for our customers in the community. And what was it last year. But the momentum Harry, what is what are some of the things that are next for model bed as and the most exciting thing that happened to me here this week was one of our early design partner. Last question for you is what's next for hyper affinity. So it really helps companies like ours that allow, you know, allow retailers or whoever is to democratise Huge increase from the last time we were all in person. the partners that we have spoke. show since Tuesday and certainly from the two of you as well. Yeah, that's a good thing to look for. We'll be right back with a volonte and our final guest of the day.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Damon BryanPERSON

0.99+

Stephane WilliamsPERSON

0.99+

Lisa MartinPERSON

0.99+

Harry GlaserPERSON

0.99+

HarryPERSON

0.99+

IndiaLOCATION

0.99+

4QUANTITY

0.99+

Silicon ValleyLOCATION

0.99+

five minutesQUANTITY

0.99+

fourQUANTITY

0.99+

ModlbitPERSON

0.99+

VegasLOCATION

0.99+

StephanePERSON

0.99+

next yearDATE

0.99+

three yearsQUANTITY

0.99+

five monthsQUANTITY

0.99+

Last yearDATE

0.99+

Hyper AffinityORGANIZATION

0.99+

last yearDATE

0.99+

twoQUANTITY

0.99+

two guysQUANTITY

0.99+

yesterdayDATE

0.99+

fiveQUANTITY

0.99+

Stefan WilliamsPERSON

0.99+

250 applicantsQUANTITY

0.99+

200QUANTITY

0.99+

20QUANTITY

0.99+

70 countriesQUANTITY

0.99+

Las VegasLOCATION

0.99+

Denise PearsonPERSON

0.99+

StefanPERSON

0.99+

five minuteQUANTITY

0.99+

threeQUANTITY

0.99+

second yearQUANTITY

0.99+

SnowflakeORGANIZATION

0.99+

this yearDATE

0.99+

todayDATE

0.99+

TuesdayDATE

0.99+

oneQUANTITY

0.99+

three guestsQUANTITY

0.98+

23 years agoDATE

0.98+

DamonPERSON

0.98+

50QUANTITY

0.98+

5 monthsQUANTITY

0.98+

Model BitORGANIZATION

0.98+

one modelQUANTITY

0.97+

40QUANTITY

0.97+

one sentenceQUANTITY

0.97+

Snow ParkTITLE

0.97+

Snowflake DamonORGANIZATION

0.97+

this weekDATE

0.96+

top threeQUANTITY

0.95+

two productionQUANTITY

0.95+

bothQUANTITY

0.94+

250 incredible applicationsQUANTITY

0.94+

Fourth annualQUANTITY

0.94+

SnowflakeEVENT

0.94+

top 50QUANTITY

0.92+

day oneQUANTITY

0.92+

VenturesORGANIZATION

0.91+

top 10QUANTITY

0.91+

aboveQUANTITY

0.9+

Did HPE GreenLake Just Set a New Bar in the On-Prem Cloud Services Market?


 

>> Welcome back to The Cube's coverage of HPE's GreenLake announcements. My name is Dave Vellante and you're watching the Cube. I'm here with Holger Mueller, who is an analyst at Constellation Research. And Matt Maccaux is the global field CTO of Ezmeral software at HPE. We're going to talk data. Gents, great to see you. >> Holger: Great to be here. >> So, Holger, what do you see happening in the data market? Obviously data's hot, you know, digital, I call it the force marks to digital. Everybody realizes wow, digital business, that's a data business. We've got to get our data act together. What do you see in the market is the big trends, the big waves? >> We are all young enough or old enough to remember when people were saying data is the new oil, right? Nothing has changed, right? Data is the key ingredient, which matters to enterprise, which they have to store, which they have to enrich, which they have to use for their decision-making. It's the foundation of everything. If you want to go into machine learning or (indistinct) It's growing very fast, right? We have the capability now to look at all the data in enterprise, which weren't able 10 years ago to do that. So data is main center to everything. >> Yeah, it's even more valuable than oil, I think, right? 'Cause with oil, you can only use once. Data, you can, it's kind of polyglot. I can go in different directions and it's amazing, right? >> It's the beauty of digital products, right? They don't get consumed, right? They don't get fired up, right? And no carbon footprint, right? "Oh wait, wait, we have to think about carbon footprint." Different story, right? So to get to the data, you have to spend some energy. >> So it's that simple, right? I mean, it really is. Data is fundamental. It's got to be at the core. And so Matt, what are you guys announcing today, and how does that play into what Holger just said? >> What we're announcing today is that organizations no longer need to make a difficult choice. Prior to today, organizations were thinking if I'm going to do advanced machine learning and really exploit my data, I have to go to the cloud. But all my data's still on premises because of privacy rules, industry rules. And so what we're announcing today, through GreenLake Services, is a cloud services way to deliver that same cloud-based analytical capability. Machine learning, data engineering, through hybrid analytics. It's a unified platform to tie together everything from data engineering to advance data science. And we're also announcing the world's first Kubernetes native object store, that is hybrid cloud enabled. Which means you can keep your data connected across clouds in a data fabric, or Dave, as you say, mesh. >> Okay, can we dig into that a little bit? So, you're essentially saying that, so you're going to have data in both places, right? Public cloud, edge, on-prem, and you're saying, HPE is announcing a capability to connect them, I think you used the term fabric. I'm cool, by the way, with the term fabric, we can, we'll parse that out another time. >> I love for you to discuss textiles. Fabrics vs. mesh. For me, every fabric breaks down to mesh if you put it on a microscope. It's the same thing. >> Oh wow, now that's really, that's too detailed for my brain, right this moment. But, you're saying you can connect all those different estates because data by its very nature is everywhere. You're going to unify that, and what, that can manage that through sort of a single view? >> That's right. So, the management is centralized. We need to be able to know where our data is being provisioned. But again, we don't want organizations to feel like they have to make the trade off. If they want to use cloud surface A in Azure, and cloud surface B in GCP, why not connect them together? Why not allow the data to remain in sync or not, through a distributed fabric? Because we use that term fabric over and over again. But the idea is let the data be where it most naturally makes sense, and exploit it. Monetization is an old tool, but exploit it in a way that works best for your users and applications. >> In sync or not, that's interesting. So it's my choice? >> That's right. Because the back of an automobile could be a teeny tiny, small edge location. It's not always going to be in sync until it connects back up with a training facility. But we still need to be able to manage that. And maybe that data gets persisted to a core data center. Maybe it gets pushed to the cloud, but we still need to know where that data is, where it came from, its lineage, what quality it has, what security we're going to wrap around that, that all should be part of this fabric. >> Okay. So, you've got essentially a governance model, at least maybe you're working toward that, and maybe it's not all baked today, but that's the north star. Is this fabric connect, single management view, governed in a federated fashion? >> Right. And it's available through the most common API's that these applications are already written in. So, everybody today's talking S3. I've got to get all of my data, I need to put it into an object store, it needs to be S3 compatible. So, we are extending this capability to be S3 native. But it's optimized for performance. Today, when you put data in an object store, it's kind of one size fits all. Well, we know for those streaming analytical capabilities, those high performance workloads, it needs to be tuned for that. So, how about I give you a very small object on the very fastest disk in your data center and maybe that cheaper location somewhere else. And so we're giving you that balance as part of the overall management estate. >> Holger, what's your take on this? I mean, Frank Slootman says we'll never, we're not going halfway house. We're never going to do on-prem, we're only in the cloud. So that basically says, okay, he's ignoring a pretty large market by choice. You're not, Matt, you must love those words. But what do you see as the public cloud players, kind of the moves on-prem, particularly in this realm? >> Well, we've seen lots of cloud players who were only cloud coming back towards on-premise, right? We call it the next generation compute platform where I can move data and workloads between on-premise and ideally, multiple clouds, right? Because I don't want to be logged into public cloud vendors. And we see two trends, right? One trend is the traditional hardware supplier of on-premise has not scaled to cloud technology in terms of big data analytics. They just missed the boat for that in the past, this is changing. You guys are a traditional player and changing this, so congratulations. The other thing, is there's been no innovation for the on-premise tech stack, right? The only technology stack to run modern application has been invested for a long time in the cloud. So what we see since two, three years, right? With the first one being Google with Kubernetes, that are good at GKE on-premise, then onto us, right? Bringing their tech stack with compromises to on-premises, right? Acknowledging exactly what we're talking about, the data is everywhere, data is important. Data gravity is there, right? It's just the network's fault, where the networks are too slow, right? If you could just move everything anywhere we want like juggling two balls, then we'd be in different place. But that's the not enough investment for the traditional IT players for that stack, and the modern stack being there. And now every public cloud player has an on-premise offering with different flavors, different capabilities. >> I want to give you guys Dave's story of kind of history and you can kind of course correct, and tell me how this, Matt, maybe fits into what's happened with customers. So, you know, before Hadoop, obviously you had to buy a big Oracle database and you know, you running Unix, and you buy some big storage subsystem if you had any money left over, you know, you maybe, you know, do some actual analytics. But then Hadoop comes in, lowers the cost, and then S3 kneecaps the entire Hadoop market, right? >> I wouldn't say that, I wouldn't agree. Sorry to jump on your history. Because the fascinating thing, what Hadoop brought to the enterprise for the first time, you're absolutely right, affordable, right, to do that. But it's not only about affordability because S3 as the affordability. The big thing is you can store information without knowing how to analyze it, right? So, you mentioned Snowflake, right? Before, it was like an Oracle database. It was Starschema for data warehouse, and so on. You had to make decisions how to store that data because compute capabilities, storage capabilities, were too limited, right? That's what Hadoop blew away. >> I agree, no schema on, right. But then that created data lakes, which create a data swamps, and that whole mess, and then Spark comes in and help clean it out, okay, fine. So, we're cool with that. But the early days of Hadoop, you had, companies would have a Hadoop monolith, they probably had their data catalog in Excel or Google sheets, right? And so now, my question to you, Matt, is there's a lot of customers that are still in that world. What do they do? They got an option to go to the cloud. I'm hearing that you're giving them another option? >> That's right. So we know that data is going to move to the cloud, as I mentioned. So let's keep that data in sync, and governed, and secured, like you expect. But for the data that can't move, let's bring those cloud native services to your data center. And so that's a big part of this announcement is this unified analytics. So that you can continue to run the tools that you want to today while bringing those next generation tools based on Apache Spark, using libraries like Delta Lake so you can go anything from Tableaux through Presto sequel, to advance machine learning in your Jupiter notebooks on-premises where you know your data is secured. And if it happens to sit in existing Hadoop data lake, that's fine too. We don't want our customers to have to make that trade off as they go from one to the other. Let's give you the best of both worlds, or as they say, you can eat your cake and have it too. >> Okay, so. Now let's talk about sort of developers on-prem, right? They've been kind of... If they really wanted to go cloud native, they had to go to the cloud. Do you feel like this changes the game? Do on-prem developers, do they want that capability? Will they lean into that capability? Or will they say no, no, the cloud is cool. What's your take? >> I love developers, right? But it's about who makes the decision, who pays the developers, right? So the CXOs in the enterprises, they need exactly, this is why we call the next-gen computing platform, that you can move your code assets. It's very hard to build software, so it's very valuable to an enterprise. I don't want to have limited to one single location or certain computing infrastructure, right? Luckily, we have Kubernetes to be able to move that, but I want to be able to deploy it on-premise if I have to. I want to deploy it, would be able to deploy in the multiple clouds which are available. And that's the key part. And that makes developers happy too, because the code you write has got to run multiple places. So you can build more code, better code, instead of building the same thing multiple places, because a little compiler change here, a little compiler change there. Nobody wants to do portability testing and rewriting, recertified for certain platforms. >> The head of application development or application architecture and the business are ultimately going to dictate that, number one. Number two, you're saying that developers shouldn't care because it can write once, run anywhere. >> That is the promise, and that's the interesting thing which is available now, 'cause people know, thanks to Kubernetes as a container platform and the abstraction which containers provide, and that makes everybody's life easier. But it goes much more higher than the Head of Apps, right? This is the digital transformation strategy, the next generation application the company has to build as a response to a pandemic, as a pivot, as digital transformation, as digital disruption capability. >> I mean, I see a lot of organizations basically modernizing by building some kind of abstraction to their backend systems, modernizing it through cloud native, and then saying, hey, as you were saying Holger, run it anywhere you want, or connect to those cloud apps, or connect across clouds, connect to other on-prem apps, and eventually out to the edge. Is that what you see? >> It's so much easier said than done though. Organizations have struggled so much with this, especially as we start talking about those data intensive app and workloads. Kubernetes and Hadoop? Up until now, organizations haven't been able to deploy those services. So, what we're offering as part of these GreenLake unified analytics services, a Kubernetes runtime. It's not ours. It's top of branch open source. And open source operators like Apache Spark, bringing in Delta Lake libraries, so that if your developer does want to use cloud native tools to build those next generation advanced analytics applications, but prod is still on-premises, they should just be able to pick that code up, and because we are deploying 100% open-source frameworks, the code should run as is. >> So, it seems like the strategy is to basically build, now that's what GreenLake is, right? It's a cloud. It's like, hey, here's your options, use whatever you want. >> Well, and it's your cloud. That's, what's so important about GreenLake, is it's your cloud, in your data center or co-lo, with your data, your tools, and your code. And again, we know that organizations are going to go to a multi or hybrid cloud location and through our management capabilities, we can reach out if you don't want us to control those, not necessarily, that's okay, but we should at least be able to monitor and audit the data that sits in those other locations, the applications that are running, maybe I register your GKE cluster. I don't manage it, but at least through a central pane of glass, I can tell the Head of Applications, what that person's utilization is across these environments. >> You know, and you said something, Matt, that struck, resonated with me, which is this is not trivial. I mean, not as simple to do. I mean what you see, you see a lot of customers or companies, what they're doing, vendors, they'll wrap their stack in Kubernetes, shove it in the cloud, it's essentially hosted stack, right? And, you're kind of taking a different approach. You're saying, hey, we're essentially building a cloud that's going to connect all these estates. And the key is you're going to have to keep, and you are, I think that's probably part of the reason why we're here, announcing stuff very quickly. A lot of innovation has to come out to satisfy that demand that you're essentially talking about. >> Because we've oversimplified things with containers, right? Because containers don't have what matters for data, and what matters for enterprise, which is persistence, right? I have to be able to turn my systems down, or I don't know when I'm going to use that data, but it has to stay there. And that's not solved in the container world by itself. And that's what's coming now, the heavy lifting is done by people like HPE, to provide that persistence of the data across the different deployment platforms. And then, there's just a need to modernize my on-premise platforms. Right? I can't run on a server which is two, three years old, right? It's no longer safe, it doesn't have trusted identity, all the good stuff that you need these days, right? It cannot be operated remotely, or whatever happens there, where there's two, three years, is long enough for a server to have run their course, right? >> Well you're a software guy, you hate hardware anyway, so just abstract that hardware complexity away from you. >> Hardware is the necessary evil, right? It's like TSA. I want to go somewhere, but I have to go through TSA. >> But that's a key point, let me buy a service, if I need compute, give it to me. And if I don't, I don't want to hear about it, right? And that's kind of the direction that you're headed. >> That's right. >> Holger: That's what you're offering. >> That's right, and specifically the services. So GreenLake's been offering infrastructure, virtual machines, IaaS, as a service. And we want to stop talking about that underlying capability because it's a dial tone now. What organizations and these developers want is the service. Give me a service or a function, like I get in the cloud, but I need to get going today. I need it within my security parameters, access to my data, my tools, so I can get going as quickly as possible. And then beyond that, we're going to give you that cloud billing practices. Because, just because you're deploying a cloud native service, if you're still still being deployed via CapEx, you're not solving a lot of problems. So we also need to have that cloud billing model. >> Great. Well Holger, we'll give you the last word, bring us home. >> It's very interesting to have the cloud qualities of subscription-based pricing maintained by HPE as the cloud vendor from somewhere else. And that gives you that flexibility. And that's very important because data is essential to enterprise processes. And there's three reasons why data doesn't go to the cloud, right? We know that. It's privacy residency requirement, there is no cloud infrastructure in the country. It's performance, because network latency plays a role, right? Especially for critical appraisal. And then there's not invented here, right? Remember Charles Phillips saying how old the CIO is? I know if they're going to go to the cloud or not, right? So, it was not invented here. These are the things which keep data on-premise. You know that load, and HP is coming on with a very interesting offering. >> It's physics, it's laws, it's politics, and sometimes it's cost, right? Sometimes it's too expensive to move and migrate. Guys, thanks so much. Great to see you both. >> Matt: Dave, it's always a pleasure. All right, and thank you for watching the Cubes continuous coverage of HPE's big GreenLake announcements. Keep it right there for more great content. (calm music begins)

Published Date : Sep 28 2021

SUMMARY :

And Matt Maccaux is the global field CTO I call it the force marks to digital. So data is main center to everything. 'Cause with oil, you can only use once. So to get to the data, you And so Matt, what are you I have to go to the cloud. capability to connect them, It's the same thing. You're going to unify that, and what, We need to be able to know So it's my choice? It's not always going to be in sync but that's the north star. I need to put it into an object store, But what do you see as for that in the past, I want to give you guys Sorry to jump on your history. And so now, my question to you, Matt, And if it happens to sit in they had to go to the cloud. because the code you write has and the business the company has to build as and eventually out to the edge. to pick that code up, So, it seems like the and audit the data that sits to have to keep, and you are, I have to be able to turn my systems down, guy, you hate hardware anyway, I have to go through TSA. And that's kind of the but I need to get going today. the last word, bring us home. I know if they're going to go Great to see you both. the Cubes continuous coverage

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

Frank SlootmanPERSON

0.99+

MattPERSON

0.99+

Matt MaccauxPERSON

0.99+

HolgerPERSON

0.99+

DavePERSON

0.99+

Holger MuellerPERSON

0.99+

twoQUANTITY

0.99+

100%QUANTITY

0.99+

Charles PhillipsPERSON

0.99+

Constellation ResearchORGANIZATION

0.99+

HPEORGANIZATION

0.99+

ExcelTITLE

0.99+

HPORGANIZATION

0.99+

todayDATE

0.99+

three yearsQUANTITY

0.99+

GreenLakeORGANIZATION

0.99+

three reasonsQUANTITY

0.99+

TodayDATE

0.99+

GoogleORGANIZATION

0.99+

two ballsQUANTITY

0.98+

firstQUANTITY

0.98+

OracleORGANIZATION

0.98+

10 years agoDATE

0.98+

EzmeralORGANIZATION

0.98+

both worldsQUANTITY

0.98+

first timeQUANTITY

0.98+

S3TITLE

0.98+

One trendQUANTITY

0.98+

GreenLake ServicesORGANIZATION

0.98+

first oneQUANTITY

0.98+

SnowflakeTITLE

0.97+

both placesQUANTITY

0.97+

KubernetesTITLE

0.97+

onceQUANTITY

0.96+

bothQUANTITY

0.96+

two trendsQUANTITY

0.96+

Delta LakeTITLE

0.95+

GoogleTITLE

0.94+

HadoopTITLE

0.94+

CapExORGANIZATION

0.93+

TableauxTITLE

0.93+

AzureTITLE

0.92+

GKEORGANIZATION

0.92+

CubesORGANIZATION

0.92+

UnixTITLE

0.92+

one single locationQUANTITY

0.91+

single viewQUANTITY

0.9+

SparkTITLE

0.86+

ApacheORGANIZATION

0.85+

pandemicEVENT

0.82+

HadoopORGANIZATION

0.81+

three years oldQUANTITY

0.8+

singleQUANTITY

0.8+

KubernetesORGANIZATION

0.74+

big wavesEVENT

0.73+

Apache SparkORGANIZATION

0.71+

Number twoQUANTITY

0.69+

Maurizio Davini, University of Pisa and Kaushik Ghosh, Dell Technologies | CUBE Conversation 2021


 

>>Hi, Lisa Martin here with the cube. You're watching our coverage of Dell technologies world. The digital virtual experience. I've got two guests with me here today. We're going to be talking about the university of Piza and how it is leaning into all flash data lakes powered by Dell technologies. One of our alumni is back MERITO, Debbie, and the CTO of the university of PISA. Maricio welcome back to the cube. Thank you. Very excited to talk to you today. CAUTI Gosha is here as well. The director of product management at Dell technologies. Kaushik. Welcome to the cube. Thank you. So here we are at this virtual event again, Maricio you were last on the cube at VMworld a few months ago, the virtual experience as well, but talk to her audience a little bit before we dig into the technology and some of these demanding workloads that the university is utilizing. Talk to me a little bit about your role as CTO and about the university. >>So my role as CTO at university of PISA is, uh, uh, regarding the, uh, data center operations and, uh, scientific computing support for these, the main, uh, occupation that, uh, that, uh, yeah. Then they support the world, saw the technological choices that university of PISA is, uh, is doing, uh, during the latest, uh, two or three years. >>Talk to me about some, so this is a, in terms of students we're talking about 50,000 or so students 3000 faculty and the campus is distributed around the town of PISA, is that correct? Maricio >>Uh, the university of PISA is sort of a, uh, town campus in the sense that we have 20 departments that are, uh, located inside the immediate eval town, uh, but due to the choices, but university of peace, I S uh, the, uh, last, uh, uh, nineties, uh, we are, uh, owner of, uh, of a private fiber network connecting all our, uh, departments and allow the templates. And so we can use the town as a sort of white board to design, uh, uh, new services, a new kind of support for teaching. Uh, and, uh, and so, >>So you've really modernized the data infrastructure for the university that was founded in the middle ages. Talk to me now about some of the workloads and that are generating massive amounts of data, and then we'll get into what you're doing with Dell technologies. >>Oh, so the university of PISA as a, uh, quite old on HPC, traditional HPC. So we S we are supporting, uh, uh, the traditional workloads from, uh, um, CAE or engineering or chemistry or oil and gas simulations. Uh, of course it during, uh, uh, the pandemic year, last year, especially, uh, we have new, uh, kind of work you'll scan, uh, summer related, uh, to the, uh, fast movement of the HPC workload from let's say, traditional HPC to AI and machine learning. And those are the, um, request that you support a lot of remote activities coming from, uh, uh, uh, distance learning, uh, to remote ties, uh, uh, laboratories or stations or whatever, most elder in presence in the past. And so the impact either on the infrastructure or, and the specialty and the storage part was a significant. >>So you talked about utilizing the high performance computing environments for awhile and for scientific computing and things. I saw a case study that you guys have done with Dell, but then during the pandemic, the challenge and the use case of remote learning brought additional challenges to your environment from that perspective, how, how were you able to transfer your curriculum to online and enable the scientists, the physicists that oil and gas folks doing research to still access that data at the speed that they needed to, >>Uh, you know, for what you got, uh, uh, uh, distance learning? Of course. So we were, uh, based on the cloud services were not provided internally by Yas. So we lie, we based on Microsoft services, so Google services and so on, but what regards, uh, internal support, uh, scientific computing was completely, uh, remote dies either on support or experience, uh, because, uh, I can, uh, I, can I, uh, bring some, uh, some examples, uh, for example, um, laboratory activities, uh, we are, the access to the laboratories, uh, was the of them, uh, as much as possible. Uh, we design a special networker to connect all the and to give the researcher the possibility of accessing the data on visit special network. So as sort of a collector of data, uh, inside our, our university network, uh, you can imagine that the, uh, for example, was, was a key factor for us because utilization was, uh, uh, for us, uh, and flexible way to deliver new services, uh, in an easy way, uh, especially if you have to, uh, have systems for remote. So, as, as I told you before about the, uh, network, as well as a white board, but also the computer infrastructure, it was VM-ware visualization and treated as a, as a sort of what we were designing with services either, either for interactive services or especially for, uh, scientific computing. For example, we have an experience with it and a good polarization of HPC workload. We start agents >>Talk to me about the storage impact, because as we know, we talk about, you know, these very demanding, unstructured workloads, AI machine learning, and that can be, those are difficult for most storage systems to handle the radio. Talk to us about why you leaned into all flash with Dell technologies and talk to us a little bit about the technologies that you've implemented. >>So, uh, if I, if I have to think about our, our storage infrastructure before the pandemic, I have to think about Iceland because our HPC workloads Moss, uh, mainly based off, uh, Isilon, uh, as a storage infrastructure, uh, together with some, uh, final defense system, as you can imagine, we were deploying in-house, uh, duty independently, especially with the explosion of the AI, with them, uh, blueprint of the storage requests change the law because of what we have, uh, uh, deal dens. And in our case, it was an, I breathed the Isilon solution didn't fit so well for HB for AI. And this is why we, uh, start with the data migration. That was, it was not really migration, but the sort of integration of the power scaler or flash machine inside our, uh, environment, because then the power scale, all flesh and especially, uh, IO in the future, uh, the MVME support, uh, is a key factor for the storage. It just support, uh, we already have experience as some of the, uh, NBME, uh, possibilities, uh, on the power PowerMax so that we have here, uh, that we use part for VDI support, uh, but off, um, or fleshly is the minimum land and EME, uh, is what we need to. >>Gotcha. Talk to me about what Dell technologies has seen the uptick in the demand for this, uh, as Maricio said, they were using Isilon before adding in power scale. What are some of the changing demands that, that Dell technologies has seen and how does technologies like how our scale and the F 900 facilitate these organizations being able to rapidly change their environment so that they can utilize and extract the value from data? >>Yeah, no, absolutely. What occupational intelligence is an area that, uh, continues to amaze me. And, uh, personally I think the, the potential here is immense. Um, uh, as Maurizio said, right, um, the, the data sets, uh, with artificial intelligence, I have, uh, grown significantly and, and not only the data has become, um, uh, become larger the models, the AI models that, that we, that are used have become more complex. Uh, for example, uh, one of the studies suggests that, uh, the, uh, that for a modeling of, uh, natural language processing, um, uh, one of the fields in AI, uh, the number of parameters used, could exceed like about a trillion in, uh, in a few years, right? So almost a size of a human brain. So, so not only that means that there's a lot of fear mounted to be, uh, data, to be processed, but, uh, by, uh, the process stored in yesterday, uh, but probably has to be done in the same amount of Dinah's before, perhaps even a smaller amount of time, right? So a larger data theme time, or perhaps even a smaller amount of time. So, absolutely. I agree. I mean, those type of, for these types of workloads, you need a storage that gives you that high-performance access, but also being able to store the store, that data is economically. >>And how does Dell technologies deliver that? The ability to scale the economics what's unique and differentiated about power skill? >>Uh, so power scale is, is, is our all flash, uh, system it's, uh, it's, uh, it's bad users, dark techno does some of the same capabilities that, uh, Isilon, um, products use used to offer, uh, one of his fault system capabilities, some of the capabilities that Maurizio has used and loved in the past, some of those, some of those same capabilities are brought forward. Now on this spar scale platform, um, there are some changes, like for example, on new Parscale's platform supports Nvidia GPU direct, right? So for, uh, artificial intelligence, uh, workloads, you do need these GPU capable machines. And, uh, and, uh, Parscale supports that those, uh, high high-performance Jupiter rec machines, uh, through, through the different technologies that we offer. And, um, the Parscale F 900, which should, which we are going to launch very soon, um, um, is, is, is our best hype, highest performance all-flash and the most economic allowed slash uh, to date. So, um, so it is, um, it not only is our fastest, but also offers, uh, the most economic, uh, most economical way of storing the data. Um, so, so ideal far for these type of high-performance workloads, like AIML, deep learning and so on. Excellent. >>So talk to me about some of the results that the university is achieving so far. I did read a three X improvement in IO performance. You were able to get nearly a hundred percent of the curriculum online pretty quickly, but talk to me about some of the other impacts that Dell technologies has helping the university to achieve. >>Oh, we had, uh, we had an old, uh, in all the Dell customer, and if you, uh, give a Luca walk, we have that inside the insomnia, our data centers. Uh, we typically joking, we define them as a sort of, uh, Dell technologies supermarket in the sense that, uh, uh, degreed part of our, our servers storage environment comes from, uh, from that technology said several generations of, uh, uh, PowerEdge servers, uh, uh, power, my ex, uh, Isaac along, uh, powers, Gale power store. So we, uh, we are, uh, um, using a lot of, uh, uh, Dell technologies here, here, and of course, uh, um, in the past, uh, our traditional, uh, workloads were well supported by that technologies. And, uh, Dell technologies is, uh, uh, driving ourselves versus, uh, the, what we call the next generation workloads, uh, because we are, uh, uh, combining gas, uh, in, um, in the transition of, uh, um, uh, the next generation of computing there, but to be OPA who, uh, to ask here, and he was walked through our research of looking for, cause if I, if I have to, to, to, to give a look to what we are, uh, doing, uh, mostly here, healthcare workloads, uh, deep learning, uh, uh, data analysis, uh, uh, image analysis in C major extraction that everything have be supported, especially from, uh, the next next generation servers typically keep the, uh, with, with GPU's. >>This is why GPU activities is, is so important for answer, but also, uh, supported on the, on the, on the networking side. But because of that, the, the, the speed, the, and the, of the storage, and must be tired to the next generation networking. Uh, low-latency high-performance because at the end of the day, you have to, uh, to bring the data in storage and DP. Can you do it? Uh, so, uh, they're, uh, one of the low latency, uh, uh, I performance, if they're connected zones is also a side effect of these new work. And of course that the college is, is, is. >>I love how you described your data centers as a Dell technologies supermarket, maybe a different way of talking about a center of excellence question. I want to ask you about, I know that the university of PISA is SCOE for Dell. Talk to me about in the last couple of minutes we have here, what that entails and how Dell helps customers become a center of excellence. >>Yeah, so Dell, um, like talked about has a lot of the Dell Dell products, uh, today, and, and, and in fact, he mentioned about the pirate servers, the power scale F 900 is, is actually based on a forehead server. So, so you can see, so a lot of these technologies are sort of in the linked with each other, they talk to each other, they will work together. Um, and, and, and that sort of helps, helps customers manage the entire, uh, ecosystem lifecycle data, life cycle together, versus as piece parts, because we have solutions that solve all aspects of, of, of the, uh, of, of, uh, of our customer like Mauricio's needs. Right. So, um, so yeah, I'm glad Maurizio is, is leveraging Dell and, um, and I'm happy we are able to help help more issue or solve solve, because, uh, all his use cases, uh, and UN >>Excellent. Maricio last question. Are you going to be using AI machine learning, powered by Dell to determine if the tower of PISA is going to continue to lean, or if it's going to stay where it is? >>Uh, the, the, the leaning tower is, uh, an engineering miracle. Uh, some years ago, uh, an engineering, uh, incredible worker, uh, was able, uh, uh, to fix them. They leaning for a while and let's open up the tower visa, stay there because he will be one of our, uh, beauty that you can come to to visit. >>And that's one part of Italy I haven't been to. So as pandemic, I gotta add that to my travel plans, MERITO and Kaushik. It's been a pleasure talking to you about how Dell is partnering with the university of PISA to really help you power AI machine learning workloads, to facilitate many use cases. We are looking forward to hearing what's next. Thanks for joining me this morning. Thank you for my guests. I'm Lisa Martin. You're watching the cubes coverage of Dell technologies world. The digital event experience.

Published Date : Jun 9 2021

SUMMARY :

We're going to be talking about the university of Piza and how it is leaning into all flash data uh, scientific computing support for these, the main, uh, uh, uh, nineties, uh, we are, uh, Talk to me now about some of the workloads and that are generating massive amounts of data, a lot of remote activities coming from, uh, uh, scientists, the physicists that oil and gas folks doing research to still access that data at the speed that the access to the laboratories, uh, was the of them, uh, Talk to me about the storage impact, because as we know, we talk about, you know, these very demanding, unstructured workloads, uh, Isilon, uh, as a storage infrastructure, uh, together with for this, uh, as Maricio said, they were using Isilon before adding in power that means that there's a lot of fear mounted to be, uh, data, to be processed, but, and the most economic allowed slash uh, to date. a hundred percent of the curriculum online pretty quickly, but talk to me about some of the other impacts the sense that, uh, uh, degreed part of our, they're, uh, one of the low latency, uh, uh, I know that the university of PISA is SCOE for Dell. a lot of the Dell Dell products, uh, today, and, and, if the tower of PISA is going to continue to lean, or if it's going to stay where it is? Uh, the, the, the leaning tower is, uh, an engineering miracle. So as pandemic, I gotta add that to my travel plans,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Lisa MartinPERSON

0.99+

MaurizioPERSON

0.99+

MERITOPERSON

0.99+

Maurizio DaviniPERSON

0.99+

MaricioPERSON

0.99+

DebbiePERSON

0.99+

DellORGANIZATION

0.99+

twoQUANTITY

0.99+

MicrosoftORGANIZATION

0.99+

University of PisaORGANIZATION

0.99+

20 departmentsQUANTITY

0.99+

GoogleORGANIZATION

0.99+

NvidiaORGANIZATION

0.99+

two guestsQUANTITY

0.99+

ItalyLOCATION

0.99+

KaushikPERSON

0.99+

PISAORGANIZATION

0.99+

three yearsQUANTITY

0.99+

CAUTI GoshaPERSON

0.99+

last yearDATE

0.99+

OneQUANTITY

0.99+

oneQUANTITY

0.99+

F 900COMMERCIAL_ITEM

0.98+

todayDATE

0.98+

MauricioPERSON

0.98+

yesterdayDATE

0.98+

pandemicEVENT

0.98+

3000 facultyQUANTITY

0.98+

about a trillionQUANTITY

0.97+

IsilonORGANIZATION

0.96+

Dell TechnologiesORGANIZATION

0.96+

SCOEORGANIZATION

0.96+

ParscaleORGANIZATION

0.96+

YasORGANIZATION

0.95+

IcelandLOCATION

0.94+

about 50,000QUANTITY

0.94+

ninetiesQUANTITY

0.93+

VMworldORGANIZATION

0.91+

MossORGANIZATION

0.89+

one partQUANTITY

0.88+

JupiterORGANIZATION

0.87+

Kaushik GhoshPERSON

0.87+

CTOPERSON

0.85+

this morningDATE

0.84+

few months agoDATE

0.8+

Gale power storeORGANIZATION

0.79+

hundred percentQUANTITY

0.76+

university of PizaORGANIZATION

0.75+

some years agoDATE

0.75+

university of PISAORGANIZATION

0.71+

A Day in the Life of an IT Admin | HPE Ezmeral Day 2021


 

>>Hi, everyone. Welcome to ASML day. My name is Yasmin Joffey. I'm the director of systems engineering for ASML at HPE. Today. We're here and joined by my colleague, Don wake, who is a technical marketing engineer who will talk to us about the date and the life of an it administrator through the lens of ASML container platform. We'll be answering your questions real time. So if you have any questions, please feel free to put your questions in the chat, and we should have some time at the end for some live Q and a. Don wants to go ahead and kick us off. >>All right. Thanks a lot, Yasir. Yeah, my name is Don wake. I'm the tech marketing guy and welcome to asthma all day, day in the life of an it admin and happy St. Patrick's day. At the same time, I hope you're wearing green virtual pinch. If you're not wearing green, don't have to look that up if you don't know what I'm scouting. So we're just going to go through some quick things. Talk about discussion of modern business. It needs to kind of set the stage and go right into a demo. Um, so what is the need here that we're trying to fulfill with, uh, ASML container platform? It's, it's all rooted in analytics. Um, modern businesses are driven by data. Um, they are also application centric and the separation of applications and data has never been more important or, or the relationship between the two applications are very data hungry. >>These days, they consume data in all new ways. The applications themselves are, are virtualized, containerized, and distributed everywhere, and optimizing every decision and every application is, is become a huge problem to tackle for every enterprise. Um, so we look at, um, for example, data science, um, as one big use case here, um, and it's, it's really a team sport and I'm today wearing the hat of perhaps, you know, operations team, maybe software engineer, guy working on, you know, continuous integration, continuous development integration with source control, and I'm supporting these data scientists, data analysts. And I also have some resource control. I can decide whether or not the data science team gets a, a particular cluster of compute and storage so that they can do their work. So this is the solution that I've been given as an it admin, and that is the ASML container platform. >>And just walking through this real quick, at the top, I'm trying to, as wherever possible, not get involved in these guys' lives. So the data engineers, scientists, app developers, dev ops guys, they all have particular needs and they can access their resources and spin up clusters, or just do work with the Jupiter notebook or run spark or Kafka or any of the, you know, popular analytics platforms by just getting in points that we can provide to them web URLs and their self service. But in the backend, I can then as the it guy makes sure the Kubernetes clusters are up and running, I can assign particular access to particular roles. I can make sure the data's well protected and I can connect them. I can import clusters from public clouds. I can, uh, you know, put my like clusters on premise if I want to. >>And I can do all this through this centralized control plane. So today I'm just going to show you I'm supporting some data scientists. So one of our very own guys is actually doing a demo right now as well, called the a day in the life of the data scientist. And he's on the opposite side, not caring about all the stuff I'm doing in the backend and he's training models and registering the models and working with data, uh, inside his, you know, Jupiter notebook, running inferences, running postman scripts. And so I'm in the background here, making sure that he's got access to his cluster storage protected, make sure it's, um, you know, his training models are up, he's got service endpoints, connecting him to, um, you know, his source control and making sure he's got access to all that stuff. So he's got like a taxi ride prediction model that he's working on and he has a Jupiter notebook and models. So why don't we, um, get hands on and I'll just jump right over it. >>It was no container platform. So this is a web UI. So this is the interface into the container platform. Our centralized control plane, I'm using my active directory credentials to log in here. >>And >>When I log in, I've also been assigned a particular role, uh, with regard to how much of the resources I can access. Now, in my case, I'm a site admin you can see right up here in the upper right hand, I'm a site admin and I have access to lots and lots of resources. And the one I'm going to be focusing on today is a Kubernetes cluster. Um, so I have a cluster I can go in here and let's say, um, we have a new data scientists come on board one. I can give him his own resources so he can do whatever he wants, use some GPU's and not affect other clusters. Um, so we have all these other clusters already created here. You can see here that, um, this is a very busy, um, you know, production system. They've got some dev clusters over here. >>I see here, we have a production cluster. So he needs to produce something for data scientists to use. It has to be well protected and, and not be treated like a development resource. So under his production cluster, I decided to create a new Kubernetes cluster. And literally I just push a button, create Kubernetes cluster once I've done that. And I'll just show you some of the screens and this is a live environment. So this is, I could actually do it all my hosts are used up right now, but I wouldn't be able to go in here and give it a name, just select, um, some hosts to use as the primary master controller and some workers answer a few more questions. And then once that's done, I have now created a special, a whole nother Kubernetes cluster, um, that I could also create tenants from. >>So tenants are really Kubernetes. Uh namespaces so in addition to taking hosts and Kubernetes clusters, I can also go to that, uh, to existing clusters and now carve out a namespace from that. So I look at some of the clusters that were already created and, um, let's see, we've got, um, we've got this year is an example of a tenant that I could have created from that production cluster. And to do that here in the namespace, I just hit create and similar to how you create a cluster. You can now carve down from a given cluster and we'll say the production cluster and give it a name and a description. I can even tell it, I want this specific one to be an AI ML project, um, which really is our ML ops license. So at the end of the day, I can say, okay, I'm going to create an ML ops tenant from that cluster that I created. >>And so I've already created it here for this demo. And I'm going to just go into that Kubernetes namespace now that we also call it tenant. I mean, it's like, multitenancy the name essentially means we're carving out resources so that somebody can be isolated from another environment. First thing I typically do. Um, and at this point I could also give access to this tenant and only this tenant to my data scientist. So the first thing I typically do is I go in here and you can actually assign users right here. So right now it's just me. But if I want it to, for example, give this, um, to Terry, I could go in here and find another user and assign him from this lead, from this list, as long as he's got the proper credentials here. So you can see here, all these other users have active directory credentials, and they, uh, when we created the cluster itself, we also made sure it integrated with our active directory, so that only authorized users can get in there. >>Let's say the first thing I want to do is make sure when I do Jupiter notebook work, or when Terry does, I'm going to connect him up straight up to the get hub repository. So he gives me a link to get hub and says, Hey man, this is all of my cluster work that I've been doing. I've got my source control there. My scripts, my Python notebooks, my Jupiter notebooks. So when I create that, I simply give him, you know, he gives me his, I create a configuration. I say, okay, here's a, here's a get repo. Here's the link to it. I can use a token, here's his username. And I can now put in that token. So this is actually a private repo and using a token, you know, standard get interface. And then the cool thing after that, you can go in here and actually copy the authorization secret. >>And this gets into the Kubernetes world. Um, you know, if you want to make sure you have secure integration with things like your source control or perhaps your active directory, that's all maintained in secrets. So you can take that secret. And when I then create his notebook, I can put that secret right in here in this, uh, launch Yammel. And I say, Hey, connect this Jupiter notebook up with this secret so he can log in. And when I've launched this Jupiter notebook cluster, this is actually now, uh, within my, my, uh, Kubernetes tenant. It is now really a pod. And if I want to, I can go right into a terminal for that, uh, Kubernetes tenant and say, coop CTL, these are standard, you know, CNCF certified Kubernetes get pods. And when I do this, it'll tell me all of the active pods and within those positive containers that I'm running. >>So I'm running quite a few pods and containers here in this, uh, artificial intelligence machine learning, um, tenant. So that's kind of cool. Also, if I wanted to, I could go straight and I can download the config for Kubernetes, uh, control. Uh well, and then I can do something like this, where on my own system where I'm more comfortable, perhaps coop CTL get pods. So this is running on my laptop and I just had to do a coop CTL refresh and give the IP address and authorization, um, information in order to connect from my laptop to that end point. So from a CIC D perspective from, you know, an it admin guides, he usually wants to use tools right on his, uh, desktop. So here am I back in my web browser, I'm also here on the dashboard of this, uh, Kubernetes, um, tenant, and I can see how it's doing. >>It looks like it's kind of busy here. I can focus specifically on a pod if I want to. I happen to know this pod is my Jupiter notebook pod. So aren't, I show how, you know, I could enable my data scientists by just giving him the, uh, URL or what we call a notebook service end points or notebook end point. And just by clicking on this URL or copying it, copying, you know, it's a link, uh, and then emailing it to them and say, okay, here's your, uh, you know, here's your duper notebook. And I say, Hey, just log in with your credentials. I've already logged in. Um, and so then he's got his Jupiter notebook here and you can see that he's connected to his GitHub repo directly. He's got all of the files that he needs to run his data science project and within here, and this is really in the data science realm, data scientists realm. >>He can see that he can have access to centralized storage and he can copy the files from his GitHub repo to that centralized storage. And, you know, these, these commands, um, are kind of cool. They're a little Jupiter magic commands, and we've got some of our own that showed that attachment to the cluster. Um, but you can see here if you run these commands, they're actually looking at the shared project repository managed by the container platform. So, you know, just to show you that again, I'll go back to the container platform. And in fact, the data scientist, uh, could do the same thing. Attitude put a notebook back to platform. So here's this project repository. So this is other big point. So now putting on my storage admin hat, you know, I've got this shared, um, storage, um, volume that is managed for me by the ESMO data fabric. >>Um, in, in here, you can see that the data scientist, um, from his get repo is able to through Jupiter notebook directly, uh, copy his code. He was able to run as Jupiter notebook and create this XG boost, uh, model. So this file can then be registered in this AIML tenant. So he can go in here and register his model. So this is, you know, this is really where the data scientist guy can self-service kick off his notebooks, even get a deployment end point so that he can then inference his cluster. So here again, another URL that you could then take this and put it into like a postman rest URL and get answers. Um, but let's say he wants to, um, he's been doing all this work and I want to make sure that his, uh, data's protected, uh, how about creating a mirror. >>So if I want to create a mirror of that data, now I go back to this other, uh, and this is the, the, uh, data fabric embedded in a very special cluster called the Picasso cluster. And it's a version of the ASML data fabric that allows you to launch what was formerly called Matt bar as a Kubernetes cluster. And when you create this special cluster, every other cluster that you create is automatically, uh, gets things like that. Tenant storage. I showed you to create a shared workspace, and it's automatically managed by this, uh, data fabric. Uh, and you're even given an end point to go into the data fabric and then use all of the awesome features of ASML data fabric. So here I can just log in here. And now I'm at the, uh, data fabric, web UI to do some data protection and mirroring. >>So >>Let's go over here. Let's say I want to, uh, create a mirror of that tenant. So I forgot to note what the name of my tenant was. I'm going to go back to my tenant, the name of the volume that I'm playing with here. So in my AIML tenant, I'm going to go to my source, control my project repository that I want to protect. And I see that the ESMO data fabric has created 10 and 30 as a volume. So I'll go back to my, um, data fabric here, and I'm going to look for 10 and 30. And if I want to, I can go into tenant 30, >>Okay. >>Down here, I can look at the usage. I can look at all of the, you know, I've used very little of the, uh, allocated storage that I want, but let's, uh, you know what, let's go ahead and create a volume to mirror that one. So very simple web UI that has said create volume. I go in here and I say, I want to do a, a tenant 30 mirror. And I say, mirror the mirror volume. Um, I want to use my Picasso cluster. I want to use tenant 30. So now that's actually looking up in the data fabric, um, database there's 10 and 30 K. So it knows exactly which one I want to use. I can go in here and I can say, you know, ext HCP, tenant, 30 mirror, you know, I can give it whatever name I want and this path here. >>And that's a whole nother, uh, demo is this could be in Tokyo. This could be mirrored to all kinds of places all over the world, because this is truly a global name, split namespace, which is a huge differentiator for us in this case, I'm creating a local mirror and that can go down here and, um, I can add, uh, audit and encryptions. I can do, um, access control. I can, you know, change permissions, you know, so full service, um, interactivity here. And of course this is using the web UI, but there's also rest API interfaces as well. So that is pretty much the, the brunt of what I wanted to show you in the demo. Um, so we got hands on and I'm just going to throw this up real quick and then come back to Yasser. See if he's got any questions he has received from anybody watching, if you have any new questions. >>Yeah. We've got a few questions. Um, we can, uh, just take some time to go, hopefully answer a few. Um, so it, it does look like you can integrate or incorporate your existing get hub, uh, to be able to, um, extract, uh, shared code or repositories. Correct? >>Yeah. So we have that built in and can either be, um, get hub or bit bucket it's, you know, pretty standard interface. So just like you can go into any given, get hub and do a clone of a, of a repo, pull it into your local environment. We integrated that directly into the gooey so that you can, uh, say to your, um, AIML tenant, uh, to your Jupiter notebook. You know, here's, here's my GitHub repo. When you open up my notebook, just connect me straight up. So it saves you some, some steps there because Jupiter notebook is designed to be integrated with get hub. So we have get hub integrated in as well or bit bucket. Right. >>Um, another question around the file system, um, has the map, our file system that was carried over, been modified in any way to run on top of Kubernetes. >>So yeah, I would say that the map, our file system data fabric, what I showed here is the Kubernetes version of it. So it gives you a lot of the same features, but if you need, um, perhaps run it on bare metal, maybe you have performance, um, concerns, um, you know, you can, uh, you can also deploy it as a separate bare metal instance of data fabric, but this is just one way that you can, uh, use it integrated directly into Kubernetes depends really the needs of, of the, uh, the user and that a fabric has a lot of different capabilities, but this is, um, it has a lot of the core file system capabilities where you can do snapshots and mirrors, and it it's of course, striped across multiple, um, multiple disks and nodes. And, uh, you know, Matt BARR data fabric has been around for years. It's, uh, and it's designed for integration with these, uh, analytic type workloads. >>Great. Um, you showed us how you can manage, um, Kubernetes clusters through the ASML container platform you buy. Um, but the question is, can you, uh, control who accesses, which tenant, I guess, namespace that you created, um, and also can you restrict or, uh, inject resource limitations for each individual namespace through the UI? >>Oh yeah. So that's, that's a great question. Yes. To both of those. So, um, as a site admin, I had lots of authority to create clusters, to go into any cluster I wanted, but typically for like the data scientist example I used, I would give him, I would create a user for him. And there's a couple of ways you can create users. Um, and it's all role-based access control. So I could create a local user and have container platform authenticate him, or I can say integrate directly with, uh, active directory or LDAP, and then even including which groups he has access to. And then in the user interface for the site admin, I could say he gets access to this tenant and only this tenant. Um, another thing you asked about is his limitations. So when you create the tenant to prevent that noisy neighbor problem, you can, um, go in and create quotas. >>So I didn't show the process of actually creating a Quentin, a tenant, but integral to that, um, flow is okay, I've defined which cluster I want to use. I defined how much memory I want to use. So there's a quota right there. You could say, Hey, how many CPU's am I taking from this pool? And that's one of the cool things about the platform is that it abstracts all that away. You don't have to really know exactly which host, um, you know, you can create the cluster and select specific hosts, but once you've created the cluster, it's not just a big pool of resources. So you can say Bob, over here, um, he's only going to get 50 of the a hundred CPU's available and he's only going to get X amount of gigabytes of memory. And he's only going to get this much storage that he can consume. So you can then safely hand off something and know they're not going to take all the resources, especially the GPU's where those will be expensive. And you want to make sure that one person doesn't hog all the resources. And so that absolutely quotas are built in there. >>Fantastic. Well, we, I think we are out of time. Um, we have, uh, a list of other questions that we will absolutely reach out and, um, get all your questions answered, uh, for those of you who ask questions in the chat. Um, Don, thank you very much. Thanks everyone else for joining Don, will this recording be made available for those who couldn't make it today? >>I believe so. Honestly, I'm not sure what the process is, but, um, yeah, it's being recorded so they must've done that for a reason. >>Fantastic. Well, Don, thank you very much for your time and thank everyone else for joining. Thank you.

Published Date : Mar 17 2021

SUMMARY :

So if you have any questions, please feel free to put your questions in the chat, don't have to look that up if you don't know what I'm scouting. you know, continuous integration, continuous development integration with source control, and I'm supporting I can, uh, you know, And so I'm in the background here, making sure that he's got access to So this is a web UI. You can see here that, um, this is a very busy, um, you know, And I'll just show you some of the screens and this is a live environment. in the namespace, I just hit create and similar to how you create a cluster. So you can see here, all these other users have active I create that, I simply give him, you know, he gives me his, I create a configuration. So you can take that secret. So this is running on my laptop and I just had to do a coop CTL refresh And just by clicking on this URL or copying it, copying, you know, it's a link, So now putting on my storage admin hat, you know, I've got this shared, So here again, another URL that you could then take this and put it into like a postman rest URL And when you create this special cluster, every other cluster that you create is automatically, And I see that the ESMO data I can look at all of the, you know, I can, you know, change permissions, Um, so it, it does look like you can integrate So just like you can go into any given, Um, another question around the file system, um, has the it has a lot of the core file system capabilities where you can do snapshots and mirrors, and also can you restrict or, uh, inject resource limitations for each So when you create the tenant to prevent So I didn't show the process of actually creating a Quentin, a tenant, but integral to that, Um, Don, thank you very much. I believe so.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
YasirPERSON

0.99+

TerryPERSON

0.99+

Don wakePERSON

0.99+

TokyoLOCATION

0.99+

50QUANTITY

0.99+

Yasmin JoffeyPERSON

0.99+

FirstQUANTITY

0.99+

two applicationsQUANTITY

0.99+

DonPERSON

0.99+

TodayDATE

0.99+

todayDATE

0.99+

St. Patrick's dayEVENT

0.98+

10QUANTITY

0.98+

bothQUANTITY

0.98+

30 K.QUANTITY

0.98+

oneQUANTITY

0.98+

KubernetesTITLE

0.98+

HPEORGANIZATION

0.97+

one personQUANTITY

0.97+

first thingQUANTITY

0.97+

YasserPERSON

0.97+

KafkaTITLE

0.97+

PythonTITLE

0.96+

ASMLORGANIZATION

0.96+

CNCFORGANIZATION

0.96+

one wayQUANTITY

0.95+

JupiterLOCATION

0.94+

ESMOORGANIZATION

0.94+

GitHubORGANIZATION

0.94+

ASMLEVENT

0.93+

BobPERSON

0.93+

Matt BARRPERSON

0.92+

this yearDATE

0.91+

JupiterORGANIZATION

0.9+

each individualQUANTITY

0.86+

30OTHER

0.85+

a hundred CPUQUANTITY

0.82+

ASMLTITLE

0.82+

2021DATE

0.8+

coopORGANIZATION

0.78+

a dayQUANTITY

0.78+

KubernetesORGANIZATION

0.75+

coupleQUANTITY

0.75+

A Day in the LifeTITLE

0.73+

an ITTITLE

0.7+

30 mirrorQUANTITY

0.69+

caseQUANTITY

0.64+

CTLCOMMERCIAL_ITEM

0.57+

few more questionsQUANTITY

0.57+

coop CTLORGANIZATION

0.55+

yearsQUANTITY

0.55+

QuentinPERSON

0.51+

30QUANTITY

0.49+

Ezmeral DayPERSON

0.48+

lotsQUANTITY

0.43+

JupiterCOMMERCIAL_ITEM

0.42+

10TITLE

0.41+

PicassoORGANIZATION

0.38+

A Day in the Life of a Data Scientist


 

>>Hello, everyone. Welcome to the a day in the life of a data science talk. Uh, my name is Terry Chang. I'm a data scientist for the ASML container platform team. And with me, I have in the chat room, they will be moderating the chat. I have Matt MCO as well as Doug Tackett, and we're going to dive straight into kind of what we can do with the asthma container platform and how we can support the role of a data scientist. >>So just >>A quick agenda. So I'm going to do some introductions and kind of set the context of what we're going to talk about. And then we're actually going to dive straight into the ASML container platforms. So we're going to walk straight into what a data scientist will do, kind of a pretty much a day in the life of the data scientists. And then we'll have some question and answer. So big data has been the talk within the last few years within the last decade or so. And with big data, there's a lot of ways to derive meaning. And then a lot of businesses are trying to utilize their applications and trying to optimize every decision with their, uh, application utilizing data. So previously we had a lot of focus on data analytics, but recently we've seen a lot of data being used for machine learning. So trying to take any data that they can and send it off to the data scientists to start doing some modeling and trying to do some prediction. >>So that's kind of where we're seeing modern businesses rooted in analytics and data science in itself is a team sport. We're seeing that it doesn't, we need more than data scientists to do all this modeling. We need data engineers to take the data, massage the data and do kind of some data manipulation in order to get it right for the data scientists. We have data analysts who are monitoring the models, and we even have the data scientists themselves who are building and iterating through multiple different models until they find a one that is satisfactory to the business needs. Then once they're done, they can send it off to the software engineers who will actually build it out into their application, whether it's a mobile app or a web app. And then we have the operations team kind of assigning the resources and also monitoring it as well. >>So we're really seeing data science as a team sport, and it does require a lot of different expertise and here's the kind of basic machine learning pipeline that we see in the industry now. So, uh, at the top we have this training environment and this is, uh, an entire loop. Uh, we'll have some registration, we'll have some inferencing and at the center of all, this is all the data prep, as well as your repositories, such as for your data, for any of your GitHub repository, things of that sort. So we're kind of seeing the machine learning industry, go follow this very basic pattern and at a high level I'll glance through this very quickly, but this is kind of what the, uh, machine learning pipeline will look like on the ASML container platform. So at the top left, we'll have our, our project depository, which is our, uh, persistent storage. >>We'll have some training clusters, we'll have a notebook, we'll have an inference deployment engine and a rest API, which is all sitting on top of the Kubernetes cluster. And the benefit of the container platform is that this is all abstracted away from the data scientist. So I will actually go straight into that. So just to preface, before we go into the data as small container platform, where we're going to look at is a machine learning example, problem that is, uh, trying to predict how long a specific taxi ride will take. So with a Jupiter notebook, the data scientists can take all of this data. They can do their data manipulation, train a model on a specific set of features, such as the location of a taxi ride, the duration of a taxi ride, and then model it to trying to figure out, you know, what, what kind of prediction we can get on a future taxi ride. >>So that's the example that we will talk through today. I'm going to hop out of my slides and jump into my web browser. So let me zoom in on this. So here I have a Jupiter environment and, um, this is all running on the container platform. All I need is actually this link and I can access my environment. So as a data scientist, I can grab this link from my it admin or my system administrator. And I could quickly start iterating and, and start coding. So on the left-hand side of the Jupiter, we actually have a file directory structure. So this is already synced up to my get repository, which I will show in a little bit on the container platform so quickly I can pull any files that are on my get hub repository. I can even push with a button here, but I can, uh, open up this Python notebook. >>And with all this, uh, unique features of the Jupiter environment, I can start coding. So each of these cells can run Python code and in specific the container at the ESMO container platform team, we've actually built our own in-house lime magic commands. So these are unique commands, um, that we can use to interact with the underlying infrastructure of the container platform. So the first line magic command that I want to mention is this command called percent attachments. When I run this command, I'll actually get the available training clusters that I can send training jobs to. So this specific notebook, uh, it's pretty much been created for me to quickly iterate and develop a model very quickly. I don't have to use all the resources. I don't have to allocate a full set of GPU boxes onto my little Jupiter environment. So with the training cluster, I can attach these individual data science notebooks to those training clusters and the data scientists can actually utilize those resources as a shared environment. >>So the, essentially the shared large eight GPU box can actually be shared. They don't have to be allocated to a single data scientist moving on. We have another line magic command, it's called percent percent Python training. This is how we're going to utilize that training cluster. So I will prepare the cell percent percent with the name of the training cluster. And this is going to tell this notebook to send this entire training cell, to be trained on those resources on that training cluster. So the data scientists can quickly iterate through a model. They can then format that model and all that code into a large cell and send it off to that training cluster. So because of that training cluster is actually located somewhere else. It has no context of what has been done locally in this notebook. So we're going to have to do and copy everything into one large cell. >>So as you see here, I'm going to be importing some libraries and I'm in a, you know, start defining some helper functions. I'm going to read in my dataset and with the typical data science modeling life cycle, we're going to have to take in the data. We're going to have to do some data pre-processing. So maybe the data scientists will do this. Maybe the data engineer will do this, but they have access to that data. So I'm here. I'm actually getting there to be reading in the data from the project repository. And I'll talk about this a little bit later with all of the clusters within the container platform, we have access to some project repository that has been set up using the underlying data fabric. So with this, I have, uh, some data preprocessing, I'm going to cleanse some of my data that I noticed that maybe something is missing or, uh, some data doesn't look funky. >>Maybe the data types aren't correct. This will all happen here in these cells. So once that is done, I can print out that the data is done cleaning. I can start training my model. So here we have to split our data, set into a test, train, uh, data split so that we have some data for actually training the model and some data to test the model. So I can split my data there. I could create my XG boost object to start doing my training and XG boost is kind of like a decision tree machine learning algorithm, and I'm going to fit my data into this, uh, XG boost algorithm. And then I'm going to do some prediction. And then in addition, I'm actually going to be tracking some of the metrics and printing them out. So these are common metrics that we, that data scientists want to see when they do their training of the algorithm. >>Just to see if some of the accuracy is being improved, if the loss is being improved or the mean absolute error. So things like that. So these are all things, data scientists want to see. And at the end of this training job, I'm going to be saving the model. So I'm going to be saving it back into the project repository in which we will have access to. And at the end, I will print out the end time so I can execute that cell. And I've already executed that cell. So you'll see all of these print statements happening here. So importing the libraries, the training was run reading and data, et cetera. All of this has been printed out from that training job. Um, and in order to access that, uh, kind of glance through that, we would get an output with a unique history URL. >>So when we send the training job to that training cluster, we'll the training cluster will send back a unique URL in which we'll use the last line magic command that I want to talk about called percent logs. So percent logs will actually, uh, parse out that response from the training cluster. And actually we can track in real time what is happening in that training job so quickly, we can see that the data scientist has a sandbox environment available to them. They have access to their get repository. They have access to a project repository in which they can read in some of their data and save the model. So very quick interactive environment for the data scientists to do all of their work. And it's all provisioned on the ASML container platform. And it's also abstracted away. So here, um, I want to mention that again, this URL is being surfaced through the container platform. >>The data scientist doesn't have to interact with that at all, but let's take, it's take a step back. Uh, this is the day to day in the life of the data scientists. Now, if we go backwards into the container platform and we're going to walk through how it was all set up for them. So here is my login page to the container platform. I'm going to log in as my user, and this is going to bring me to the, uh, view of the, uh, Emma lops tenant within the container platform. So this is where everything has been set up for me, the data scientist doesn't have to see this if they don't need to, but what I'll walk through now is kind of the topics that I mentioned previously that we would go back into. So first is the project repository. So this project deposited comes with each tenant that is created on the platform. >>So this is a more, nothing more than a shared collaborative workspace environment in which data scientist or any data scientist who is allocated to this tenant. They have this politics client that can visually see all their data of all, all of their code. And this is actually taking a piece of the underlying data fabric and using that for your project depository. So you can see here, I have some code I can create and see my scoring script. I can see the models that have been created within this tenant. So it's pretty much a powerful tool in which you can store your code store any of your data and have the ability to read and write from any of your Jupiter environments or any of your created clusters within this tenant. So a very cool ad here in which you can, uh, quickly interact with your data. >>The next thing I want to show is the source control. So here is where you would plug in all of your information for your source control. And if I edit this, you guys will actually see all the information that I've passed in to configure the source control. So on the backend, the container platform will take these credentials and connect the Jupiter notebooks you create within this tenant to that get repository. So this is the information that I've passed in. If GitHub is not of interest, we also have support for bit bucket here as well. So next I want to show you guys that we do have these notebook environments. So, um, the notebook environment was created here and you can see that I have a notebook called Teri notebook, and this is all running on the Kubernetes environment within the container platform. So either the data scientists can come here and create their notebook or their project admin can create the notebook. >>And all you'd have to do is come here to this notebook end points. And this, the container platform will actually map the container platform to a specific port in which you can just give this link to the data scientists. And this link will actually bring them to their own Jupiter environment and they can start doing all of their model just as I showed in that previous Jupiter environment. Next I want to show the training cluster. This is the training cluster that was created in which I can attach my notebook to start utilizing those training clusters. And then the last thing I want to show is the model, the deployment cluster. So once that model has been saved, we have a model registry in which we can register the model into the platform. And then the last step is to create a deployment clusters. So here on my screen, I have a deployment cluster called taxi deployment. >>And then all these serving end points have been configured for me. And most importantly, this endpoint model. So the deployment cluster is actually a wrap the, uh, train model with the flask wrapper and add a rest endpoint to it so quickly. I can operationalize my model by taking this end point and creating a curl command, or even a post request. So here I have my trusty postman tool in which I can format a post request. So I've taken that end point from the container platform. I've formatted my body, uh, right here. So these are some of the features that I want to send to that model. And I want to know how long this specific taxi ride at this location at this time of day would take. So I can go ahead and send that request. And then quickly I will get an output of the ride. >>Duration will take about 2,600 seconds. So pretty much we've walked through how a data scientists can quickly interact with their notebook. They can train their model. And then coming into the platform, we saw the project repository, we saw the source control. We can register the model within the platform, and then quickly we can operationalize that model with our deployment cluster, uh, and have our model up and running and available for inference. So that wraps up the demo. Uh, I'm gonna pass it back to Doug and Matt and see if they want to come off mute and see if there are any questions, Matt, Doug, you there. Okay. >>Yeah. Hey, Hey Terry, sorry. Sorry. Just had some trouble getting off mute there. Uh, no, that was a, that was an excellent presentation. And I think there are generally some questions that come up when I talk to customers around how integrated into the Kubernetes ecosystem is this capability and where does this sort of Ezreal starts? And the open source, uh, technologies like, um, cube flow as an example, uh, begin. >>Yeah, sure. Matt. So this is kind of one layer up. We have our Emma LOBs tenant and this is all running on a piece of a Kubernetes cluster. So if I log back out and go into the site admin view, this is where you would see all the Kubernetes clusters being created. And it's actually all abstracted away from the data scientists. They don't have to know Kubernetes. They just interact with the platform if they want to. But here in the site admin view, I had this Kubernetes dashboard and here on the left-hand side, I have all my Kubernetes sections. So if I just add some compute hosts, whether they're VMs or cloud compute hosts, like ETQ hosts, we can have these, uh, resources abstracted away from us to then create a Kubernetes cluster. So moving on down, I have created this Kubernetes cluster utilizing those resources. >>Um, so if I go ahead and edit this cluster, you'll actually see that have these hosts, which is just a click and a click and drop method. I can move different hosts to then configure my Kubernetes cluster. Once my Kubernetes cluster is configured, I can then create Kubernetes tenant or in this case, it's a namespace. So once I have this namespace available, I can then go into that tenant. And as my user, I don't actually see that it is running on Kubernetes. So in addition with our ML ops tenants, you have the ability to bootstrap cute flow. So queue flow is a open source machine learning framework that is run on Kubernetes, and we have the ability to link that up as well. So, uh, coming back to my Emma lops tenant, I can log in what I showed is the ASML container platform version of Emma flops. But you see here, we've also integrated QP flow. So, uh, very, uh, a nod to, uh, HPS contribution to, you know, utilizing open source. Um, it's actually all configured within our platform. So, um, hopefully, >>Yeah, actually, Tara, can you hear me? It's Doug. So there were a couple of other questions actually about key flare that came in. I wonder whether you could just comment on why we've chosen cube flow. Cause I know there was a question about ML flow in stead and what the differences between ML flow and coop flow. >>Yeah, sure. So the, just to reiterate, there are some questions about QP flow and I'm just, >>Yeah, so obviously one of, uh, one of the people watching saw the queue flow dashboard there, I guess. Um, and so couldn't help but get excited about it. But there was another question about whether, you know, ML flow versus cube flow and what the difference was between them. >>Yeah. So with flow, it's, it's an open source framework that Google has developed. It's a very powerful framework that comes with a lot of other unique tools and Kubernetes. So with Q flow, you really have the ability to launch other notebooks. You have the ability to utilize different Kubernetes operators like TensorFlow and PI torch. You can utilize a lot of the, some of the frameworks within Q4 to do training like Q4 pipelines, which visually allow you to see your training jobs, uh, within the queue flow. It also has a plethora of different serving mechanisms, such as Seldin, uh, for, you know, deploying your, your machine learning models. You have Ks serving, you have TF serving. So Q4 is very, it's a very powerful tool for data scientists to utilize if they want a full end to end open source and know how to use Kubernetes. So it's just a, another way to do your machine learning model development and right with ML flow, it's actually a different piece of the machine learning pipeline. So ML flow mainly focuses on model experimentation, comparing different models, uh, during the training and it off it can be used with Q4. >>The complimentary Terry I think is what you're saying. Sorry. I know we are dramatically running out of time now. So that was really fantastic demo. Thank you very much, indeed. >>Exactly. Thank you. So yeah, I think that wraps it up. Um, one last thing I want to mention is there is this slide that I want to show in case you have any other questions, uh, you can visit hp.com/asml, hp.com/container platform. If you have any questions and that wraps it up. So thank you guys.

Published Date : Mar 17 2021

SUMMARY :

I'm a data scientist for the ASML container platform team. So I'm going to do some introductions and kind of set the context of what we're going to talk about. the models, and we even have the data scientists themselves who are building and iterating So at the top left, we'll have our, our project depository, which is our, And the benefit of the container platform is that this is all abstracted away from the data scientist. So that's the example that we will talk through today. So the first line magic command that I want to mention is this command called percent attachments. So the data scientists can quickly iterate through a model. So maybe the data scientists will do this. So once that is done, I can print out that the data is done cleaning. So I'm going to be saving it back into the project repository in which we will So here, um, I want to mention that again, this URL is being So here is my login page to the container So this is a more, nothing more than a shared collaborative workspace environment in So on the backend, the container platform will take these credentials and connect So once that model has been saved, we have a model registry in which we can register So I've taken that end point from the container platform. So that wraps up the demo. And the open source, uh, technologies like, um, cube flow as an example, So moving on down, I have created this Kubernetes cluster So once I have this namespace available, So there were a couple of other questions actually So the, just to reiterate, there are some questions about QP flow and I'm just, But there was another question about whether, you know, ML flow versus cube flow and So with Q flow, you really have the ability to launch So that was really fantastic demo. So thank you guys.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DougPERSON

0.99+

Doug TackettPERSON

0.99+

Terry ChangPERSON

0.99+

TerryPERSON

0.99+

TaraPERSON

0.99+

MattPERSON

0.99+

PythonTITLE

0.99+

GoogleORGANIZATION

0.99+

Matt MCOPERSON

0.99+

JupiterLOCATION

0.99+

KubernetesTITLE

0.99+

first lineQUANTITY

0.98+

eachQUANTITY

0.98+

GitHubORGANIZATION

0.98+

todayDATE

0.98+

firstQUANTITY

0.98+

about 2,600 secondsQUANTITY

0.97+

Q4TITLE

0.97+

A Day in the Life of a Data ScientistTITLE

0.97+

hp.com/asmlOTHER

0.97+

last decadeDATE

0.97+

one layerQUANTITY

0.95+

hp.com/containerOTHER

0.92+

single dataQUANTITY

0.91+

EmmaPERSON

0.91+

one large cellQUANTITY

0.91+

each tenantQUANTITY

0.88+

oneQUANTITY

0.84+

one last thingQUANTITY

0.81+

Q flowTITLE

0.8+

EmmaTITLE

0.8+

ESMOORGANIZATION

0.76+

last few yearsDATE

0.74+

one ofQUANTITY

0.73+

dayQUANTITY

0.72+

eight GPUQUANTITY

0.7+

SeldinTITLE

0.69+

Q4DATE

0.67+

percent percentOTHER

0.65+

EzrealORGANIZATION

0.65+

some questionsQUANTITY

0.65+

ASMLTITLE

0.65+

ASMLORGANIZATION

0.61+

peopleQUANTITY

0.49+

ETQTITLE

0.46+

TeriORGANIZATION

0.4+

EmmaORGANIZATION

0.35+

Io-Tahoe Episode 5: Enterprise Digital Resilience on Hybrid and Multicloud


 

>>from around the globe. It's the Cube presenting enterprise. Digital resilience on hybrid and multi cloud Brought to You by Iota Ho. Hello, everyone, and welcome to our continuing Siri's covering data automation brought to you by Io Tahoe. Today we're gonna look at how to ensure enterprise resilience for hybrid and multi cloud. Let's welcome in age. Eva Hora, who is the CEO of Iota A J. Always good to see you again. Thanks for coming on. >>Great to be back. David Pleasure. >>And he's joined by Fozzy Coons, who is a global principal architect for financial services. The vertical of financial services. That red hat. He's got deep experiences in that sector. Welcome, Fozzie. Good to see you. >>Thank you very much. Happy to be here. >>Fancy. Let's start with you. Look, there are a lot of views on cloud and what it is. I wonder if you could explain to us how you think about what is a hybrid cloud and and how it works. >>Sure, yes. So the hybrid cloud is a 90 architecture that incorporates some degree off workload, possibility, orchestration and management across multiple clouds. Those clouds could be private cloud or public cloud or even your own data centers. And how does it all work? It's all about secure interconnectivity and on demand. Allocation of resources across clouds and separate clouds can become hydrate when they're similarly >>interconnected. And >>it is that interconnectivity that allows the workloads workers to be moved and how management can be unified in off the street. You can work and how well you have. These interconnections has a direct impact on how well your hybrid cloud will work. >>Okay, so we'll fancy staying with you for a minute. So in the early days of Cloud that turned private Cloud was thrown a lot around a lot, but often just meant virtualization of an on PREM system and a network connection to the public cloud. Let's bring it forward. What, in your view, does a modern hybrid cloud architecture look like? >>Sure. So for modern public clouds, we see that, um, teams organizations need to focus on the portability off applications across clouds. That's very important, right? And when organizations build applications, they need to build and deploy these applications as small collections off independently, loosely coupled services, and then have those things run on the same operating system which means, in other words, running it on Lenox everywhere and building cloud native applications and being able to manage and orchestrate thes applications with platforms like KUBERNETES or read it open shit, for example. >>Okay, so that Z, that's definitely different from building a monolithic application that's fossilized and and doesn't move. So what are the challenges for customers, you know, to get to that modern cloud? Aziz, you've just described it. Is it skill sets? Is that the ability to leverage things like containers? What's your view there? >>So, I mean, from what we've seen around around the industry, especially around financial services, where I spent most of my time, we see that the first thing that we see is management right now because you have all these clouds and all these applications, you have a massive array off connections off interconnections. You also have massive array off integrations, possibility and resource allocations as well, and then orchestrating all those different moving pieces. Things like storage networks and things like those are really difficult to manage, right? That's one. What s O Management is the first challenge. The second one is workload, placement, placement. Where do you place this? How do you place this cloud? Native applications. Do you or do you keep on site on Prem? And what do you put in the cloud? That is the the the other challenge. The major one. The third one is security. Security now becomes the key challenge and concern for most customers. And we could talk about how hundreds? Yeah, >>we're definitely gonna dig into that. Let's bring a J into the conversation. A J. You know, you and I have talked about this in the past. One of the big problems that virtually every companies face is data fragmentation. Um, talk a little bit about how I owe Tahoe unifies data across both traditional systems legacy systems. And it connects to these modern I t environments. >>Yeah, sure, Dave. I mean, fancy just nailed it. There used to be about data of the volume of data on the different types of data. But as applications become or connected and interconnected at the location of that data really matters how we serve that data up to those those app. So working with red hat in our partnership with Red Hat being able Thio, inject our data Discovery machine learning into these multiple different locations. Would it be in AWS on IBM Cloud or A D. C p R. On Prem being able thio Automate that discovery? I'm pulling that. That single view of where is all my data then allows the CEO to manage cast that can do things like one. I keep the data where it is on premise or in my Oracle Cloud or in my IBM cloud on Connect. The application that needs to feed off that data on the way in which you do that is machine learning. That learns over time is it recognizes different types of data, applies policies to declassify that data. Andi and brings it all together with automation. >>Right? And that's one of the big themes and we've talked about this on earlier episodes. Is really simplification really abstracting a lot of that heavy lifting away so we can focus on things A. J A. Z. You just mentioned e nifaz e. One of the big challenges that, of course, we all talk about his governance across thes disparity data sets. I'm curious as your thoughts. How does Red Hat really think about helping customers adhere to corporate edicts and compliance regulations, which, of course, are are particularly acute within financial services. >>Oh, yeah, Yes. So for banks and the payment providers, like you've just mentioned their insurers and many other financial services firms, Um, you know, they have to adhere Thio standards such as a PC. I. D. S s in Europe. You've got the G g d p g d p r, which requires strange and tracking, reporting documentation. And you know, for them to to remain in compliance and the way we recommend our customers to address these challenges is by having an automation strategy. Right. And that type of strategy can help you to improve the security on compliance off the organization and reduce the risk after the business. Right. And we help organizations build security and compliance from the start without consulting services residencies. We also offer courses that help customers to understand how to address some of these challenges. And that's also we help organizations build security into their applications without open sources. Mueller, where, um, middle offerings and even using a platform like open shift because it allows you to run legacy applications and also continue rights applications in a unified platform right And also that provides you with, you know, with the automation and the truly that you need to continuously monitor, manage and automate the systems for security and compliance >>purposes. Hey, >>Jay, anything. Any color you could add to this conversation? >>Yeah, I'm pleased. Badly brought up Open shift. I mean, we're using open shift to be able. Thio, take that security application of controls to to the data level. It's all about context. So, understanding what data is there being able to assess it to say who should have access to it. Which application permission should be applied to it. Um, that za great combination of Red Hat tonight. Tahoe. >>But what about multi Cloud? Doesn't that complicate the situation even even further? Maybe you could talk about some of the best practices to apply automation across not only hybrid cloud, but multi >>cloud a swell. Yeah, sure. >>Yeah. So the right automation solution, you know, can be the difference between, you know, cultivating an automated enterprise or automation caress. And some of the recommendations we give our clients is to look for an automation platform that can offer the first thing is complete support. So that means have an automation solution that provides that provides, um, you know, promotes I t availability and reliability with your platform so that you can provide, you know, enterprise great support, including security and testing, integration and clear roadmaps. The second thing is vendor interoperability interoperability in that you are going to be integrating multiple clouds. So you're going to need a solution that can connect to multiple clouds. Simples lee, right? And with that comes the challenge off maintain ability. So you you you're going to need to look into a automation Ah, solution that that is easy to learn or has an easy learning curve. And then the fourth idea that we tell our customers is scalability in the in the hybrid cloud space scale is >>is >>a big, big deal here, and you need a to deploy an automation solution that can span across the whole enterprise in a constituent, consistent manner, right? And then also, that allows you finally to, uh, integrate the multiple data centers that you have, >>So A J I mean, this is a complicated situation, for if a customer has toe, make sure things work on AWS or azure or Google. Uh, they're gonna spend all their time doing that, huh? What can you add really? To simplify that that multi cloud and hybrid cloud equation? >>Yeah. I could give a few customer examples here Warming a manufacturer that we've worked with to drive that simplification Onda riel bonuses for them is has been a reduction cost. We worked with them late last year to bring the cost bend down by $10 million in 2021 so they could hit that reduced budget. Andre, What we brought to that was the ability thio deploy using open shift templates into their different environments. Where there is on premise on bond or in as you mentioned, a W s. They had G cps well, for their marketing team on a cross, those different platforms being out Thio use a template, use pre built scripts to get up and running in catalog and discover that data within minutes. It takes away the legacy of having teams of people having Thio to jump on workshop cause and I know we're all on a lot of teens. The zoom cause, um, in these current times, they just sent me is in in of hours in the day Thio manually perform all of this. So yeah, working with red hat applying machine learning into those templates those little recipes that we can put that automation toe work, regardless of which location the data is in allows us thio pull that unified view together. Right? >>Thank you, Fozzie. I wanna come back to you. So the early days of cloud, you're in the big apple, you know, financial services. Really well. Cloud was like an evil word within financial services, and obviously that's changed. It's evolved. We talked about the pandemic, has even accelerated that, Um And when you really, you know, dug into it when you talk to customers about their experiences with security in the cloud it was it was not that it wasn't good. It was great, whatever. But it was different. And there's always this issue of skill, lack of skills and multiple tools suck up teams, they're really overburdened. But in the cloud requires new thinking. You've got the shared responsibility model you've got obviously have specific corporate requirements and compliance. So this is even more complicated when you introduce multiple clouds. So what are the differences that you can share from your experience is running on a sort of either on Prem or on a mono cloud, um, or, you know, and versus across clouds. What? What? What do you suggest there? >>Yeah, you know, because of these complexities that you have explained here, Miss Configurations and the inadequate change control the top security threats. So human error is what we want to avoid because is, you know, as your clouds grow with complexity and you put humans in the mix, then the rate off eras is going to increase, and that is going to exposure to security threat. So this is where automation comes in because automation will streamline and increase the consistency off your infrastructure management. Also application development and even security operations to improve in your protection, compliance and change control. So you want to consistently configure resources according to a pre approved um, you know, pre approved policies and you want to proactively maintain a to them in a repeatable fashion over the whole life cycle. And then you also want to rapid the identified system that require patches and and reconfiguration and automate that process off patching and reconfiguring so that you don't have humans doing this type of thing, right? And you want to be able to easily apply patches and change assistant settings. According Thio, Pre defined, based on like explained before, you know, with the pre approved policies and also you want is off auditing and troubleshooting, right? And from a rate of perspective, we provide tools that enable you to do this. We have, for example, a tool called danceable that enables you to automate data center operations and security and also deployment of applications and also obvious shit yourself, you know, automates most of these things and obstruct the human beings from putting their fingers on, causing, uh, potentially introducing errors right now in looking into the new world off multiple clouds and so forth. The difference is that we're seeing here between running a single cloud or on prem is three main areas which is control security and compliance. Right control here it means if your on premise or you have one cloud, um, you know, in most cases you have control over your data and your applications, especially if you're on Prem. However, if you're in the public cloud, there is a difference there. The ownership, it is still yours. But your resources are running on somebody else's or the public clouds. You know, e w s and so forth infrastructure. So people that are going to do this need to really especially banks and governments need to be aware off the regulatory constraints off running, uh, those applications in the public cloud. And we also help customers regionalize some of these choices and also on security. You will see that if you're running on premises or in a single cloud, you have more control, especially if you're on Prem. You can control this sensitive information that you have, however, in the cloud. That's a different situation, especially from personal information of employees and things like that. You need to be really careful off that. And also again, we help you rationalize some of those choices. And then the last one is compliant. Aziz. Well, you see that if you're running on Prem or a single cloud, um, regulations come into play again, right? And if you're running a problem, you have control over that. You can document everything you have access to everything that you need. But if you're gonna go to the public cloud again, you need to think about that. We have automation, and we have standards that can help you, uh, you know, address some of these challenges for security and compliance. >>So that's really strong insights, Potsie. I mean, first of all, answerable has a lot of market momentum. Red hats in a really good job with that acquisition, your point about repeatability is critical because you can't scale otherwise. And then that idea you're you're putting forth about control, security compliance It's so true is I called it the shared responsibility model. And there was a lot of misunderstanding in the early days of cloud. I mean, yeah, maybe a W s is gonna physically secure the, you know, s three, but in the bucket. But we saw so many Miss configurations early on. And so it's key to have partners that really understand this stuff and can share the experiences of other clients. So this all sounds great. A j. You're sharp, you know, financial background. What about the economics? >>You >>know, our survey data shows that security it's at the top of the spending priority list, but budgets are stretched thin. E especially when you think about the work from home pivot and and all the areas that they had toe the holes that they had to fill their, whether it was laptops, you know, new security models, etcetera. So how do organizations pay for this? What's the business case look like in terms of maybe reducing infrastructure costs so I could, you know, pay it forward or there's a There's a risk reduction angle. What can you share >>their? Yeah. I mean, the perspective I'd like to give here is, um, not being multi cloud is multi copies of an application or data. When I think about 20 years, a lot of the work in financial services I was looking at with managing copies of data that we're feeding different pipelines, different applications. Now what we're saying I talk a lot of the work that we're doing is reducing the number of copies of that data so that if I've got a product lifecycle management set of data, if I'm a manufacturer, I'm just gonna keep that in one location. But across my different clouds, I'm gonna have best of breed applications developed in house third parties in collaboration with my supply chain connecting securely to that. That single version of the truth. What I'm not going to do is to copy that data. So ah, lot of what we're seeing now is that interconnectivity using applications built on kubernetes. Um, that decoupled from the data source that allows us to reduce those copies of data within that you're gaining from the security capability and resilience because you're not leaving yourself open to those multiple copies of data on with that. Couldn't come. Cost, cost of storage on duh cost of compute. So what we're seeing is using multi cloud to leverage the best of what each cloud platform has to offer That goes all the way to Snowflake and Hiroko on Cloud manage databases, too. >>Well, and the people cost to a swell when you think about yes, the copy creep. But then you know when something goes wrong, a human has to come in and figured out um, you brought up snowflake, get this vision of the data cloud, which is, you know, data data. I think this we're gonna be rethinking a j, uh, data architectures in the coming decade where data stays where it belongs. It's distributed, and you're providing access. Like you said, you're separating the data from the applications applications as we talked about with Fozzie. Much more portable. So it Z really the last 10 years will be different than the next 10 years. A. >>J Definitely. I think the people cast election is used. Gone are the days where you needed thio have a dozen people governing managing black policies to data. Ah, lot of that repetitive work. Those tests can be in power automated. We've seen examples in insurance were reduced teams of 15 people working in the the back office China apply security controls compliance down to just a couple of people who are looking at the exceptions that don't fit. And that's really important because maybe two years ago the emphasis was on regulatory compliance of data with policies such as GDP are in CCP a last year, very much the economic effect of reduce headcounts on on enterprises of running lean looking to reduce that cost. This year, we can see that already some of the more proactive cos they're looking at initiatives such as net zero emissions how they use data toe under understand how cape how they can become more have a better social impact. Um, and using data to drive that, and that's across all of their operations and supply chain. So those regulatory compliance issues that may have been external we see similar patterns emerging for internal initiatives that benefiting the environment, social impact and and, of course, course, >>great perspectives. Yeah, Jeff Hammer, Bucker once famously said, The best minds of my generation are trying to get people to click on ads and a J. Those examples that you just gave of, you know, social good and moving. Uh, things forward are really critical. And I think that's where Data is gonna have the biggest societal impact. Okay, guys, great conversation. Thanks so much for coming on the program. Really appreciate your time. Keep it right there from, or insight and conversation around, creating a resilient digital business model. You're watching the >>Cube digital resilience, automated compliance, privacy and security for your multi cloud. Congratulations. You're on the journey. You have successfully transformed your organization by moving to a cloud based platform to ensure business continuity in these challenging times. But as you scale your digital activities, there is an inevitable influx of users that outpaces traditional methods of cybersecurity, exposing your data toe underlying threats on making your company susceptible toe ever greater risk to become digitally resilient. Have you applied controls your data continuously throughout the data Lifecycle? What are you doing to keep your customer on supply data private and secure? I owe Tahoe's automated, sensitive data. Discovery is pre programmed with over 300 existing policies that meet government mandated risk and compliance standards. Thes automate the process of applying policies and controls to your data. Our algorithm driven recommendation engine alerts you to risk exposure at the data level and suggests the appropriate next steps to remain compliant on ensure sensitive data is secure. Unsure about where your organization stands In terms of digital resilience, Sign up for a minimal cost commitment. Free data Health check. Let us run our sensitive data discovery on key unmapped data silos and sources to give you a clear understanding of what's in your environment. Book time within Iot. Tahoe Engineer Now >>Okay, let's now get into the next segment where we'll explore data automation. But from the angle of digital resilience within and as a service consumption model, we're now joined by Yusuf Khan, who heads data services for Iot, Tahoe and Shirish County up in. Who's the vice president and head of U. S. Sales at happiest Minds? Gents, welcome to the program. Great to have you in the Cube. >>Thank you, David. >>Trust you guys talk about happiest minds. This notion of born digital, foreign agile. I like that. But talk about your mission at the company. >>Sure. >>A former in 2011 Happiest Mind is a born digital born a child company. The reason is that we are focused on customers. Our customer centric approach on delivering digitals and seamless solutions have helped us be in the race. Along with the Tier one providers, Our mission, happiest people, happiest customers is focused to enable customer happiness through people happiness. We have Bean ranked among the top 25 i t services company in the great places to work serving hour glass to ratings off 41 against the rating off. Five is among the job in the Indian nineties services company that >>shows the >>mission on the culture. What we have built on the values right sharing, mindful, integrity, learning and social on social responsibilities are the core values off our company on. That's where the entire culture of the company has been built. >>That's great. That sounds like a happy place to be. Now you said you had up data services for Iot Tahoe. We've talked in the past. Of course you're out of London. What >>do you what? Your >>day to day focus with customers and partners. What you focused >>on? Well, David, my team work daily with customers and partners to help them better understand their data, improve their data quality, their data governance on help them make that data more accessible in a self service kind of way. To the stakeholders within those businesses on dis is all a key part of digital resilience that will will come on to talk about but later. You're >>right, e mean, that self service theme is something that we're gonna we're gonna really accelerate this decade, Yussef and so. But I wonder before we get into that, maybe you could talk about the nature of the partnership with happiest minds, you know? Why do you guys choose toe work closely together? >>Very good question. Um, we see Hyo Tahoe on happiest minds as a great mutual fit. A Suresh has said, uh, happiest minds are very agile organization um, I think that's one of the key things that attracts their customers on Io. Tahoe is all about automation. Uh, we're using machine learning algorithms to make data discovery data cataloging, understanding, data done. See, uh, much easier on. We're enabling customers and partners to do it much more quickly. So when you combine our emphasis on automation with the emphasis on agility that happiest minds have that that's a really nice combination work works very well together, very powerful. I think the other things that a key are both businesses, a serious have said, are really innovative digital native type type companies. Um, very focused on newer technologies, the cloud etcetera on. Then finally, I think they're both Challenger brands on happiest minds have a really positive, fresh ethical approach to people and customers that really resonates with us at Ideo Tahoe to >>great thank you for that. So Russia, let's get into the whole notion of digital resilience. I wanna I wanna sort of set it up with what I see, and maybe you can comment be prior to the pandemic. A lot of customers that kind of equated disaster recovery with their business continuance or business resilient strategy, and that's changed almost overnight. How have you seen your clients respond to that? What? I sometimes called the forced march to become a digital business. And maybe you could talk about some of the challenges that they faced along the way. >>Absolutely. So, uh, especially during this pandemic, times when you say Dave, customers have been having tough times managing their business. So happiest minds. Being a digital Brazilian company, we were able to react much faster in the industry, apart from the other services company. So one of the key things is the organisation's trying to adopt onto the digital technologies. Right there has bean lot off data which has been to manage by these customers on There have been lot off threats and risk, which has been to manage by the CEO Seo's so happiest minds digital resilient technology, right where we bring in the data. Complaints as a service were ableto manage the resilience much ahead off other competitors in the market. We were ableto bring in our business continuity processes from day one, where we were ableto deliver our services without any interruption to the services. What we were delivered to our customers So that is where the digital resilience with business community process enabled was very helpful for us. Toe enable our customers continue their business without any interruptions during pandemics. >>So I mean, some of the challenges that customers tell me they obviously they had to figure out how to get laptops to remote workers and that that whole remote work from home pivot figure out how to secure the end points. And, you know, those were kind of looking back there kind of table stakes, But it sounds like you've got a digital business. Means a data business putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe the philosophy you have toward digital resilience in the specific approach you take with clients? >>Absolutely. They seen any organization data becomes. The key on that, for the first step is to identify the critical data. Right. So we this is a six step process. What we following happiest minds. First of all, we take stock off the current state, though the customers think that they have a clear visibility off their data. How are we do more often assessment from an external point off view on see how critical their data is, then we help the customers to strategies that right. The most important thing is to identify the most important critical herself. Data being the most critical assert for any organization. Identification off the data's key for the customers. Then we help in building a viable operating model to ensure these identified critical assets are secure on monitor dearly so that they are consumed well as well as protected from external threats. Then, as 1/4 step, we try to bring in awareness, toe the people we train them >>at >>all levels in the organization. That is a P for people to understand the importance off the digital ourselves and then as 1/5 step, we work as a back up plan in terms of bringing in a very comprehensive and a holistic testing approach on people process as well as in technology. We'll see how the organization can withstand during a crisis time, and finally we do a continuous governance off this data, which is a key right. It is not just a one step process. We set up the environment, we do the initial analysis and set up the strategy on continuously govern this data to ensure that they are not only know managed will secure as well as they also have to meet the compliance requirements off the organization's right. That is where we help organizations toe secure on Meet the regulations off the organizations. As for the privacy laws, so this is a constant process. It's not on one time effort. We do a constant process because every organization goes towards their digital journey on. They have to face all these as part off the evolving environment on digital journey. And that's where they should be kept ready in terms off. No recovering, rebounding on moving forward if things goes wrong. >>So let's stick on that for a minute, and then I wanna bring yourself into the conversation. So you mentioned compliance and governance when when your digital business, you're, as you say, you're a data business, so that brings up issues. Data sovereignty. Uh, there's governance, this compliance. There's things like right to be forgotten. There's data privacy, so many things. These were often kind of afterthoughts for businesses that bolted on, if you will. I know a lot of executives are very much concerned that these air built in on, and it's not a one shot deal. So do you have solutions around compliance and governance? Can you deliver that as a service? Maybe you could talk about some of the specifics there, >>so some of way have offered multiple services. Tow our customers on digital against. On one of the key service is the data complaints. As a service here we help organizations toe map the key data against the data compliance requirements. Some of the features includes in terms off the continuous discovery off data right, because organizations keep adding on data when they move more digital on helping the helping and understanding the actual data in terms off the residents of data, it could be a heterogeneous data soldiers. It could be on data basis, or it could be even on the data legs. Or it could be a no even on compromise all the cloud environment. So identifying the data across the various no heterogeneous environment is very key. Feature off our solution. Once we identify classify this sensitive data, the data privacy regulations on the traveling laws have to be map based on the business rules So we define those rules on help map those data so that organizations know how critical their digital assets are. Then we work on a continuous marching off data for anomalies because that's one of the key teachers off the solution, which needs to be implemented on the day to day operational basis. So we're helping monitoring those anomalies off data for data quality management on an ongoing basis. On finally, we also bringing the automated data governance where we can manage the sensory data policies on their later relationships in terms off mapping on manage their business roots on we drive reputations toe Also suggest appropriate actions to the customers. Take on those specific data sets. >>Great. Thank you, Yousef. Thanks for being patient. I want to bring in Iota ho thio discussion and understand where your customers and happiest minds can leverage your data automation capability that you and I have talked about in the past. I'm gonna be great if you had an example is well, but maybe you could pick it up from there, >>John. I mean, at a high level, assertions are clearly articulated. Really? Um, Hyoty, who delivers business agility. So that's by, um accelerating the time to operationalize data, automating, putting in place controls and actually putting helping put in place digital resilience. I mean way if we step back a little bit in time, um, traditional resilience in relation to data often met manually, making multiple copies of the same data. So you have a d b A. They would copy the data to various different places, and then business users would access it in those functional style owes. And of course, what happened was you ended up with lots of different copies off the same data around the enterprise. Very inefficient. ONDA course ultimately, uh, increases your risk profile. Your risk of a data breach. Um, it's very hard to know where everything is. And I realized that expression. They used David the idea of the forced march to digital. So with enterprises that are going on this forced march, what they're finding is they don't have a single version of the truth, and almost nobody has an accurate view of where their critical data is. Then you have containers bond with containers that enables a big leap forward so you could break applications down into micro services. Updates are available via a p I s on. So you don't have the same need thio to build and to manage multiple copies of the data. So you have an opportunity to just have a single version of the truth. Then your challenge is, how do you deal with these large legacy data states that the service has been referring Thio, where you you have toe consolidate and that's really where I attack comes in. Um, we massively accelerate that process of putting in a single version of the truth into place. So by automatically discovering the data, discovering what's dubica? What's redundant? Uh, that means you can consolidate it down to a single trusted version much more quickly. We've seen many customers have tried to do this manually, and it's literally taken years using manual methods to cover even a small percentage of their I T estates. With our tire, you could do it really very quickly on you can have tangible results within weeks and months on Ben, you can apply controls to the data based on context. So who's the user? What's the content? What's the use case? Things like data quality validations or access permissions on. Then, once you've done there. Your applications and your enterprise are much more secure, much more resilient. As a result, you've got to do these things whilst retaining agility, though. So coming full circle. This is where the partnership with happiest minds really comes in as well. You've got to be agile. You've gotta have controls. Um, on you've got a drug toward the business outcomes. Uh, and it's doing those three things together that really deliver for the customer. >>Thank you. Use f. I mean you and I. In previous episodes, we've looked in detail at the business case. You were just talking about the manual labor involved. We know that you can't scale, but also there's that compression of time. Thio get to the next step in terms of ultimately getting to the outcome. And we talked to a number of customers in the Cube, and the conclusion is, it's really consistent that if you could accelerate the time to value, that's the key driver reducing complexity, automating and getting to insights faster. That's where you see telephone numbers in terms of business impact. So my question is, where should customers start? I mean, how can they take advantage of some of these opportunities that we've discussed today. >>Well, we've tried to make that easy for customers. So with our Tahoe and happiest minds, you can very quickly do what we call a data health check. Um, this is a is a 2 to 3 week process, uh, to really quickly start to understand on deliver value from your data. Um, so, iota, who deploys into the customer environment? Data doesn't go anywhere. Um, we would look at a few data sources on a sample of data. Onda. We can very rapidly demonstrate how they discovery those catalog e on understanding Jupiter data and redundant data can be done. Um, using machine learning, um, on how those problems can be solved. Um, And so what we tend to find is that we can very quickly, as I say in the matter of a few weeks, show a customer how they could get toe, um, or Brazilian outcome on then how they can scale that up, take it into production on, then really understand their data state? Better on build. Um, Brasiliense into the enterprise. >>Excellent. There you have it. We'll leave it right there. Guys, great conversation. Thanks so much for coming on the program. Best of luck to you and the partnership Be well, >>Thank you, David Suresh. Thank you. Thank >>you for watching everybody, This is Dave Volonte for the Cuban are ongoing Siris on data automation without >>Tahoe, digital resilience, automated compliance, privacy and security for your multi cloud. Congratulations. You're on the journey. You have successfully transformed your organization by moving to a cloud based platform to ensure business continuity in these challenging times. But as you scale your digital activities, there is an inevitable influx of users that outpaces traditional methods of cybersecurity, exposing your data toe underlying threats on making your company susceptible toe ever greater risk to become digitally resilient. Have you applied controls your data continuously throughout the data lifecycle? What are you doing to keep your customer on supply data private and secure? I owe Tahoe's automated sensitive data. Discovery is pre programmed with over 300 existing policies that meet government mandated risk and compliance standards. Thes automate the process of applying policies and controls to your data. Our algorithm driven recommendation engine alerts you to risk exposure at the data level and suggests the appropriate next steps to remain compliant on ensure sensitive data is secure. Unsure about where your organization stands in terms of digital resilience. Sign up for our minimal cost commitment. Free data health check. Let us run our sensitive data discovery on key unmapped data silos and sources to give you a clear understanding of what's in your environment. Book time within Iot. Tahoe Engineer. Now. >>Okay, now we're >>gonna go into the demo. We want to get a better understanding of how you can leverage open shift. And I owe Tahoe to facilitate faster application deployment. Let me pass the mic to Sabetta. Take it away. >>Uh, thanks, Dave. Happy to be here again, Guys, uh, they've mentioned names to be the Davis. I'm the enterprise account executive here. Toyota ho eso Today we just wanted to give you guys a general overview of how we're using open shift. Yeah. Hey, I'm Noah Iota host data operations engineer, working with open ship. And I've been learning the Internets of open shift for, like, the past few months, and I'm here to share. What a plan. Okay, so So before we begin, I'm sure everybody wants to know. Noel, what are the benefits of using open shift. Well, there's five that I can think of a faster time, the operation simplicity, automation control and digital resilience. Okay, so that that's really interesting, because there's an exact same benefits that we had a Tahoe delivered to our customers. But let's start with faster time the operation by running iota. Who on open shift? Is it faster than, let's say, using kubernetes and other platforms >>are >>objective iota. Who is to be accessible across multiple cloud platforms, right? And so by hosting our application and containers were able to achieve this. So to answer your question, it's faster to create and use your application images using container tools like kubernetes with open shift as compared to, like kubernetes with docker cry over container D. Okay, so we got a bit technical there. Can you explain that in a bit more detail? Yeah, there's a bit of vocabulary involved, uh, so basically, containers are used in developing things like databases, Web servers or applications such as I have top. What's great about containers is that they split the workload so developers can select the libraries without breaking anything. And since Hammond's can update the host without interrupting the programmers. Uh, now, open shift works hand in hand with kubernetes to provide a way to build those containers for applications. Okay, got It s basically containers make life easier for developers and system happens. How does open shift differ from other platforms? Well, this kind of leads into the second benefit I want to talk about, which is simplicity. Basically, there's a lot of steps involved with when using kubernetes with docker. But open shift simplifies this with their source to image process that takes the source code and turns it into a container image. But that's not all. Open shift has a lot of automation and features that simplify working with containers, an important one being its Web console. Here. I've set up a light version of open ship called Code Ready Containers, and I was able to set up her application right from the Web console. And I was able to set up this entire thing in Windows, Mac and Lennox. So its environment agnostic in that sense. Okay, so I think I've seen the top left that this is a developers view. What would a systems admin view look like? It's a good question. So here's the administrator view and this kind of ties into the benefit of control. Um, this view gives insights into each one of the applications and containers that are running, and you could make changes without affecting deployment. Andi can also, within this view, set up each layer of security, and there's multiple that you can prop up. But I haven't fully messed around with it because with my luck, I'd probably locked myself out. So that seems pretty secure. Is there a single point security such as you use a log in? Or are there multiple layers of security? Yeah, there are multiple layers of security. There's your user login security groups and general role based access controls. Um, but there's also a ton of layers of security surrounding like the containers themselves. But for the sake of time, I won't get too far into it. Okay, eso you mentioned simplicity In time. The operation is being two of the benefits. You also briefly mention automation. And as you know, automation is the backbone of our platform here, Toyota Ho. So that's certainly grabbed my attention. Can you go a bit more in depth in terms of automation? Open shift provides extensive automation that speeds up that time the operation. Right. So the latest versions of open should come with a built in cryo container engine, which basically means that you get to skip that container engine insulation step and you don't have to, like, log into each individual container host and configure networking, configure registry servers, storage, etcetera. So I'd say, uh, it automates the more boring kind of tedious process is Okay, so I see the iota ho template there. What does it allow me to do? Um, in terms of automation in application development. So we've created an open shift template which contains our application. This allows developers thio instantly, like set up our product within that template. So, Noah Last question. Speaking of vocabulary, you mentioned earlier digital resilience of the term we're hearing, especially in the banking and finance world. Um, it seems from what you described, industries like banking and finance would be more resilient using open shift, Correct. Yeah, In terms of digital resilience, open shift will give you better control over the consumption of resource is each container is using. In addition, the benefit of containers is that, like I mentioned earlier since Hammond's can troubleshoot servers about bringing down the application and if the application does go down is easy to bring it back up using templates and, like the other automation features that open ship provides. Okay, so thanks so much. Know us? So any final thoughts you want to share? Yeah. I just want to give a quick recap with, like, the five benefits that you gained by using open shift. Uh, the five are timeto operation automation, control, security and simplicity. You could deploy applications faster. You could simplify the workload you could automate. A lot of the otherwise tedious processes can maintain full control over your workflow. And you could assert digital resilience within your environment. Guys, >>Thanks for that. Appreciate the demo. Um, I wonder you guys have been talking about the combination of a Iot Tahoe and red hat. Can you tie that in subito Digital resilience >>Specifically? Yeah, sure, Dave eso when we speak to the benefits of security controls in terms of digital resilience at Io Tahoe, we automated detection and apply controls at the data level, so this would provide for more enhanced security. >>Okay, But so if you were trying to do all these things manually. I mean, what what does that do? How much time can I compress? What's the time to value? >>So with our latest versions, Biota we're taking advantage of faster deployment time associated with container ization and kubernetes. So this kind of speeds up the time it takes for customers. Start using our software as they be ableto quickly spin up io towel on their own on premise environment are otherwise in their own cloud environment, like including aws. Assure or call GP on IBM Cloud a quick start templates allow flexibility deploy into multi cloud environments all just using, like, a few clicks. Okay, so so now just quickly add So what we've done iota, Who here is We've really moved our customers away from the whole idea of needing a team of engineers to apply controls to data as compared to other manually driven work flows. Eso with templates, automation, previous policies and data controls. One person can be fully operational within a few hours and achieve results straight out of the box on any cloud. >>Yeah, we've been talking about this theme of abstracting the complexity. That's really what we're seeing is a major trend in in this coming decade. Okay, great. Thanks, Sabina. Noah, How could people get more information or if they have any follow up questions? Where should they go? >>Yeah, sure. They've. I mean, if you guys are interested in learning more, you know, reach out to us at info at iata ho dot com to speak with one of our sales engineers. I mean, we love to hear from you, so book a meeting as soon as you can. All >>right. Thanks, guys. Keep it right there from or cube content with.

Published Date : Jan 27 2021

SUMMARY :

Always good to see you again. Great to be back. Good to see you. Thank you very much. I wonder if you could explain to us how you think about what is a hybrid cloud and So the hybrid cloud is a 90 architecture that incorporates some degree off And it is that interconnectivity that allows the workloads workers to be moved So in the early days of Cloud that turned private Cloud was thrown a lot to manage and orchestrate thes applications with platforms like Is that the ability to leverage things like containers? And what do you put in the cloud? One of the big problems that virtually every companies face is data fragmentation. the way in which you do that is machine learning. And that's one of the big themes and we've talked about this on earlier episodes. And that type of strategy can help you to improve the security on Hey, Any color you could add to this conversation? is there being able to assess it to say who should have access to it. Yeah, sure. the difference between, you know, cultivating an automated enterprise or automation caress. What can you add really? bond or in as you mentioned, a W s. They had G cps well, So what are the differences that you can share from your experience is running on a sort of either And from a rate of perspective, we provide tools that enable you to do this. A j. You're sharp, you know, financial background. know, our survey data shows that security it's at the top of the spending priority list, Um, that decoupled from the data source that Well, and the people cost to a swell when you think about yes, the copy creep. Gone are the days where you needed thio have a dozen people governing managing to get people to click on ads and a J. Those examples that you just gave of, you know, to give you a clear understanding of what's in your environment. Great to have you in the Cube. Trust you guys talk about happiest minds. We have Bean ranked among the mission on the culture. Now you said you had up data services for Iot Tahoe. What you focused To the stakeholders within those businesses on dis is of the partnership with happiest minds, you know? So when you combine our emphasis on automation with the emphasis And maybe you could talk about some of the challenges that they faced along the way. So one of the key things putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe for the first step is to identify the critical data. off the digital ourselves and then as 1/5 step, we work as a back up plan So you mentioned compliance and governance when when your digital business, you're, as you say, So identifying the data across the various no heterogeneous environment is well, but maybe you could pick it up from there, So you don't have the same need thio to build and to manage multiple copies of the data. and the conclusion is, it's really consistent that if you could accelerate the time to value, to really quickly start to understand on deliver value from your data. Best of luck to you and the partnership Be well, Thank you, David Suresh. to give you a clear understanding of what's in your environment. Let me pass the mic to And I've been learning the Internets of open shift for, like, the past few months, and I'm here to share. into each one of the applications and containers that are running, and you could make changes without affecting Um, I wonder you guys have been talking about the combination of apply controls at the data level, so this would provide for more enhanced security. What's the time to value? a team of engineers to apply controls to data as compared to other manually driven work That's really what we're seeing I mean, if you guys are interested in learning more, you know, reach out to us at info at iata Keep it right there from or cube content with.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Jeff HammerPERSON

0.99+

JohnPERSON

0.99+

Eva HoraPERSON

0.99+

David SureshPERSON

0.99+

SabinaPERSON

0.99+

DavePERSON

0.99+

Yusuf KhanPERSON

0.99+

EuropeLOCATION

0.99+

LondonLOCATION

0.99+

2021DATE

0.99+

twoQUANTITY

0.99+

AWSORGANIZATION

0.99+

Dave VolontePERSON

0.99+

SiriTITLE

0.99+

GoogleORGANIZATION

0.99+

FozziePERSON

0.99+

2QUANTITY

0.99+

fiveQUANTITY

0.99+

David PleasurePERSON

0.99+

iata ho dot comORGANIZATION

0.99+

JayPERSON

0.99+

FiveQUANTITY

0.99+

six stepQUANTITY

0.99+

five benefitsQUANTITY

0.99+

15 peopleQUANTITY

0.99+

YousefPERSON

0.99+

$10 millionQUANTITY

0.99+

This yearDATE

0.99+

first stepQUANTITY

0.99+

Ideo TahoeORGANIZATION

0.99+

last yearDATE

0.99+

AndrePERSON

0.99+

hundredsQUANTITY

0.99+

OneQUANTITY

0.99+

one cloudQUANTITY

0.99+

2011DATE

0.99+

TahoeORGANIZATION

0.99+

TodayDATE

0.99+

NoelPERSON

0.99+

Red HatORGANIZATION

0.99+

PremORGANIZATION

0.99+

todayDATE

0.99+

tonightDATE

0.99+

Io TahoeORGANIZATION

0.99+

second benefitQUANTITY

0.99+

oneQUANTITY

0.99+

Iota A J.ORGANIZATION

0.99+

one stepQUANTITY

0.99+

bothQUANTITY

0.98+

third oneQUANTITY

0.98+

SirisTITLE

0.98+

AzizPERSON

0.98+

red hatORGANIZATION

0.98+

each layerQUANTITY

0.98+

both businessesQUANTITY

0.98+

fourth ideaQUANTITY

0.98+

appleORGANIZATION

0.98+

1/5 stepQUANTITY

0.98+

Toyota HoORGANIZATION

0.98+

first challengeQUANTITY

0.98+

41QUANTITY

0.98+

azureORGANIZATION

0.98+

Io TahoePERSON

0.98+

One personQUANTITY

0.98+

one locationQUANTITY

0.98+

singleQUANTITY

0.98+

NoahPERSON

0.98+

over 300 existing policiesQUANTITY

0.98+

Iot TahoeORGANIZATION

0.98+

ThioPERSON

0.98+

LenoxORGANIZATION

0.98+

two years agoDATE

0.98+

A. J A. Z.PERSON

0.98+

single pointQUANTITY

0.98+

first thingQUANTITY

0.97+

YussefPERSON

0.97+

JupiterLOCATION

0.97+

second thingQUANTITY

0.97+

three thingsQUANTITY

0.97+

about 20 yearsQUANTITY

0.97+

single cloudQUANTITY

0.97+

FirstQUANTITY

0.97+

SureshPERSON

0.97+

3 weekQUANTITY

0.97+

each containerQUANTITY

0.97+

each cloud platformQUANTITY

0.97+

Yusef Khan & Suresh Kanniappan | Io Tahoe Enterprise Digital Resilience on Hybrid & Multicloud


 

>>from around the globe. It's the Cube presenting enterprise, Digital resilience on hybrid and multi cloud Brought to You by Iota Ho. Okay, let's now get into the next segment where we'll explore data automation. But from the angle of digital resilience within and as a service consumption model, we're now joined by Yusuf Khan, who heads data services for Iota Ho and Shirish County. Up in Who's the vice president and head of U. S. Sales at happiest Minds. Gents, welcome to the program. Great to have you in the Cube. >>Thank you, David. >>Stretch. You guys talk about happiest minds. This notion of born digital, foreign agile. I like that. But talk about your mission at the company. >>Sure. A former in 2011 Happiest minds Up Born digital born a child company. >>The >>reason is that we are focused on customers. Our customer centric approach on delivering digitals and seamless solutions have helped us be in the race. Along with the Tier one providers, our mission, happiest people, happiest customers is focused to enable customer happiness through people happiness. We have Bean ranked among the top 25 I t services company in the great places to work serving hour glass to ratings off 4.1 against the rating off five is among the job in the Indian nineties services company that >>shows the >>mission on the culture. What we have built on the values, right sharing, mindful, integrity, learning and social on social responsibilities are the core values off our company on. That's where the entire culture of the company has been built. >>That's great. That sounds like a happy place to be. Now you have you head up data services for Iot Tahoe. We've talked in the past. Of course you're out of London. What do you what's your day to day focus with customers and partners? What you focused on? >>Well, David, my team work daily with customers and partners to help them better understand their data, improve their data quality, their data governance on help them make that data more accessible in a self service kind of way. To the stakeholders within those businesses on dis is all a key part of digital resilience that will will come on to talk about but later. You're >>right, e mean, that self service theme is something that we're gonna we're gonna really accelerate this decade, Yussef and so. But I wonder before we get into that, maybe you could talk about the nature of the partnership with happiest minds. You know, why do you guys choose toe work closely together? >>Very good question. Um, we see Io Tahoe on Happiest minds as a great mutual fit. A Suresh has said happiest minds are very agile organization. Um, I think that's one of the key things that attracts their customers on Io. Tahoe is all about automation. We're using machine learning algorithms to make data discovery data cataloging, understanding, data, redundancy, uh, much easier on. We're enabling customers and partners to do it much more quickly. So when you combine our emphasis on automation with the emphasis on agility, the happiest minds have that. That's a really nice combination. Work works very well together, very powerful. I think the other things that a key are both businesses, a serious have said are really innovative digital native type type companies. Um, very focused on newer technologies, the cloud etcetera, uh, on. Then finally, I think that both challenger brands Andi happiest minds have a really positive, fresh ethical approach to people and customers that really resonates with us that I have tied to its >>great thank you for that. So Russia, Let's get into the whole notion of digital resilience. I wanna I wanna sort of set it up with what I see. And maybe you can comment be prior to the pandemic. A lot of customers that kind of equated disaster recovery with their business continuance or business resilient strategy, and that's changed almost overnight. How have you seen your clients respond to that? What? I sometimes called the forced march to become a digital business. And maybe you could talk about some of the challenges that they faced along the way. >>Absolutely. So, uh, especially during this pandemic times when you see Dave customers have been having tough times managing their business. So happiest minds. Being a digital Brazilian company, we were able to react much faster in the industry, apart from the other services company. So one of the key things is the organizations trying to adopt onto the digital technologies right there has bean lot off data which has been to managed by these customers on. There have been lot off threats and risk, which has been to manage by the CEO Seo's so happiest minds digital resilient technology fight the where we're bringing the data complaints as a service, we were ableto manage the resilience much ahead off other competitors in the market. We were ableto bring in our business community processes from day one, where we were ableto deliver our services without any interruption to the services what we were delivering to our customers. >>So >>that is where the digital resilience with business community process enabled was very helpful for us who enable our customers continue there business without any interruptions during pandemics. >>So, I mean, some of the challenges that that customers tell me they obviously had to figure out how to get laptops to remote workers and that that whole remote, you know, work from home pivot figure out how to secure the end points. And, you know, those were kind of looking back there kind of table stakes, but it sounds like you've got a digital business means a data business putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe the philosophy you have toward digital resilience in the specific approach you take with clients? >>Absolutely. They seen any organization data becomes. The key on this for the first step is to identify the critical data. Right. So we this is 1/6 process. What we following happiest minds. First of all, we take stock off the current state, though the customers think that they have a clear visibility off their data. How are we do more often assessment from an external point off view on See how critical their data is? Then we help the customers to strategies that right the most important thing is to identify the most important critical herself. Data being the most critical assault for any organization. Identification off the data's key for the customers. Then we help in building a viable operating model to ensure these identified critical assets are secure on monitor dearly so that they are consumed well as well as protected from external threats. Then, as 1/4 step, we try to bring in awareness, toe the people we train them at all levels in the organization. That is a P for people to understand the importance off the residual our cells. And then as 1/5 step, we work as a back up plan in terms of bringing in a very comprehensive and the holistic testing approach on people process as well as in technology. We'll see how the organization can withstand during a crisis time. And finally we do a continuous governance off this data, which is a key right. It is not just a one step process. We set up the environment. We do the initial analysis and set up the strategy on continuously govern this data to ensure that they are not only know managed will secure as well as they also have to meet the compliance requirements off the organization's right. That is where we help organizations toe secure on Meet the regulations off the organizations. As for the privacy laws, >>so >>this is a constant process. It's not on one time effort. We do a constant process because every organization goes towards the digital journey on. They have to face all these as part off the evolving environment on digital journey, and that's where they should be kept ready in terms off. No recovering, rebounding on moving forward if things goes wrong. >>So let's stick on that for a minute, and then I wanna bring yourself into the conversation. So you mentioned compliance and governance. When? When your digital business. Here, as you say, you're a data business. So that brings up issues. Data sovereignty. Uh, there's governance, this compliance. There's things like right to be forgotten. There's data privacy, so many things. These were often kind of afterthoughts for businesses that bolted on, if you will. I know a lot of executives are very much concerned that these air built in on, and it's not a one shot deal. So do you have solutions around compliance and governance? Can you deliver that as a service? Maybe you could talk about some of the specifics there, >>so some of way have offered multiple services. Tow our customers on digital race against. On one of the key service is the data complaints. As a service here we help organizations toe map the key data against the data compliance requirements. Some of the features includes in terms off the continuous discovery off data right, because organizations keep adding on data when they move more digital on helping the helping and understanding the actual data in terms off the residents of data, it could be a heterogeneous data sources. It could be on data basis or it could be even on the data lakes. Or it could be or no even on compromise, all the cloud environment. So identifying the data across the various no heterogeneous environment is very key. Feature off our solution. Once we identify, classify this sensitive data, the data privacy regulations on the traveling laws have to be map based on the business rules. So we define those rules on help map those data so that organizations know how critical their digital assets are. Then we work on a continuous marching off data for anomalies because that's one of the key teachers off the solution, which needs to be implemented on the day to day operational basis. So we're helping monitoring those anomalies off data for data quality management on an ongoing basis. And finally we also bringing the automatic data governance where we can manage the sensory data policies on their data relationships in terms off, mapping on manage their business rules on we drive reputations toe also suggest appropriate actions to the customers. Take on those specific data sets. >>Great. Thank you, Yousef. Thanks for being patient. I want to bring in Iota ho thio discussion and understand where your customers and happiest minds can leverage your data automation capability that you and I have talked about in the past. And I'm gonna be great if you had an example is well, but maybe you could pick it up from there. >>Sure. I mean, at a high level, assertions are clearly articulated. Really? Um, Iota, who delivers business agility. So that's by, um, accelerating the time to operationalize data, automating, putting in place controls and ultimately putting, helping put in place digital resilience. I mean, way if we step back a little bit in time, um, traditional resilience in relation to data are often met manually, making multiple copies of the same data. So you have a DB A. They would copy the data to various different places on business. Users would access it in those functional style owes. And of course, what happened was you ended up with lots of different copies off the same data around the enterprise. Very inefficient. Onda course ultimately, uh, increases your risk profile. Your risk of a data breach. Um, it's very hard to know where everything is, and I realized that expression they used David, the idea of the forced march to digital. So with enterprises that are going on this forced march, what they're finding is they don't have a single version of the truth, and almost nobody has an accurate view of where their critical data is. Then you have containers bond with containers that enables a big leap forward so you could break applications down into micro services. Updates are available via a P I s. And so you don't have the same need to build and to manage multiple copies of the data. So you have an opportunity to just have a single version of the truth. Then your challenge is, how do you deal with these large legacy data states that the service has been referring Thio, where you you have toe consolidate, and that's really where I Tahoe comes in. Um, we massively accelerate that process of putting in a single version of the truth into place. So by automatically discovering the data, um, discovering what's duplicate what's redundant, that means you can consolidate it down to a single trusted version much more quickly. We've seen many customers have tried to do this manually, and it's literally taken years using manual methods to cover even a small percentage of their I T estates with a tire. You could do it really very quickly on you can have tangible results within weeks and months. Um, and then you can apply controls to the data based on context. So who's the user? What's the content? What's the use case? Things like data quality validations or access permissions on. Then once you've done there, your applications and your enterprise are much more secure, much more resilient. As a result, you've got to do these things whilst retaining agility, though. So coming full circle. This is where the partnership with happiest minds really comes in as well. You've got to be agile. You've gotta have controls, um, on you've got a drug towards the business outcomes and it's doing those three things together that really deliver for the customer. Thank >>you. Use f. I mean you and I. In previous episodes, we've looked in detail at the business case. You were just talking about the manual labor involved. We know that you can't scale, but also there's that compression of time. Thio get to the next step in terms of ultimately getting to the outcome and we talked to a number of customers in the Cube. And the conclusion is really consistent that if you could accelerate the time to value, that's the key driver reducing complexity, automating and getting to insights faster. That's where you see telephone numbers in terms of business impact. So my question is, where should customers start? I mean, how can they take advantage of some of these opportunities that we've discussed >>today? Well, we've tried to make that easy for customers. So with our Tahoe and happiest minds, you can very quickly do what we call a data health check on. Dis is a is a 2 to 3 weeks process are two Really quickly start to understand and deliver value from your data. Um, so, iota, who deploys into the customer environment? Data doesn't go anywhere. Um, we would look at a few data sources on a sample of data Onda. We can very rapidly demonstrate how date discovery those catalog e understanding Jupiter data and redundant data can be done. Um, using machine learning, um, on how those problems can be solved. Um, and so what we tend to find is that we can very quickly as I say in a matter of a few weeks, show a customer how they could get toe, um, or Brazilian outcome on. Then how they can scale that up, take it into production on, then really understand their data state Better on build resilience into the enterprise. >>Excellent. There you have it. We'll leave it right there. Guys. Great conversation. Thanks so much for coming on the program. Best of luck to you in the partnership. Be well. >>Thank you, David. Sorry. Thank you. Thank >>you for watching everybody, This is Dave Volonte for the Cuban Are ongoing Siris on data Automation without Tahoe.

Published Date : Jan 27 2021

SUMMARY :

Great to have you in the Cube. But talk about your mission at the company. digital born a child company. I t services company in the great places to work serving hour glass to ratings mission on the culture. What do you what's your day to day focus To the stakeholders within those businesses on dis is all a key part of digital of the partnership with happiest minds. So when you combine our emphasis I sometimes called the forced march to become a digital business. So one of the key things that is where the digital resilience with business community process enabled was very putting data at the core, I like to say, but so I wonder if you could talk a little bit more about maybe for the first step is to identify the critical data. They have to face all these as part off the evolving environment So do you have solutions around compliance and governance? So identifying the data across the various no heterogeneous is well, but maybe you could pick it up from there. So by automatically discovering the data, um, And the conclusion is really consistent that if you could accelerate the time to value, So with our Tahoe and happiest minds, you can very quickly do what we call Best of luck to you in the partnership. Thank you. you for watching everybody, This is Dave Volonte for the Cuban Are ongoing Siris on data Automation without

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Yusuf KhanPERSON

0.99+

Yusef KhanPERSON

0.99+

2QUANTITY

0.99+

LondonLOCATION

0.99+

Suresh KanniappanPERSON

0.99+

YousefPERSON

0.99+

one stepQUANTITY

0.99+

Dave VolontePERSON

0.99+

first stepQUANTITY

0.99+

2011DATE

0.99+

1/5 stepQUANTITY

0.99+

4.1QUANTITY

0.99+

YussefPERSON

0.99+

Iot TahoeORGANIZATION

0.99+

bothQUANTITY

0.99+

both businessesQUANTITY

0.98+

oneQUANTITY

0.98+

twoQUANTITY

0.98+

fiveQUANTITY

0.98+

singleQUANTITY

0.98+

DavePERSON

0.98+

1/6QUANTITY

0.98+

todayDATE

0.97+

3 weeksQUANTITY

0.97+

SureshPERSON

0.97+

JupiterLOCATION

0.96+

Io TahoeORGANIZATION

0.96+

one shotQUANTITY

0.96+

single versionQUANTITY

0.96+

RussiaLOCATION

0.96+

1/4 stepQUANTITY

0.96+

FirstQUANTITY

0.96+

SirisTITLE

0.96+

TahoePERSON

0.94+

CubeORGANIZATION

0.93+

IotaORGANIZATION

0.92+

day oneQUANTITY

0.9+

one timeQUANTITY

0.88+

Iota HoORGANIZATION

0.87+

three thingsQUANTITY

0.85+

BrazilianOTHER

0.84+

Tier oneQUANTITY

0.84+

forcedEVENT

0.82+

Shirish CountyLOCATION

0.81+

SeoPERSON

0.81+

CubanOTHER

0.81+

TahoeORGANIZATION

0.73+

BeanPERSON

0.72+

IotaTITLE

0.69+

pandemicEVENT

0.67+

U. S. SalesORGANIZATION

0.66+

top 25 I tQUANTITY

0.64+

ThioPERSON

0.61+

IoORGANIZATION

0.57+

IndianOTHER

0.55+

teachersQUANTITY

0.55+

AndiPERSON

0.54+

minuteQUANTITY

0.53+

CEOPERSON

0.52+

OndaLOCATION

0.51+

CubeCOMMERCIAL_ITEM

0.45+

serviceQUANTITY

0.45+

marchEVENT

0.44+

ninetiesDATE

0.41+

Fernando Brandao, AWS & Richard Moulds, AWS Quantum Computing | AWS re:Invent 2020


 

>>From around the globe. It's the cube with digital coverage of AWS reinvent 2020, sponsored by Intel and AWS. >>Welcome back to the queue. It's virtual coverage of Avis reinvent 2020 I'm John furry, your host. Um, this is a cute virtual we're here. Not in, in remote. We're not in person this year, so we're doing the remote interviews. And then this segment is going to build on the quantum conversation we had last year, Richard moles, general manager of Amazon bracket and aid was quantum computing and Fernando Brandao head of quantum algorithms at AWS and Brent professor of theoretical physics at Caltech. Fernando, thanks for coming on, Richard. Thanks for joining us. >>You're welcome to be here. >>So, Fernando, first of all, love your title, quantum algorithms. That's the coolest title I've heard so far and you're pretty smart because you're a theoretical professor of physics at Caltech. So, um, which I'd never be able to get into, but I wish I could get into there someday, but, uh, thanks for coming on. Um, quantum has been quite the rage and you know, there's a lot of people talking about it. Um, it's not ready for prime time. Some say it's moving faster than others, but where are we on quantum right now? What are, what are you, what are you seeing Fernanda where the quantum, where are peg us in the evolution of, of, uh, where we are? >>Um, yeah, what quantum, uh, it's an emerging and rapidly developing fields. Uh, but we are see where are you on, uh, both in terms of, uh, hardware development and in terms of identifying the most impactful use cases of one company. Uh, so, so it's, it's, it's early days for everyone and, and we have like, uh, different players and different technologies that are being sport. And I think it's, it's, it's early, but it's exciting time to be doing quantum computing. And, uh, and it's very interesting to see the interest in industry growing and, and customers. Uh, for example, Casa from AWS, uh, being, uh, being willing to take part in this journey with us in developmental technology. >>Awesome. Richard, last year we talked to bill Vass about this and he was, you know, he set expectations really well, I thought, but it was pretty much in classic Amazonian way. You know, it makes the announcement a lot of progress then makes me give us the update on your end. You guys now are shipping brackets available. What's the update on your end and Verner mentioned in his keynote this week >> as well. Yeah, it was a, it was great until I was really looking at your interview with bill. It was, uh, that was when we launched the launch the service a year ago, almost exactly a year ago this week. And we've come a long way. So as you mentioned, we've, uh, we've, uh, we've gone to general availability with the service now that that happened in August. So now a customer can kind of look into the, uh, to the bracket console and, uh, installed programming concept computers. You know, there's, uh, there's tremendous excitement obviously, as, as you mentioned, and Fernando mentioned, you know, quantum computers, uh, we think >>Have the potential to solve problems that are currently, uh, uh, unsolvable. Um, the goal of bracket is to fundamentally give customers the ability to, uh, to go test, uh, some of those notions to explore the technology and to just start planning for the future. You know, our goal was always to try and solve some of the problems that customers have had for, you know, gee, a decade or so now, you know, they tell us from a variety of different industries, whether it's drug discovery or financial services, whether it's energy or there's chemical engineering, machine learning, you know, th the potential for quantum computer impacts may industries could potentially be disruptive to those industries. And, uh, it's, it's essential that customers can can plan for the future, you know, build their own internal resources, become experts, hire the right staff, figure out where it might impact their business and, uh, and potentially disrupt. >>So, uh, you know, in the past they're finding it hard to, to get involved. You know, these machines are very different, different technologies building in different ways of different characteristics. Uh, the tooling is very disparate, very fragmented. Historically, it's hard for companies to get access to the machines. These tend to be, you know, owned by startups or in, you know, physics labs or universities, very difficult to get access to these things, very different commercial models. Um, and, uh, as you, as you suggested, a lot of interests, a lot of hype, a lot of claims in the industry, customers want to cut through all that. They want to understand what's real, uh, what they can do today, uh, how they can experiment and, uh, and get started. So, you know, we see bracket as a catalyst for innovation. We want to bring together end-users, um, consultants, uh, software developers, um, providers that want to host services on top of bracket, try and get the industry, you know, rubbing along them. You spoke to lots of Amazonians. I'm sure you've heard the phrase innovation flywheel, plenty of times. Um, we see the same approach that we've used successfully in IOT and robotics and machine learning and apply that same approach to content, machine learning software, to quantum computing, and to learn, to bring it together. And, uh, if we get the tooling right, and we make it easy, um, then we don't see any reason why we can't, uh, you know, rapidly try and move this industry forward. And >>It was fun areas where there's a lot of, you know, intellectual computer science, um, technology science involved in super exciting. And Amazon's supposed to some of that undifferentiated heavy. >>That's what I am, you know, it's like, >>There's a Maslow hierarchy of needs in the tech industry. You know, people say, Oh, why five people freak out when there's no wifi? You know, you can't get enough compute. Right. So, you know, um, compute is one of those things with machine learning is seeing the benefits and quantum there's so much benefits there. Um, and you guys made some announcements at, at re-invent, uh, around BRACA. Can you share just quickly share some of those updates, Richard? >>Sure. I mean, it's the way we innovate at AWS. You know, we, we start simple and we, and we build up features. We listen to customers and we learn as we go along, we try and move as quickly as possible. So since going public in, uh, in, in August, we've actually had a string of releases, uh, pretty consistent, um, delivering new features. So we try to tie not the integration with the platform. Customers have told us really very early on that they, they don't just want to play with the technology. They want to figure out how to, how to envisage a production quantum computing service, how it might look, you know, in the context of a broad cloud platform with AWS. So we've, uh, we launched some integration with, uh, other AWS capabilities around security, managing limits, quotas, tagging resources, that type of thing, things that are familiar to, uh, to, to, to current AWS users. >>Uh, we launched some new hardware. Uh, all of our partners D-Wave launched some, uh, uh, you know, a 5,000 cubit machine, uh, just in September. Uh, so we made that available on bracket the same day that they launched that hardware, which was very cool. Um, you know, we've made it, uh, we've, we've made it easier for researchers. We've been, you know, impressed how many academics and researchers have used the service, not just large corporations. Um, they want to have really deep access to these machines. They want to program these things at a low level. So we launched some features, uh, to enable them to do their research, but reinvent, we were really focused on two things, um, simulators and making it much easier to use, uh, hybrid systems systems that, uh, incorporate classical compute, traditional digital computing with quantum machinery, um, in the vein that follow some of the liens that we've seen, uh, in machine learning. >>So, uh, simulators are important. They're a very important part of, uh, learning how to use concepts, computers. They're always available 24, seven they're super convenient to use. And of course they're critical in verifying the accuracy of the results that we get from quantum hardware. When we launched the service behind free simulator for customers to help debug their circuits and experiments quickly, um, but simulating large experiments and large systems is a real challenge on classical computers. You know, it, wasn't hard on classical. Uh, then you wouldn't need a quantum computer. That's the whole point. So running large simulations, you know, is expensive in terms of resources. It's complicated. Uh, we launched a pretty powerful simulator, uh, back in August, which we thought at the time was always powerful managed. Quantum stimulates circuit handled 34 cubits, and it reinvented last week, we launched a new simulator, which actually the first managed simulator to use tensor network technology. >>And it can run up to 50 cubits. So we think is, we think is probably the most powerful, uh, managed quantum simulator on the market today. And customers can flip easily between either using real quantum hardware or either of our, uh, stimulators just by changing a line of code. Um, the other thing we launched was the ability to run these hybrid systems. You know, quantum computers will get more, no don't get onto in a moment is, uh, today's computers are very imperfect, you know, lots of errors. Um, we working, obviously the industry towards fault-tolerant machines and Fernando can talk about some research papers that were published in that area, but right now the machines are far from perfect. And, uh, and the way that we can try to squeeze as much value out of these devices today is to run them in tandem with classical systems. >>We think of the notion of a self-learning quantum algorithm, where you use a classical optimization techniques, such as we see machine learning to tweak and tune the parameters of a quantum algorithm to try and iterate and converge on the best answer and try and overcome some of these issues surrounding errors. That's a lot of moving parts to orchestrate for customers, a lot of different systems, a lot of different programming techniques. And we wanted to make that much easier. We've been impressed with a, a, an open projects, been around for a couple of years, uh, called penny lane after the Beatles song. And, um, so we wanted to double down on that. We were getting a lot of positive feedback from customers about the penny lane talk it, so we decided to, uh, uh, make it a first class citizen on bracket, make it available as a native feature, uh, in our, uh, in our Jupiter notebooks and our tutorials learning examples, um, that open source project has very similar, um, guiding principles that we do, you know, it's open, it's cross platform, it's technology agnostic, and we thought he was a great fit to the service. >>So we, uh, we announced that and made it available to customers and, uh, and, and, uh, already getting great feedback. So, uh, you know, finishing the finishing the year strongly, I think, um, looking forward to 2021, you know, looking forward to some really cool technology it's on the horizon, uh, from a hardware point of view, making it easy to use, um, you know, and always, obviously trying to work back from customer problems. And so congratulations on the success. I'm sure it's not hard to hire people interested, at least finding qualified people it'd be different, but, you know, sign me up. I love quantum great people, Fernando real quick, understanding the relationship with Caltech unique to Amazon. Um, tell us how that fits into the, into this, >>Uh, right. John S no, as I was saying, it's it's early days, uh, for, for quantum computing, uh, and to make progress, uh, in abreast, uh, put together a team of experts, right. To work both on, on find new use cases of quantum computing and also, uh, building more powerful, uh, quantum hardware. Uh, so the AWS center for quantum computing is based at Caltech. Uh, and, and this comes from the belief of AWS that, uh, in quantum computing is key to, uh, to keep close, to stay close of like fresh ideas and to the latest scientific developments. Right. And Caltech is if you're near one computing. So what's the ideal place for doing that? Uh, so in the center, we, we put together researchers and engineers, uh, from computer science, physics, and other subjects, uh, from Amazon, but also from all the academic institutions, uh, of course some context, but we also have Stanford and university of Chicago, uh, among others. So we broke wrongs, uh, in the beauty for AWS and for quantum computer in the summer, uh, and under construction right now. Uh, but, uh, as we speak, John, the team is busy, uh, uh, you know, getting stuff in, in temporary lab space that we have at cottage. >>Awesome. Great. And real quick, I know we've got some time pressure here, but you published some new research, give a quick a plug for the new research. Tell us about that. >>Um, right. So, so, you know, as part of the effort or the integration for one company, uh, we are developing a new cubix, uh, which we choose a combination of acoustic and electric components. So this kind of hybrid Aquacel execute, it has the promise for a much smaller footprint, think about like a few microliters and much longer storage times, like up to settlements, uh, which, which is a big improvement over the scale of the arts sort of writing all export based cubits, but that's not the whole story, right? On six, if you have a good security should make good use of it. Uh, so what we did in this paper, they were just put out, uh, is, is a proposal for an architecture of how to build a scalable quantum computer using these cubits. So we found from our analysis that we can get more than a 10 X overheads in the resources required from URI, a universal thought around quantum computer. >>Uh, so what are these resources? This is like a smaller number of physical cubits. Uh, this is a smaller footprint is, uh, fewer control lines in like a smaller approach and a consistent, right. And, and these are all like, uh, I think this is a solid contribution. Uh, no, it's a theoretical analysis, right? So, so the, uh, the experimental development has to come, but I think this is a solid contribution in the big challenge of scaling up this quantum systems. Uh, so, so, so John, as we speak like, uh, data blessed in the, for quantum computing is, uh, working on the experimental development of this, uh, a highly adequacy architecture, but we also keep exploring other promising ways of doing scalable quantum computers and eventually, uh, to bring a more powerful computer resources to AWS customers. >>It's kind of like machine learning and data science, the smartest people work on it. Then you democratize that. I can see where this is going. Um, Richard real quick, um, for people who want to get involved and participate or consume, what do they do? Give us the playbook real quick. Uh, so simple, just go to the AWS console and kind of log onto the, to the bracket, uh, bracket console, jump in, you know, uh, create, um, create a Jupiter notebook, pull down some of our sample, uh, applications run through the notebook and program a quantum computer. It's literally that simple. There's plenty of tutorials. It's easy to get started, you know, classic cloud style right now from commitment. Jump in, start simple, get going. We want you to go quantum. You can't go back, go quantum. You can't go back to regular computing. I think people will be running concert classical systems in parallel for quite some time. So yeah, this is the, this is definitely not a one way door. You know, you go explore quantum computing and see how it fits into, uh, >>You know, into the, into solving some of the problems that you wanted to solve in the future. But definitely this is not a replacement technology. This is a complimentary technology. >>It's great. It's a great innovation. It's kind of intoxicating technically to get, think about the benefits Fernando, Richard, thanks for coming on. It's really exciting. I'm looking forward to keeping up keeping track of the progress. Thanks for coming on the cube coverage of reinvent, quantum computing going the next level coexisting building on top of the shoulders of other giant technologies. This is where the computing wave is going. It's different. It's impacting people's lives. This is the cube coverage of re-invent. Thanks for watching.

Published Date : Dec 16 2020

SUMMARY :

It's the cube with digital coverage of AWS And then this segment is going to build on the quantum conversation we had last Um, quantum has been quite the rage and you know, Uh, but we are see where are you on, uh, both in terms of, uh, hardware development and Richard, last year we talked to bill Vass about this and he was, you know, he set expectations really well, there's, uh, there's tremendous excitement obviously, as, as you mentioned, and Fernando mentioned, Have the potential to solve problems that are currently, uh, uh, unsolvable. So, uh, you know, in the past they're finding it hard to, to get involved. It was fun areas where there's a lot of, you know, intellectual computer science, So, you know, um, compute is one of those things how it might look, you know, in the context of a broad cloud platform with AWS. uh, uh, you know, a 5,000 cubit machine, uh, just in September. So running large simulations, you know, is expensive in terms of resources. And, uh, and the way that we can try to you know, it's open, it's cross platform, it's technology agnostic, and we thought he was a great fit to So, uh, you know, finishing the finishing the year strongly, but also from all the academic institutions, uh, of course some context, but we also have Stanford And real quick, I know we've got some time pressure here, but you published some new research, uh, we are developing a new cubix, uh, which we choose a combination of acoustic So, so the, uh, the experimental development has to come, to the bracket, uh, bracket console, jump in, you know, uh, create, You know, into the, into solving some of the problems that you wanted to solve in the future. It's kind of intoxicating technically to get, think about the benefits Fernando,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Fernando BrandaoPERSON

0.99+

AWSORGANIZATION

0.99+

RichardPERSON

0.99+

CaltechORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

Richard MouldsPERSON

0.99+

SeptemberDATE

0.99+

John SPERSON

0.99+

JohnPERSON

0.99+

FernandoPERSON

0.99+

BrentPERSON

0.99+

AugustDATE

0.99+

last weekDATE

0.99+

VernerPERSON

0.99+

2021DATE

0.99+

StanfordORGANIZATION

0.99+

sixQUANTITY

0.99+

last yearDATE

0.99+

last yearDATE

0.99+

34 cubitsQUANTITY

0.99+

a year agoDATE

0.99+

firstQUANTITY

0.99+

five peopleQUANTITY

0.99+

IntelORGANIZATION

0.99+

FernandaPERSON

0.98+

5,000 cubitQUANTITY

0.98+

todayDATE

0.98+

two thingsQUANTITY

0.98+

bothQUANTITY

0.97+

oneQUANTITY

0.97+

this weekDATE

0.96+

sevenQUANTITY

0.96+

D-WaveORGANIZATION

0.95+

Richard molesPERSON

0.95+

this yearDATE

0.95+

bill VassPERSON

0.94+

up to 50 cubitsQUANTITY

0.94+

24QUANTITY

0.93+

one wayQUANTITY

0.93+

a year ago this weekDATE

0.89+

AquacelORGANIZATION

0.89+

Avis reinvent 2020TITLE

0.88+

one companyQUANTITY

0.87+

BeatlesORGANIZATION

0.86+

AWS Quantum ComputingORGANIZATION

0.8+

BRACALOCATION

0.76+

a decadeQUANTITY

0.76+

computingEVENT

0.75+

couple of yearsQUANTITY

0.75+

10 XQUANTITY

0.74+

more thanQUANTITY

0.73+

re:Invent 2020TITLE

0.62+

playbookCOMMERCIAL_ITEM

0.62+

JupiterORGANIZATION

0.6+

waveEVENT

0.55+

ChicagoLOCATION

0.54+

MaslowORGANIZATION

0.52+

pennyTITLE

0.49+

A Brief History of Quasi Adaptive NIZKs


 

>>Hello, everyone. This is not appropriate to lapse of America. I'm going to talk about the motivation. For zero knowledge goes back to the heart off, winding down identity, ownership, community and control. Much of photography exists today to support control communications among individuals in the one world. We also consider devices as extensions of individuals and corporations as communities. Here's hoping you're not fit in this picture. What defines the boundary off an individual is the ability to hold a secret with maybe, it says, attached to the ownership. Off some ethic, we want the ability to use the secret to prove ownership of this asset. However, giving up the secret itself essentially announced ownership since then, anybody else can do the same. Dear Knowledge gives us tools to prove ownership without revealing the secret. The notion of proving ownership off a digital object without revealing it sounds very paradoxical outside the model off. So it gives us a surprise when this motion was formalized and constructed by Goldwasser Miccoli and back off in the late eighties, we'll focus on the non interactive >>version of Siri, a knowledge our music in the >>stock, which was first developed by blow Tillman and Peggy, where the general it can span multiple rounds of communications music only allows a single message to be trusted. No, let's get into some technical details for musics. The objective of for music is to show that an object X, which you can think off as the public footprint, often asset, belonging clan and the language without revealing its witness. W, which you can think off as the Future Analytics team consists off three algorithms, video proof and very. The key generation process is executed by a trusted third party and the very opposite, resulting in a common >>random string, or steers, which is made public. The >>true vendor produces a proof by based on the CIA's X and the very fine with the checks. The proof against X and accepts or rejects music off course has to satisfy some properties. We needed to be correct, which basically says that when everyone follows the protocol correctly on, so we can expect, we need to be thought, which says that a false statement cannot be proven. The channel is a trickier properly to form this. How do we capture the intuition behind saying that the proof there is no knowledge of the witness. One way to capture that is to imagine their tools is the real world where the proof is calculated. Using the witness on there's a simulation worth where the proof is calculated without a witness. To make this possible, the simulator may have some extra information about the CIA's, which is independent off the objectives. The property then requires that it is not possible to effectively distinguish these words Now. It is especially challenging to construct music's compared to encryption signature schemes, in particular in signature schemes. The analog off the Hoover can use a secret, and in any case, the analog off the very fire can use a secret. But in is it's none of the crew layer and the verifier can hold a secret. Yeah, in this talk, I'm going to focus on linear subspace languages. This class is the basis of hardness. >>Assumptions like GH and deliver >>on has proved extremely useful in crypto constructions. This is how we express DD it and dealing as linear software. We will use additive notation on express the spirit logs as the near group actions on coop elements. You think the syntax we can write down Deitch on dealing Jupiter's very naturally a zoo witness sector times a constant electric so we can view the language as being penetrated by a constant language. Metrics really was hard by many groups in our instructions. What does it mean? S while uh, Standard group allows traditions and explain it off by in your group also allows one modification In such groups, we can state various in yourself facing elections. The DDN is the simplest one. It assumes that sampling a one dimensional space is indistinguishable from something full professional. The decisional linear assumption assumes the theme from tours is three dimensional spaces generalizing the sequence of Presumptions. The scaling the resumption asks to distinguish between gay damaged examples and full it and >>examples from a K plus one national space. >>Right, So I came up with a breakthrough. Is the construction in Europe 2008 in particular? There? Music for many years Off Spaces was the first efficient >>construction based on idiots and gear. Structurally, >>it consisted of two parts Our commitment to the witness Andre question proof part and going how the witness actually corresponds to the object. The number of elements in the proof is linear in the number >>of witnesses on the number of elements in the object. >>The question remains to build even shorter visits. The Sierras itself seemed to provide some scoop Rosa Russo fix. See how that works for an entire class of languages? Maybe there's a way to increase proof efficiency on the cost of having had Taylor Sierra's for each year. This is what motivates quality and after six, where we let the solace depend on the language itself. In particular, we didn't require the discrete logs of the language constants to generate this, Yes, but we did require this constant student generated from witness sample distributions. This still turns out to be sufficient for many applications. The construction achieved a perfect knowledge, which was universally in the sense that the simulator was independent. However, soundness is competition. So here's how the construction differed from roots high at a very high level, the language constants are embedded into the CIA s in such a way that the object functions as it's only so we end up not needing any separate commitment in the perfect sense. Our particular construction also needed fewer elements in the question proof, as there On the flip side, the CIA's blows up quadratic instead of constant. Let's get into the detail construction, which is actually present with this script. Let the language apparently trace by Giovanni tricks with the witness changing over time, we sat down and matrices >>D and B with appropriate damages. >>Then we construct the public series into what C. S. D is meant to be used. By the way. On it is constructed by >>multiplying the language matrix with D and being worse, Sierra's V is the part that is meant to be used by the very fair, and it is constructed using details be on be embedded in teaching. >>Now let's say you're asked to computer proof for a candidate X with fitness number we computed simply as a product of the witness with CSP. The verification of the truth is simply taking with the pairing off the candidate and the proof with the Sierras. Seeming threats is equal to zero. If you look carefully. Sierra's V essentially embedded in G to the kernel of the Matrix, owned by the language metrics here and so to speak. This is what is responsible for the correctness. The zero knowledge property is also straightforward, >>given the trapdoor matrices, D and B. Now, >>when corrected journalism relatively simple to prove proving illnesses strictly The central observation is that, given CSP, there is still enough entropy. >>India and me to >>random I seriously in particular Sierra's we Can we expand it to have an additional component with a random sample from the kernel allows it. This transformation is purely statistical. No, we essentially invented idiots are killing their talent in the era of kernel part in this transform sitting within show that an alleged proof on a bad candidate and we used to distinguish whether a subspace sample was used for a full space >>sample was used at the challenge. The need >>to have the kernel of the language in this city. That's the technical >>reason why we need the language to come from a witness. Sample. >>Uh, let's give a simple illustration >>of the system on a standard Diffie Hellman, which g one with the hardness assumption being idiot. >>So the language is defined by G one elements small D, E and F, with pupils off the phone due to the W. After that ugly, the CIA is is generated as follows example D and >>B from random on Compute Sierra speak as due to the day after the being verse and Sierra's V as G to do to do the big on day two of the video. The >>proof of the pupil >>detail that I do after the bill is computed using W. As Sierra Speed race to the party. I know that this is just a single element in the group. The verification is done by bearing the Cooper and the proof with the Sierras VMS and then checking in quality. The >>similar can easily compute the proof using trapdoors demand without knowing that what we are expecting. People leave a Peter's die and reduce the roof size, the constant under a given independent of the number of witnesses and object dimensions. Finally, at Cryptocurrency 14 we optimize the proof toe, one group >>element under the idiots. In both the works, the theorists was reduced to linear sites. The >>number of bearings needed for ratification was also industry in years. This is the crypto Ford in construction in action, the construction skeleton remains more or less the famous VR turkey. But the core observation was that many of the Sierras elements could were anomaly. Comite. While still >>maintaining some of this, these extra random items are depicted in red in this side. >>This round of combination of the Sierras elements resulted in a reduction of boat, Bruce says, as also the number of clearings required for education in Europe in 2015 kills, and we came up with a beautiful >>interpretation of skill sets based on the concept of small predictive hash functions. >>This slide is oversimplified but illustrated, wanting, uh, this system has four collecting >>puzzle pieces. The goodness of the language metrics okay again and a key Haider when >>the hidden version of the key is given publicly in the Sears. Now, when we have a good object, the pieces fit together nicely into detectable. However, when we have a bad object, the pieces no longer fit and it becomes >>infeasible to come up with convincing. Zero knowledge is demonstrable by giving the key to the simulator on observing that the key is independent of the language metrics. >>Through the years, we have extended >>enhanced not mind to be six system, especially with our collaborators, Masayuki Abby Koko Jr. Born on U. >>N. Based on your visits, we were able to construct very efficient, identity based encryption structure, resulting signatures >>public verifiable CCS, secure encryption, nine signatures, group signatures, authorities, key extremes and so on. >>It has also been gratifying to see the community make leaps and bounces ideas and also use queuing visits in practical limits. Before finishing off, I wanted to talk to you a little bit about >>some exciting activities going on Hyper ledger, which is relevant for photographers. Hyper >>Leisure is an open source community for enterprise. Great. It's hosted by the minute formation on enjoys participation from numerous industry groups. Uh, so difficult funded to efforts in Africa, we have versa, which is poised to be the crypto home for all. Blocking it and practice a platform for prospecting transactions are part of the legs on the slide here, >>we would love participation from entity inference. So >>that was a brief history of your analytics. Thanks for giving me the opportunity. And thanks for listening

Published Date : Sep 21 2020

SUMMARY :

an individual is the ability to hold a secret with maybe, it says, the public footprint, often asset, belonging clan and the language without The is it's none of the crew layer and the verifier can hold a secret. The scaling the resumption asks to distinguish between Is the construction in Europe 2008 construction based on idiots and gear. in the proof is linear in the number the discrete logs of the language constants to generate this, Yes, By the way. Sierra's V is the part that is meant to be used by the very fair, owned by the language metrics here and so to speak. The central observation is that, given CSP, there is still enough entropy. to distinguish whether a subspace sample was used for a full space The need That's the technical reason why we need the language to come from a witness. of the system on a standard Diffie Hellman, which g one with the hardness So the language is defined by G one elements small D, E and F, B from random on Compute Sierra speak as due to the day after the and the proof with the Sierras VMS and then checking in quality. similar can easily compute the proof using trapdoors demand without In both the works, the theorists was reduced to linear This is the crypto Ford in construction in action, the construction skeleton in this side. The goodness of the language metrics okay the hidden version of the key is given publicly in the Sears. giving the key to the simulator on observing that the key is independent enhanced not mind to be six system, especially with our collaborators, N. Based on your visits, we were able to construct very efficient, authorities, key extremes and so on. It has also been gratifying to see the community make leaps and bounces ideas and some exciting activities going on Hyper ledger, which is relevant for photographers. on the slide here, we would love participation from entity inference. Thanks for giving me the opportunity.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
BrucePERSON

0.99+

2015DATE

0.99+

AfricaLOCATION

0.99+

CIAORGANIZATION

0.99+

SiriTITLE

0.99+

EuropeLOCATION

0.99+

Masayuki Abby Koko Jr.PERSON

0.99+

each yearQUANTITY

0.99+

bothQUANTITY

0.99+

firstQUANTITY

0.99+

GiovanniPERSON

0.99+

2008DATE

0.99+

kernelTITLE

0.99+

late eightiesDATE

0.99+

six systemQUANTITY

0.99+

two partsQUANTITY

0.98+

Goldwasser MiccoliPERSON

0.98+

AmericaLOCATION

0.98+

one worldQUANTITY

0.98+

PeterPERSON

0.98+

JupiterLOCATION

0.98+

single elementQUANTITY

0.97+

Diffie HellmanPERSON

0.97+

One wayQUANTITY

0.96+

nine signaturesQUANTITY

0.96+

todayDATE

0.95+

sixQUANTITY

0.93+

SierraTITLE

0.93+

SierraPERSON

0.93+

Rosa RussoPERSON

0.92+

PeggyPERSON

0.92+

SierrasLOCATION

0.91+

oneQUANTITY

0.9+

SearsORGANIZATION

0.89+

zeroQUANTITY

0.88+

one nationalQUANTITY

0.87+

Cryptocurrency 14ORGANIZATION

0.86+

single messageQUANTITY

0.84+

one groupQUANTITY

0.82+

CooperPERSON

0.81+

U.LOCATION

0.78+

day twoQUANTITY

0.76+

Zero knowledgeQUANTITY

0.72+

DeitchPERSON

0.71+

FordORGANIZATION

0.68+

AnalyticsORGANIZATION

0.67+

IndiaLOCATION

0.65+

blowPERSON

0.64+

TillmanPERSON

0.59+

TaylorORGANIZATION

0.55+

one elementsQUANTITY

0.52+

Hyper ledgerTITLE

0.5+

SierraORGANIZATION

0.5+

SierrasTITLE

0.39+

SierrasCOMMERCIAL_ITEM

0.36+

Ajay Vohora, Io-Tahoe | SmartData Marketplaces


 

>> Narrator: From around the globe, it's theCUBE. With digital coverage of smart data marketplaces. Brought to you by Io-Tahoe. >> Digital transformation has really gone from a buzzword to a mandate, but digital business is a data business. And for the last several months we've been working with Io-Tahoe on an ongoing content series, focused on smart data and automation to drive better insights and outcomes, essentially putting data to work. And today we're going to do a deeper dive on automating data discovery. And one of the thought leaders in this space is Ajay Vohora, who's the CEO of Io-Tahoe. Once again, joining me, Ajay good to see you. Thanks for coming on. >> Great to be here, David, thank you. >> So let's, let's start by talking about some of the business realities and what are the economics that are driving automated data discovery? Why is that so important? >> Yeah, on this one, David it's a number of competing factors. We've got the reality of data which may be sensitive. So there's control. Three other elements wanting to drive value from that data to innovation. You can't really drive a lot of value without exchanging data. So the ability to exchange data and to manage those cost overheads and data discovery is at the root of managing that in an automated way to classify that data and set some policies to put that automation in place. >> Yeah, look, we have a picture of this. If we could bring it up guys, cause I want to, Ajay, help the audience understand kind of where data discovery fits in here. This is, as we talked about, this is a complicated situation for a lot of customers. They've got variety of different tools and you've really laid it out nicely here in this diagram. So, take us through sort of where that piece fits. >> Yeah, I mean, we're at the right hand side of this exchange, you know. We're really now in a data driven economy that is everything's connected through APIs that we consume online through mobile apps. And what's not apparent is the chain of activities and tasks that have to go into serving that data to an API at the outset. They may be many legacy systems, technologies, platforms On-premise, in cloud, hybrid, you name it and across those silos, getting to a unified view is the heavy lifting. I think we've seen some, some great impacts that BI tools, such as Power BI, Tableau, Looker, and so on, and Qlik have had, and they're in our ecosystem on visualizing Data and, you know, CEOs, managers, people that are working in companies day-to-day get a lot of value from saying, "What's the real time activity? "What was the trend over this month versus last month?" The tools to enable that, you know, we hear a lot of good things that we're doing with Snowflake, MongoDB on the public Cloud platforms, GCP Azure about enabling building those pipelines to feed into those analytics. But what often gets hidden is how do you source that data that could be locked into a mainframe, a data warehouse, IOT data, and pull over all of that together. And that is the reality of that is it's a lot of heavy lifting. It's hands on work that can be time consuming. And the issue there is that data may have value. It might have potential to have an impact on the top line for a business, on outcomes for consumers, but you're never really sure unless you've done the investigation, discovered it, unified that, and be able to serve that through to other technologies. >> Guys, if you would bring that picture back up again, because Ajay you made a point and I want to land on that for a second. There's a lot of manual curating. An example would be the data catalog. You know, data scientists complain all the time that they're manually wrangling data. And so you're trying to inject automation into the cycle. And then the other piece that I want you to address is the importance of APIs. You really can't do this without an architecture that allows you to connect things together that sort of enables some of the automation. >> Yep, I mean, I'll take that in two parts, David, the APIs, so virtual machines connected by APIs, business rules, and business logic driven by APIs, applications, so everything across the stack from infrastructure down to the network, hardware is all connected through APIs and the work of serving data through to an API, building those pipelines, is often miscalculated, just how much manual effort that takes and that manual effort, we've got a nice list here of what we automate down at the bottom, those tasks of indexing, labeling, mapping across different legacy systems, all of that takes away from the job of a data scientist or data engineer, looking to produce value, monetize data, and to help that business convey to consumers. >> Yeah, it's that top layer that the business sees, of course, there's a lot of work that has to go into achieving that. I want to talk about some of the key tech trends that you're seeing. And one of the things that we talk about a lot is metadata. The importance of metadata, you know, can't be understated. What are some of the big trends that you're seeing metadata and others? >> Yeah, I'll summarize it as five. There's a trend now look at metadata more holistically across the enterprise. And that really makes sense from trying to look across different data silos and apply a policy to manage that data. So that's the control piece. That's that lever. The other side, sometimes competing with that control around sensitive data around managing the cost of data is innovation. Innovation being able to speculate and experiment and try things out where you don't really know what the outcome is if you're a data scientist and engineer, you've got a hypothesis and therefore you've got that tension between control over data and innovation and driving value from it. So enterprise wide metadata management is really helping to unlock where might that latent value be across that sets of data. The other piece is adaptive data governance. Those controls that stick from the data policemen, data stewards, where they're trying to protect the organization, protect the brand, protect consumers data necessary, but in different use cases, you might want to nuance and apply a different policy to govern that data relevant to the context where you might have data that is less sensitive, that can be used for innovation and adapting the style of governance to fit the context is another trend that we're seeing coming up here. A few others is where we're sitting quite extensively in working with automating data discovery. We're now breaking that down into what can we direct? What do we know is a business outcome is a known upfront objective and direct that data discovery to towards that. And that means applying our algorithms around technology and our tools towards solving a known problem. The other one is autonomous data discovery. And that means, you know, trying to allow background processes to understand what changes are happening with data over time, flagging those anomalies. And the reason that's important is when you look over a length of time to see different spikes, different trends and activity, that's really giving a data ops team the ability to manage and calibrate how they're applying policies and controls the data. And the last two, David, that we're seeing is this huge drive towards self-service. So re-imagining how to apply policy data governance into the hands of a data consumer inside a business, or indeed the consumer themselves, to self-service if they're a banking customer or healthcare customer and the policies and the controls and rules, making sure that those are all in place to adaptively serve those data marketplaces that when are involved in creating. >> I want to ask you about the autonomous data discovering, the adaptive data governance, is the problem we're addressing there one of quality, in other words, machines are better than humans are at doing this? Is it one of scale? That humans just don't don't scale that well? Is it both? Can you add some color to that? >> Yeah, honestly, it's the same equation that existed 10 years ago, 20 years ago, it's being exacerbated, but it's that equation of how do I control all the things that I need to protect? How do I enable innovation where it is going to deliver business value? How do I exchange data between a customer, somebody in my supply chain safely, and do all of that whilst managing the fourth leg, which is cost overheads. There's not an open checkbook here. I've got to figure out if I'm the CIO and CDO, how I do all of this within a fixed budget. So those aspects have always been there, now with more choices, infrastructure in the Cloud, API driven applications, On-premises, and that is expanding the choices that a business has and how they put their data to work. It's also then creating a layer of management and data governance that really has to now manage those four aspects, control, innovation, exchange of data, and the cost overhead. >> That top layer of the first slide that we showed was all about the business value. So, I wonder if we could drill into the business impact a little bit. What are your customers seeing specifically in terms of the impact of all this automation on their business? >> Yeah, so we've had some great results. I think a few of the biggest have been helping customers move away from manually curating their data and their metadata. It used to be a time where if data initiatives or data governance initiatives, there'd be teams of people manually feeding a data catalog. And it's great to have that inventory of classified data to be able to understand single version of the truth, but having 10, 15 people manually process that, keep it up to date, when it's moving feet, the reality of it is what's true about data today, add another few sources and a few months time to your business, start collaborating with new partners, suddenly the landscape has changed. The amount of work has gone up, but what we're finding is through automating, creating that data discovery, feeding our data catalog, that's releasing a lot more time for our customers to spend on innovating and managing their data. A couple of others is around self service data analytics, moving the choices of what data might have business value into the hands of business users and data consumers to have faster cycle times around generating insights. And we're really helping them by automating the creation of those data sets that are needed for that. And the last piece, I'd have to say where we're seeing impacts more recently is in the exchange of data. There are a number of marketplaces out there who are now being compelled to become more digital, to rewire their business processes and everything from an RPA initiative to automation involving digital transformation is having CIOs, chief data officers and enterprise architects rethink how do they, how do they rewire the pipelines for their data to feed that digital transformation? >> Yeah, to me, it comes down to monetization. Now, of course, that's for a for-profit industry. For non-profits, for sure, the cost cutting or in the case of healthcare, which we'll talk about in a moment, I mean, it's patient outcomes, but the job of a Chief Data Officer has gone from data quality and governance and compliance to really figuring out how data can be monetized, not necessarily selling the data, but how it contributes to the monetization of the company. And then really understanding specifically for that organization, how to apply that. And that is a big challenge. We sort of chatted about 10 years ago, the early days of a dupe. And then 1% of the companies had enough engineers to figure it out, but now the tooling is available. The technology is there and the practices are there. And that really, to me is the bottom line, Ajay, is it's show me the money. >> Absolutely. It's definitely is focusing in on the single view of that customer and where we're helping there is to pull together those disparate, siloed sources of data to understand what are the needs of the patient, of the broker of the, if it's insurance? What are the needs of the supply chain manager, if it's manufacturing? And providing that 360 view of data is helping to see, helping that individual unlock the value for the business. So data's providing the lens provided, you know which data it is that can assist in doing that. >> And, you know, you mentioned RPA before, I had an RPA customer tell me she was a Six Sigma expert and she told me, "We would never try to apply Six Sigma "to a business process, "but with RPA we can do so very cheaply." Well, what that means is lower costs. It means better employee satisfaction and really importantly, better customer satisfaction and better customer outcomes. Let's talk about healthcare for a minute because it's a really important industry. It's one that is ripe for disruption and has really been, up until recently, pretty slow to adopt a lot of the major technologies that have been made available. But what are you seeing in terms of this theme we're using a putting data to work in healthcare specifically? >> Yeah, I mean, health care's has had a lot thrown at it. There's been a lot of change in terms of legislation recently, particularly in the U.S. market, in other economies, healthcare is on a path to becoming more digital. And part of that is around transparency of price. So, to be operating effectively as a healthcare marketplace, being able to have that price transparency around what an elective procedure is going to cost before taking that step forward. It's super important to have an informed decision around that. So if we look at the U.S., for example, we've seen that healthcare costs annually have risen to $4 trillion, but even with all of that cost, we have healthcare consumers who are reluctant sometimes to take up healthcare even if they have symptoms. And a lot of that is driven through not knowing what they're opening themselves up to. And, you know, I think David, if you or I were to book travel a holiday, maybe, or trip, we'd want to know what we're in for, what we're paying for upfront. But sometimes in healthcare that choice, the option might be the plan, but the cost that comes with it isn't. So recent legislation in the U.S. is certainly helpful to bring forward that price transparency. The underlying issue there though is the disparate different format types of data that are being used from payers, patients, employers, different healthcare departments to try and make that work. And where we're helping on that aspect in particular related to price transparency is to help make that data machine readable. So, sometimes with data, the beneficiary might be a person, but in a lot of cases, now we're seeing the ability to have different systems interact and exchange data in order to process the workflow to generate online lists of pricing from a provider that's been negotiated with a payer is really an enabling factor. >> So guys, I wonder if you could bring up the next slide, which is kind of the nirvana. So, if you saw the previous slide that the middle there was all different shapes and presumably to disparate data, this is the outcome that you want to get, where everything fits together nicely. And you've got this open exchange. It's not opaque as it is today. It's not bubble gum, band-aids and duct tape, but describe this sort of outcome that you're trying to achieve and maybe a little bit about what it's going to take to get there. >> Ajay: Yeah, that that's the culmination of a number of things. It's making sure that the data is machine readable, making it available to APIs, that could be RPA tools. We're working with technology companies that employ RPA for healthcare, and specifically to manage that patient and payer data to bring that together. In our data discovery, what we're able to do is to classify that data and have it made available to a downstream tool technology or person to apply that, that workflow to the data. So this looks like nirvana, it looks like utopia, but it's, you know, the end objective of a journey that we can see in different economies, that are at different stages of maturity in turning healthcare into a digital service even so that you can consume it from where you live, from home with telemedicine and tele care. >> Yeah, so, and this is not just for healthcare, but you know, you want to achieve that self-service data marketplace in virtually any industry. You're working with TCS, Tata Consulting Services to achieve this. You know, a company like Io-Tahoe has to have partnerships with organizations that have deep industry expertise. Talk about your relationship with TCS and what you guys are doing specifically in this regard. >> Yeah, we've been working with TCS now for a long while and we'll be announcing some of those initiatives here where we're now working together to reach their customers where they've got a brilliant framework of business, 4.0, where they're re-imagining with the clients, how their business can operate with AI, with automation and become more agile and digital. Our technology, now, the reams of patients that we have in our portfolio, being able to apply that at scale, on a global scale across industries, such as banking, insurance and healthcare is really allowing us to see a bigger impact on consumer outcomes, patient outcomes. And the feedback from TCS is that we're really helping in those initiatives remove that friction. They talk a lot about data friction. I think that's a polite term for the image that we just saw with the disparate technologies that the legacy that has built up. So if we want to create a transformation, having that partnership with TCS across industries is giving us that reach and that impact on many different people's day-to-day jobs and lives. >> Let's talk a little bit about the Cloud. It's a topic that we've hit on quite a bit here in this content series. But, but you know, the Cloud companies, the big hyper-scalers, they've put everything into the Cloud, right? But customers are more circumspect than that. But at the same time, machine intelligence, ML, AI, the Cloud is a place to do a lot of that. That's where a lot of the innovation occurs. And so what are your thoughts on getting to the Cloud, putting data to work, if you will, with machine learning, stuff that you're doing with AWS, what's your fit there? >> Yeah, we, David, we work with all of the Cloud platforms, Microsoft Azure, GCP, IBM, but we're expanding our partnership now with AWS. And we're really opening up the ability to work with their Greenfield accounts, where a lot of that data, that technology is in their own data centers at the customer. And that's across banking, healthcare, manufacturing, and insurance. And for good reason, a lot of companies that have taken the time to see what works well for them with the technologies that the Cloud providers are offering, and a lot of cases, testing services or analytics using the Cloud to move workloads to the Cloud to drive data analytics is a real game changer. So there's good reason to maintain a lot of systems On-premise. If that makes sense from a cost, from a liability point of view and the number of clients that we work with that do have, and will keep their mainframe systems when in Cobra is no surprise to us, but equally they want to tap into technologies that AWS has such as SageMaker. The issue is as a Chief Data Officer, I didn't have the budget to move everything to the Cloud they want, I might want to show some results first upfront to my business users and work closely with my Chief Marketing Officer to look at what's happening in terms of customer trends and customer behavior> What are the customer outcomes, patient outcomes and partner outcomes that you can achieve through analytics, data science? So, working with AWS and with clients to manage that hybrid topology of some of that data being in the Cloud, being put to work with AWS SageMaker and Io-Tahoe being used to identify where is the data that needs to be amalgamated and curated to provide the dataset for machine learning, advanced analytics to have an impact for the business. >> So what are the critical attributes of what you're looking at to help customers decide what to move and what the keep if you will? >> Well, one of the quickest outcomes that we help customers achieve is to buy that business glossary, you know, that the items of data, that means something to them across those different silos and pull all of that together into a unified view. Once they've got that data engineer working with a business manager to think through, how do we want to create this application? Now, what is the churn model, the loyalty or the propensity model that we want to put in place here? How do we use predictive analytics to understand what needs for a patient that sort of innovation is what we're unlocking, applying a tools such as SageMaker on AWS to then do the computation and to build those models to deliver that outcome is across that value chain. And it goes back to the first picture that we put up, David, you know, the outcome is that API on the back of it, you've got a machine learning model that's been developed in a tool such as Databricks or Jupiter notebook. That data has to be sourced from somewhere. Somebody has to say that, "Yep, "You've got permission to do what you're trying to do without falling foul "of any compliance around data." And it all goes back to discovering that data, classifying it, indexing it in an automated way to cut those timelines down to hours and days. >> Yeah, it's the innovation part of your data portfolio, if you will, that you're going to put into the Cloud, apply tools like SageMaker and others, your tool Azure. I mean, whatever your favorite tool is, you don't care. The customer's going to choose that. And you know, the Cloud vendors, maybe they want you to use their tool, but they're making their marketplaces available to everybody, but it's that innovation piece, the ones that you, where you want to apply that self-service data marketplace to, and really drive, as I said before, monetization, All right, give us your final thoughts. Ajay, bring us home. >> So final thoughts on this, David, is at the moment, we're seeing a lot of value in helping customers discover their data using automation, automatically curating a data catalog. And that unified view is then being put to work through our API is having an open architecture to plug in whatever tool technology our clients have decided to use. And that open architecture is really feeding into the reality of what CIOs and Chief Data Officers are managing, which is a hybrid On-premise Cloud approach to use best of breed. But business users wanting to use a particular technology to get their business outcome, having the flexibility to do that no matter where your data is sitting On-premise, on Cloud is where self-service comes in so that sales service view of what data I can plug together, jive exchange, monetizing that data is where we're starting to see some real traction with customers. Now accelerating, becoming more digital to serve their own customers. >> Yeah, we really have seen a cultural mind shift going from sort of complacency, and obviously COVID has accelerated this, but the combination of that cultural shift, the Cloud machine intelligence tools give me a lot of hope that the promises of big data will ultimately be lived up to in this next 10 years. So Ajay Vohora, thanks so much for coming back on theCUBE. You're a great guest and appreciate your insights. >> Appreciate it, David. See you next time. >> All right, keep it right there, everybody, right back after this short break. (techno music)

Published Date : Sep 17 2020

SUMMARY :

Brought to you by Io-Tahoe. and automation to drive So the ability to exchange data help the audience understand and tasks that have to go into serving is the importance of APIs. all of that takes away from the job that has to go into achieving that. And that means, you know, and that is expanding the choices in terms of the impact And the last piece, I'd have to say And that really, to me is the bottom line, of the broker of the, of the major technologies that choice, the option might be the plan, that the middle there Ajay: Yeah, that that's the culmination has to have partnerships that the legacy that has built up. on getting to the Cloud, of some of that data being in the Cloud, that means something to them to apply that self-service having the flexibility to do that that the promises of big data See you next time. right back after this short break.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Ajay VohoraPERSON

0.99+

TCSORGANIZATION

0.99+

AWSORGANIZATION

0.99+

Io-TahoeORGANIZATION

0.99+

$4 trillionQUANTITY

0.99+

Tata Consulting ServicesORGANIZATION

0.99+

fiveQUANTITY

0.99+

IBMORGANIZATION

0.99+

two partsQUANTITY

0.99+

bothQUANTITY

0.99+

MicrosoftORGANIZATION

0.99+

first pictureQUANTITY

0.99+

fourth legQUANTITY

0.99+

AjayPERSON

0.99+

Io-TahoePERSON

0.99+

oneQUANTITY

0.99+

20 years agoDATE

0.99+

U.S.LOCATION

0.99+

10 years agoDATE

0.98+

Three other elementsQUANTITY

0.98+

360 viewQUANTITY

0.98+

1%QUANTITY

0.98+

last monthDATE

0.98+

first slideQUANTITY

0.97+

todayDATE

0.97+

Power BITITLE

0.97+

CobraLOCATION

0.96+

DatabricksORGANIZATION

0.96+

10, 15 peopleQUANTITY

0.96+

single viewQUANTITY

0.95+

Six SigmaORGANIZATION

0.95+

GCP AzureTITLE

0.95+

single versionQUANTITY

0.94+

CloudTITLE

0.94+

TableauTITLE

0.92+

AzureTITLE

0.88+

MongoDBTITLE

0.86+

about 10 years agoDATE

0.84+

COVIDTITLE

0.83+

firstQUANTITY

0.82+

SnowflakeTITLE

0.81+

GCPORGANIZATION

0.81+

twoQUANTITY

0.81+

LookerTITLE

0.79+

SageMakerTITLE

0.78+

GreenfieldORGANIZATION

0.78+

next 10 yearsDATE

0.74+

Six SigmaTITLE

0.7+

this monthDATE

0.67+

JupiterORGANIZATION

0.63+

QlikTITLE

0.62+

AWS SageMakerORGANIZATION

0.61+

a secondQUANTITY

0.55+

SageMakerORGANIZATION

0.54+

Ajay Vohora 9 9 V1


 

>>from around the globe. It's the Cube with digital coverage of smart data. Marketplace is brought to You by Io Tahoe Digital transformation is really gone from buzzword to a mandate. Additional businesses, a data business. And for the last several months, we've been working with Iot Tahoe on an ongoing content. Serious, serious, focused on smart data and automation to drive better insights and outcomes, essentially putting data to work. And today we're gonna do a deeper dive on automating data Discovery. And one of the thought leaders in this space is a J ahora who is the CEO of Iot. Tahoe's once again joining Me A J Good to see you. Thanks for coming on. >>A great to be here, David. Thank you. >>So let's start by talking about some of the business realities. And what are the economics that air? That air driving, automated data Discovery? Why is that so important? >>Yeah, and on this one, David, it's It's a number of competing factors we've got. The reality is data which may be sensitive, so this control on three other elements are wanting to drive value from that data. So innovation, you can't really drive a lot of value without exchanging data. So the ability to exchange data and to manage those costs, overheads and data discovery is at the roots of managing that in an automated way to classify that data in sets and policies to put that automation in place. >>Yeah. Okay, look, we have a picture of this. We could bring it up, guys, because I want oh, A j help the audience. Understand? Unaware data Discovery fits in here. This is as we talked about this, a complicated situation for a lot of customers. They got a variety of different tools, and you really laid it out nicely here in this diagram. So take us through. Sort of where that he spits. >>Yeah. I mean, where at the right hand side, This exchange. You know, we're really now in a data driven economy that is, everything's connected through AP, eyes that we consume on mine free mobile relapse. And what's not a parent is the chain of activities and tasks that have to go into serving that data two and eight p. I. At the outset, there may be many legacy systems, technologies, platforms on premise and cloud hybrids. You name it. Andi across those silos. Getting to a unified view is the heavy lifting. I think we've seen Cem some great impacts that be I titles such as Power Bi I tableau looker on DSO on in Clear. Who had Andi there in our ecosystem on visualising Data and CEO's managers, people that are working in companies day to day get a lot of value from saying What's the was the real time activity? What was the trend over this month? First his last month. The tools to enable that you know, we here, Um, a lot of good things are work that we're doing with snowflake mongo db on the public cloud platforms gcpd as your, um, about enabling building those pay planes to feed into those analytics. But what often gets hidden is have you sauce that data that could be locked into a mainframe, a data warehouse? I ot data on DPA, though, that all of that together that is the reality of that is it's it's, um, it's a lot of heavy lifting It z hands on what that, um, can be time consuming on the issue There is that data may have value. It might have potential to have an impact on the on the top line for a business on outcomes for consumers. But you never any sure unless you you've done the investigation discovered it unified that Onda and be able to serve that through to other technologies. >>Guys have. You would bring that picture back up again because A. J, you made a point, and I wanna land on that for a second. There's a lot of manual curating. Ah, an example would be the data catalogue if they decide to complain all the time that they're manually wrangling data. So you're trying to inject automation in the cycle, and then the other piece that I want you to addresses the importance of AP eyes. You really can't do this without an architecture that allows you to connect things together. That sort of enables some of the automation. >>Yeah, I mean, I don't take that in two parts. They would be the AP eyes so virtual machines connected by AP eyes, um, business rules and business logic driven by AP eyes applications. So everything across the stack from infrastructure down to the network um, hardware is all connected through AP eyes and the work of serving data three to an MP I Building these pipelines is is often, um, miscalculated. Just how much manual effort that takes and that manual ever. We've got a nice list here of what we automate down at the bottom. Those tasks of indexing, labeling, mapping across different legacy systems. Um, all of that takes away from the job of a data scientist today to engineer it, looking to produce value monetize data on day two to help their business day to conceive us. >>Yes. So it's that top layer that the business sees, of course, is a lot of work that has to go went into achieving that. I want to talk about some of the key tech trends that you're seeing and one of the things that we talked about a lot of metadata at the importance of metadata. It can't be understated. What are some of the big trends that you're seeing metadata and others? >>Yeah, I'll summarize. It is five. There's trains now, look, a metadata more holistically across the enterprise, and that really makes sense from trying. Teoh look across different data silos on apply, um, a policy to manage that data. So that's the control piece. That's that lever the other side's on. Sometimes competing with that control around sense of data around managing the costs of data is innovation innovation, being able to speculate on experiment and trying things out where you don't really know what the outcome is. If you're a data scientist and engineer, you've got a hypothesis. And now, before you got that tension between control over data on innovation and driving value from it. So enterprise wide manage data management is really helping to enough. Where might that latent value be across that sets of data? The other piece is adaptive data governance. Those controls that that that stick from the data policemen on day to steer its where they're trying to protect the organization, protect the brand, protect consumers data is necessary. But in different use cases, you might want to nuance and apply a different policy to govern that data run of into the context where you may have data that is less sensitive. Um, that can me used for innovation. Andi. Adapting the style of governance to fit the context is another trend that we're seeing coming up here. A few others is where we're sitting quite extensively and working with automating data discovery. We're now breaking that down into what can we direct? What do we know is a business outcome is a known up front objective on direct that data discovery to towards that. And that means applying around with Dems run technology and our tools towards solving a known problem. The other one is autonomous data discovery. And that means, you know, trying to allow background processes do winds down what changes are happening with data over time flagging those anomalies. And the reason that's important is when you look over a length of time to see different spikes, different trends and activity that's really giving a day drops team the ability to to manage and calibrate how they're applying policies and controls today. There, in the last two David that we're seeing is this huge drive towards self service so reimagining how to play policy data governance into the hands off, um, a day to consumer inside a business or indeed, the consumer themselves. The South service, um, if their banking customer or healthcare customer and the policies and the controls and rules, making sure that those are all in place to adaptive Lee, um, serve those data marketplaces that, um when they're involved in creating, >>I want to ask you about the autonomous data discovering the adaptive data. Governance is the is the problem where addressing their one of quality. In other words, machines air better than humans are doing this. Is that one of scale that humans just don't don't scale that well, is it? Is it both? Can you add some color to that >>yet? Honestly, it's the same equation that existed 10 years ago, 20 years ago. It's It's being exacerbated, but it's that equation is how do I control both things that I need to protect? How do we enable innovation where it is going to deliver business value? Had to exchange data between a customer, somebody in my supply chains safely. And all of that was managing the fourth that leg, which is cost overheads. You know, there's no no can checkbook here. I've got a figure out. If only see io and CDO how I do all of this within a fixed budget so that those aspects have always been there. Now, with more choices. Infrastructure in the cloud, um, NPR driven applications own promise. And that is expanding the choices that a a business has and how they put mandated what it's also then creating a layer off management and data governance that really has to now, uh, manage those full wrath space control, innovation, exchange of data on the cost overhead. >>That that top layer of the first slide that we showed was all about business value. So I wonder if we could drill into the business impact a little bit. What do your customers seeing you know, specifically in terms of the impact of all this automation on their business? >>Yeah, so we've had some great results. I think view the biggest Have Bean helping customers move away from manually curating their data in their metadata. It used to be a time where for data quality initiatives or data governance initiative that be teams of people manually feeding a data Cavallo. And it's great to have the inventory of classified data to be out to understand single version of the trees. But in a having 10 15 people manually process that keep it up to date when it's moving feet. The reality of it is what's what's true about data today? and another few sources in a few months. Time to your business on start collaborating with new partners. Suddenly the landscape has changed. The amount of work is gonna But the, um, what we're finding is through automating creating that data discovery feeding a dent convoke that's releasing a lot more time for our CAS. Mr Spend on innovating and managing their data. A couple of others is around cell service data and medics moving the the choices of what data might have business value into the hands of business users and and data consumers to They're faster cycle times around generating insights. Um, we really helping that by automating the creation of those those data sets that are needed for that. And in the last piece, I'd have to say where we're seeing impacts. A more recently is in the exchange of data. There are a number of marketplaces out there who are now being compelled to become more digital to rewire their business processes. Andi. Everything from an r p a initiative. Teoh automation involving digital transformation is having, um, see iose Chief data officers Andi Enterprise architects rethink how do they how they re worthy pipelines? But they dated to feed that additional transformation. >>Yeah, to me, it comes down to monetization. Of course, that's for for profit in industry, from if nonprofits, for sure, the cost cutting or, in the case of healthcare, which we'll talk about in a moment. I mean, it's patient outcomes. But you know, the the job of ah, chief data officer has gone from your data quality and governance and compliance to really figuring out how data and be monetized, not necessarily selling the data, but how it contributes for the monetization of the company and then really understanding specifically for that organization how to apply that. And that is a big challenge. We chatted about it 10 years ago in the early days of a Duke. And then, you know, 1% of the companies had enough engineers to figure it out. But now the tooling is available, the technology is there and the the practices air there, and that really to me, is the bottom line. A. J is it says to show me the money. >>Absolutely. It's is definitely then six sing links is focusing in on the saying over here, that customer Onda, where we're helping there is dio go together. Those disparities siloed source of data to understand what are the needs of the patient of the broker of the if it's insurance? Ah, one of the needs of the supply chain manager If its manufacturing onda providing that 3 60 view of data, um is helping to see helping that individual unlock the value for the business. Eso data is providing the lens, provided you know which data it is that can God assist in doing that? >>And you know, you mentioned r p A. Before an r p A customer tell me she was a six Sigma expert and she told me we would never try to apply six segment to a business process. But with our P A. We can do so very cheaply. Well, what that means is lower costs means better employee satisfaction and, really importantly, better customer satisfaction and better customer outcomes. Let's talk about health care for a minute because it's a really important industry. It's one that is ripe for disruption on has really been up until recently, pretty slow. Teoh adopt ah, lot of the major technologies that have been made available, but come, what are you seeing in terms of this theme, we're using a putting data to work in health care. Specific. >>Yeah, I mean, healthcare's Havlat thrown at it. There's been a lot of change in terms of legislation recently. Um, particularly in the U. S. Market on in other economies, um, healthcare ease on a path to becoming more digital on. Part of that is around transparency of price, saying to be operating effectively as a health care marketplace, being out to have that price transparency, um, around what an elective procedure is going to cost before taking that that's that forward. It's super important to have an informed decision around there. So we look at the US, for example. We've seen that health care costs annually have risen to $4 trillion. But even with all of that on cost, we have health care consumers who are reluctant sometimes to take up health care if they even if they have symptoms on a lot of that is driven through, not knowing what they're opening themselves up to. Andi and I think David, if you are, I want to book, travel, holiday, maybe, or trip. We want to know what what we're in for what we're paying for outfront, but sometimes in how okay, that choice, the option might be their plan, but the cost that comes with it isn't so recent legislation in the US Is it certainly helpful to bring for that tryst price, transparency, the underlying issue there? There is the disparity. Different formats, types of data that being used from payers, patients, employers, different healthcare departments try and make that make that work. And when we're helping on that aspect in particular related to track price transparency is to help make that date of machine readable. So sometimes with with data, the beneficiary might be on a person. I've been a lot of cases now we're seeing the ability to have different systems, interact and exchange data in order to process the workflow. To generate online at lists of pricing from a provider that's been negotiated with a payer is, um, is really a neighboring factor. >>So, guys, I wonder if you bring up the next slide, which is kind of the Nirvana. So if you if you saw the previous slide that the middle there was all different shapes and presumably to disparage data, this is that this is the outcome that you want to get. Everything fits together nicely and you've got this open exchange. It's not opaque as it is today. It's not bubble gum band aids and duct tape, but but but described this sort of outcome the trying to achieve and maybe a little bit about what gonna take to get there. >>Yeah, that's a combination of a number of things. It's making sure that the data is machine readable. Um, making it available to AP eyes that could be our ph toes. We're working with technology companies that employ R P. A full health care. I'm specifically to manage that patient and pay a data. Teoh, bring that together in our data Discovery. What we're able to do is to classify that data on having made available to eight downstream tour technology or person to imply that that workflow to to the data. So this looks like nirvana. It looks like utopia. But it's, you know, the end objective of a journey that we can see in different economies there at different stages of maturity, in turning healthcare into a digital service, even so that you could consume it from when you live from home when telling medicine. Intellicast >>Yes, so And this is not just health care but you wanna achieve that self service doing data marketplace in virtually any industry you working with TCS, Tata Consultancy Services Toe Achieve this You know, if you are a company like Iota has toe have partnerships with organizations that have deep industry expertise Talk about your relationship with TCS and what you guys are doing specifically in this regard. >>Yeah, we've been working with TCS now for room for a long while. Andi will be announcing some of those initiatives here where we're now working together to reach their customers where they've got a a brilliant framework of business for that zero when there re imagining with their clients. Um, how their business cause can operate with ai with automation on, become more agile in digital. Um, our technology, the dreams of patients that we have in our portfolio being out to apply that at scale on the global scale across industries such as banking, insurance and health care is is really allowing us to see a bigger impact on consumer outcomes. Patient outcomes And the feedback from TCS is that we're really helping in those initiatives remove that friction. They talk a lot about data. Friction. Um, I think that's a polite term for the the image that we just saw with the disparity technologies that the legacy that has built up. So if we want to create a transformation, Um, having a partnership with TCS across Industries is giving us that that reach and that impacts on many different people's day to day jobs and knives. >>Let's talk a little bit about the cloud. It's It's a topic that we've hit on quite a bit here in this in this content Siri's. But But you know, the cloud companies, the big hyper scale should put everything into the cloud, right? But but customers are more circumspect than that. But at the same time, machine intelligence M. L. A. The cloud is a place to do a lot of that. That's where a lot of the innovation occurs. And so what are your thoughts on getting to the cloud? Ah, putting dated to work, if you will, with machine learning stuff you're doing with aws. What? You're fit there? >>Yeah, we we and David. We work with all of the cloud platforms. Mike stuffed as your G, c p IBM. Um, but we're expanding our partnership now with AWS Onda we really opening up the ability to work with their Greenfield accounts, where a lot of that data that technology is in their own data centers at the customer, and that's across banking, health care, manufacturing and insurance. And for good reason. A lot of companies have taken the time to see what works well for them, with the technologies that the cloud providers ah, are offered a offering in a lot of cases testing services or analytics using the cloud to move workloads to the cloud to drive Data Analytics is is a real game changer. So there's good reason to maintain a lot of systems on premise. If that makes sense from a cost from a liability point of view on the number of clients that we work with, that do have and we will keep their mainframe systems within kobo is is no surprise to us, but equally they want to tap into technologies that AWS have such a sage maker. The issue is as a chief data officer, I don't have the budget to me, everything to the cloud day one, I might want to show some results. First upfront to my business users Um, Onda worked closely with my chief marketing officer to look at what's happening in terms of customer trains and customer behavior. What are the customer outcomes? Patient outcomes and partner at comes I can achieve through analytics data signs. So I, working with AWS and with clients to manage that hybrid topology of some of that data being, uh, in the cloud being put to work with AWS age maker on night, I hope being used to identify where is the data that needs to bay amalgamated and curated to provide the data set for machine learning advanced and medics to have an impact for the business. >>So what are the critical attributes of what you're looking at to help customers decide what what to move and what to keep, if you will. >>Well, what one of the quickest outcomes that we help custom achieve is to buy that business blustery. You know that the items of data that means something to them across those different silos and pour all of that together into a unified view once they've got that for a data engineer working with a a business manager to think through how we want to create this application. There was the turn model, the loyalty or the propensity model that we want to put in place here. Um, how do we use predictive and medics to understand what needs are for a patient, that sort of innovation is what we're looking applying the tools such a sagemaker, uh, night to be west. So they do the the computation and to build those models to deliver the outcome is is across that value chain, and it goes back to the first picture that we put up. David, you know the outcome Is that a P I On the back of it, you've got the machine learning model that's been developed in That's always such as data breaks. But with Jupiter notebook, that data has to be sourced from somewhere. Somebody has to say that yet you've got permission to do what you're trying to do without falling foul of any compliance around data. Um, it'll goes back to discovering that data, classifying it, indexing it in an automated way to cut those timelines down two hours and days. >>Yeah, it's the it's the innovation part of your data portfolio, if you will, that you're gonna put into the cloud. Apply tools like sage maker and others. You told the jury. Whatever your favorite tool is, you don't care. The customer's gonna choose that and hear the cloud vendors. Maybe they want you to use their tool, but they're making their marketplaces available to everybody. But it's it's that innovation piece, the ones that you where you want to apply that self service data marketplace to and really drive. As I said before monetization. All right, give us your final thoughts. A. J bring us home. >>So final thoughts on this David is that at the moment we're seeing, um, a lot of value in helping customers discover that day the using automation automatically curating a data catalogue, and that unified view is then being put to work through our A B. I's having an open architecture to plug in whatever tool technology our clients have decided to use, and that open architecture is really feeding into the reality of what see Iose in Chief Data Officers of Managing, which is a hybrid on premise cloud approach. Do you suppose to breed Andi but business users wanting to use a particular technology to get their business outcome having the flexibility to do that no matter where you're dating. Sitting on Premise on Cloud is where self service comes in that self service. You of what data I can plug together, Dr Exchange. Monetizing that data is where we're starting to see some real traction. Um, with customers now accelerating becoming more digital, uh, to serve their own customers, >>we really have seen a cultural mind shift going from sort of complacency. And obviously, cove, it has accelerated this. But the combination of that cultural shift the cloud machine intelligence tools give give me a lot of hope that the promises of big data will ultimately be lived up to ah, in this next next 10 years. So a J ahora thanks so much for coming back on the Cube. You're you're a great guest. And ah, appreciate your insights. >>Appreciate, David. See you next time. >>All right? And keep it right there. Very right back. Right after this short break

Published Date : Sep 9 2020

SUMMARY :

And for the last several months, we've been working with Iot Tahoe on an ongoing content. A great to be here, David. So let's start by talking about some of the business realities. So the ability to exchange and you really laid it out nicely here in this diagram. tasks that have to go into serving that data two and eight p. addresses the importance of AP eyes. So everything across the stack from infrastructure down to the network um, What are some of the big trends that you're the costs of data is innovation innovation, being able to speculate Governance is the is and data governance that really has to now, uh, manage those full wrath space control, the impact of all this automation on their business? And in the last piece, I'd have to say where we're seeing in the case of healthcare, which we'll talk about in a moment. Eso data is providing the lens, provided you know Teoh adopt ah, lot of the major technologies that have been made available, that choice, the option might be their plan, but the cost that comes with it isn't the previous slide that the middle there was all different shapes and presumably to disparage into a digital service, even so that you could consume it from Yes, so And this is not just health care but you wanna achieve that self service the image that we just saw with the disparity technologies that the legacy Ah, putting dated to work, if you will, with machine learning stuff A lot of companies have taken the time to see what works well for them, to move and what to keep, if you will. You know that the items of data that means something to The customer's gonna choose that and hear the cloud vendors. the flexibility to do that no matter where you're dating. that cultural shift the cloud machine intelligence tools give give me a lot of hope See you next time. And keep it right there.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Ajay VohoraPERSON

0.99+

AWSORGANIZATION

0.99+

IBMORGANIZATION

0.99+

TCSORGANIZATION

0.99+

MikePERSON

0.99+

AndiPERSON

0.99+

IotaORGANIZATION

0.99+

Iot TahoeORGANIZATION

0.99+

IotORGANIZATION

0.99+

$4 trillionQUANTITY

0.99+

SiriTITLE

0.99+

first pictureQUANTITY

0.99+

fiveQUANTITY

0.99+

two partsQUANTITY

0.99+

two hoursQUANTITY

0.99+

USLOCATION

0.99+

bothQUANTITY

0.99+

Tata Consultancy ServicesORGANIZATION

0.99+

10 years agoDATE

0.99+

first slideQUANTITY

0.99+

1%QUANTITY

0.99+

both thingsQUANTITY

0.98+

FirstQUANTITY

0.98+

oneQUANTITY

0.98+

SpendPERSON

0.98+

fourthQUANTITY

0.97+

U. S.LOCATION

0.97+

IosePERSON

0.97+

todayDATE

0.97+

six sing linksQUANTITY

0.96+

A. JPERSON

0.96+

20 years agoDATE

0.96+

three other elementsQUANTITY

0.96+

day twoQUANTITY

0.95+

twoQUANTITY

0.95+

last monthDATE

0.94+

DukeORGANIZATION

0.93+

zeroQUANTITY

0.91+

covePERSON

0.9+

AndiORGANIZATION

0.9+

OndaORGANIZATION

0.88+

10 15 peopleQUANTITY

0.87+

single versionQUANTITY

0.86+

this monthDATE

0.83+

CDOTITLE

0.83+

eightQUANTITY

0.81+

TahoeORGANIZATION

0.79+

eight downstreamQUANTITY

0.77+

AWS OndaORGANIZATION

0.77+

Io TahoeORGANIZATION

0.77+

next next 10 yearsDATE

0.76+

dayQUANTITY

0.73+

CubeCOMMERCIAL_ITEM

0.72+

six segmentQUANTITY

0.71+

last several monthsDATE

0.71+

threeQUANTITY

0.68+

EsoORGANIZATION

0.6+

Have BeanORGANIZATION

0.6+

3 60 viewQUANTITY

0.58+

JORGANIZATION

0.55+

GreenfieldORGANIZATION

0.49+

ahoraPERSON

0.48+

JPERSON

0.48+

SouthORGANIZATION

0.47+

six SigmaQUANTITY

0.46+

Stuti Deshpande, AWS | Smart Data Marketplaces


 

>> Announcer: From around the globe it's theCUBE with digital coverage of smart data marketplaces brought to you by Io Tahoe. >> Hi everybody, this is Dave Vellante. And welcome back. We've been talking about smart data. We've been hearing Io Tahoe talk about putting data to work and keep heart of building great data outcomes is the Cloud of course, and also Cloud native tooling. Stuti Deshpande is here. She's a partner solutions architect for Amazon Web Services and an expert in this area. Stuti, great to see you. Thanks so much for coming on theCUBE. >> Thank you so much for having me here. >> You're very welcome. So let's talk a little bit about Amazon. I mean, you have been on this machine learning journey for quite sometime. Take us through how this whole evolution has occurred in technology over the period of time. Since the Cloud really has been evolving. >> Amazon in itself is a company, an example of a company that has gotten through a multi year machine learning transformation to become the machine learning driven company that you see today. They have been improvising on original personalization model using robotics to all different women's centers, developing a forecasting system to predict the customer needs and improvising on that and reading customer expectations on convenience, fast delivery and speed, from developing natural language processing technology for end user infraction, to developing a groundbreaking technology such as Prime Air jobs to give packages to the customers. So our goal at Amazon With Services is to take this rich expertise and experience with machine learning technology across Amazon, and to work with thousands of customers and partners to handle this powerful technology into the hands of developers or data engineers of all levels. >> Great. So, okay. So if I'm a customer or a partner of AWS, give me the sales pitch on why I should choose you for machine learning. What are the benefits that I'm going to get specifically from AWS? >> Well, there are three main reasons why partners choose us. First and foremost, we provide the broadest and the deepest set of machine learning and AI services and features for your business. The velocity at which we innovate is truly unmatched. Over the last year, we launched 200 different services and features. So not only our pace is accelerating, but we provide fully managed services to our customers and partners who can easily build sophisticated AI driven applications and utilizing those fully managed services began build and train and deploy machine learning models, which is both valuable and differentiating. Secondly, we can accelerate the adoption of machine learning. So as I mentioned about fully managed services for machine learning, we have Amazon SageMaker. So SageMaker is a fully managed service that are any developer of any level or a data scientist can utilize to build complex machine learning, algorithms and models and deploy that at scale with very less effort and a very less cost. Before SageMaker, it used to take so much of time and expertise and specialization to build all these extensive models, but SageMaker, you can literally build any complex models within just a time of days or weeks. So to increase it option, AWS has acceleration programs just in a solution maps. And we also have education and training programs such as DeepRacer, which are enforces on enforcement learning and Embark, which actually help organization to adopt machine learning very readily. And we also support three major frameworks that just tensive no charge, or they have separate teams who are dedicated to just focus on all these frameworks and improve the support of these frameworks for a wide variety of workloads. And finaly, we provide the most comprehensive platform that is optimized for machine learning. So when you think about machine learning, you need to have a data store where you can store your training sets, your test sets, which is highly reliable, highly scalable, and secure data store. Most of our customers want to store all of their data and any kind of data into a centralized repository that can be treated at the central source of fraud. And in this case from the Amazon Esri data store to build and endurance machine learning workflow. So we believe that we provide this capability of having the most comprehensive platform to build the machine learning workflow from internally. >> Great. Thank you for that. So I wanted, my next question is, this is a complicated situation for a lot of customers. You know, having the technology is one thing, but adoption is sort of everything. So I wonder if you could paint a picture for us and help us understand, how you're helping customers think about machine learning, thinking about that journey and maybe give us the context of what the ecosystem looks like? >> Sure. If someone can put up the belt, I would like to provide a picture representation of how AWS and fusion machine learning as three layers of stack. And moving on to next bill, I can talk about the bottom there. And bottom there as you can see over this screen, it's basically for advanced technologists advanced data scientists who are machine learning practitioners who work at the framework level. 90% of data scientists use multiple frameworks because multiple frameworks are adjusted and are suitable for multiple and different kinds of workloads. So at this layer, we provide support for all of the different types of frameworks. And the bottom layer is only for the advanced scientists and developers who are actually actually want to build, train and deploy these machine learning models by themselves and moving onto the next level, which is the middle layer. This layer is only suited for non-experts. So here we have seen Jamaica where it provides a fully managed service there you can build, tune, train and deploy your machine learning models at a very low cost and with very minimal efforts and at a higher scale, it removes all the complexity, having a thing and guess guesswork from this stage of machine learning and Amazon SageMaker has been the scene that will change. Many of our customers are actually standardizing on top off Amazon SageMaker. And then I'm moving on to the next layer, which is the top most layer. We call this as AI services because this may make the human recognition. So all of the services mentioned here such as Amazon Rekognition, which is basically a deep learning service optimized for image and video analysis. And then we have Amazon Polly, which can do the text to speech from Russian and so on and so forth. So these are the AI services that can be embedded into the application so that the end user or the end customer can build AI driven applications. >> Love it. Okay. So you've got the experts at the bottom with the frameworks, the hardcore data scientists, you kind of get the self driving machine learning in the middle, and then you have all the ingredients. I'm like an AI chef or a machine learning chef. I can pull in vision, speech, chatbots, fraud detection, and sort of compile my own solutions that's cool. We hear a lot about SageMaker studio. I wonder if you could tell us a little bit more, can we double click a little bit on SageMaker? That seems to be a pretty important component of that stack that you just showed us. >> I think that was an absolutely very great summarization of all the different layers of machine unexpected. So thank you for providing the gist of that. Of course, I'll be really happy to talk about Amazon SageMaker because most of our customers are actually standardizing on top of SageMaker. That is spoken about how machine learning traditionally has so many complications and it's very complex and expensive and I traded process, which makes it even harder because they don't know integrated tools or if you do the traditional machine learning all kind of deployment, there are no integrated tools for the entire workflow process and deployment. And that is where SageMaker comes into the picture. SageMaker removes all the heaviness thing and complexities from each step of the deployment of machine learning workflow, how it solves our challenges by providing all of the different components that are optimized for every stage of the workflow into one single tool set. So that models get to production faster and with much less effort and at a lower cost. We really continue to add important (indistinct) leading to Amazon SageMaker. I think last year we announced 50 cubic litres in this far SageMaker being improvised it's features and functionalities. And I would love to call out a couple of those here, SageMaker notebooks, which are just one thing, the prominent notebooks that comes along with easy two instances, I'm sorry for quoting Jarvin here is Amazon Elastic Compute Instances. So you just need to have a one thing deployment and you have the entire SageMaker Notebook Interface, along with the Elastic Compute Instances running that gives you the faster time to production. If you're a machine, if you are a data scientist or a data engineer who worked extensively for machine learning, you must be aware about building training datasets is really complex. So there we have on his own ground truth, that is only for building machine learning training data sets, which can reduce your labeling cost by 70%. And if you perform machine learning and other model technology in general, there are some workflows where you need to do inferences. So there we have inference, Elastic Inference Incense, which you can reduce the cost by 75% by adding a little GP acceleration. Or you can reduce the cost by adding managed squad training, utilizing easy to spot instances. So there are multiple ways that you can reduce the costs and there are multiple ways there you can improvise and speed up your machine, learning deployment and workflow. >> So one of the things I love about, I mean, I'm a prime member who is not right. I love to shop at Amazon. And what I like about it is the consumer experience. It kind of helps me find things that maybe I wasn't aware of, maybe based on other patterns that are going on in the buying community with people that are similar. If I want to find a good book. It's always gives me great reviews and recommendations. So I'm wondering if that applies to sort of the tech world and machine learning, are you seeing any patterns emerge across the various use cases, you have such scale? What can you tell us about that? >> Sure. One of the battles that we have seen all the time is to build scalable layer for any kind of use case. So as I spoke before that as much, I'm really looking to put their data into a single set of depository where they have the single source of truth. So storing of data and any kind of data at any velocity into a single source of would actually help them build models who run on these data and get useful insights out of it. So when you speak about an entry and workflow, using Amazon SageMaker along bigger, scalable analytical tool is actually what we have seen as one of the factors where they can perform some analysis using Amazon SageMaker and build predictive models to say samples, if you want to take a healthcare use case. So they can build a predictive model that can victimize the readmissions of using Amazon SageMaker. So what I mean, to say is, by not moving data around and connecting different services to the same set of source of data, that's tumor avoid creating copies of data, which is very crucial when you are having training data set and test data sets with Amazon SageMaker. And it is highly important to consider this. So the pattern that we have seen is to utilize a central source of depository of data, which could be Amazon Extra. In this scenario, scalable analytical layer along with SageMaker. I would have to code at Intuit for a success story over here. I'm using sandwich, a Amazon SageMaker Intuit had reviews the machine learning deployment time by 90%. So I'm quoting here from six months to one week. And if you think about a healthcare industry, there hadn't been a shift from reactive to predictive care. So utilizing predictive models to accelerate research and discovery of new drugs and new treatments. And you've also observed that nurses were supported by AI tools increase their, their productivity has increased by 50%. I would like to say that one of our customers are really diving deep into the AWS portfolio of machine learning and AI services and including transcribed medical, where they are able to provide some insights so that their customers are getting benefits from them. Most of their customers are healthcare providers and they are able to give some into insights so that they can create some more personalized and improvise patient care. So there you have the end user benefits as well. One of the patterns that I have, I can speak about and what we have seen as well, appearing a predictive model with real time integration into healthcare records will actually help their healthcare provider customers for informed decision making and improvising the personalized patient care. >> That's a great example, several there. And I appreciate that. I mean, healthcare is one of those industries that is just so right for technology ingestion and transformation, that is a great example of how the cloud has really enabled really. I mean, I'm talking about major changes in healthcare with proactive versus reactive. We're talking about lower costs, better health, longer lives is really inspiring to see that evolve. We're going to watch it over the next several years. I wonder if we could close in the marketplace. I've had the pleasure of interviewing Dave McCann, a number of times. He and his team have built just an awesome capability for Amazon and its ecosystem. What about the data products, whether it's SageMaker or other data products in the marketplace, what can you tell us? >> Sure. Either of this market visits are interesting thing. So let me first talk about the AWS marketplace of what, AWS marketplace you can browse and search for hundreds of machine learning algorithms and machine learning, modern packages in a broad range of categories that this company provision, fixed analysis, voice answers, email, video, and it says predictive models and so on and so forth. And all of these models and algorithms can be deployed to a Jupiter notebook, which comes as part of the SageMaker that form. And you can integrate all of these different models and algorithms into our fully managed service, which is Amazon SageMaker to Jupiter notebooks, Sage maker, STK, and even command as well. And this experience is followed by either of those marketplace catalog and API. So you get the same benefits as any other marketplace products, the just seamless deployments and consolidate it. So you get the same benefits as the products and the invest marketplace for your machine learning algorithms and model packages. And this is really important because these can be darkly integrated into our SageMaker platform. And I don't even be honest about the data products as well. And I'm really happy to provide and code one of the example over here in the interest of cooler times and because we are in unprecedented times over here we collaborated with our partners to provide some data products. And one of them is data hub by tablet view that gives you the time series data of phases and depth data gathered from multiple trusted sources. And this is to provide better and informed knowledge so that everyone who was utilizing this product can make some informed decisions and help the community at the end. >> I love it. I love this concept of being able to access the data, algorithms, tooling. And it's not just about the data, it's being able to do something with the data and that we've been talking about injecting intelligence into those data marketplaces. That's what we mean by smart data marketplaces. Stuti Deshpande, thanks so much for coming to theCUBES here, sharing your knowledge and tell us a little bit about AWS. There's a pleasure having you. >> It's my pleasure too. Thank you so much for having me here. >> You're very welcome. And thank you for watching. Keep it right there. We will be right back right after this short break. (soft orchestral music)

Published Date : Sep 3 2020

SUMMARY :

brought to you by Io Tahoe. and keep heart of building in technology over the period of time. and to work with thousands What are the benefits that I'm going to and improve the support of these So I wonder if you could paint So all of the services mentioned here in the middle, and then you So that models get to production faster and machine learning, are you So the pattern that we of how the cloud has and code one of the example And it's not just about the data, Thank you so much for having me here. And thank you for watching.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Dave VellantePERSON

0.99+

Dave McCannPERSON

0.99+

Stuti DeshpandePERSON

0.99+

AWSORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

StutiPERSON

0.99+

90%QUANTITY

0.99+

50%QUANTITY

0.99+

JarvinPERSON

0.99+

75%QUANTITY

0.99+

Amazon Web ServicesORGANIZATION

0.99+

200 different servicesQUANTITY

0.99+

FirstQUANTITY

0.99+

70%QUANTITY

0.99+

one weekQUANTITY

0.99+

six monthsQUANTITY

0.99+

hundredsQUANTITY

0.99+

SageMakerTITLE

0.99+

each stepQUANTITY

0.99+

last yearDATE

0.99+

firstQUANTITY

0.99+

oneQUANTITY

0.99+

JamaicaLOCATION

0.98+

IntuitORGANIZATION

0.98+

bothQUANTITY

0.97+

two instancesQUANTITY

0.97+

SecondlyQUANTITY

0.97+

Io TahoePERSON

0.97+

OneQUANTITY

0.96+

single sourceQUANTITY

0.96+

Prime AirCOMMERCIAL_ITEM

0.94+

one thingQUANTITY

0.92+

todayDATE

0.92+

Elastic ComputeTITLE

0.92+

three main reasonsQUANTITY

0.92+

single setQUANTITY

0.9+

DeepRacerTITLE

0.89+

single toolQUANTITY

0.87+

50 cubic litresQUANTITY

0.87+

Elastic ComputeTITLE

0.86+

RekognitionTITLE

0.86+

Amazon With ServicesORGANIZATION

0.82+

JupiterORGANIZATION

0.81+

three layersQUANTITY

0.79+

SageORGANIZATION

0.78+

Will Nowak, Dataiku | AWS re:Invent 2019


 

>>long from Las Vegas. It's the Q covering a ws re invent 2019. Brought to you by Amazon Web service is and in along with its ecosystem partners. >>Hey, welcome back to the Cube. Lisa Martin at AWS Reinvent 19. This is Day three of the Cubes coverage. We have two sets here. Lots of cute content are joined by Justin Warren, the founder and chief analyst at Pivot nine. Justin. How's it going? Great, right? You still have a voice? Three days? >>Just barely. I've been I've been trying to take care of it. >>Impressed. And you probably have talked to at least half of the 65,000 attendees. >>I'm trying to talk to as many as I can. >>Well, we're gonna talk to another guy here. Joining us from data ICU is well, Novak, the solutions architect will be the Cube. >>Thanks for having me. >>You have a good voice too. After a three day is that you >>have been doing the best I can. >>Yeah, he's good. So did ICU. Interesting name. Let's start off by sharing with our audience. Who did a coup is and what you guys do in technology. >>Yes. So the Entomology of date ICU. It's like hi cooze for data. So we say we take your data and, you know, we make poetry out of it. Make your data so beautiful. Wow, Now, But for those who are unaware Day like it was an enterprise data science platform. Eso we provide a collaborative environment for we say coders and clickers kind of business analyst and native data scientists to make use of organizations, data bill reports and Bill productive machine learning base models and deploy them. >>I'm only the guy's been around around for eight years. Eight years. Okay, >>so start up. Still >>mourning the cloud, the opportunity there That data is no longer a liability. It's an asset or should be. >>So we've been server based from the start, which is one of our differentiators. And so by that we see ourselves as a collaborative platform. Users access it through a Web browser, log into a shared space and share code, can share visual recipes, as we call them to prepare data. >>Okay, so what customers using the platform to do with machine learning is pretty hot at the moment. I think it might be nearing the peak of the life cycle pretty hot. Yeah, what a customer is actually actually doing on the platform, >>you know, So we really focus on enabling the enterprise. So, for example, G has been a customer for some time now, and Sergey is a great prototypical example on that. They have many disparate use cases, like simple things like doing customer segmentation for, you know, marketing campaigns but also stuff like Coyote predicted maintenance. So use cases kind of run the gamut, and so did ICU. Based on open source, we're enabling all of G's users to come into a centralized platform, access their data manipulated for whatever purposes. Maybe >>nobody talked about marketing campaigns for a second. I'm wondering. Are, is their integration with serum technologies? Or how would a customer like wanting to understand customer segmentation or had a segment it for marketing campaign? How would they work in conjunction with a serum and data ICU, for example? >>It's a great question. So again, us being a platform way sit on a single server, something like an Amazon ec2 instance, and then we make connections into an organization's data sources. So if using something like Salesforce weaken seamlessly, pull in data from Salesforce Yuka manipulated in date ICU, but the same time. Maybe also have some excel file someone you know me. I can bring that into my data to work environment. And I also have a red shift data table. All those things would come into the same environment. I can visualize. I can analyze, and I can prepare the data. I see. >>So you tell you it's based on open source? I'm a longtime fan of over. It's always been involved in it for longer than I care to remember. Actually, that's an interesting way t base your product on that. So maybe talk us through how you how you came to found the company based on basic an open source. What? What led to that choice? What? What was that decision based on? >>Yeah, for sure. So you talked about how you know the hype cycle? A. I saw how hot is a I and so I think again, our founders astutely recognize that this is a very fast moving place to be. And so I'm kind of betting on one particular technology can be risky. So instead, by being a platform, we say, like sequel has been the data transformation language do jour for many days now. So, of course, that you can easily write Sequel and a lot of our visual data Transformations are based on the sequel language, but also something like Python again. It's like the language de jour for machine law machine learning model building right now, so you can easily code in python. Maintain your python libraries in date, ICU And so by leveraging open source, we figured we're making our clients more future proof as long as they're staying in date ICU. But using data ICU to leverage the best in breed and open source, they'll always be kind of where they want to be in the technological landscape by supposed to locked into some tech that is now out of date. >>What's been the appetite for making data beautiful for a legacy enterprise, like a G E that's been around for a very long time versus a more modern either. Born in the Cloud er's our CEO says, reborn in the cloud. What are some of the differences but also similarities that you see in terms of we have to be able to use emerging tech. Otherwise someone's gonna come in behind us and replace us. >>Yeah, I mean, I think it's complicated in that there's still a lot of value to be had in someone says, like a bar chart you can rely on right, So it's maybe not sexy. But having good reporting and analytics is something that both you know, 200 year old enterprise organizations and data native organizations startups needs. At the same time, building predicted machine learning models and deploying those is rest a p i n points that developers can use in your organization to provide a data driven product for your consumers. Like that's amore advanced use case that everyone kind of wants to be a part of again data. Who's a nice tool, which says Maybe you don't have developers who are very fluent in turning out flashed applications. We could give you a place to build a predictive model and deploy that predictive model, saving you time to write all that code on the back end. >>One of the themes of the show has been transformation, so it sounds like data ICU would be It's something that you can dip your toes in and start to get used to using. Even if you're not particularly familiar with Time machine learning model a model building. >>Yeah, that's exactly right. So a big part of our product and encourage watchers to go try it out themselves and go to our website. Download a free version pretrial, but is enablement. So if you're the most sophisticated applied math PhD there is, like, Who's a great environment for you to Code and Bill predictive models. If you never built the machine learning model before you can use data ICU to run visual machine learning recipes, we call them, and also we give you documentation, which is, Hey, this is a random forest model. What is a random forest model? We'll tell you a little bit about it. And that's another thing that some of these enterprises have really appreciated about date I could. It is helping up skill there user base >>in terms of that transformation theme that Justin just mention which we're hearing a lot about, not visit this show. It's a big thing, but we hear it all the time, right? But in terms of customers transformation, journey, whatever you wanna call it, cloud is gonna be an essential enabler of being able to really love it value from a I. So I'm just wondering from a strategic positioning standpoint. Is did ICU positioned as a facilitator or as fuel for a cloud transformation that on enterprise would undergo >>again? Yes, great point. So for us, I can't take the credit. This credit goes to our founders, but we've thought from the start the clouds and exciting proposition Not everyone is. They're still in 2019. Most people, if not all of them, want to get there. Also, people want too many of our clients want the multi cloud on a day. Like who says, If you want to be on prim, if you want to be in a single cloud subscription. If you want to be multi cloud again as a platform, we're just gonna give you connection to your underlying infrastructure. You could use the infrastructure that you like and just use our front end to help your analyst get value. They can. I >>think I think a lot of vendors across the entire ecosystem around to say the customer choice is really important, and the customers, particularly enterprise customers, want to be able to have lots of different options, and not all of them will be ready to go completely. All in on cloud today. They made it may take them years, possibly decades, to get there. So having that choice is like it's something that it would work with you today and we'll work with you tomorrow, depending on what choices you make. >>It's exactly right. Another thing we've seen a lot of to that day, like who helps with and whether it's like you or other tools. Like, of course, you want best in breed, but you also want particularly for a large enterprise. You don't want people operating kind of in a wild West, particularly in like the ML data science space. So you know we integrate with Jupiter notebooks, but some of our clients come to us initially. Just have I won't say rogues that has a negative connotation. But maybe I will say Road road data Scientists are just tapping into some day the store. They're using Jupiter notebooks to build a predictive model, but then to actually production allies that to get sustainable value out of it like it's to one off and so having a centralized platform like date ICU, where you can say this is where we're going to use our central model depository, that something where businesses like they can sleep easier at night because they know where is my ML development happening? It's happening in one ecosystem. What tools that happening with, well, best in breed of open source. So again, you kind of get best of both worlds like they like you. >>It sounds like it's more about the operations of machine learning. It is really, really important rather than just. It's the pure technology. Yes, that's important as well, and you need to have the data Sinus to build it, but having something that allows you to operationalize it so that you can just bake it into what we do every day as a business. >>Yeah, I think in a conference like this all about tech, it's easy to forget what we firmly believe, which is a I and maybe tech. More broadly, it's still human problems at the core, right? Once you get the tech right, the code runs corrected. The code is written correctly. Therefore, like human interactions, project management model deployment in an organization. These are really hard, human centered problems, but so having tech that enables that human centric collaboration helps with that, we find >>Let's talk about some of the things that we can't ever go to an event and not talk about. Nut is respected data quality, reliability and security. Understood? I could facilitate those three cornerstones. >>Yeah, sure. So, again, viewers, I would encourage you to check out the date. ICU has some nice visual indications of data quality. So an analyst or data scientists and come in very easily understand, you know, is this quality to conform to the standards that my organization has set and what I mean by standards that could be configured. Right? So does this column have the appropriate schema? Does it have the appropriate carnality? These are things that an individual might decide to use on then for security. So Data has its own security mechanisms. However, we also to this point about incorporating best Retek. We'll work with whatever underlying security mechanisms organizations organizations have in place. So, for instance, if you're using a W s, you have, I am rolls to manage your security. Did ICU comport those that apply those to the date ICU environment or using something like on prime miss, uh, duke waken you something like Kerberos has the technology to again manage access to resources. So we're taking the best in breed that this organization already has invested time, energy and resources into and saying We're not trying to compete with them but rather were trying to enable organizations to use these technologies efficiently. >>Yeah, I like that consistency of customer choice. We spoke about that just before. I'm seeing that here with their choices around. Well, if you're on this particular platform will integrate with whatever the tools are there. People underestimate how important that is for enterprises, that it has to be ahead. Virginia's environment, playing well with others is actually quite important. >>Yeah, I don't know that point. Like the combination of heterogeneity but also uniformity. It's a hard balance to strike, and I think it's really important, giving someone a unified environment but still choice. At the same time. A good restaurant or something like you won't be able to pick your dish, but you want to know that the entire quality is high. And so having that consistent ecosystem, I think, really helps >>what are, in your opinion, some of the next industries that you see there really right to start Really leveraging machine learning to transfer You mentioned g e a very old legacy business. If we think of you know what happened with the ride hailing industry uber, for example, or fitness with Saletan or pinchers with visible Serge, what do you think is the next industry? That's like you guys taking advantage of machine learning will completely transform this and our lives. >>I mean, the easy answer that I'll give because it's easy to say it's gonna transform. But hard to operationalize is health care, right? So there is structured data, but the data quality is so desperate and had a row genius s, I think you know, if organizations in a lot of this again it's a human centered problem. If people could decide on data standards and also data privacy is, of course, a huge issue. We talked about data security internally, but also as a customer. What day to do I want you know, this hospital, this health care provider, to have access to that human issues we have to result but conditional on that being resolved that staring out a way to anonymous eyes data and respect data privacy but have consistent data structure. And we could say, Hey, let's really set these a I M L models loose and figure out things like personalized medicine which were starting to get to. But I feel like there's still a lot of room to go. That >>sounds like it's exciting time to be in machine learning. People should definitely check out products such as Dead Rock you and see what happens. >>Last question for you is so much news has come out in the last three days. It's mind boggling sum of the takeaways, that of some of the things that you've heard from Andy Jassy to border This'll Morning. >>Yeah, I think a big thing for me, which was something for me before this week. But it's always nice to hear an Amazon reassures the concept of white box. Aye, aye. We've been talking about that a date ICU for some time, but everyone wants performance A. I R ml solutions, but increasing. There's a really appetite publicly for interpret ability, and so you have to be responsible. You have to have interpret belay I and so it's nice to hear a leader like Amazon echo that day like you. That's something we've been talking about since our start. >>A little bit validating them for data ICU, for sure, for sure. Well, thank you for joining. Just to be on the kid, the suffering. And we appreciate it. Appreciate it. All right. For my co host, Justin Warren, I'm Lisa Martin and your work to the Cube from Vegas. It's AWS reinvent 19.

Published Date : Dec 5 2019

SUMMARY :

Brought to you by Amazon Web service by Justin Warren, the founder and chief analyst at Pivot nine. I've been I've been trying to take care of it. And you probably have talked to at least half of the 65,000 attendees. Well, we're gonna talk to another guy here. After a three day is that you Who did a coup is and what you guys do in technology. you know, we make poetry out of it. I'm only the guy's been around around for eight years. so start up. mourning the cloud, the opportunity there That data is no longer a And so by that we see ourselves as a collaborative platform. actually doing on the platform, like simple things like doing customer segmentation for, you know, marketing campaigns but Are, is their integration with serum Maybe also have some excel file someone you know me. So maybe talk us through how you how you came to found the company based on basic So, of course, that you can easily write Sequel and a lot of our visual data Transformations What are some of the differences but also similarities that you see in terms of we have to be had in someone says, like a bar chart you can rely on right, So it's maybe not sexy. One of the themes of the show has been transformation, so it sounds like data ICU would be It's something that you can dip your we call them, and also we give you documentation, which is, Hey, this is a random forest model. transformation, journey, whatever you wanna call it, cloud is gonna be an essential as a platform, we're just gonna give you connection to your underlying infrastructure. So having that choice is like it's something that it would work with you today and we'll work with you tomorrow, So you know we integrate with Jupiter notebooks, but some of our clients come to us initially. to operationalize it so that you can just bake it into what we do every day as a business. Yeah, I think in a conference like this all about tech, it's easy to forget what we firmly Let's talk about some of the things that we can't ever go to an event and not talk about. like on prime miss, uh, duke waken you something like Kerberos has the technology to again Yeah, I like that consistency of customer choice. A good restaurant or something like you won't be able to pick your dish, If we think of you know what happened with the ride hailing industry uber, for example, What day to do I want you know, such as Dead Rock you and see what happens. Last question for you is so much news has come out in the last three days. There's a really appetite publicly for interpret ability, and so you have to be responsible. thank you for joining.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Justin WarrenPERSON

0.99+

Lisa MartinPERSON

0.99+

2019DATE

0.99+

JustinPERSON

0.99+

Andy JassyPERSON

0.99+

Las VegasLOCATION

0.99+

Will NowakPERSON

0.99+

AmazonORGANIZATION

0.99+

Eight yearsQUANTITY

0.99+

pythonTITLE

0.99+

200 yearQUANTITY

0.99+

PythonTITLE

0.99+

VegasLOCATION

0.99+

AWSORGANIZATION

0.99+

echoCOMMERCIAL_ITEM

0.99+

SergeyPERSON

0.99+

todayDATE

0.99+

tomorrowDATE

0.99+

NovakPERSON

0.99+

two setsQUANTITY

0.99+

Three daysQUANTITY

0.99+

VirginiaLOCATION

0.98+

DataikuPERSON

0.98+

bothQUANTITY

0.98+

Dead RockTITLE

0.97+

single serverQUANTITY

0.97+

both worldsQUANTITY

0.97+

three dayQUANTITY

0.97+

SergePERSON

0.96+

oneQUANTITY

0.96+

single cloudQUANTITY

0.96+

RetekORGANIZATION

0.95+

uberORGANIZATION

0.95+

SalesforceORGANIZATION

0.95+

a dayQUANTITY

0.93+

Day threeQUANTITY

0.93+

OneQUANTITY

0.91+

65,000 attendeesQUANTITY

0.91+

This'll MorningTITLE

0.9+

CoyoteORGANIZATION

0.89+

Amazon WebORGANIZATION

0.89+

KerberosORGANIZATION

0.88+

decadesQUANTITY

0.88+

one ecosystemQUANTITY

0.87+

ec2TITLE

0.85+

last three daysDATE

0.82+

three cornerstonesQUANTITY

0.79+

GORGANIZATION

0.79+

19QUANTITY

0.78+

eight yearsQUANTITY

0.74+

CubeORGANIZATION

0.74+

this weekDATE

0.73+

EsoORGANIZATION

0.72+

G EORGANIZATION

0.7+

Pivot nineORGANIZATION

0.69+

excelTITLE

0.67+

SaletanPERSON

0.59+

CubesORGANIZATION

0.57+

secondQUANTITY

0.57+

YukaCOMMERCIAL_ITEM

0.53+

halfQUANTITY

0.5+

JupiterORGANIZATION

0.48+

Invent 2019EVENT

0.46+

Reinvent 19EVENT

0.39+

inventEVENT

0.24+

Bill Vass, AWS | AWS re:Invent 2019


 

>> Announcer: Live from Las Vegas, it's theCUBE! Covering AWS re:Invent 2019. Brought to you by Amazon Web Services and Intel. Along with it's ecosystem partners. >> Okay, welcome back everyone. It's theCUBE's live coverage here in Las Vegas for Amazon Web Series today, re:Invent 2019. It's theCUBE's seventh year covering re:Invent. Eight years they've been running this event. It gets bigger every year. It's been a great wave to ride on. I'm John Furrier, my cohost, Dave Vellante. We've been riding this wave, Dave, for years. It's so exciting, it gets bigger and more exciting. >> Lucky seven. >> This year more than ever. So much stuff is happening. It's been really exciting. I think there's a sea change happening, in terms of another wave coming. Quantum computing, big news here amongst other great tech. Our next guest is Bill Vass, VP of Technology, Storage Automation Management, part of the quantum announcement that went out. Bill, good to see you. >> Yeah, well, good to see you. Great to see you again. Thanks for having me on board. >> So, we love quantum, we talk about it all the time. My son loves it, everyone loves it. It's futuristic. It's going to crack everything. It's going to be the fastest thing in the world. Quantum supremacy. Andy referenced it in my one-on-one with him around quantum being important for Amazon. >> Yes, it is, it is. >> You guys launched it. Take us through the timing. Why, why now? >> Okay, so the Braket service, which is based on quantum notation made by Dirac, right? So we thought that was a good name for it. It provides for you the ability to do development in quantum algorithms using gate-based programming that's available, and then do simulation on classical computers, which is what we call our digital computers today now. (men chuckling) >> Yeah, it's a classic. >> These are classic computers all of a sudden right? And then, actually do execution of your algorithms on, today, three different quantum computers, one that's annealing and two-bit gate-based machines. And that gives you the ability to test them in parallel and separate from each other. In fact, last week, I was working with the team and we had two machines, an ion trap machine and an electromagnetic tunneling machine, solving the same problem and passing variables back and forth from each other, you could see the cloud watch metrics coming out, and the data was going to an S3 bucket on the output. And we do it all in a Jupiter notebook. So it was pretty amazing to see all that running together. I think it's probably the first time two different machines with two different technologies had worked together on a cloud computer, fully integrated with everything else, so it was pretty exciting. >> So, quantum supremacy has been a word kicked around. A lot of hand waving, IBM, Google. Depending on who you talk to, there's different versions. But at the end of the day, quantum is a leap in computing. >> Bill: Yes, it can be. >> It can be. It's still early days, it would be day zero. >> Yeah, well I think if you think of, we're about where computers were with tubes if you remember, if you go back that far, right, right? That's about where we are right now, where you got to kind of jiggle the tubes sometimes to get them running. >> A bug gets in there. Yeah, yeah, that bug can get in there, and all of those kind of things. >> Dave: You flip 'em off with a punch card. Yeah, yeah, so for example, a number of the machines, they run for four hours and then they come down for a half hour for calibration. And then they run for another four hours. So we're still sort of at that early stage, but you can do useful work on them. And more mature systems, like for example D-Wave, which is annealer, a little different than gate-based machines, is really quite mature, right? And so, I think as you go back and forth between these machines, the gate-based machines and annealers, you can really get a sense for what's capable today with Braket and that's what we want to do is get people to actually be able to try them out. Now, quantum supremacy is a fancy word for we did something you can't do on a classical computer, right? That's on a quantum computer for the first time. And quantum computers have the potential to exceed the processing power, especially on things like factoring and other things like that, or on Hamiltonian simulations for molecules, and those kids of things, because a quantum computer operates the way a molecule operates, right, in a lot of ways using quantum mechanics and things like that. And so, it's a fancy term for that. We don't really focus on that at Amazon. We focus on solving customer's problems. And the problem we're solving with Braket is to get them to learn it as it's evolving, and be ready for it, and continue to develop the environment. And then also offer a lot of choice. Amazon's always been big on choice. And if you look at our processing portfolio, we have AMD, Intel x86, great partners, great products from them. We have Nvidia, great partner, great products from them. But we also have our Graviton 1 and Graviton 2, and our new GPU-type chip. And those are great products, too, I've been doing a lot on those, as well. And the customer should have that choice, and with quantum computers, we're trying to do the same thing. We will have annealers, we will have ion trap machines, we will have electromagnetic machines, and others available on Braket. >> Can I ask a question on quantum if we can go back a bit? So you mentioned vacuum tubes, which was kind of funny. But the challenge there was with that, it was cooling and reliability, system downtime. What are the technical challenges with regard to quantum in terms of making it stable? >> Yeah, so some of it is on classical computers, as we call them, they have error-correction code built in. So you have, whether you know it or not, there's alpha particles that are flipping bits on your memory at all times, right? And if you don't have ECC, you'd get crashes constantly on your machine. And so, we've built in ECC, so we're trying to build the quantum computers with the proper error correction, right, to handle these things, 'cause nothing runs perfectly, you just think it's perfect because we're doing all the error correction under the covers, right? And so that needs to evolve on quantum computing. The ability to reproduce them in volume from an engineering perspective. Again, standard lithography has a yield rate, right? I mean, sometimes the yield is 40%, sometimes it's 20%, sometimes it's a really good fab and it's 80%, right? And so, you have a yield rate, as well. So, being able to do that. These machines also generally operate in a cryogenic world, that's a little bit more complicated, right? And they're also heavily affected by electromagnetic radiation, other things like that, so you have to sort of faraday cage them in some cases, and other things like that. So there's a lot that goes on there. So it's managing a physical environment like cryogenics is challenging to do well, having the fabrication to reproduce it in a new way is hard. The physics is actually, I shudder to say well understood. I would say the way the physics works is well understood, how it works is not, right? No one really knows how entanglement works, they just knows what it does, and that's understood really well, right? And so, so a lot of it is now, why we're excited about it, it's an engineering problem to solve, and we're pretty good at engineering. >> Talk about the practicality. Andy Jassy was on the record with me, quoted, said, "Quantum is very important to Amazon." >> Yes it is. >> You agree with that. He also said, "It's years out." You said that. He said, "But we want to make it practical "for customers." >> We do, we do. >> John: What is the practical thing? Is it just kicking the tires? Is it some of the things you mentioned? What's the core goal? >> So, in my opinion, we're at a point in the evolution of these quantum machines, and certainly with the work we're doing with Cal Tech and others, that the number of available cubits are starting to increase at an astronomic rate, a Moore's Law kind of of rate, right? Whether it's, no matter which machine you're looking at out there, and there's about 200 different companies building quantum computers now, and so, and they're all good technology. They've all got challenges, as well, as reproducibility, and those kind of things. And so now's a good time to start learning how to do this gate-based programming knowing that it's coming, because quantum computers, they won't replace a classical computer, so don't think that. Because there is no quantum ram, you can't run 200 petabytes of data through a quantum computer today, and those kind of things. What it can do is factoring very well, or it can do probability equations very well. It'll have affects on Monte Carlo simulations. It'll have affects specifically in material sciences where you can simulate molecules for the first time that you just can't do on classical computers. And when I say you can't do on classical computers, my quantum team always corrects me. They're like, "Well, no one has proven "that there's an algorithm you can run "on a classical computer that will do that yet," right? (men chuckle) So there may be times when you say, "Okay, I did this on a quantum computer," and you can only do it on a quantum computer. But then someone's very smart mathematician says, "Oh, I figured out how to do it on a regular computer. "You don't need a quantum computer for that." And that's constantly evolving, as well, in parallel, right? And so, and that's what's that argument between IBM and Google on quantum supremacy is that. And that's an unfortunate distraction in my opinion. What Google did was quite impressive, and if you're in the quantum world, you should be very happy with what they did. They had a very low error rate with a large number of cubits, and that's a big deal. >> Well, I just want to ask you, this industry is an arms race. But, with something like quantum where you've got 200 companies actually investing in it so early days, is collaboration maybe a model here? I mean, what do think? You mentioned Cal Tech. >> It certainly is for us because, like I said, we're going to have multiple quantum computers available, just like we collaborate with Intel, and AMD, and the other partners in that space, as well. That's sort of the nice thing about being a cloud service provider is we can give customers choice, and we can have our own innovation, plus their innovations available to customers, right? Innovation doesn't just happen in one place, right? We got a lot of smart people at Amazon, we don't invent everything, right? (Dave chuckles) >> So I got to ask you, obviously, we can take cube quantum and call it cubits, not to be confused with theCUBE video highlights. Joking aside, classical computers, will there be a classical cloud? Because this is kind of a futuristic-- >> Or you mean a quantum cloud? >> Quantum cloud, well then you get the classic cloud, you got the quantum cloud. >> Well no, they'll be together. So I think a quantum computer will be used like we used to use a math coprocessor if you like, or FPGAs are used today, right? So, you'll go along and you'll have your problem. And I'll give you a real, practical example. So let's say you had a machine with 125 cubits, okay? You could just start doing some really nice optimization algorithms on that. So imagine there's this company that ships stuff around a lot, I wonder who that could be? And they need to optimize continuously their delivery for a truck, right? And that changes all the time. Well that algorithm, if you're doing hundreds of deliveries in a truck, it's very complicated. That traveling salesman algorithm is a NP-hard problem when you do it, right? And so, what would be the fastest best path? But you got to take into account weather and traffic, so that's changing. So you might have a classical computer do those algorithms overnight for all the delivery trucks and then send them out to the trucks. The next morning they're driving around. But it takes a lot of computing power to do that, right? Well, a quantum computer can do that kind of problemistic or deterministic equation like that, not deterministic, a best-fit algorithm like that, much faster. And so, you could have it every second providing that. So your classical computer is sending out the manifests, interacting with the person, it's got the website on it. And then, it gets to the part where here's the problem to calculate, we call it a shot when you're on a quantum computer, it runs it in a few seconds that would take an hour or more. >> It's a fast job, yeah. >> And it comes right back with the result. And then it continues with it's thing, passes it to the driver. Another update occurs, (buzzing) and it's just going on all the time. So those kind of things are very practical and coming. >> I've got to ask for the younger generations, my sons super interested as I mentioned before you came on, quantum attracts the younger, smart kids coming into the workforce, engineering talent. What's the best path for someone who has an either advanced degree, or no degree, to get involved in quantum? Is there a certain advice you'd give someone? >> So the reality is, I mean, obviously having taken quantum mechanics in school and understanding the physics behind it to an extent, as much as you can understand the physics behind it, right? I think the other areas, there are programs at universities focused on quantum computing, there's a bunch of them. So, they can go into that direction. But even just regular computer science, or regular mechanical and electrical engineering are all neat. Mechanical around the cooling, and all that other stuff. Electrical, these are electrically-based machines, just like a classical computer is. And being able to code at low level is another area that's tremendously valuable right now. >> Got it. >> You mentioned best fit is coming, that use case. I mean, can you give us a sense of a timeframe? And people will say, "Oh, 10, 15, 20 years." But you're talking much sooner. >> Oh, I don't, I think it's sooner than that, I do. And it's hard for me to predict exactly when we'll have it. You can already do, with some of the annealing machines, like D- Wave, some of the best fit today, right? So it's a matter of people want to use a quantum computer because they need to do something fast, they don't care how much it costs, they need to do something fast. Or it's too expensive to do it on a classical computer, or you just can't do it at all on a classical computer. Today, there isn't much of that last one, you can't do it at all, but that's coming. As you get to around 52, 50, 52 cubits, it's very hard to simulate that on a classical computer. You're starting to reach the edge of what you can practically do on a classical computer. At about 125 cubits, you probably are at a point where you can't just simulate it anymore. >> But you're talking years, not decades, for this use case? >> Yeah, I think you're definitely talking years. I think, and you know, it's interesting, if you'd asked me two years ago how long it would take, I would've said decades. So that's how fast things are advancing right now, and I think that-- >> Yeah, and the computers just getting faster and faster. >> Yeah, but the ability to fabricate, the understanding, there's a number of architectures that are very well proven, it's just a matter of getting the error rates down, stability in place, the repeatable manufacturing in place, there's a lot of engineering problems. And engineering problems are good, we know how to do engineering problems, right? And we actually understand the physics, or at least we understand how the physics works. I won't claim that, what is it, "Spooky action at a distance," is what Einstein said for entanglement, right? And that's a core piece of this, right? And so, those are challenges, right? And that's part of the mystery of the quantum computer, I guess. >> So you're having fun? >> I am having fun, yeah. >> I mean, this is pretty intoxicating, technical problems, it's fun. >> It is. It is a lot of fun. Of course, the whole portfolio that I run over at AWS is just really a fun portfolio, between robotics, and autonomous systems, and IOT, and the advanced storage stuff that we do, and all the edge computing, and all the monitor and management systems, and all the real-time streaming. So like Kinesis Video, that's the back end for the Amazon ghost stores, and working with all that. It's a lot of fun, it really is, it's good. >> Well, Bill, we need an hour to get into that, so we may have to come up and see you, do a special story. >> Oh, definitely! >> We'd love to come up and dig in, and get a special feature program with you at some point. >> Yeah, happy to do that, happy to do that. >> Talk some robotics, some IOT, autonomous systems. >> Yeah, you can see all of it around here, we got it up and running around here, Dave. >> What a portfolio. >> Congratulations. >> Alright, thank you so much. >> Great news on the quantum. Quantum is here, quantum cloud is happening. Of course, theCUBE is going quantum. We've got a lot of cubits here. Lot of CUBE highlights, go to SiliconAngle.com. We got all the data here, we're sharing it with you. I'm John Furrier with Dave Vellante talking quantum. Want to give a shout out to Amazon Web Services and Intel for setting up this stage for us. Thanks to our sponsors, we wouldn't be able to make this happen if it wasn't for them. Thank you very much, and thanks for watching. We'll be back with more coverage after this short break. (upbeat music)

Published Date : Dec 4 2019

SUMMARY :

Brought to you by Amazon Web Services and Intel. It's so exciting, it gets bigger and more exciting. part of the quantum announcement that went out. Great to see you again. It's going to be the fastest thing in the world. You guys launched it. It provides for you the ability to do development And that gives you the ability to test them in parallel Depending on who you talk to, there's different versions. It's still early days, it would be day zero. we're about where computers were with tubes if you remember, can get in there, and all of those kind of things. And the problem we're solving with Braket But the challenge there was with that, And so that needs to evolve on quantum computing. Talk about the practicality. You agree with that. And when I say you can't do on classical computers, But, with something like quantum and the other partners in that space, as well. So I got to ask you, you get the classic cloud, you got the quantum cloud. here's the problem to calculate, we call it a shot and it's just going on all the time. quantum attracts the younger, smart kids And being able to code at low level is another area I mean, can you give us a sense of a timeframe? And it's hard for me to predict exactly when we'll have it. I think, and you know, it's interesting, Yeah, and the computers Yeah, but the ability to fabricate, the understanding, I mean, this is and the advanced storage stuff that we do, so we may have to come up and see you, and get a special feature program with you Yeah, happy to do that, Talk some robotics, some IOT, Yeah, you can see all of it We got all the data here, we're sharing it with you.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JohnPERSON

0.99+

IBMORGANIZATION

0.99+

Dave VellantePERSON

0.99+

Amazon Web ServicesORGANIZATION

0.99+

two machinesQUANTITY

0.99+

AmazonORGANIZATION

0.99+

Cal TechORGANIZATION

0.99+

AMDORGANIZATION

0.99+

AndyPERSON

0.99+

BillPERSON

0.99+

Andy JassyPERSON

0.99+

EinsteinPERSON

0.99+

John FurrierPERSON

0.99+

40%QUANTITY

0.99+

DavePERSON

0.99+

Bill VassPERSON

0.99+

GoogleORGANIZATION

0.99+

20%QUANTITY

0.99+

NvidiaORGANIZATION

0.99+

IntelORGANIZATION

0.99+

80%QUANTITY

0.99+

last weekDATE

0.99+

AWSORGANIZATION

0.99+

an hourQUANTITY

0.99+

four hoursQUANTITY

0.99+

200 companiesQUANTITY

0.99+

10QUANTITY

0.99+

Las VegasLOCATION

0.99+

two-bitQUANTITY

0.99+

15QUANTITY

0.99+

TodayDATE

0.99+

125 cubitsQUANTITY

0.99+

200 petabytesQUANTITY

0.99+

20 yearsQUANTITY

0.99+

two different machinesQUANTITY

0.99+

oneQUANTITY

0.99+

50QUANTITY

0.99+

two different technologiesQUANTITY

0.99+

Eight yearsQUANTITY

0.98+

first timeQUANTITY

0.98+

Monte CarloTITLE

0.98+

todayDATE

0.98+

two years agoDATE

0.98+

52 cubitsQUANTITY

0.97+

BraketORGANIZATION

0.97+

x86COMMERCIAL_ITEM

0.97+

This yearDATE

0.96+

next morningDATE

0.96+

about 125 cubitsQUANTITY

0.95+

Graviton 1COMMERCIAL_ITEM

0.95+

DiracORGANIZATION

0.95+

Graviton 2COMMERCIAL_ITEM

0.94+

about 200 different companiesQUANTITY

0.93+

three different quantum computersQUANTITY

0.93+

Moore's LawTITLE

0.91+

seventh yearQUANTITY

0.9+

decadesQUANTITY

0.87+

secondsQUANTITY

0.86+

every secondQUANTITY

0.85+

re:EVENT

0.82+

half hourQUANTITY

0.81+

Yaron Haviv, Iguazio | KubeCon + CloudNativeCon NA 2019


 

>>Live from San Diego, California at the cube covering to clock in cloud native con brought to you by red hat, the cloud native computing foundation and its ecosystem Marsh. >>Welcome back. This is the cubes coverage of CubeCon cloud date of con 2019 in San Diego, 12,000 in attendance. I'm just two minute and my cohost is John trier. And welcome back to the program. A multi-time cube alumni. You're on Aviv, who is the CTO and cofounder of a Gwoza. We've had quite a lot of, you know, founders, CTOs, you know, their big brains at this show, your own. So you know, let, let, let's start, you know, there's, there's really a gathering, uh, there's a lot of effort building out, you know, a very complicated ecosystem. Give us first, kind of your overall impressions of the show in this ecosystem. Yeah, so we're very early on on Desecco system. We were one of the first in the first batch of CNCF members when there were a few dozens of those. Not like a thousand of those. Uh, so I've been, I've been to all those shows. >>Uh, we're part of the CNCF committees for different things. And any initiating, I think this has become much more mainstream. I told you before, it's sort of the new van world. You know, I lot a lot more, uh, all day infrastructure vendors along with middleware and application vendor are coming here. All right, so, so one of the things we like having you on the program you're on is you don't pull any punches. So we've seen certain waves of technology come with big promise and fall short, you know, big data was going to allow us to leverage everything and you know, large percentage of, uh, solutions, you know, had to stop or be pulled back. Um, give us, what's the cautionary tale that we should learn and make sure that we don't repeat, you know, so I've been a CTO for many years in different companies and, and what everyone used to say about it, I'm always right. >>I'm only one year off usually. I'm usually a little more optimistic. So, you know, we've been talking about Cloudera and Hadoop world sort of going down and Kubernetes and cloud services, essentially replacing them. We were talking about it four years ago and what do you see that's actually happening? You know, with the collapse of my par and whore, then we're going to Cloudera things are going down, customer now Denon guys, we need equivalent solution for Kubernetes. We're not going to maintain two clusters. So I think in general we've been, uh, picking on many of those friends. We've, we've invented serverless before it was even called serverless with, with nuclear and now we're expanding it further and now we see the new emerging trends really around machine learning and AI. That's sort of the big thing. I'm surprised, you know, that's our space where essentially you're doing a data science platform as a service fully automated around serverless constructs so people can, can develop things really, really quickly. >>And what I see that, you know, third of the people I talk to are, have some relations to machine learning and AI. Yeah. Maybe explain that for our audience a little bit. Because when, you know, Kubernetes first started very much an infrastructure discussion, but the last year or two, uh, very much application specific, we hear many people talking about those data use cases, AI and ML early days. But you know how, how does that fit into the overall? It's simple. You know there, if you're moving to the cloud are two workloads. There is lift and shift workloads and there are new workloads. Okay, lift and ship. Why? Why bother moving them to Kubernetes? Okay, so you end up with new workloads. Everyone is trying to be cloud native server, elastic services and all that. Everyone has to feed data and machine learning into those new applications. This is why you see those trends that talk about old data integration, various frameworks and all that in that space. >>So I don't think it's by coincidence. I think it's, that's because new applications incorporate the intelligence. That's why you hear a lot of the talk about those things. What I loved about the architecture, what you just said is like people don't want to run into another cluster. I don't want to run two versions of Kubernetes, you know, if I'm moving there you, because you, but you're still built on that, that kind of infrastructure framework and, and knowledge of, of how to do serverless and how to make more nodes and fewer nodes and persistent storage and all that sort of good stuff and uh, and, and run TensorFlow and run, you know, all these, all these big data apps. But you can, um, you can talk about that just as a, as a, the advantage to your customer cause you could, it seems like you could, you could run it on top of GKE. >>You could run it on prem. I could run my own Coobernetti's you could, you could just give me a, uh, so >> we, we say Kubernetes is not interesting. I didn't know. I don't want anyone to get offended. Okay. But Kubernetes is not the big deal. The big deal is organizations want to be competitive in this sort of digital world. They need to build new applications. Old ones are sort of in sort of a maintenance mode. And the big point is about delivering new application with elastic scaling because your, your customers may, may be a million people behind some sort of, uh, you know, uh, app. Okay. Um, so that's the key thing and Kubernetes is a way to deliver those microservices. But what we figured out, it's still very complicated for people. Okay. Especially in, in the data science work. Uh, he takes him a few weeks to deliver a model on a Jupiter notebook, whatever. >>And then productizing it is about the year. That's something we've seen between six months to a year to productize things that are relatively simple. Okay. And that's because people think about the container, the TensorFlow, the Kuda driver, whatever, how to scale it, how to make it perform, et cetera. So let's, we came up with is traditionally there's a notion of serverless, which is abstraction with very slow performance, very limited set of use cases. We sell services about elastic scaling paper, use, full automation around dev ops and all that. Okay. Why cannot apply to other use cases are really high concurrency, high-speed batch, no distributed training, distributed workload. Because we're coming, if you know my background, you know, been beeping in Mellanox and other high-performance companies. So where I have a, we have a high performance DNA so we don't know how to build things are extremely slow. >>It sort of irritates me. So the point is that how can we apply this notion of abstraction and scaling and all that to variety of workloads and this is essentially what it was. It is a combination of high speed data technology for like, you know, moving data around on between those function and extremely high speed set though functions that work on the different domains of data collection and ingestion, data analytics, you know, machine learning, training and CIN learning model serving. So a customer can come on on our platform and we have testimonials around that, that you know, things that they thought about building on Amazon or even on prem for months and months. They'd built in our platform in few weeks with fewer people because the focus is on building the application. The focus is not about joining your Kubernetes. Now we go to customers, some of them are large banks, et cetera. >>They say, Alrighty, likes Kubernetes, we have our own Kubernetes. So you know what, we don't butter. Initially we, we used to bring our own Kubernetes, but then you know, I don't mind, you know, we do struggle sometimes because our level of expertise in Coobernetti's is way more sophisticated than what they have to say. Okay, we've installed Kubernetes and we come with our software stack. No you didn't, you know, you didn't configure the security, they didn't configure ingress, et cetera. So sometimes it's easier for us to bring, but we don't want him to get into this sort of tension with it. Our focus is to accelerate development on the new application that are intelligent, you know, move applications from, if you think of the traditional data analytics and data science, it's about reporting and what people want to do. And some applications we've announced this week and application around real time cyber collection, it's being used in some different governments is that you can collect a lot of information, SMS, telephony, video, et cetera. >>And in real time you could detect terrorists. Okay. So those application requires high concurrency always on rolling upgrades, things that weren't there in the traditional BI, Oracle, you know, kind of reporting. So you have this wave of putting intelligence into more highly concurrent online application. It requires all the dev ops sort of aspects, but all the data analytics and machine learning aspects to to come to come along. Alright. So speaking of those workloads for, for machine learning, uh, cube flow is a project, uh, moving the, moving in that space along it. Give us the update there. Yeah. So, so there is sort of a rising star in the Kubernetes community around how to automate machine learning workflows. That's cube flow. Uh, I'm personally, I one of the committers and killed flow and what we've done, because it's very complicated cause Google developed the cube cube flow as one of the services on, on a GKE. >>Okay. And the tweaked everything. It works great in GK, even that it's relatively new technology and people want to move around it in a more generic. So one of the things in our platform is a managed cube flow that works natively with all the rest of the solutions. And other thing that we've done is we make it, we made it fully. So instead of queue flow approach is very con, you know, Kubernetes oriented containers, the ammos, all that. Uh, in our flavor of Coupa we can just create function and you just like chain functions and you click and it runs. Just, you've mentioned a couple of times, uh, how does serverless, as you defined it, fit in with, uh, Coobernetti's? Is that working together just functions on top or I'm just trying to make here, >> you'll, you'll hear different things. I think when most people say serverless, they mean sort of front end application things that are served low concurrency, a Terra, you know, uh, when we mean serverless, it's, we have eight different engines that each one is very good in, in different, uh, domain like distributed deep learning, you know, distributed machine learning, et cetera. >>And we know how to fit the thing into any workloads. So for me, uh, we deliver the elastic scaling, the paper use and the ease of use of sort of no dev ops across all the eight workloads that we're addressing. For most people it's like a single Dreek phony. And I think really that the future is, is moving to that. And if you think about serverless, there's another aspect here which is very important for machine learning and Israel's ability. I'm not going to develop any algorithm in the world. Okay. There are a bunch of companies or users or developers that can develop an algorithm and I can just consume it. So the future in data science but not just data science is essentially to have like marketplaces of algorithms premade or analytic tools or maybe even vendors licensing their technology through sort of prepackaged solution. >>So we're a great believer of forget about the infrastructure, focus on the business components and Daisy chain them in to a pipeline like UFO pipeline and run them. And that will allow you most reusability that, you know, lowest amount of cost, best performance, et cetera. That's great. I just want to double click on the serverless idea one more time, but, so you're, you're developing, it's an architectural pattern, uh, and you're developing these concepts yourself. You're not actually, sometimes the concept gets confused with the implementations of other people's serverless frameworks or things like that. Is that, is that correct? I think there are confusion. I'm getting asked a lot of times. How do you compare your technology compared to let's say a? You've heard the term gay native is just a technology or open FAS or, yeah. Hold on. Pfizer's a CGIs or Alito. An open community is very nice for hobbies, but if you're an enterprise and it's security, Eldep integration, authentication for anything, you need DUIs, you need CLI, you need all of those things. >>So Amazon provides that with Lambda. Can you compare Lambda to K native? No. Okay. Native is, I need to go from get and build and all that. Serverless is about taking a function and clicking and deploying. It's not about building. And the problem is that this conference is about people, it people in crowd for people who like to build. So they, they don't like to get something that work. They want to get the build the Lego building blocks so they can play. So in our view, serverless is not open FAS or K native. Okay. It's something that you click and it works and have all the enterprise set of features. We've extended it to different levels of magnitude of performance. I'll give you an anecdote. I did a comparison for our customer asking me the same question, not about Canadian, but this time Lambda. How do you guys compare with London? >>Know Nokia is extremely high performance. You know we are doing up to 400,000 events on a single process and the customer said, you know what, I have a use case. I need like 5,000 events per second. How do you guys compare a total across all my functions? How do you compare against Lambda? We went into, you know the price calculator, 5,000 events per second on Lambda. That's $50,000 okay. $50,000 we do about, let's say even in simple function, 60,000 per process, $500 VM on Amazon, $500 VM on Amazon with our technology stick, 2000 transactions per second, 5,000 events per second on Lambda. That's 50,000. Okay. 100 times more expensive. So it depends on the design point. We designed our solution to be extremely efficient, high concurrency. If you just need something to do a web hook, use Lambda, you know, if you are trying to build a high concurrency application efficient, you know, an enterprise application on it, on a serverless architecture construct come to us. >>Yeah. So, so just a, I'll pause at this for you because a, it reminds me what you were talking about about the builders here in the early days of VMware to get it to work the way I wanted to. People need to participate and build it and there's the Ikea effect. If I actually helped build it a little bit, I like it more to get to the vast majority, uh, to uh, adopt those things. It needs to become simplified and I can't have, you know, all the applications move over to this environment if I have to constantly tweak that. Everything. So that's the trend we've been really seeing this year is some of that simplification needs to get there. There's focus on, you know, the operators, the day two operations, the applications so that anybody can get there without having to build themselves. So we know there's still work to be done. >>Um, but if we've crossed the chasm and we want the majority to now adopt this, it can't be that I have to customize it. It needs to be more turnkey. Yeah. And I think it's a friendly and attitude between what you'll see in Amazon reinvent in couple of weeks. And then what you see here, because there is those, the focus of we're building application a what kind of tools and the Jess is gonna just launch today on the, on the floor. Okay. So we can just consume it and build our new application. They're not thinking, how did Andy just, he built his tools. Okay. And I think that's the opposite here is like how can you know Ali's is still working inside underneath dude who cares about his team. You know, you care about having connectivity between two points and and all that. How do you implement it that, you know, let someone else take care of it and then you can apply your few people that you have on solving your business problem, not on infrastructure. >>You know, I just met a guy, came to our booth, we've seen our demo. Pretty impressive how we rise people function and need scales and does everything automatically said we want to build something like you're doing, you know, not really like only 10% of what you just showed me. And we have about six people and for three months where it just like scratching our head. I said, okay, you can use our platform, pay us some software license and now you'll get, you know, 10 times more functionality and your six people can do something more useful. Says right, let's do a POC. So, so that's our intention and I think people are starting to get it because Kubernetes is not easy. Again, people tell me we installed Kubernete is now installed your stack and then they haven't installed like 20% of all the things that you need to stop so well your own have Eve always pleasure to catch up with you. Thanks for the all the updates and I know we'll catch up with you again soon. Sure. All right. For John Troyer, I'm Stu Miniman. We'll be back with more coverage here from CubeCon cloud date of con in San Diego. Thanks for watching the cube.

Published Date : Nov 20 2019

SUMMARY :

clock in cloud native con brought to you by red hat, the cloud native computing foundation So you know, All right, so, so one of the things we like having you on the program you're on is you don't pull any punches. I'm surprised, you know, that's our space where essentially you're doing a data science platform as a service And what I see that, you know, third of the people I talk to are, have some relations to machine learning you know, if I'm moving there you, because you, but you're still built on that, that kind of infrastructure I could run my own Coobernetti's you could, you could just give me a, uh, so sort of, uh, you know, uh, app. Because we're coming, if you know my background, you know, been beeping in Mellanox and other high-performance companies. and we have testimonials around that, that you know, things that they thought about building on Amazon or even I don't mind, you know, we do struggle sometimes because our level of expertise in Coobernetti's is Oracle, you know, kind of reporting. you know, Kubernetes oriented containers, the ammos, all that. in different, uh, domain like distributed deep learning, you know, distributed machine learning, And if you think about serverless, most reusability that, you know, lowest amount of cost, best performance, It's something that you click and it works and have all the enterprise set of features. a web hook, use Lambda, you know, if you are trying to build a high concurrency application you know, all the applications move over to this environment if I have to constantly tweak that. And I think that's the opposite here is like how can you know Ali's is still working inside I said, okay, you can use our platform, pay us some software license and now you'll get, you know,

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
$50,000QUANTITY

0.99+

John TroyerPERSON

0.99+

John trierPERSON

0.99+

$500QUANTITY

0.99+

Stu MinimanPERSON

0.99+

AndyPERSON

0.99+

NokiaORGANIZATION

0.99+

AmazonORGANIZATION

0.99+

three monthsQUANTITY

0.99+

10 timesQUANTITY

0.99+

two pointsQUANTITY

0.99+

San DiegoLOCATION

0.99+

50,000QUANTITY

0.99+

GoogleORGANIZATION

0.99+

six monthsQUANTITY

0.99+

six peopleQUANTITY

0.99+

San Diego, CaliforniaLOCATION

0.99+

two minuteQUANTITY

0.99+

KuberneteTITLE

0.99+

Yaron HavivPERSON

0.99+

20%QUANTITY

0.99+

100 timesQUANTITY

0.99+

KubernetesTITLE

0.99+

LambdaTITLE

0.99+

IguazioPERSON

0.99+

one yearQUANTITY

0.99+

OracleORGANIZATION

0.99+

PfizerORGANIZATION

0.99+

firstQUANTITY

0.99+

four years agoDATE

0.99+

CNCFORGANIZATION

0.99+

two clustersQUANTITY

0.98+

12,000QUANTITY

0.98+

KubeConEVENT

0.98+

CubeConEVENT

0.98+

JessPERSON

0.97+

a yearQUANTITY

0.97+

LegoORGANIZATION

0.97+

last yearDATE

0.97+

CloudNativeConEVENT

0.97+

first batchQUANTITY

0.97+

each oneQUANTITY

0.97+

todayDATE

0.96+

DeseccoORGANIZATION

0.96+

weeksQUANTITY

0.96+

5,000 events per secondQUANTITY

0.96+

AliPERSON

0.96+

two versionsQUANTITY

0.96+

oneQUANTITY

0.96+

two workloadsQUANTITY

0.95+

10%QUANTITY

0.95+

twoQUANTITY

0.94+

MellanoxORGANIZATION

0.94+

dozensQUANTITY

0.94+

GwozaORGANIZATION

0.94+

5,000 events per secondQUANTITY

0.94+

singleQUANTITY

0.93+

thirdQUANTITY

0.93+

up to 400,000 eventsQUANTITY

0.93+

60,000 per processQUANTITY

0.92+

this yearDATE

0.91+

this weekDATE

0.91+

a million peopleQUANTITY

0.9+

EvePERSON

0.9+

5,000 events per secondQUANTITY

0.9+

DenonORGANIZATION

0.89+

2000 transactions per secondQUANTITY

0.88+

AlitoORGANIZATION

0.87+

AvivPERSON

0.85+

about six peopleQUANTITY

0.85+

CoobernettiORGANIZATION

0.85+

eight workloadsQUANTITY

0.84+

red hatORGANIZATION

0.83+

HadoopTITLE

0.82+

ClouderaORGANIZATION

0.81+

thousandQUANTITY

0.79+

CanadianLOCATION

0.79+

Renaud Gaubert, NVIDIA & Diane Mueller, Red Hat | KubeCon + CloudNativeCon NA 2019


 

>>Live from San Diego, California It's the Q covering Koopa and Cloud Native Cot brought to you by Red Cloud, Native Computing Pounding and its ecosystem March. >>Welcome back to the Cube here at Q. Khan Club native Khan, 2019 in San Diego, California Instrumental in my co host is Jon Cryer and first of all, happy to welcome back to the program. Diane Mueller, who is the technical of the tech lead of cloud native technology. I'm sorry. I'm getting the wrong That's director of community development Red Hat, because renew. Goodbye is the technical lead of cognitive technologies at in video game to the end of day one. I've got three days. I gotta make sure >>you get a little more Red Bull in the conversation. >>All right, well, there's definitely a lot of energy. Most people we don't even need Red Bull here because we're a day one. But Diane, we're going to start a day zero. So, you know, you know, you've got a good group of community of geeks when they're like Oh, yeah, let me fly in a day early and do like 1/2 day or full day of deep dives. There So the Red Hat team decided to bring everybody on a boat, I guess. >>Yeah. So, um, open ships Commons gathering for this coup con we hosted at on the inspiration Hornblower. We had about 560 people on a boat. I promised them that it wouldn't leave the dock, but we deal still have a little bit of that weight going on every time one of the big military boats came by. And so people were like a little, you know, by the end of the day, but from 8 a.m. in the morning till 8 p.m. In the evening, we just gathered had some amazing deep dives. There was unbelievable conversations onstage offstage on we had, ah, wonderful conversation with some of the new Dev ops folks that have just come on board. That's a metaphor for navigation and Coop gone. And and for events, you know, Andrew Cliche for John Willis, the inevitable Crispin Ella, who runs Open Innovation Labs, and J Bloom have all just formed the global Transformation Office. I love that title on dhe. They're gonna be helping Thio preach the gospel of Cultural Dev ops and agile transformation from a red hat office From now going on, there was a wonderful conversation. I felt privileged to actually get to moderate it and then just amazing people coming forward and sharing their stories. It was a great session. Steve Dake, who's with IBM doing all the SDO stuff? Did you know I've never seen SDO done so well, Deployment explains so well and all of the contents gonna be recorded and up on Aaron. We streamed it live on Facebook. But I'm still, like reeling from the amount of information overload. And I think that's the nice thing about doing a day zero event is that it's a smaller group of people. So we had 600 people register, but I think was 560 something. People show up and we got that facial recognition so that now when they're traveling through the hallways here with 12,000 other people, that go Oh, you were in the room. I met you there. And that's really the whole purpose for comments. Events? >>Yeah, I tell you, this is definitely one of those shows that it doesn't take long where I say, Hey, my brain is full. Can I go home. Now. You know I love your first impressions of Q Khan. Did you get to go to the day zero event And, uh, what sort of things have you been seeing? So >>I've been mostly I went to the lightning talks, which were amazing. Anything? Definitely. There. A number of shout outs to the GPU one, of course. Uh, friend in video. But I definitely enjoyed, for example, of the amazing D. M s one, the one about operators. And generally all of them were very high quality. >>Is this your first Q? Khan, >>I've been there. I've been a year. This is my third con. I've been accused in Europe in the past. Send you an >>old hat old hand at this. Well, before we get into the operator framework and I wanna love to dig into this, I just wanted to ask one more thought. Thought about open shift, Commons, The Commons in general, the relationship between open shift, the the offering. And then Okay, the comments and okay, D and then maybe the announcement about about Okay. Dee da da i o >>s. Oh, a couple of things happened yesterday. Yesterday we dropped. Okay, D for the Alfa release. So anyone who wants to test that out and try it out it's an all operators based a deployment of open shift, which is what open ship for is. It's all a slightly new architectural deployment methodology based on the operator framework, and we've been working very diligently. Thio populate operator hub dot io, which is where all of the upstream projects that have operators like the one that Reynolds has created for in the videos GP use are being hosted so that anyone could deploy them, whether on open shift or any kubernetes so that that dropped. And yesterday we dropped um, and announced Open Sourcing Quay as project quay dot io. So there's a lot of Io is going on here, but project dia dot io is, um, it's a fulfillment, really, of a commitment by Red Hat that whenever we do an acquisition and the poor folks have been their acquired by Cora West's and Cora Weston acquired by Red Hat in an IBM there. And so in the interim, they've been diligently working away to make the code available as open source. And that hit last week and, um, to some really interesting and users that are coming up and now looking forward to having them to contribute to that project as well. But I think the operator framework really has been a big thing that we've been really hearing, getting a lot of uptake on. It's been the new pattern for deploying applications or service is on getting things beyond just a basic install of a service on open shift or any kubernetes. And that's really where one of the exciting things yesterday on we were talking, you know, and I were talking about this earlier was that Exxon Mobil sent a data scientist to the open ship Commons, Audrey Resnick, who gave this amazing presentation about Jupiter Hub, deeper notebooks, deploying them and how like open shift and the advent of operators for things like GP use is really helping them enable data scientists to do their work. Because a lot of the stuff that data signs it's do is almost disposable. They'll run an experiment. Maybe they don't get the result they want, and then it just goes away, which is perfect for a kubernetes workload. But there are other things you need, like a Jeep use and work that video has been doing to enable that on open shift has been just really very helpful. And it was It was a great talk, but we were talking about it from the first day. Signs don't want to know anything about what's under the hood. They just want to run their experiments. So, >>you know, let's like to understand how you got involved in the creation of the operator. >>So generally, if we take a step back and look a bit at what we're trying to do is with a I am l and generally like EJ infrastructure and five G. We're seeing a lot of people. They're trying to build and run applications. Whether it's in data Center at the and we're trying to do here with this operator is to bring GPS to enterprise communities. And this is what we're working with. Red Hat. And this is where, for example, things like the op Agrestic A helps us a lot. So what we've built is this video Gee, few operator that space on the upper air sdk where it wants us to multiple phases to in the first space, for example, install all the components that a data scientist were generally a GPU cluster of might want to need. Whether it's the NVIDIA driver, the container runtime, the community's device again feast do is as you go on and build an infrastructure. You want to be able to have the automation that is here and, more importantly, the update part. So being able to update your different components, face three is generally being able to have a life cycle. So as you manage multiple machines, these are going to get into different states. Some of them are gonna fail, being able to get from these bad states to good states. How do you recover from them? It's super helpful. And then last one is monitoring, which is being able to actually given sites dr users. So the upper here is decay has helped us a lot here, just laying out these different state slips. And in a way, it's done the same thing as what we're trying to do for our customers. The different data scientists, which is basically get out of our way and allow us to focus on core business value. So the operator, who basically takes care of things that are pretty cool as an engineer I lost due to your election. But it doesn't really help me to focus on like my core business value. How do I do with the updates, >>you know? Can I step back one second, maybe go up a level? The problem here is that each physical machine has only ah limited number of NVIDIA. GPU is there and you've got a bunch of containers that maybe spawning on different machines. And so they have to figure out, Do I have a GPU? Can I grab one? And if I'm using it, I assume I have to reserve it and other people can't use and then I have to give it up. Is that is that the problem we're solving here? So this is >>a problem that we've worked with communities community so that like the whole resource management, it's something that is integrated almost first class, citizen in communities, being able to advertise the number of deep, use their your cluster and used and then being able to actually run or schedule these containers. The interesting components that were also recently added are, for example, the monitoring being able to see that a specific Jupiter notebook is using this much of GP utilization. So these air supercool like features that have been coming in the past two years in communities and which red hat has been super helpful, at least in these discussions pushing these different features forward so that we see better enterprise support. Yeah, >>I think the thing with with operators and the operator lifecycle management part of it is really trying to get to Day two. So lots of different methodologies, whether it's danceable or python or job or or UH, that's helm or anything else that can get you an insult of a service or an application or something. And in Stan, she ate it. But and the operator and we support all of that with SD case to help people. But what we're trying to do is bridge the to this day to stuff So Thea, you know, to get people to auto pilot, you know, and there's a whole capacity maturity model that if you go to operator hab dot io, you can see different operators are a different stages of the game. So it's been it's been interesting to work with people to see Theo ah ha moment when they realize Oh, I could do this and then I can walk away. And then if that pod that cluster dies, it'll just you know, I love the word automatically, but they, you know, it's really the goal is to help alleviate the hands on part of Day two and get more automation into the service's and applications we deploy >>right and when they when they this is created. Of course it works well with open shift, but it also works for any kubernetes >>correct operator. HAB Daddio. Everything in there runs on any kubernetes, and that's really the goal is to be ableto take stuff in a hybrid cloud model. You want to be able to run it anywhere you want, so we want people to be unable to do it anywhere. >>So if this really should be an enabler for everything that it's Vinny has been doing to be fully cloud native, Yes, >>I think completely arable here is this is a new attack. Of course, this is a bit there's a lot of complexity, and this is where we're working towards is reducing the complexity and making true that people there. Dan did that a scientist air machine learning engineers are able to focus on their core business. >>You watch all of the different service is in the different things that the data scientists are using. They don't I really want to know what's under under the hood. They would like to just open up a Jupiter Hub notebook, have everything there. They need, train their models, have them run. And then after they're done, they're done and it goes away. And hopefully they remember to turn off the Jeep, use in the woods or wherever it is, and they don't keep getting billed for it. But that's the real beauty of it is that they don't have to worry so much anymore about that. And we've got a whole nice life cycle with source to image or us to I. And they could just quickly build on deploy its been, you know, it's near and dear to my heart, the machine learning the eyesight of stuff. It is one of the more interesting, you know, it's the catchy thing, but the work was, but people are really doing it today, and it's been we had 23 weeks ago in San Francisco, we had a whole open ship comments gathering just on a I and ML and you know, it was amazing to hear. I think that's the most redeeming thing or most rewarding thing rather for people who are working on Kubernetes is to have the folks who are doing workloads come and say, Wow, you know, this is what we're doing because we don't get to see that all the time. And it was pretty amazing. And it's been, you know, makes it all worthwhile. So >>Diane Renaud, thank you so much for the update. Congratulations on the launch of the operators and look forward to hearing more in the future. >>All right >>to >>be here >>for John Troy runs to minimum. More coverage here from Q. Khan Club native Khan, 2019. Thanks for watching. Thank you.

Published Date : Nov 20 2019

SUMMARY :

Koopa and Cloud Native Cot brought to you by Red Cloud, California Instrumental in my co host is Jon Cryer and first of all, happy to welcome back to the program. There So the Red Hat team decided to bring everybody on a boat, And that's really the whole purpose for comments. Did you get to go to the day zero event And, uh, what sort of things have you been seeing? But I definitely enjoyed, for example, of the amazing D. I've been accused in Europe in the past. The Commons in general, the relationship between open shift, And so in the interim, you know, let's like to understand how you got involved in the creation of the So the operator, who basically takes care of things that Is that is that the problem we're solving here? added are, for example, the monitoring being able to see that a specific Jupiter notebook is using this the operator and we support all of that with SD case to help people. Of course it works well with open shift, and that's really the goal is to be ableto take stuff in a hybrid lot of complexity, and this is where we're working towards is reducing the complexity and It is one of the more interesting, you know, it's the catchy thing, but the work was, Congratulations on the launch of the operators and look forward for John Troy runs to minimum.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Audrey ResnickPERSON

0.99+

Andrew ClichePERSON

0.99+

Diane MuellerPERSON

0.99+

Steve DakePERSON

0.99+

IBMORGANIZATION

0.99+

Jon CryerPERSON

0.99+

Exxon MobilORGANIZATION

0.99+

Diane RenaudPERSON

0.99+

EuropeLOCATION

0.99+

John TroyPERSON

0.99+

San FranciscoLOCATION

0.99+

1/2 dayQUANTITY

0.99+

Red HatORGANIZATION

0.99+

San Diego, CaliforniaLOCATION

0.99+

firstQUANTITY

0.99+

J BloomPERSON

0.99+

DianePERSON

0.99+

2019DATE

0.99+

Open Innovation LabsORGANIZATION

0.99+

yesterdayDATE

0.99+

Red CloudORGANIZATION

0.99+

560QUANTITY

0.99+

NVIDIAORGANIZATION

0.99+

600 peopleQUANTITY

0.99+

three daysQUANTITY

0.99+

John WillisPERSON

0.99+

8 a.m.DATE

0.99+

Crispin EllaPERSON

0.99+

JeepORGANIZATION

0.99+

San Diego, CaliforniaLOCATION

0.99+

Cora WestORGANIZATION

0.99+

YesterdayDATE

0.99+

last weekDATE

0.99+

SDOTITLE

0.99+

DanPERSON

0.99+

8 p.m.DATE

0.98+

23 weeks agoDATE

0.98+

first impressionsQUANTITY

0.98+

one secondQUANTITY

0.98+

Q. Khan ClubORGANIZATION

0.98+

oneQUANTITY

0.98+

RenauPERSON

0.98+

Red BullORGANIZATION

0.98+

ReynoldsPERSON

0.97+

AaronPERSON

0.97+

Day twoQUANTITY

0.97+

MarchDATE

0.96+

third con.QUANTITY

0.96+

first spaceQUANTITY

0.96+

first dayQUANTITY

0.95+

VinnyPERSON

0.95+

Cora WestonORGANIZATION

0.94+

ThioPERSON

0.94+

CloudORGANIZATION

0.93+

FacebookORGANIZATION

0.92+

first classQUANTITY

0.92+

todayDATE

0.9+

about 560 peopleQUANTITY

0.9+

JupiterLOCATION

0.89+

each physical machineQUANTITY

0.88+

12,000 otherQUANTITY

0.88+

day zeroQUANTITY

0.88+

D. MPERSON

0.87+

CloudNativeCon NA 2019EVENT

0.87+

d GaubertPERSON

0.87+

TheaPERSON

0.86+

pythonTITLE

0.84+

Native Computing PoundingORGANIZATION

0.83+

a dayQUANTITY

0.79+

day zeroEVENT

0.78+

day oneQUANTITY

0.78+

KoopaORGANIZATION

0.76+

one more thoughtQUANTITY

0.74+

KhanPERSON

0.72+

CommonsORGANIZATION

0.72+

KubeCon +EVENT

0.72+

Jupiter HubORGANIZATION

0.71+

Seth Juarez, Microsoft | Microsoft Ignite 2019


 

>>Live from Orlando, Florida. It's the cube covering Microsoft ignite brought to you by Cohesity. >>Good afternoon everyone and welcome back to the cubes live coverage of Microsoft ignite 26,000 people here at this conference at the orange County convention center. I'm your host, Rebecca Knight, alongside my cohost Stu Miniman. We are joined by Seth Juarez. He is the cloud developer advocate at Microsoft. Thank you so much for coming on the show. >>Glad to be here. You have such a lovely sad and you're lovely people. We just met up. You don't know any better? No. Well maybe after after the end of the 15 minutes we'll have another discussion. >>You're starting off on the right foot, so tell us a little bit about what you do. You're also a host on channel nine tell us about your role as a, as a cloud developer. >>So a cloud advocate's job is primarily to help developers be successful on Azure. My particular expertise lies in AI and machine learning and so my job is to help developers be successful with AI in the cloud, whether it be developers, data scientists, machine learning engineers or whatever it is that people call it nowadays. Because you know how the titles change a lot, but my job is to help them be successful and sometimes what's interesting is that sometimes our customers can't find success in the cloud. That's actually a win for me too because then I have a deep integration with the product group and my job is to help them understand from a customer perspective what it is they need and why. So I'm like the ombudsman so to speak because the product groups are the product groups. I don't report up to them. So I usually go in there and I'm like, Hey, I don't report to any of you, but this is what the customers are saying. >>We are very keen on being customer centered and that's why I do what I do. >> Seth, I have to imagine when you're dealing with customers, some of that skills gap and learning is something that they need to deal with. You know, we've been hearing for a long time, you know, there's not enough data scientists, you know, we need to learn these environments. Satya Nadella spent a lot of time talking about the citizen developers out there. So you know H bring us inside the customers you're talking to, you know, kind of, where do you usually start and you know, how do they pull the right people in there or are they bringing in outside people a little bit? Great organization, great question. It turns out that for us at Microsoft we have our product groups and then right outside we have our advocates that are very closely aligned to the product groups. >>And so anytime we do have an interaction with a customer, it's for the benefit of all the other customers. And so I meet with a lot of customers and I don't, I'm to get to talk about them too much. But the thing is I go in there, I see what they're doing. For example, one time I went to the touring Institute in the UK. I went in there and because I'm not there to sell, I'm there to figure out like what are you trying to do and does this actually match up? It's a very different kind of conversation and they'd tell me about what they're working on. I tell them about how we can help them and then they tell me where the gaps are or where they're very excited and I take both of those pieces of feedback to the, to the product group and they, they just love being able to have someone on the ground to talk to people because sometimes you know, when work on stuff you get a little siloed and it's good to have an ombudsman so to speak, to make sure that we're doing the right thing for our customers. >>As somebody that works on AI. You must've been geeking out working, working with the Turing Institute though. Oh yeah. Those people are absolutely wonderful and it was like as I was walking in, a little giddy, but the problems that they're facing in AI are very similar. The problems that people at the other people doing and that are in big organizations, other organizations are trying to onboard to AI and try to figure out, everyone says I need to be using this hammer and they're trying to hammer some screws in with the hammer. So it's good to figure out when it's appropriate to use AI and when it isn't. And I also have customers with that >>and I'm sure the answer is it depends in terms of when it's appropriate, but do you have any sort of broad brush advice for helping an organization determine is is this a job for AI? Absolutely. >>That's uh, it's a question I get often and developers, we have this thing called the smell that tells us if a code smell, we have a code smell tells us, maybe we should refactor, maybe we should. For me, there's this AI smell where if you can't precisely figure out the series of steps to execute an algorithm and you're having a hard time writing code, or for example, if every week you need to change your if L statements or if you're changing numbers from 0.5 to 0.7 and now it works, that's the smell that you should think about using AI or machine learning, right? There's also a set of a class of algorithms that, for example, AI, it's not that we've solved, solved them, but they're pretty much solved. Like for example, detecting what's in an image, understanding sentiment and text, right? Those kinds of problems we have solutions for that are just done. >>But if you have a code smell where you have a lot of data and you don't want to write an algorithm to solve that problem, machine learning and AI might be the solution. Alright, a lot of announcements this week. Uh, any of the highlights for from your area. We last year, AI was mentioned specifically many times now with you know, autonomous systems and you know it feels like AI is in there not necessarily just you know, rubbing AI on everything. >> I think it's because we have such a good solution for people building custom machine learning that now it's time to talk about the things you can do with it. So we're talking about autonomous systems. It's because it's based upon the foundation of the AI that we've already built. We released something called Azure machine learning, a set of tools called in a studio where you can do end and machine learning. >>Because what what's happening is most data scientists nowadays, and I'm guilty of this myself, we put stuff in things called Jupiter notebooks. We release models, we email them to each other, we're emailing Python files and that's kinda like how programming was in 1995 and now we're doing is we're building a set of tools to allow machine learning developers to go end to end, be able to see how data scientists are working and et cetera. For example, let's just say you're a data scientist. Bill. Did an awesome job, but then he goes somewhere else and Sally who was absolutely amazing, comes in and now she's the data scientist. Usually Sally starts from zero and all of the stuff that bill did is lost with Azure machine learning. You're able to see all of your experiments, see what bill tried, see what he learned and Sally can pick right up and go on. And that's just doing the experiments. Now if you want to get machine learning models into production, we also have the ability to take these models, version them, put them into a CIC, D similar process with Azure dev ops and machine learning. So you can go from data all the way to machine learning in production very easily, very quickly and in a team environment, you know? And that's what I'm excited about mostly. >>So at a time when AI and big and technology companies in general are under fire and not, Oh considered to not always have their users best interests at heart. I'd like you to talk about the Microsoft approach to ethical AI and responsible AI. >>Yeah, I was a part of the keynote. Scott Hanselman is a very famous dab and he did a keynote and I got to form part of it and one of the things that we're very careful even on a dumb demo or where he was like doing rock paper, scissors. I said, and Scott, we were watching you with your permission to see like what sequence of throws you were doing. We believe that through and through all the way we will never use our customers' data to enhance any of our models. In fact, there was a time when we were doing like a machine learning model for NLP and I saw the email thread and it's like we don't have language food. I don't remember what it was. We don't have enough language food. Let's pay some people to ethically source this particular language data. We will never use any of our customer's data and I've had this question asked a lot. >>Like for example, our cognitive services which have built in AI, we will never use any of our customer's data to build that neither. For example, if we have, for example, we have a custom vision where you upload your own pictures, those are your pictures. We're never going to use them for anything. And anything that we do, there's always consent and we want to make sure that everyone understands that AI is a powerful tool, but it also needs to be used ethically. And that's just on how we use data for people that are our customers. We also have tools inside of Azure machine learning to get them to use AI. Ethically. We have tools to explain models. So for example, if you very gender does the model changes prediction or if you've very class or race, is your model being a little iffy? We allow, we have those tools and Azure machine learning, so our customers can also be ethical with the AI they build on our platform. So we have ethics built into how we build our models and we have ethics build into how our customers can build their models too, which is to me very. >>And is that a selling point? Are customers gravitating? I mean we've talked a lot about it on the show. About the, the trust that customers have in Microsoft and the image that Microsoft has in the industry right now. But the idea that it is also trying to perpetuate this idea of making everyone else more ethical. Do you think that that is one of the reasons customers are gravitate? >>I hope so. And as far as a selling point, I absolutely think it's a selling point, but we've just released it and so I'm going to go out there and evangelize the fact that not only are we as tickle with what we do in AI, but we want our customers to be ethical as well. Because you know, trust pays, as Satya said in his keynote, tra trust the enhancer in the exponent that allows tech intensity to actually be tech intensity. And we believe that through and through not only do believe it for ourselves, but we want our customers to also believe it and see the benefits of having trust with our customers. One of the things we, we talked to Scott Hanselman a little bit yesterday about that demo is the Microsoft of today isn't just use all the Microsoft products, right? To allow you to use, you know, any tool, any platform, you know, your own environment, uh, to tell us how that, that, that plays into your world. >>It's, you know, like in my opinion, and I don't know if it's the official opinion, but we are in the business of renting computer cycles. We don't care how you use them, just come into our house and use them. You wanna use Java. We've recently announced a tons of things with spraying. We're become an open JDK contributor. You know, one of my colleagues, we're very hard on that. I work primarily in Python because it's machine learning. I have a friend might call a friend and colleague, David Smith who works in our, I have other colleagues that work in a number of different languages. We don't care. What we are doing is we're trying to empower every organization and every person on the planet to achieve more where they are, how they are, and hopefully bring a little bit of of it to our cloud. >>What are you doing that, that's really exciting to you right now? I know you're doing a new.net library. Any other projects that are sparking your end? >>Yeah, so next week I'm going to France and this is before anyone's going to see this and there is a, there is a company, I think it's called surf, I'll have to look it up and we'll put it in the notes, but they are basically trying to use AI to be more environmentally conscious and they're taking pictures of trash and rivers and they're using AI to figure out where it's coming from so they can clean up environment. I get to go over there and see what they're doing, see how I can help them improvement and promote this kind of ethical way of doing AI. We also do stuff with snow leopards. I was watching some Netflix thing with my kids and we were watching snow leopards and there was like two of them. Like this is impressive because as I'm watching this with my kids, I'm like, Hey we are at Microsoft, we're helping this population, you know, perpetuate with AI. >>And so those are the things it's actually a had had I've seen on TV is, you know, rather than spending thousands of hours of people out there, the AI can identify the shape, um, you know, through the cameras. So they're on a, I love that powerful story to explain some of those pieces as opposed to it. It's tough to get the nuance of what's happening here. Absolutely. With this technology, these models are incredibly easy to build on our platform. And, and I and I st fairly easy to build with what you have. We love people use TensorFlow, use TensorFlow, people use pie torch. That's great cafe on it. Whatever you want to use. We are happy to let you use a rent out our computer cycles because we want you to be successful. Maybe speak a little bit of that when you talk about, you know, the, the cloud, one of the things is to democratize, uh, availability of this. >>There's usually free tiers out there, especially in the emerging areas. Uh, you know, how, how is Microsoft helping to get that, that compute and that world technology to people that might not have had it in the past? I was in, I was in Peru a number of years ago and I and I had a discussion with someone on the channel nine show and it was absolutely imp. Like I under suddenly understood the value of this. He said, Seth, if I wanted to do a startup here in Peru, right, and it was a capital Peru, like a very industrialized city, I would have to buy a server. It would come from California on a boat. It would take a couple of months to get here and then it would be in a warehouse for another month as it goes through customs. And then I would have to put it into a building that has a C and then I could start now sat with a click of a button. >>I can provision an entire cluster of machines on Azure and start right now. That's what, that's what the cloud is doing in places like Peru and places that maybe don't have a lot of infrastructure. Now infrastructure is for everyone and maybe someone even in the United States, you know, in a rural area that doesn't, they can start up their own business right now anywhere. And it's not just because it's Peru, it's not just because it's some other place that's becoming industrialized. It's everywhere. Because any kid with a dream can spin up an app service and have a website done in like five minutes. >>So what does this mean? I mean, as you said, any, any kid, any person or rural area, any developing country, what does this mean in five or 10 years from now in terms of the future of commerce and work and business? >>Honestly, some people feel like computers are art, stealing, you know, human engineering. I think they are really augmenting it. Like for example, I don't have to, if I want to know something for her. Back when, when I was a kid, I had to, if I want to know something, sometimes I had to go without knowing where like I guess we'll never know. Right? And then five years later we're like, okay, we found out it was that a character on that show, you know? And now we just look at our phone. It's like, Oh, you were wrong. And I like not knowing that I'm wrong for a lot longer, you know what I'm saying? But nowadays with our, with our phones and with other devices, we have information readily available so that we can make appropriate response, appropriate answers to questions that we have. AI is going to help us with that by augmenting human ingenuity, by looking at the underlying structure. >>We can't, for example, if you look at, if you look at an Excel spreadsheet, if it's like five rows and maybe five columns, you and I as humans can look at and see a trend. But what if it's 10 million rows and 5,000 columns? Our ingenuity has been stretched too far, but with computers now we can aggregate, we can do some machine learning models, and then we can see the patterns that the computer found aggregated, and now we can make the decisions we could make with five columns, five rows, but it's not taking our jobs. It's augmenting our capacity to do the right thing. >>Excellent. We'll assess that. Thank you so much for coming on the Cuba. Really fun conversation. >>Glad to be here. Thanks for having me. >>Alright, I'm Rebecca Knight for Stu minimun. Stay tuned for more of the cubes live coverage of Microsoft ignite.

Published Date : Nov 6 2019

SUMMARY :

Microsoft ignite brought to you by Cohesity. Thank you so much for coming on the show. Glad to be here. You're starting off on the right foot, so tell us a little bit about what you do. So I'm like the ombudsman so to speak because the product groups are the product groups. You know, we've been hearing for a long time, you know, there's not enough data scientists, they just love being able to have someone on the ground to talk to people because sometimes you know, And I also have customers with that and I'm sure the answer is it depends in terms of when it's appropriate, but do you have any sort of broad brush if every week you need to change your if L statements or if you're changing numbers from 0.5 to 0.7 many times now with you know, autonomous systems and you know it feels like AI is to talk about the things you can do with it. So you can go from data all the way to machine learning in I'd like you to talk about the Microsoft approach to ethical AI and responsible AI. I said, and Scott, we were watching you with your permission to see For example, if we have, for example, we have a custom vision where you upload your own pictures, Do you think that that is one of the reasons customers are gravitate? any platform, you know, your own environment, uh, to tell us how that, We don't care how you use them, just come into our house What are you doing that, that's really exciting to you right now? we're helping this population, you know, perpetuate with AI. And, and I and I st fairly easy to build with what you have. Uh, you know, how, how is Microsoft helping to get that, that compute and that world technology to you know, in a rural area that doesn't, they can start up their own business right now anywhere. Honestly, some people feel like computers are art, stealing, you know, We can't, for example, if you look at, if you look at an Excel spreadsheet, if it's like five rows and maybe five Thank you so much for coming on the Cuba. Glad to be here. Alright, I'm Rebecca Knight for Stu minimun.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SallyPERSON

0.99+

Rebecca KnightPERSON

0.99+

ScottPERSON

0.99+

David SmithPERSON

0.99+

PeruLOCATION

0.99+

Seth JuarezPERSON

0.99+

CaliforniaLOCATION

0.99+

FranceLOCATION

0.99+

1995DATE

0.99+

Satya NadellaPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Turing InstituteORGANIZATION

0.99+

10 million rowsQUANTITY

0.99+

Scott HanselmanPERSON

0.99+

UKLOCATION

0.99+

Stu MinimanPERSON

0.99+

United StatesLOCATION

0.99+

five minutesQUANTITY

0.99+

five rowsQUANTITY

0.99+

5,000 columnsQUANTITY

0.99+

last yearDATE

0.99+

yesterdayDATE

0.99+

five columnsQUANTITY

0.99+

Orlando, FloridaLOCATION

0.99+

SatyaPERSON

0.99+

JavaTITLE

0.99+

next weekDATE

0.99+

ExcelTITLE

0.99+

PythonTITLE

0.99+

SethPERSON

0.99+

CubaLOCATION

0.99+

BillPERSON

0.99+

todayDATE

0.99+

26,000 peopleQUANTITY

0.99+

oneQUANTITY

0.99+

five years laterDATE

0.98+

this weekDATE

0.98+

bothQUANTITY

0.98+

15 minutesQUANTITY

0.98+

OneQUANTITY

0.97+

0.7QUANTITY

0.97+

AzureTITLE

0.96+

JDKTITLE

0.96+

thousands of hoursQUANTITY

0.95+

10 yearsQUANTITY

0.94+

fiveQUANTITY

0.93+

NetflixORGANIZATION

0.92+

0.5QUANTITY

0.91+

zeroQUANTITY

0.91+

TensorFlowTITLE

0.9+

orange County convention centerLOCATION

0.84+

snow leopardsTITLE

0.84+

nine showQUANTITY

0.76+

number of years agoDATE

0.73+

NLPORGANIZATION

0.72+

two of themQUANTITY

0.7+

billPERSON

0.67+

monthsQUANTITY

0.66+

StuORGANIZATION

0.65+

thingsQUANTITY

0.61+

igniteTITLE

0.6+

CohesityORGANIZATION

0.59+

coupleQUANTITY

0.54+

Yaron Haviv, Iguazio | CUBEConversation, April 2019


 

>> From our studios in the heart of Silicon Valley. HOLLOWAY ALTO, California It is a cube conversation. >> Hello and welcome to Cube conversations. I'm James Kabila's lead analyst at Wicked Bond. Today we've got an excellent guest. Who's a Cube alumnus? Par excellence. It's your own Haviv who is the founder and CEO of a guajillo. Hello. You're wrong. Welcome in. I think you're you're coming in from Tel Aviv. If I'm not mistaken, >> right? Really? Close the deal of any thanks from my seeing you again. >> Yeah. Nice to see you again. So I'm here in our Palo Alto studios. And so I'm always excited when I can hear your own and meet with your room because he always has something interesting in new to share. But what they're doing in the areas of cloud and serve earless and really time streaming analytics And now, data science. I wasn't aware of how deeply they're involved in the whole data Science pipelines, so ah, your own. This is great to have you. So my first question really is. Can you sketch out? What are the emerging marketplace requirements that USA gua Si are seeing in the convergence of all these spaces? Especially riel time streaming analytics edge computing server lis and data science and A I can you give us a sort of ah broad perspective and outlook on the convergence and really the new opportunities or possibilities that the convergence of those technologies enable for enterprises that are making deep investments. >> Yeah, so I think we were serving dissipated. What's happening now? We just call them different names will probably get into into this discussion in a minute. I think what you see is the traditional analytics and even data scientist Science was starting at sort of a research labs, people exploring cancer, expressing, you know, impact. Whether on, you know, people's moved its era. And now people are trying to make real or a Y from a guy in their assigned, so they have to plug it within business applications. Okay, so it's not just a veil. A scientist Inning the silo, you know, with a bunch of large that he got from his friends, the data engineer in the scan them and Derrickson Namesake runs to the boss and says, You know what? You know, we could have made some money in a year ago. We've done something so that doesn't make a lot of impact on the business, where the impact on the business is happening is when you actually integrate a I in jackpot in recommendation engines in doing predictive analytics on analyzing failures and saving saving failures on, you know, saving people's life. Those kind of use cases. Doctors are the ones that record a tighter integration between the application and the data and algorithms that come from the day I. And that's where we started to think about our platform. Way worked on a real time data, which is where you know, when you're going into more production environment of not fatal accident. Very good, very fast integration with data. And we have this sort of fast computation layer, which was a one micro services, and now everyone talks about micro services. We sort of started with this area, and that is allowing people to build those intelligent application that are integrated into the business applications. And the biggest challenges they see today for organizations is moving from this process of books on research, on data in a historical data and translating that into a visit supplication or into impact on business application. This is where people can spend the year. You know, I've seen the tweet saying with build a machine learning model in, like, a few weeks. And now we've waited eleven months for the product ization. So that artifact, >> Yes, that's what we're seeing it wicked bomb. Which is that A. I is the heart of modern applications in business and the new generation of application developers, in many ways, our data scientists, or have you know, lovers the skills and tools for data science. Now, looking at a glass zeros portfolio, you evolve so rapidly and to address a broader range of use cases I've seen. And you've explained it over the years that in position to go, as well as being a continuous data platform and intelligent edge platform, a surveillance platform. And now I see that you're a bit of a data science workbench or pipeline tooling. Clever. Could you connect these dots here on explain what is a guajillo fully >> role, Earl? Nice mark things for this in technology that we've built, OK, just over the years, you know, people, four years when we started, So we have to call it something else. Well, that I thought that analytic sort of the corporate state of science. And when we said continued analytics, we meant essentially feeding data and running, some of them speaking some results. This is the service opposed to the trend of truth which was dating the lady Throw data in and then you run the batch that analytic and they're like, Do you have some insight? So continue statistics was served a term that we've came up with a B, not the basket. You know, describe that you're essentially thinking, needing from different forces crunching it, Prue algorithms and generating triggers and actions are responsible user requests. Okay on that will serve a pretty unique and serve the fireman here in this industry even before they called it streaming or in a real time, data science or whatever. Now, if you look at our architecture are architecture, as I explained before, is comprised of three components. The first event is a real time, full time model database. You know, you know about it really exceptional in his performance and its other capabilities. The second thing is a pursue miss engine that allows us to essentially inject applications. Various guys, initially we started with application. I sense you do analytics, you know, grouping joining, you know, correlating. And then we start just adding more functions and other things like inference, saying humans recognitions and analysis. It's Arab is we have dysfunction engine. It allows us a lot of flexibility and find the really fast for the engine on a really fast data there endure it, remarkable results and then this return calling this turn this micro assume it's finger serve Ellis who certainly even where have the game of this or service gang. And the third element of our platform is a sense she having a fully manage, passed a platform where a ll those micro services our data and it threw a self service into face surfing over there is a mini cloud. You know, we've recently the last two years we've shifted to working with coronaries versus using our own A proprietary micro spurs does or frustration originally. So we went into all those three major technologies. Now, those pit into different application when they're interesting application. If you think about edge in the engine in serving many clouds, you need variety of data, sources and databases. With you, no problem arose streaming files. Terra. We'LL support all of them when our integrated the platform and then you need to go micro services that developed in the cloud and then just sort of shift into the enforcement point in the edge. And you need for an orchestration there because you want to do suffer upgrades, you need to protect security. So having all the integrated separated an opportunity for us to work with providers of agin, you may have noticed our joint announcement with Google around solution for hedge around retailers and an i O. T. We've made some announcement with Microsoft in the fast. We're going to do some very interesting announcement very soon. We've made some joint that nonsense with Samsung and in video, all around those errands, we continue. It's not that we're limited to EJ just what happens because we have extremely high density data platform, very power of fish and very well integrated. It has a great feat in the India, but it's also the same platform that we sell in. The cloud is a service or we sell two on from customers s so they can run. The same things is in the clouds, which happens to be the fastest, most real time platform on the Advantage service. An essential feature cannot just ignore. >> So you're wrong. Europe. Yeah, Iguazu is a complete cloud, native development and run time platform. Now serve earless in many ways. Seems to be the core of your capability in your platform. New Cleo, which is your technology you've open sourced. It's bill for Prem bays to private clouds. But also it has is extensible to be usable in broader hybrid cloud scenarios. Now, give us a sense for how nuclear and civilised functions become valuable or useful for data science off or for executing services or functions of data of the data science pipeline kick you connect the dots of nuclear and data science and a I from the development standpoint >> church. So So I think you know, the two pillars that we have technology that the most important ones are the data. You know, we have things like twelve batons on our data engine is very high performance and nuclear functions, and also they're very well integrated because usually services stateless. So you know, you you end up. If you want to practice that they have some challenges with service with No, no, you can't. You stay for use cases. You can mount files. You have real time connections to data, so that makes it a lot more interesting than just along the functions. The other thing, with no clothes that is extremely high performance has about two hundred times faster than land. So that means that you can actually go and build things like the stream processing and joins in real time all over practice, their base activities. You can just go and do collectors. We call them those like things. Go fetch information from whether services from routers for the X cybersecurity analysis for all sorts of sensors. So those functions are becoming like, you know, those nanobots technology of off the movies is that you just send them over to go and do things for you, whether it's the daily collection and crunching, whether it's the influencing engines, those things that, for example, get a picture of very put the model, decide what's in the picture, and that this is where we're really comes into play. They nothing important you see now an emergence off a service patterns in data science. So there are many companies that do like mother influencing as a service city what they do, they launch an end point of your eleven point and serve runs the model inside you send the Vector America values and get back in the Americans and their conversion. It's not really different and service it just wait more limited because I don't just want to send a vector off numbers because usually I understand really like a geo location of my cellphone, which are user I D. And I need dysfunction to cross correlated with other information about myself with the location. Then came commendation of which a product they need to buy. So and then those functions also have all sorts of dependency exam on different packages. Different software environment, horribles, build structures, all those. This is really where service technologies are much more suitable now. It's interesting that if you'LL go to Amazon, they have a product called Sage Maker. I'm sure yes, which is dinner, then a science block. Okay, now sage mint for although you would say that's a deal use case for after Onda functions actually don't use Amazon London functions in sage maker, and you ask yourself, Why aren't they using Lambda Stage Maker just telling you, you know you could use Lambda is a blue logic around sage maker. And that's because because London doesn't feed the use case. Okay, because lambda doesn't it is not capable of storing large content and she learning miles could be hundreds of megabytes or Landa is extremely slow. So you cannot do hi concurrency influencing with will land the function so essentially had to create another surveillance and college with a different name. Although if they just would have approved Landa, maybe it was one or a Swiss are So we're looking, We've took it, were taken the other approach We don't have the resources that I have so we created a monster virus Engine one servant attention does batch Frost is saying scream processing, consort, lots of data, even rocketeer services to all the different computation pattern with a single engine. And that's when you started taking all this trend because that's about yeah, we need two version our code. We need to, you know, record all our back into dependencies. And although yes, service doesn't so if we just had to go and tied more into the existing frameworks and you've looked at our frantically product called Tokyo Jupiter, which is essentially a scientist, right, some code in his data's passport book and then in clicks. One command called nuclear Deploy, it automatically compiles, is their science artifact in notebooks, that server and converted into a real hand function that can listen in on your next city. People can listen on streams and keep the scheduled on various timing. It could do magic. So many other things. So, and the interesting point is that if you think about their scientists there, not the farmers, because they should be a scientist on this's means that they actually have a bigger barrier to write in code. So if you serve in this framework that also automates the law daughter scaling the security provisioning of data, the versions of everything in fact fantasies, they just need to focus on writing other them's. It's actually a bigger back for the book. Now, if you just take service into them, Epstein's and they will tell you, Yeah, you know, we know how to explain, Doctor. We know all those things, so they're very their eyes is smaller than the value in the eyes of their scientists. So that's why we're actually seeing this appeal that those those people that essentially focus in life trying math and algorithms and all sorts of those sophisticated things they don't want to deal with. Coding and maintenance are refreshed. And by also doing so by oppression analyzing their cool for service, you can come back to market. You can address calle ability to avoid rewriting of code. All those big challenges the organizations are facing. >> You're gonna have to ask you, that's great. You have the tools to build, uh, help customers build serve Ellis functions for and so forth inside of Jupiter notebooks. And you mentioned Sage Maker, which is in a WS solution, which is up in coming in terms of supporting a full data science tool chain for pipeline development. You know, among teams you have a high profile partnerships with Microsoft and Google and Silver. Do you incorporate or integrator support either of these cloud providers own data science workbench offerings or third party offerings from? There's dozens of others in this space. What are you doing in terms of partnerships in that area? >> Yeah, obviously we don't want to lock us out from any of those, and, you know, if someone already has his work bench that I don't know my customers say they were locking me into your world back in our work when things are really cool because like our Jupiter is connected for real time connections to the database. And yes, serve other cool features that sentir getting like a huge speed boost we have. But that's on A with an within vigna of round Heads and Integration, which reviews are creating a pool of abuse from each of one of the data scientist running on African essentially launch clubs on this full of civilians whose off owning the abuse, which are extremely expensive, is you? No. But what we've done is because of her. The technology beside the actual debate engine is open source. We can accept it essentially just going any sold packages. And we demonstrate that to Google in danger. The others we can essentially got just go and load a bunch of packages into their work match and make it very proposed to what we provide in our manage platform. You know, not with the same performance levels. Well, functionality wise, the same function. >> So how can you name some reference customers that air using a guajillo inside a high performance data science work flows is ah, are you Are there you just testing the waters in that market for your technology? Your technology's already fairly mature. >> That says, I told you before, although you know, sort of changed messaging along the lines. We always did the same thing. So when we were continuous analytics and we've spoken like a year or two ago both some news cases that we Iran like, you know, tell cooperators and running really time, you know, health, a predictive health, monitoring their networks and or killing birds and those kind of things they all use algorithms. You control those those positions. We worked with Brian nailing customers so we can feed a lot of there there in real time maps and do from detection. And another applications are on all those things that we've noticed that all of the use cases that we're working with involved in a science in some cases, by the way, because of sort of politics that with once we've said, we have analytics for continuous analytics, we were serving send into sent into the analytic schools with the organization, which more focused on survey data warehouse because I know the case is still serve. They were saying, and I do. And after the people that build up can serve those data science applications and serve real time. Aye, aye. OK, Ianto. Business applications or more, the development and business people. This is also why we sort of change are our name, because we wanted to make it very clear that we're aren't the carnage is about building a new applications. It's not about the warehousing or faster queries. On a day of Eros is about generating value to the business, if you ask it a specific amplification. And we just announced two weeks in the investment off Samsung in Iguazu, former that essentially has two pillars beyond getting a few million dollars, It says. One thing is that they're adopted. No cure. Is there a service for the internal clouds on the second one is, we're working with them on a bunch of us, Della sighs. Well, use case is one of them was even quoted in enough would make would be There are no I can not say, but says she knows our real business application is really a history of those that involves, you know, in in intercepting data from your sister's customers, doing real time on analytics and responding really quickly. One thing that we've announced it because of youse off nuclear sub picture. We're done with inferior we actually what were pulled their performance. >> You're onto you see if you see a fair number of customers embedding machine learning inside of Realtor time Streaming stream computing back ones. This is the week of Flink forward here in San San Francisco. I I was at the event earlier this week and I I saw the least. They're presenting a fair amount of uptake of ml in sight of stream computing. Do you see that as being a coming meet Mainstream best practice. >> Streaming is still the analytics bucket. OK, because what we're looking for is a weakness which are more interactive, you know, think about like, uh, like a chatterbox or like doing a predictive analytic. It's all about streaming. Streaming is still, you know, it's faster flow data, but it's still, sir has delay the social. It's not responses, you know. It's not the aspect of legacy. Is that pickle in streaming? Okay, the aspect of throughput is is higher on streaming, but not necessarily the response that I think about sparks streaming. You know, it's good at crossing a lot of data. It's definitely not good at three to one on would put spark as a way to respond to user request on the Internet S O. We're doing screaming, and we see that growth. But think where we see the real growth is panic to reel of inches. The ones with the customer logs in and sends a request or working with telcos on scenarios where conditions of LA car, if the on the tracks and they settled all sorts of information are a real time invent train. Then the customer closer says, I need a second box and they could say No, this guy needs to go away to that customer because how many times you've gotten technician coming to your house and said I don't have that more exactly. You know, they have to send a different guy. So they were. How do you impact the business on three pillars of business? Okay, the three pillars are one is essentially improving your china Reducing the risk is essentially reducing your calls. Ask him. The other one is essentially audio, rap or customer from a more successful. So this is around front and application and whether it's box or are doing, you know our thing or those kind of us kisses. And also under you grow your market, which is a together on a recommendation in at this time. So all those fit you if you want, have hey, I incorporated in your business applications. In few years you're probably gonna be dead. I don't see any bits of sustained competition without incorporating so ability to integrate really real data with some customer data and essentially go and react >> changes. Something slightly you mentioned in video as a partner recently, Of course, he announced that few weeks ago. At their event on, they have recently acquired Melon ox, and I believe you used to be with Melon Axe, so I'd like to get your commentary on that acquisition or merger. >> Right? Yes, yes, I was VP Data Center man Ox. Like my last job, I feel good friends off off the Guider, including the CEO and the rest of the team with medicines. And last week I was in Israel's with talk to the media. Kansas. Well, I think it's a great merger if you think about men in Ox Head as sort of the best that breaking and storage technology answer Silicon Side and the video has the best view technologies, man. It's also acquired some compute cheap technologies, and they also very, very nice. Photonics technologies and men are today's being by all the club providers. Remiss Troll was essentially only those technical engagement would like the seizures and you know the rest of the gas. So now VP running with the computation engine in and minerals coming, we serve the rest of the pieces were our storage and make them a very strong player. And I think it's our threatens intel because think about it until they haven't really managed to high speed networking recently. They haven't really managed to come with Jiffy use at your combat and big technology, and so I think that makes a video, sort of Ah, pretty. You know, vendor and suspect. >> And another question is not related to that. But you're in Tel Aviv, Israel. And of course, Israel is famous for the start ups in the areas of machine learning. And so, especially with a focus on cyber security of the Israel, is like near the top of the world in terms of just the amount of brainpower focused on cyber security there. What are the hot ML machine? Learning related developments or innovations you see, coming out of Israel recently related to cybersecurity and distributed cloud environments, anything in terms of just basic are indeed technology that we should all be aware of that will be finding its way into mainstream Cloud and Cooper Netease and civilised environments. Going forward, your thoughts. >> Yes, I think there are different areas, you know, The guys in Israel also look at what happens in sort of the U. S. And their place in all the different things. I think with what's unique about us is a small country is always trying to think outside of the box because we know we cannot compete in a very large market. It would not have innovation. So that's what triggers this ten of innovation part because of all this tippy expects in the country. And also there's a lot of cyber, you know, it's time. I think I've seen one cool startup. There's also backed by our VC selling. Serve, uh, think about like face un recognition, critical technology off sent you a picture and make it such that you machine learning will not be able to recognize Recognize that, you know, sort of out of the cyber attack for image recognition. So that's something pretty unique that I've heard. But there are other starts working on all the aspects on their ops and information in our animal and also cyber automated cyber security and hope. Curious aspect. >> Right, Right. Thank you very much. Your own. This has been an excellent conversation, and we've really enjoyed hearing your comments. And Iguazu. It was a great company. Quite quite an innovator is always a pleasure to have you on the Cube. With that, I'm going to sign off. This is James Kabila's with wicked bond with your own haviv on dh er we bid You all have a good day. >> Thank you.

Published Date : Apr 4 2019

SUMMARY :

From our studios in the heart of Silicon Valley. It's your own Haviv Close the deal of any thanks from my seeing you again. new opportunities or possibilities that the convergence of those technologies enable for A scientist Inning the silo, you know, with a bunch of large that Which is that A. I is the heart of modern applications built, OK, just over the years, you know, people, four years when we started, of data of the data science pipeline kick you connect the dots of nuclear and data science and a I from So, and the interesting point is that if you think You know, among teams you have a high profile partnerships with Microsoft and, you know, if someone already has his work bench that I don't know my customers say they were locking me are you Are there you just testing the waters in that market for your technology? you know, in in intercepting data from your sister's customers, This is the week of Flink forward here in San San Francisco. And also under you grow your market, which is a together Melon ox, and I believe you used to be with Melon Axe, so I'd like to get your commentary on that acquisition Well, I think it's a great merger if you think about men in in terms of just the amount of brainpower focused on cyber security there. And also there's a lot of cyber, you know, it's time. Quite quite an innovator is always a pleasure to have you on the Cube.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
MicrosoftORGANIZATION

0.99+

SamsungORGANIZATION

0.99+

IsraelLOCATION

0.99+

GoogleORGANIZATION

0.99+

San San FranciscoLOCATION

0.99+

April 2019DATE

0.99+

James KabilaPERSON

0.99+

IguazuLOCATION

0.99+

Silicon ValleyLOCATION

0.99+

eleven monthsQUANTITY

0.99+

AmazonORGANIZATION

0.99+

Tel AvivLOCATION

0.99+

Yaron HavivPERSON

0.99+

Wicked BondORGANIZATION

0.99+

two weeksQUANTITY

0.99+

twelve batonsQUANTITY

0.99+

Palo AltoLOCATION

0.99+

first questionQUANTITY

0.99+

HavivPERSON

0.99+

three pillarsQUANTITY

0.99+

third elementQUANTITY

0.99+

last weekDATE

0.99+

twoQUANTITY

0.99+

BrianPERSON

0.99+

KansasLOCATION

0.99+

TodayDATE

0.99+

WSORGANIZATION

0.99+

JupiterLOCATION

0.99+

ErosORGANIZATION

0.98+

bothQUANTITY

0.98+

IndiaLOCATION

0.98+

OxORGANIZATION

0.98+

second thingQUANTITY

0.98+

hundreds of megabytesQUANTITY

0.98+

VP Data CenterORGANIZATION

0.98+

oneQUANTITY

0.98+

earlier this weekDATE

0.98+

second boxQUANTITY

0.98+

EuropeLOCATION

0.98+

U. S.LOCATION

0.98+

four yearsQUANTITY

0.98+

two pillarsQUANTITY

0.98+

IguazuPERSON

0.98+

Melon AxeORGANIZATION

0.97+

two versionQUANTITY

0.97+

todayDATE

0.97+

first eventQUANTITY

0.97+

Tel Aviv, IsraelLOCATION

0.97+

eachQUANTITY

0.96+

threeQUANTITY

0.95+

One commandQUANTITY

0.95+

EllisPERSON

0.94+

a year agoDATE

0.94+

IguazioPERSON

0.94+

DellaPERSON

0.94+

FlinkORGANIZATION

0.94+

One thingQUANTITY

0.94+

LandaTITLE

0.94+

second oneQUANTITY

0.93+

three major technologiesQUANTITY

0.93+

few weeks agoDATE

0.92+

SilverORGANIZATION

0.92+

single engineQUANTITY

0.92+

a year orDATE

0.91+

about two hundred timesQUANTITY

0.9+

two agoDATE

0.88+

Derrickson NamesakeORGANIZATION

0.88+

EarlPERSON

0.88+

GuiderORGANIZATION

0.88+

HOLLOWAY ALTO, CaliforniaLOCATION

0.86+

AmericansLOCATION

0.86+

three componentsQUANTITY

0.85+

FrostPERSON

0.81+

Lambda Stage MakerTITLE

0.8+

Tokyo JupiterORGANIZATION

0.8+

chinaLOCATION

0.78+

ArabOTHER

0.77+

dozensQUANTITY

0.77+

EpsteinPERSON

0.76+

last two yearsDATE

0.75+

David Aronchick, Microsoft | KubeCon 2018


 

I'm from Seattle Washington it's the cube covering Gube Khan and cloud native Khan North America 2018 brought to you by Red Hat the cloud native computing foundation and its ecosystem partners ok welcome back everyone we are here live with cube covers three days with wall-to-wall coverage here at coop con cloud native con 2018 in Seattle I'm John fer with the cubes to Minutemen here breaking it down we're at day two we've got a lot of action David Ronn chick who's the head of open source ml strategy at Azure at Microsoft Microsoft Azure formerly of Google now at Microsoft welcome back to the cube we had a great chat at Copenhagen good to see you great to see you too thank you so much for having me you've been there from day one it's still kind of day one in Korea is still growing you got a new gig here at Microsoft formerly at Google you had a great talk at Google next by the way which we watched and and caught on online you just you're still doing the same thing think of me to explain kind of what the new job is what your focus is absolutely so in many ways I'm doing a very similar job to the one I was doing at Google except now across all of Asher you know when you look at machine learning today the truth of the matter is is it is about open source it's about pulling in the best from academia and open source contributors developers across the spectrum and while I was at Google I was able to launch the cube flow project which solves the very specific but very important problem now that you look at Azure a company that is growing excuse me a division that is growing extremely quickly and looking to expand their overall open source offerings make investments work with partners and projects and make sure that that researchers and customers are able to get to machine learning solutions very quickly I'm coming in to help them think about how to make those investments and accelerate customers overall time to solutions so both on the commercial side Asscher which is got a business objective to make money but also open source how is it still open source for you is it all open sores or is it crossing a little bit of bulk just quickly clarify that yeah there's no question um you know obviously as you as a business they pay me a salary and and we're gonna have a great first party solution for all of these very things but the reality is much like kubernetes has both a commercial offering and an open-source offering I think that all the major cloud providers will have that kind of duality they'll work in open source and and you can measure you know how many contributions and what they're doing in the open source projects but then they'll also have hosted and other versions that make it easier for customers to migrate their data and adopt some of these new so you know one of the things that's interesting on that point is this a super important point is that open source community that's here with kubernetes around kubernetes it's all kind of upstream kind of concept but the downstream impacts our IT and your classic developer so you have your open source yeah and a thing going on that's the core of this community an event the IT investments are shifting in 2019 we are seeing the trend of somewhat radical but certainly a reimagining of the IT I mean certainly you guys have gone cloud at Azure has seen that that result absolutely good pick up by customers office 365 that's now a SAS that's now now you've got cloud you have cloud scale this is what machine learning is really shining so I the question to you is what do you think is gonna be the big impact of 2019 to IT investment strategies in terms of what they how they procure and consume technology how they build their apps with the new goodness coming in from kubernetes etc absolutely um you know I remember back in the day you know I was an IT admin myself and and I carried a pager for literally when you know a machine went down or a power supply went out or this Ram was bad or something like that today if you went to even the most sophisticated IT shop they would be like what are you crazy you you should never carry a pager for that you should have a system that understands it's ok if something that low-level goes out that's exactly what kubernetes provided it provided this abstraction layer on top of this so if you went down kubernetes knew had a reschedule a pod and move things back and forth taking that one step further now into machine learning unfortunately today people are carrying pagers for the equivalent of if a power supply goes out or something goes wrong it's still way too low-level we're asking data scientists ml engineers to think about how to provision pods how'd it work on drivers how to do all these very very low-level things with things like kubernetes with things like hume flow you're now able to give higher level abstraction so a data scientist can in and you know open up their Jupiter notebook work on the model see how it works and when they're done they hit a button and it will provision out all the machines necessary all the drivers all the everything spin it up run that training job and bring it back and shut everything down so they won't wonder if you can help expand on that a little bit more so you know what one of the things that that's great about kubernetes is it can live in a diverse amount of infrastructure one of the biggest challenges with machine learning is you know where's my data how do I get to the right place where do I do the training you know we've spending a lot a couple of years looking at you know edge and you know what's the connectivity and how we're gonna do this you help just kind of pan us picture the landscape and what do we have solved and what are we working at trying to get put together yeah you know I think that's a really excellent question today there's so much focus on well are you gonna choose pi torch or tensorflow CNT k MX net you know numpy scikit-learn there are a bunch of really great frameworks out there done in the open source and we're really excited but the reality is when you look at the overall landscape that's just 5% of the work that the average data scientist goes through exactly your point how do I get my data in how do I transform it how do I visualize it generate statistics on it make sure that it's not biased towards certain populations and then once I'm done training how do I roll it out to production and monitor it and log and all these things and that's really what we're talking about that's what we tried to get work on when it comes to cute flow is is to think about this in a much broader sense and so you take things like data the reality is you can't beat the speed of light if I have a petabyte of data here it's gonna take a long time to move it over there and so you're gonna be really thoughtful about those kind of things i I'm very hopeful that academic research and and industry will figure out ways to reduce the amount of data and make it much much more sane in overall addressing this problem and make it easier to train in various locations but the reality is is I think you're ultimately gonna have models and training and inference move to many many different locations and so you'll do inference at the edge on my phone or on a you know little Bluetooth device in the corner of my house saying whether or not it's too hot or too cold we're gonna need that kind of intelligence and we're gonna do that kind of training and data collection at the edge do you see a landscape evolving where you have specialty ml for instance like the big caution in IOT is move you know compute to the data yeah reads that latency you see machine learning models moving around at code so I can throw a machine learning at a problem and there's that and that is that what kubernetes fits and I'm trying to put together a mental model of how to think about how ml scales yeah what's your vision on that how do you see that evolving yeah absolutely I think that you know going back to what we talked about at the beginning we're really moving to much more of a solution driven architecture today ml you know is great and the academic research is phenomenal but it is academic research it didn't really start to take off until people invented things are you know creating things like image Nets and mobile net and things like that that did very important things like object detection but then people that you know commercial researchers were able to take that and move that into locations where people actually need it in I think you will continue to see that that migration I don't think you're gonna have single ml models that do a hundred different things you're gonna have a single ml model that does a vertical specific thing anomaly detection in whatever factories and you're gonna use that in a whole variety of locations rather than trying to you know develop 1 ml model to solve them all so it's application specific or vertical alright so that means the data is super important quality data clean data is clean results dirty date bad result absolutely right people have been in this kind of virtuous circle of cleaning data you know you guys know at Google certainly Microsoft as well you know datum data quality is critical but you got the horizontally scalable cloud but you need specialism around the data and for them ml how do you see that is that I mean obviously sounds like the right architecture this is where the finesse is and the nuance I don't see that so you know you you bring up a really interesting point today the the biggest problem is is how much data there is right it's not a matter of whether or not you're able to process it you are but but it's so easy to get lost caught and little anomalies you know if you have a petabyte of data and whatever a megabyte of it is the thing that's causing your model to go sideways that's really hard to detect I think what you're seeing right now is a lot of academic research which I'm very optimistic about that will ultimately reduce that that will both call out hey this particular data is smells kind of weird maybe take a closer look at this or you will see a smaller need for training you know where it was once a petabyte you're able to train on just 10 gigabytes I'm very optimistic that both of those things happen and as you start to get to that you get better signal-to-noise and you start saying oh in fact this is questionable data let's move that off to the side or spend more time on it rather than what happens today which is oh I got this model and it works pretty well I'm just going to throw everything at it and trying you know get some answer out and then we'll go from there and that's with a lot of false positives come in all absolutely all right so take the next level here at Kubb con cloud native con in this community where kubernetes is the center of all these sets of services and building blocks where's the ML action what if I Michelle wanna jump in this community I'm watching this with hey you know what I got Amazon Web Services reinvent just pumping up a lot of MLA I you know stage maker and a bunch of other things what's going on in this community where are the projects what are the notable things where can I jump in and engage what's the what's that what's that map look like I don't know yeah absolutely so obviously I'm pretty biased you know I helped start cube flow we're very very excited about that the cube flows one yeah absolutely but let me speak a little bit more broadly kubernetes gives you this wonderful platform highly scalable incredibly portable and and I can't overstate how valuable that portability is the reality is is that customers have we talked about data a bunch already they have data on Prem they've data in cloud hey cloud B it's everywhere they want to bring it together they want to bring the the training and the inference to where the data is kubernetes solves that for you it gives you portability and lets you abstract away the underlying stuff it gives you great scalability and reliability and it lets you compose these highly complex pipelines together that let you do real training anywhere rather than having to take all your data and move it through cloud and train on a single VM that you're not sure whether or not it's been updated or not this is the way to go versus the old way which was what cuz that's an easier way orchestrating and managing that what was the alternative the alternative was you built it yourself you you piece together a whole bunch of solutions you wired it together you made sure that this service over here had the right user account to access the data that that service over there was outputting it was just a crazy time now you use kubernetes constructs use first-class objects you extend the native kubernetes api and it works on your laptop and it works on Cloud a and B and on pram and wherever you need it that's the magic basically absolutely so multi cloud has come up a lot hybrid clouds the buzzword of the year I call that the 2000 18 maybe 19 buzzword but I think the real end game and all this is what from a customer standpoint that we are reporting a silk'n angle on the cube is choice yeah multi vendor is the new multi cloud is the multi clouds the modern version of the old multi vendor comes yes which basically is choice absolutely so how does kubernetes fit into the multi cloud why is that good for the industry and what's your take on that can you share your perspective absolutely so when you go and look at the recent right scale reports 81 percent of enterprises today are multi cloud . 81 percent and not just one cloud there they're on five different clouds that could be on pram could be multi zone could be Google or Amazon or a Salesforce you name how you define cloud they're spreading they're doing it because that kind of portability is right for their business kubernetes gives you the opportunity to operate in an abstraction layer that works across all of these clouds so whether or not you're on your laptop and you're using docker or mini cube you're on your private training rig whether that you go to Google cloud or as you're on Google clouds you can eat user you have a KS these you're able to build C I'd CD systems continuous delivery systems that that use common kubernetes constructs I want to roll this application out I want there to be seven pods I wanted to have an endpoint that looks like this and that works anywhere you have a kubernetes conformant cluster and when it gets to really complex apps like machine learning you're able to do that it even a higher level using constructs like cube flow and all the many many packages that go into coop load we have Nvidia contributing and we have you know Intel and I mean just countless Cisco I you know I hesitate to keep naming names because I'll be here all day but you know we have literally over Cisco's rays tailwind Francisco they're gonna have Network forever everybody wins at the the CI CD sides for developers one common construct the network guys get more programming because if you decompose an application absolutely the network ties it together yes everybody wins in the stack absolutely I think I breed is really interesting you know hybrid kind of gets a dirty word people like oh my god you know why would you ever deploy to multiple clouds why would you ever spread across multiple clouds and that I agree with a true hybrid deployment today isn't well I'm gonna take my app and I'm gonna spread it across six different locations in fact what you really want to do is have isolated deployments to each place that it enables you in a single button deploy to all three of these locations but to isolate them to have this particular application go and if you know AWS hasn't added GCP is there or if GCB does manage asher is there and you can do that very readily or you can bring it closed for geographic reasons or legal reasons or whatever it might be those kind of flexibility that ability to take a single construct of your application and deploy it to each one of these locations not spreading them but in fact just giving you that flexibility gives you pricing power gives you flexibility and lets you take advantage of the operating model if the if the if the ICD is common and that's the key value right there absolutely right David thanks so much coming on cue as usual great commentary great insight there there from the beginning just final question predictions for 2019 I think kubernetes what's gonna happen in 2019 with kubernetes what's your prediction well III think I think you've heard this message over and over again you're seeing kubernetes become boring and and that is incredibly powerful the the stability the flexibility people are building enormous businesses on top of it but not just that they're also continuing to build things like the the custom resource definition which lets you extend kubernetes in a safe and secure way and that's incredibly important that means you don't have to go and check in code into the main tree in order to make extension you're able to build on top of it and you're seeing more and more businesses build eight solutions customer focus solutions well next time we get together I want to do a drill down on the what the word stack means I heard me say kubernetes stack I'm like yeah I think that you love the stack words let a stack anymore sets the services David thanks so much come on I appreciate it here the queue coverage live here in Seattle for coop con cloud native found I'm John Fourier was too many men we back with more after this short break

Published Date : Dec 12 2018

SUMMARY :

really shining so I the question to you

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
DavidPERSON

0.99+

Red HatORGANIZATION

0.99+

2019DATE

0.99+

SeattleLOCATION

0.99+

AmazonORGANIZATION

0.99+

KoreaLOCATION

0.99+

81 percentQUANTITY

0.99+

GoogleORGANIZATION

0.99+

David AronchickPERSON

0.99+

John FourierPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Amazon Web ServicesORGANIZATION

0.99+

CiscoORGANIZATION

0.99+

AWSORGANIZATION

0.99+

David Ronn chickPERSON

0.99+

Seattle WashingtonLOCATION

0.99+

todayDATE

0.99+

three daysQUANTITY

0.99+

NvidiaORGANIZATION

0.98+

CopenhagenLOCATION

0.98+

10 gigabytesQUANTITY

0.98+

1 mlQUANTITY

0.97+

81 percentQUANTITY

0.97+

Kubb conORGANIZATION

0.97+

AsscherORGANIZATION

0.96+

John ferPERSON

0.96+

oneQUANTITY

0.96+

bothQUANTITY

0.95+

a couple of yearsQUANTITY

0.95+

IntelORGANIZATION

0.94+

each oneQUANTITY

0.94+

2018DATE

0.94+

each placeQUANTITY

0.93+

six different locationsQUANTITY

0.93+

day twoQUANTITY

0.92+

one cloudQUANTITY

0.91+

singleQUANTITY

0.9+

one stepQUANTITY

0.89+

five different cloudsQUANTITY

0.89+

single buttonQUANTITY

0.89+

kubernetesORGANIZATION

0.89+

day oneQUANTITY

0.89+

single VMQUANTITY

0.88+

SalesforceORGANIZATION

0.87+

AzureTITLE

0.85+

threeQUANTITY

0.84+

5% ofQUANTITY

0.83+

single mlQUANTITY

0.82+

single ml modelQUANTITY

0.81+

JupiterLOCATION

0.81+

IOTTITLE

0.8+

coop conORGANIZATION

0.78+

petabyteQUANTITY

0.75+

thingsQUANTITY

0.75+

KubeCon 2018EVENT

0.75+

Google cloudTITLE

0.74+

AsherORGANIZATION

0.73+

first partyQUANTITY

0.72+

AzureORGANIZATION

0.71+

a hundred different thingsQUANTITY

0.7+

coop con cloudORGANIZATION

0.7+

a petabyteQUANTITY

0.67+

seven podsQUANTITY

0.67+

petabyte of dataQUANTITY

0.66+

19DATE

0.66+

solutionsQUANTITY

0.63+

North AmericaLOCATION

0.62+

Madhu Kochar, IBM, Susan Wegner, Deutsche Telekom | IBM CDO Fall Summit 2018


 

>> Live from Boston, it's theCUBE covering IBM Chief Data Officer Summit. Brought to you by IBM. >> Welcome back everyone to theCUBE's live coverage of the IBM CDO Summit here in beautiful Boston, Massachusetts. I'm your host, Rebecca Knight, along with my co-host Paul Gillin. We have two guests for this segment, we have Susan Wagner, who is the VP Data Artificial Intelligence and Governance at Deutsche Telekom and Madhu Kochar, whose the Vice President Analytics Product Development at IBM. Thank you so much for coming on the show. >> Thank you. >> Happy to be here. Susan you're coming to us from Berlin, tell us a little bit about what you it's a relatively new job title and Paul was marveling before the cameras are rolling. Do you have artificial intelligence in your job title? Tell us a little bit about what you do at Deutsche Telekom. >> So we have a long history, working with data and this is a central role in the headquarter guiding the different data and artificial intelligence activities within Deutsche Telekom. So we have different countries, different business units, we have activities there. We have already use case catalog of 300,000 cases there and from a central point we are looking at it and saying, how are we able really to get the business benefit out of it. So we are looking at the different product, the different cases and looking for some help for the business units, how to scale things. For example, we have a case we implemented in one of our countries, it was about a call center to predict if someone calls the call center, if this is a problem, we would never have(laughing) at Deutsche Telekom but it could happen and then we open a ticket and we are working on it and then we're closing that ticket and but the problem is not solved, so the ticket comes again and the customer will call again and this is very bad for us bad for the customer and we did on AI project, there predicting what kind of tickets will come back in future and this we implemented in a way that we are able to use it not only in one country, but really give it to the next country. So our other business units other countries can take the code and use it in another country. That's one example. >> Wow. >> How would you define artificial intelligence? There's someone who has in your job-- (laughing) >> That's sometimes very difficult question I must admit. I'm normally if I would say from a scientific point, it's really to have a machine that works and feels and did everything like a human. If you look now at the hype, it's more about how we learn, how we do things and not about I would say it's about robotic and stuff like that but it's more how we are learning and the major benefit we are getting now out of artificial intelligence is really that we are able now to really work on data. We have great algorithm and a lot of progress there and we have the chips that develops so far that we are able to do that. It's far away from things like a little kid can do because little kid can just, you show them an apple and then it knows an apple is green. It's were-- >> A little kid can't open a support ticket. (laughing) >> Yeah, but that's very special, so in where we special areas, we are already very, very good in things, but this is an area, for example, if you have an (mumbles) who is able like we did to predict this kind of tickets this agreement is not able at the moment to say this as an apple and this is an orange, so you need another one. So we are far away from really having something like a general intelligence there. >> Madhu do I want to bring you into this conversation. (laughing) And a little bit just in terms of what Susan was saying the sort of the shiny newness of it all. Where do you think we are in terms of thinking about the data getting in the weeds of the data and then also sort of the innovations that we saw, dream about really impacting the bottom line and making the customer experience better and also the employee experience better? >> Yeah, so from IBM perspective, especially coming from data and analytics, very simple message, right? We have what we say your letter to AI. Everybody like Susan and every other company who is part of doing any digital transformation or modernization is talking about Ai. So our message is very simple, in order to get to the letter of AI, the most critical part is that you have access to data, right? You can trust your data, so this way you can start using it in terms of building models, not just predictive models but prescriptive and diagnostics. Everything needs to kind of come together, right? So that is what we are doing in data analytics. Our message is very, very simple. The innovations are coming in from the perspectives of machine learning, deep learning and making and to me that all equates to automation, right? A lot of this stuff data curation, I think you can Susan, tell how long and how manual the data curation aspects can be. Now with machine learning, getting to your latter of AI, You can do this in a matter of hours, right? And you can get to your business users, you can if your CHARM model, If your clients are not happy, your fraud, you have to detect in your bank or retail industry, it just applies to all the industry. So there is tons of innovation happening. We just actually announced a product earlier called IBM Cloud Private for Data. This is our the analytics platform which is ready with data built in governance to handle all your data curation and be building models which you can test it out, have all the DevOps and push it into production. Really, really trying to get clients like Deutsche Telekom to get their journey there faster. Very simple-- >> We've heard from many of our guests today about the importance of governance, of having good quality data before you can start building anything with it. What was that process like? How is the... what is the quality of data like at Deutsche Telekom and what work did it take to get it in that condition. >> So data quality is a major issue everywhere, because as Madhu that this is one of the essential things to really get into learning, if you want to learn, you need the data and we have in the different countries, different kind of majorities and what we are doing at the moment is that we are really doing it case by case because you cannot do everything from the beginning, so you start with one of the cases looking what to do there? How to define the quality? And then if the business asked for the next case, then you can integrate that, so you have the business impact, you have demand from the business and then you can integrate the data quality there and we are doing it really step by step because to bring it to the business from the beginning, it's very, very difficult. >> You mentioned, one of the new products that you announced just today, what are some of the-- (laughing) >> We announced it in may. >> Oh, okay, I'm sorry. >> It's okay still new. >> In terms of the other innovations in the pipeline, what I mean this is such a marvelous and exciting time for technology. What are some of the most exciting developments that you see? >> I think the most exciting, especially if I talk about what I do day out everything revolves around metadata, right? Used to be not a very sticky term, but it is becoming quite sexy all over again, right? And all the work in automatic metadata generation, understanding the lineage where the data is coming from. How easy, we can make it to the business users, then all the machine learning algorithms which we are doing in terms of our prescriptive models and predictive, right? Predictive maintenance is such a huge thing. So there's a lot of work going on there and then also one of the aspects is how do you build once and run anywhere, right? If you really look at the business data, it's behind the firewalls, Is in multicloud. How do you bring solutions which are going to be bringing all the data? Doesn't matter where it resides, right? And so there's a lot of innovation like that which we are working and bringing in onto our platform to make it really simple story make data easy access which you can trust. >> One of the remarkable things about machine learning is that the leading libraries have all been open source, Google, Facebook, eBay, others have open source their libraries. What impact do you think that has had on the speed with which machine learning is developed? >> Just amazing, right. I think that gives us that agility to quickly able to use it, enhance it, give it back to the community. That has been the one of the tenants for, I think that how everybody's out there, moving really really fast. Open source is going to play a very critical role for IBM, and we're seeing that with many of our clients as well. >> What tools are you using? >> We're using different kind of tools that depending on the departments, so the data scientists like to use our patents. (laughing) They are always use it, but we are using a lot like the Jupiter notebook, for example, to have different kind of code in there. We have in one of our countries, the classical things like thus there and the data scientists working with that one or we have the Cloud-R workbench to really bringing things into the business. We have in some business-- >> Data science experience. >> IBM, things integrated, so it it really depends a little bit on the different and that's a little bit the challenge because you really have to see how people working together and how do we really get the data, the models the sharing right. >> And then also the other challenges that all the CDOs face that we've been talking about today, the getting by in the-- >> Yes. >> The facing unrealistic expectations of what data can actually do. I mean, how would you describe how you are able to work with the business side? As a chief working in the chief data office. >> Yeah, so what I really like and what I'm always doing with the business that we are going to the business and doing really a joint approach having a workshop together like the design thinking workshop with the business and the demand has to come from the business. And then you have really the data scientists in there the data engineers best to have the operational people in there and even the controlling not all the time, but that it's really clear that all people are involved from the beginning and then you're really able to bring it into production. >> That's the term of DataOps, right? That's starting to become a big thing. DevOps was all about to agility. Now DataOps bring all these various groups together and yeah I mean that's how you we really move forward. >> So for organizations so that's both of you for organizations that are just beginning to go down the machine learning path that are excited by everything you've been hearing here. What advice would you have for them? They're just getting started. >> I think if you're just getting started to me, the long pole item is all about understanding where your data is, right? The data curation. I have seen over and over again, everybody's enthusiastic. They love the technology, but the... It just doesn't progress fast enough because of that. So invest in tooling where they have automation with machine learning where they can quickly understand it, right? Data virtualization, nobody's going to move data, right? They're sitting in bedrock systems access to that which I call dark data, is important because that is sometimes your golden nugget because that's going to help you make the decisions. So to me that's where I would focus first, everything else around it just becomes a lot easier. >> Great. >> So-- >> Do you have a best practice too? Yeah. >> Yeah. Focus on really bringing quick impact on some of the cases because they're like the management needs success, so you need some kind of quick access and then really working on the basics like Madhu said, you need to have access of the data because if you don't start work on that it will take you every time like half a year. We have some cases where we took finance department half a year to really get all that kind of data and you have to sharpen that for the future, but you need the fast equipments. You need to do both. >> Excellent advice. >> Right, well Susan and Madhu thank you so much for coming on theCUBE, it's been great having you. >> Thank you. >> Thank you. >> I'm Rebecca Knight for Paul Gillin we will have more from theCUBE's live coverage of the IBM CDO just after this. (upbeat music)

Published Date : Nov 15 2018

SUMMARY :

Brought to you by IBM. Thank you so much for coming on the show. tell us a little bit about what you bad for the customer and we did are learning and the major benefit we are getting now A little kid can't open a support ticket. but this is an area, for example, if you have an (mumbles) and making the customer experience better and be building models which you can test it out, before you can start building anything with it. the business impact, you have demand from the business In terms of the other innovations in the pipeline, one of the aspects is how do you build once is that the leading libraries have all been open source, That has been the one of the tenants for, I think that how departments, so the data scientists like to use our patents. the challenge because you really have to see how I mean, how would you describe and the demand has to come from the business. and yeah I mean that's how you we really move forward. So for organizations so that's both of you They love the technology, but the... Do you have a best practice too? and you have to sharpen that for the future, Right, well Susan and Madhu thank you so much I'm Rebecca Knight for Paul Gillin we will have more

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
SusanPERSON

0.99+

Susan WagnerPERSON

0.99+

MadhuPERSON

0.99+

Rebecca KnightPERSON

0.99+

Paul GillinPERSON

0.99+

IBMORGANIZATION

0.99+

Deutsche TelekomORGANIZATION

0.99+

BerlinLOCATION

0.99+

PaulPERSON

0.99+

FacebookORGANIZATION

0.99+

GoogleORGANIZATION

0.99+

Madhu KocharPERSON

0.99+

Susan WegnerPERSON

0.99+

two guestsQUANTITY

0.99+

eBayORGANIZATION

0.99+

300,000 casesQUANTITY

0.99+

bothQUANTITY

0.99+

half a yearQUANTITY

0.99+

BostonLOCATION

0.99+

todayDATE

0.99+

one countryQUANTITY

0.99+

Boston, MassachusettsLOCATION

0.99+

oneQUANTITY

0.98+

IBM CDO SummitEVENT

0.96+

one exampleQUANTITY

0.96+

firstQUANTITY

0.94+

OneQUANTITY

0.92+

appleORGANIZATION

0.92+

IBM Chief Data Officer SummitEVENT

0.9+

mayDATE

0.9+

IBM CDO Fall Summit 2018EVENT

0.89+

theCUBEORGANIZATION

0.86+

one ofQUANTITY

0.76+

one of our countriesQUANTITY

0.72+

Cloud Private forTITLE

0.69+

DevOpsTITLE

0.66+

JupiterLOCATION

0.62+

aspectsQUANTITY

0.58+

Vice PresidentPERSON

0.58+

essentialQUANTITY

0.54+

casesQUANTITY

0.52+

tenantsQUANTITY

0.46+

CDOCOMMERCIAL_ITEM

0.42+

DataOpsTITLE

0.4+

DataOpsORGANIZATION

0.37+

Thomas LaRock, SolarWinds | Microsoft Ignite 2018


 

(music) >> Live from Orlado, Florida, it's theCUBE. Covering Microsoft Ignite. Brought to you by Cohesity. and theCube's ecosystem partners. >> Welcome back, everyone, to theCube's live coverage of Microsoft Ignite. Happy hour has started. The crowd is roaring. I'm your host Rebecca Knight, along with my cohost, Stu Miniman. We are joined by Thomas LaRock. >> He is the Head Geek at SolarWinds. Thanks so much for coming on the show. >> Thanks for having me. >> Great title: Head Geek >> Yes. >> So, tell our viewers a little bit about what - tell us about SolarWinds and also about what you do. >> SolarWinds is a company that offers about forty different products to help with your enterprise infrastructure monitoring. Really unify management of your systems. Been in the business for about twenty years and I've been with them for about eight now. Head Geek is really, uh, you can equate it to being a technical evangelist. >> Okay. So you're out there trying to win the hearts and minds, trying to tell everyone what you do. >> Yes, I need you all to love me. (laughing) and love my products. >> So, Thomas, and for those who don't already follow you on Twitter, you're a SQL rockstar. >> Yes, yes [Stu] - I need to say, "thank you," because you helped connect me with a lot of the community here, especially on the data side of the house. You and I have known each other for a bunch of years. You're a Microsoft MVP. So maybe give us a little bit of community aspect: what it means to be a Microsoft MVP for those who don't know. You're an evangelist in this space and you've been on this show many times. >> I usually don't talk about myself a lot, but sure. (Rebecca laughing) Let's go for it. I've been a Microsoft data platform MVP for about 10 year now. And it was intresting when you reached out, looking to get connected. I was kind of stunned by how many people I actually knew or knew how to get in touch with for you. I help you line up, I guess, a handful of people to be on the show because you were telling me you hadn't been here at Microsoft Ignite and I just thought, "well I know people," and they should know Stu, and we should get them connected so that you guys can have some good conversations. But, yeah, it's been a wild ride for me those ten years where Microsoft awards people MVP designation. It's kind of being an evangelist for Microsoft and some of the great stuff that they've been doing over the past ten years. >> It's a phenomenal program. Most people in the technology industry know the Microsoft MVP program. I was a Vmware expert for a number of years. Many of the things were patterned off of that. John Troyer is a friend of mine. He said that was one the things he looked at. Sytrics has programs like this. Many of the vendors here have evangelists or paragons showing that technology out here. Alight. So talk a little bit about community. Talk about database space. Data and databases have been going through such, you know, explosion of what's going on out there, right? SQL's still around. It's not all cosmos and, you know, microservices-based, cloud, native architecture. >> So the SQL Server box product is still around, but what I think is more amazing to me has been the evalution of...Let's take for example, one of the announcements today, the big data cluster. So, it's essentially a container that's going to run SQL servers, Spark and Hadoop, all in one. Basically, a pod that will get deployed by kubernetes. When you wrap all that together, what you start to realize is that the pattern that Microsoft has been doing for the past few years, which is, essentially, going to where the people are. What I mean is: you have in the open-source world, you have people and developers that have embraced things like DevOps much faster than what the Windows developers have been doing. So instead of taking your time trying to drag all these people where you want them to be, they've just start building all the cool stuff where all the cool kids already are, and everybody's just going to gravitate. Data has gravity, right? So, you're building these things, and people are going to follow it. Now, it's not that they're expecting to sell a billion dollars woth of licenses. No. They just need to be a part of the conversation. So if you're a company that's using those technologies, now all of a sudden, it's like, this is an option. Are you interested in it? Microsoft is the company that's best poised to bring enterprises to the cloud. Amazon has a huge share. We all know that, but Microsoft's already that platform of choice for these enterprises. Microsoft is going to be the one to help them get to the cloud. [Stu]- Thomas, Explain what you mean by that because the strength I look at Microsoft is look, they've got your application. Business productivity: that's where they are. Apologize for cutting you off there. Is that what you mean? The applications are changing and you trusted Microsoft and the application and therefore, that's a vendor of choice. >> Absolutely. If it's already your vendor of choice then, I don't want to say, "Lock in," but if it's already your preference and if they can help get to the cloud, or in the hybrid situation or just lift and shift and just get there, then that's the one you going to want to do it. Everything they're building and all the services they're providing... At the end of the day, they and Amazon, they're the new electric company. They want data. That's the electricity. They don't care how you get it, but between... even Vmware. Between Amazon, Vmware and Microsoft, they're going to be the ones to help... They're going to be your infrastructure companies. Microsoft-managed desktop now. We'll manage your laptop for you. >> Everything that they're doing essentially like, don't even need my own IT department. Microsoft's going to be the largest MSP in history, right? That's where they're headed. They're going to manage everything for you. The data part of it, of course for me, I just love talking about data. But the data part of it...Data is essential to everything we do. It's all about the data. They're doing their best to manage it and secure it. Security is a huge thing. There were some security announcements today as well, which were awesome. The advanced threat detection, the protection that they have. I'm always amazed when I walk through the offering they have for SQL injection protection. I try and ask people, "Who's right now monitoring for SQL injection?" And they're like, "We're not doing that." For fifteen dollars a month, you could do this for your servers. They're like, "that's amazing what they're offerening." Why wouldn't you want that as a service? Why wouldn't you sign-up tomorrow for this stuff? So, I get excited about it. I think all this stuff they're building is great. The announcements today were great. I think they have more coming out over the next couple days. Or at least in the sessions, we'll start seeing a lot of hands-on stuff. I'm excited for it. >> So when you were talking about Microsoft being the automatic vendor of choice. Why wouldn't you? You treated it as a no brainer. What does Microsoft need to do to make sure customers feel that way too? >> I think Microsoft is going to do that... How I would do that. A couple ways. One, at the end of the day, Microsoft wants what we all want, what I want, is they want happy customers. So they're going to do whatever it takes so their customers are happy. So one way you do that is you get a lot of valuable feedback from customers. So, one thing Microsoft has done in the past is they've increased the amount telemetry they're collecting from their products. So they know the usage. They know what the customers want. They know what the customers need. But they also collect simple voice to the customer. You're simply asking the customer, "What do you want?" And you're doing everything you can to keep them happy. And you're finding out where the struggles are. You're helping them solve those problems. How do you not earn trust as a result of all that, right? I think that's the avenue they've been doing for, at least, ten years. Well, let's say, eight years. That's the avenue and the approach they've been doing. I'd say it's been somewhat successful. >> Thomas, as our team was preparing for this show, we understand that Microsoft has a lot of strengths, but if I look at the AI space, Microsoft is not the clear leader today. Um, we think that some of the connections that Microsoft has, everything that you said, down to the desktop. Heck, even in the consumer space, they're down to the Xbox. There's a lot of reasons why Microsoft... You can say, "Here's a path of how Microsoft could become. You know number one, number two in the AI space over time. But, we're listening to things, like the Open Data Initiative that they announced today, which, obviously, Microsoft's working with a lot of partners out there, but it's a big ecosystem. Data plays everywhere. I mean, Google obviously has strong play in data. We've talked plenty about Amazon. What does Microsoft need to do to take the strength that they have in data move forward in AI and become even stronger player in the marketplace? >> So, AI, itself, is kind of that broad term. I mean, AI is a simple if-then statement. It doesn't really have to do anything, right? So let's talk about machine learning, predictive analytics, or even deep learning. That's really the are that we're talking about. What does Microsoft have to do? Well, they have to offer the services. But they don't have offer, say, new things. They just have to offer things that already exist. For example, the idea of, um, incorperating Jupiter notebooks into the Azure Data Studio. So if that could be achieved, you know, now you're bringing the workspaces people are using into the Microsoft platform a little bit, making it a little bit easier. So instead of these people in these enterprises... They already trust Microsoft. They already have the tools. But I got to go use these other things. Well, eventually, those other things come into the Microsoft tools, and now you don't have to use that other stuff either. I would talk about the ability to publish these models as a service. I've done the Academy program. I've earned a few certifications on some of this stuff. I was amazed at how easy it was with a few clicks, you know, published as a service as an API. It's sitting there. I sent in my data and I get back a result, a prediction. I was like, that was really easy. So I know they're not the leaders, but they're making it easy, especially for somebody like me who can start at zero and get to where I need to be. They made it incredibly easy and in some cases, it was intuative. I'm like, oh, I know what to do next with this widgit I'm building. I think it will take time for them to kind of get all that stuff in place. I don't know how long. But does Microsoft have to be the leader in AI? They have the Cognitive Toolkit. They have all that stuff with Cortana. They have the data. I think the customers are coming along. I think they get there just by attrition. I'm not sure there's something they're going to build where everybody just says, "There it is." Except there's the Quantum stuff. And last year's announcement of Quantum, I thought was one of the most stunning things. It just hit me. I had no idea working on it. So, who knows? A year from now there could be something similar to that type of announcement, where we're like, now I get it, now I got to go have this thing. I don't think we all need, you know, a hotdog not hotdog app, which seems to be the bulk of the examples out there. Some of the image classification stuff that you have out there is fabulous. There are a lot of use cases for it. Um, I'm not sure how they get there. But, I do think eventually over time, the platform that they offer, they do get just through attrition. >> One of the things you brought up earlier in this conversation was the Open Source Initiative and Stu, we had expressed a bit of skepticism that it's still going to take three to five years, for, really, customers to see the value of this. But once...The announcement was made today, so now we're going to go forward with this Initiative. What do you see as the future? >> Yeah, I was trying to, even, figure it out. So it sounds like the three companies are sharing data with each other. They pledged to be open. So if you buy one of their products, that data can seamlessly go into that other product is what it sounded like. And they were open, if I heard it right, they were open to partnering with other companies as well. >> Correct. >> Yes. Yes. >> Other vendors or customers, even that could tie in into these APIs, doing everything that they're doing. Open data models. >> Speaking as a data guy, that means if I trust one, I have to trust them all. (Stu Laughing) >> Right? So I don't know. I have trus&t issues. (Rebecca laughing) >> Clearly. >> I'm a DBA, by heart, so I have trust issues. I need to know a little more about it, but on the surface, just the words, "open data," sound great. I just don't know the practical, uh, practicality of it. It sounds like it's a way for people, or these companies, to partner with each other to get more of your data into their platform and their infrastructure. >> Yeah. I think next time we have Thomas on, we're going to spend some time talking about the dark side of data. >> Yes, indeed. >> We can talk dark data. Oh, sure. (Rebecca laughing) >> Well, Thomas, it was so much fun having you on this show and I should just plug your book. You are the author of "DBA Survivor." >> I am. Yes. It was a little book. So being a DBA, uh, I had some challenges in my role and I decided, as my friend Kevin Kline put it to me, he goes, "You should write the book you wish had written for you and handed to you on day zero of being a DBA." And I said, "Oh." It took m&e, I think, like, three weeks. It was just so easyto write all of that. >> It just flowed (laughing.) >> It was just stuff I had to say. But, yeah, thank you. >> Excellent. I'm Rebecca Knight for Stu Miniman. We will have more from theCUBE's live coverage of Microsoft Ignite coming up in just a little bit. (music playing)

Published Date : Sep 24 2018

SUMMARY :

Brought to you by Cohesity. to theCube's live coverage of He is the Head Geek at SolarWinds. and also about what you do. Been in the business trying to tell everyone what you do. Yes, I need you all to love me. So, Thomas, and for those especially on the data side of the house. and some of the great stuff Many of the things were be the one to help them the ones to help... the protection that they have. about Microsoft being the So they're going to do whatever it takes Microsoft is not the clear leader today. I don't think we all need, you know, One of the things you So it sounds like the three doing everything that they're I have to trust them all. I have trus&t issues. I just don't know the practical, the dark side of data. We can talk dark data. You are the author of "DBA Survivor." the book you wish had written It was just stuff I had to say. I'm Rebecca Knight for Stu Miniman.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
Rebecca KnightPERSON

0.99+

MicrosoftORGANIZATION

0.99+

Kevin KlinePERSON

0.99+

AmazonORGANIZATION

0.99+

ThomasPERSON

0.99+

GoogleORGANIZATION

0.99+

threeQUANTITY

0.99+

Stu MinimanPERSON

0.99+

John TroyerPERSON

0.99+

RebeccaPERSON

0.99+

VmwareORGANIZATION

0.99+

three companiesQUANTITY

0.99+

Thomas LaRockPERSON

0.99+

SolarWindsORGANIZATION

0.99+

ten yearsQUANTITY

0.99+

CortanaTITLE

0.99+

eight yearsQUANTITY

0.99+

todayDATE

0.99+

tomorrowDATE

0.99+

last yearDATE

0.99+

oneQUANTITY

0.99+

XboxCOMMERCIAL_ITEM

0.99+

five yearsQUANTITY

0.99+

OneQUANTITY

0.98+

about twenty yearsQUANTITY

0.98+

SQLTITLE

0.98+

about forty different productsQUANTITY

0.97+

StuPERSON

0.97+

CohesityORGANIZATION

0.97+

about 10 yearQUANTITY

0.96+

three weeksQUANTITY

0.96+

theCubeORGANIZATION

0.95+

about eightQUANTITY

0.95+

SytricsORGANIZATION

0.95+

Stu LaughingPERSON

0.95+

WindowsTITLE

0.94+

Orlado, FloridaLOCATION

0.94+

fifteen dollars a monthQUANTITY

0.94+

theCUBEORGANIZATION

0.94+

Azure Data StudioTITLE

0.93+

TwitterORGANIZATION

0.93+

one wayQUANTITY

0.93+

DevOpsTITLE

0.92+

billion dollarsQUANTITY

0.92+

CUBE Highlights | theCUBE NYC 2018


 

the foundation of having that data management platform is absolutely fundamental and necessary to do good machinery without good data without good data management you can't do good ml or AI sounds sort of simple but very true the dupes of all the nature and velocity of data has evolved in the last five screens taking over the world and being in charge of you and us being dominated by them as often we say in culture now it's about having this really beautiful interface between technology objects I can take the traditional tools using like Jupiter no spark tensorflow you know those packages with kubernetes on top of the databases of service and some object stores I have a much easier stack to work in able everyone to make data-driven decisions but make sure that they're interpreting that data in the right way right give them enough guidance don't let them just kind of attack the well [Music]

Published Date : Sep 19 2018

**Summary and Sentiment Analysis are not been shown because of improper transcript**

ENTITIES

EntityCategoryConfidence
NYCLOCATION

0.87+

theCUBEORGANIZATION

0.83+

five screensQUANTITY

0.79+

JupiterTITLE

0.66+

CUBEORGANIZATION

0.65+

2018DATE

0.5+

Yaron Haviv, Iguazio | theCUBE NYC 2018


 

Live from New York It's theCUBE! Covering theCUBE New York City 2018 Brought to you by Silicon Angle Media and it's ecosystem partners >> Hey welcome back and we're live in theCUBE in New York city. It's our 2nd day of two days of coverage CUBE NYC. The hashtag CUBENYC Formerly Big data NYC renamed because it's about big data, it's about the server, it's about Cooper _________'s multi-cloud data. It's all about data, and that's the fundamental change in the industry. Our next guest is Yaron Haviv, who's the CTO of Iguazio, key alumni, always coming out with some good commentary smart analysis. Kind of a guest host as well as an industry participant supplier. Welcome back to theCUBE. Good to see you. >> Thank you John. >> Love having you on theCUBE because you always bring some good insight and we appreciate that. Thank you so much. First, before we get into some of the comments because I really want to delve into comments that David Richards said a few years ago, CEO of RenDisco. He said, "Cloud's going to kill Hadoop". And people were looking at him like, "Oh my God, who is this heretic? He's crazy. What is he talking about?" But you might not need Hadoop, if you can run server less Spark, Tensorflow.... You talk about this off camera. Is Hadoop going to be the open stack of the big data world? >> I don't think cloud necessary killed Hadoop, although it is working on that, you know because you go to Amazon and you know, you can consume a bunch of services and you don't really need to think about Hadoop. I think cloud native serve is starting to kill Hadoop, cause Hadoop is three layers, you know, it's a file system, it's DFS, and then you have server scheduling Yarn, then you have applications starting with map produce and then you evolve into things like Spark. Okay, so, file system I don't really need in the cloud. I use Asfree, I can use a database as a service, as you know, pretty efficient way of storing data. For scheduling, Kubernetes is a much more generic way of scheduling workloads and not confined to Spark and specific workloads. I can run with Dancerflow, I can run with data science tools, etc., just containerize. So essentially, why would I need Hadoop? If I can take the traditional tools people are now evolving in and using like Jupiter Notebooks, Spark, Dancerflow, you know, those packages with Kubernetes on top of a database as a service and some object store, I have a much easier stack to work with. And I could mobilize that whether it's in the cloud, you know on different vendors. >> Scale is important too. How do you scale it? >> Of course, you have independent scaling between data and computation, unlike Hadoop. So I can just go to Google, and use Vquery, or use, you know, DynamoDB on Amazon or Redchick, or whatever and automatically scale it down and then, you know >> That's a unique position, so essentially, Hadoop versus Kubernetes is a top-line story. And wouldn't that be ironic for Google, because Google essentially created Map Produce and Coudera ran with it and went public, but when we're talking about 2008 timeframe, 2009 timeframe, back when ventures with cloud were just emerging in the mainstream. So wouldn't it be ironic Kubernetes, which is being driven by Google, ends up taking over Hadoop? In terms of running things on Kubernetes and cloud eight on Visa Vis on premise with Hadoop. >> The poster is tend to give this comment about Google, but essentially Yahoo started Hadoop. Google started the technology  and couple of years after Hadoop started, with Google they essentially moved to a different architecture, with something called Percolator. So Google's not too associated with Hadoop. They're not really using this approach for a long time. >> Well they wrote the map-produced paper and the internal conversations we report on theCUBE about Google was, they just let that go. And Yahoo grabbed it. (cross-conversation) >> The companies that had the most experience were the first to leave. And I think it may respect what you're saying. As the marketplace realizes the outcomes of the dubious associate with, they will find other ways of achieving those outcomes. It might be more depth. >> There's also a fundamental shift in the consumption where Hadoop was about a ranking pages in a batch form. You know, just collecting logs and ranking pages, okay. The chances that people have today revolve around applying AI to business application. It needs to be a lot more concurring, transactional, real-time ish, you know? It's nothing to do with Hadoop, okay? So that's why you'll see more and more workers, mobilizing different black server functions, into service pre-canned services, etc. And Kubernetes playing a good role here is providing the trend. Transport for migrating workloads across cloud providers, because I can use GKE, the Google Kubenetes, or Amazon Kubernetes, or Azure Kubernetes, and I could write a similar application and deploy it on any cloud, or on Clam on my own private cluster. It makes the infrastructure agnostic really application focused. >> Question about Kubernetes we heard on theCUBE earlier, the VP of Project BlueData said that Kubernetes ecosystem and community needs to do a better job with Stapla, they nailed Stapflalis, Stafle application support is something that they need help on. Do you agree with that comment, and then if so, what alternatives do you have for customers who care about Stafe? >> They should use our product (laughing) >> (mumbling) Is Kubernetes struggling there? And if so, talk about your product >> So, I think that our challenge is rounded that there are many solutions in that. I think that they are attacking it from a different approach Many of them are essentially providing some block storage to different containers on really cloud 90. What you want to be able is to have multiple containers access the same data. That means either sharing through file systems, for objects or through databases because one container is generating, for example, ingestion or __________. Another container is manipulating that same data. A third container may look for something in the data, and generate a trigger or an action. So you need shared access to data from those containers. >> The rest of the data synchronizes all three of those things. >> Yes because the data is the form of state. The form of state cannot be associated with the same container, which is what most of where I am very active and sincere in those committees, and you have all the storage guys in the committees, and they think the block story just drag solution. Cause they still think like virtual machines, okay? But the general idea is that if you think about Kubernetes is like the new OS, where you have many processes, they're just scattered around. In OS, the way for us to share state between processes an OS, is whether through files, or through databases, in those form. And that's really what >> Threads and databases as a positive engagement. >> So essentially I gave maybe two years ago, a session at KubeCon in Europe about what we're doing on storing state. It's really high-performance access from those container processes to our database. Impersonate objects, files, streams or time series data, etc And then essentially, all those workloads just mount on top of and we can all share stape. We can even control the access for each >> Do you think you nailed the stape problem? >> Yes, by the way, we have a managed service. Anyone could go today to our cloud, to our website, that's in our cloud. It gets it's own Kubernetes cluster, a provision within less than 10 minutes, five to 10 minutes. With all of those services pre-integrated with Spark, Presto, ______________, real-time, these services functions. All that pre-configured on it's own time. I figured all of these- >> 100% compatible with Kubernetes, it's a good investment >> Well we're just expanding it to Kubernetes stripes, now it's working on them, Amazon Kubernetes, EKS I think, we're working on AKS and GK. We partner with Azure and Google. And we're also building an ad solution that is essentially exactly the same stock. Can run on an edge appliance in a factory. You can essentially mobilize data and functions back and forth. So you can go and develop your work loads, your application in the cloud, test it under simulation, push a single button and teleport the artifacts into the edge factory. >> So is it like a real-time Kubernetes? >> Yes, it's a real-time Kubernetes. >> If you _______like the things we're doing, it's all real-time. >> Talk about real-time in the database world because you mentioned time-series databases. You give objects store versus blog. Talk about time series. You're talking about data that is very relevant in the moment. And also understanding time series data. And then, it's important post-event, if you will, meaning How do you store it? Do you care? I mean, it's important to manage the time series. At the same time, it might not be as valuable as other data, or valuable at certain points and time, which changes it's relationship to how it's stored and how it's used. Talk about the dynamic of time series.. >> Figured it out in the last six or 12 months that since real-time is about time series. Everything you think about real-time censored data, even video is a time-series of frames, okay And what everyone wants to do is just huge amount of time series. They want to cross-correlate it, because for example, you think about stock tickers you know, the stock has an impact from news feeds or Twitter feeds, or of a company or a segment. So essentially, what they need to do is something called multi-volume analysis of multiple time series to be able to extract some meaning, and then decide if you want to sell or buy a stock, as in vacation example. And there is a huge gap in the solution in that market, because most of the time series databases were designed for operational databases, you know, things that monitor apps. Nothing that injects millions of data points per second, and cross-correlates and run real-time AI analytics. Ah, so we've essentially extended because we have a programmable database essentially under the hoop. We've extended it to support time series data with about 50 to 1 compression ratio, compared to some other solutions. You know we've break with the customer, we've done sizing, they told them us they need half a pitabyte. After a small sizing exercise, about 10 to 20 terabytes of storage for the same data they stored in Kassandra for 500 terabytes. No huge ingestion rates, and what's very important, we can do an in-flight with all those cross-correlations, so, that's something that's working very well for us. >> This could help on smart mobility. Kenex 5G comes on, certainly. Intelligent edge. >> So the customers we have, these cases that we applied right now is in financial services, two or three main applications. One is tick data and analytics, everyone wants to be smarter learning on how to buy and sell stocks or manage risk, the second one is infrastructure, monitoring, critical infrastructure, monitoring is SLA monitoring is be able to monitor network devices, latencies, applications, you now, transaction rate, or that, be able to predict potential failures or escalation We have similar applications; we have about three Telco customers using it for real-time time. Series analytics are metric data, cybersecurity attacks, congestion avoidance, SLA management, and also automotive. Fleet management, file linking, they are also essentially feeding huge data sets of time series analytics. They're running cross-correlation and AI logic, so now they can generate triggers. Now apply to Hadoop. What does Hadoop have anything to do with those kinds of applications? They cannot feed huge amounts of datasets, they cannot react in real-time, doesn't store time-series efficiently. >> Hapoop (laughing) >> You said that. >> Yeah. That's good. >> One, I know we don't have a lot of time left. We're running out of time, but I want to make sure we get this out here. How are you engaging with customers? You guys got great technical support. We can vouch for the tech chops that you guys have. We seen the solution. If it's compatible to Kubernetes, certainly this is an alternative to have really great analytical infrastructure. Cloud native, goodness of your building, You do PFC's, they go to your website, and how do you engage, how do you get deals? How do people work with you? >> So because now we have a cloud service, so also we engage through the cloud. Mainly, we're going after customers and leads, or from webinars and activities on the internet, and we sort of follow-up with those customers, we know >> Direct sales? >> Direct sales, but through lead generation mechanism. Marketplace activity, Amazon, Azure, >> Partnerships with Azure and Google now. And Azure joint selling activities. They can actually resale and get compensated. Our solution is an edge for Azure. Working on similar solution for Google. Very focused on retailers. That's the current market focus of since you think about stores that have a single supermarket will have more than a 1,000 cameras. Okay, just because they're monitoring shelves in real-time, think about Amazon go, kind of replication. Real-time inventory management. You cannot push a 1,000 camera feeds into the cloud. In order to analyze it then decide on inventory level. Proactive action, so, those are the kind of applications. >> So bigger deals, you've had some big deals. >> Yes, we're really not a raspberry pie-kind of solution. That's where the bigger customers >> Got it. Yaron, thank you so much. The CTO of Iguazio Check him out. It's actually been great commentary. The Hadoop versus Kubernetes narrative. Love to explore that further with you. Stay with us for more coverage after this short break. We're live in day 2 of CUBE NYC. Par Strata, Hadoop Strata, Hadoop World. CUBE Hadoop World, whatever you want to call it. It's all because of the data. We'll bring it to ya. Stay with us for more after this short break. (upbeat music)

Published Date : Sep 13 2018

SUMMARY :

It's all about data, and that's the fundamental change Love having you on theCUBE because you always and then you evolve into things like Spark. How do you scale it? and then, you know and cloud eight on Visa Vis on premise with Hadoop. Google started the technology and couple of years and the internal conversations we report on theCUBE The companies that had the most experience It's nothing to do with Hadoop, okay? and then if so, what alternatives do you have for So you need shared access to data from those containers. The rest of the data synchronizes is like the new OS, where you have many processes, We can even control the access for each Yes, by the way, we have a managed service. So you can go and develop your work loads, your application If you And then, it's important post-event, if you will, meaning because most of the time series databases were designed for This could help on smart mobility. So the customers we have, and how do you engage, how do you get deals? and we sort of follow-up with those customers, we know Direct sales, but through lead generation mechanism. since you think about stores that have Yes, we're really not a raspberry pie-kind of solution. It's all because of the data.

SENTIMENT ANALYSIS :

ENTITIES

EntityCategoryConfidence
JohnPERSON

0.99+

Lisa MartinPERSON

0.99+

Ed MacoskyPERSON

0.99+

Darren AnthonyPERSON

0.99+

Yaron HavivPERSON

0.99+

Mandy DollyPERSON

0.99+

Mandy DhaliwalPERSON

0.99+

David RichardsPERSON

0.99+

Suzi JewettPERSON

0.99+

AmazonORGANIZATION

0.99+

AWSORGANIZATION

0.99+

John FurrierPERSON

0.99+

HPORGANIZATION

0.99+

twoQUANTITY

0.99+

2.9 timesQUANTITY

0.99+

DarrenPERSON

0.99+

GoogleORGANIZATION

0.99+

SuziPERSON

0.99+

Silicon Angle MediaORGANIZATION

0.99+

RenDiscoORGANIZATION

0.99+

2009DATE

0.99+

Suzie JewittPERSON

0.99+

HPEORGANIZATION

0.99+

2022DATE

0.99+

YahooORGANIZATION

0.99+

LisaPERSON

0.99+

2008DATE

0.99+

AKSORGANIZATION

0.99+

Las VegasLOCATION

0.99+

500 terabytesQUANTITY

0.99+

60%QUANTITY

0.99+

2021DATE

0.99+

HadoopTITLE

0.99+

1,000 cameraQUANTITY

0.99+

oneQUANTITY

0.99+

18,000 customersQUANTITY

0.99+

fiveQUANTITY

0.99+

AmsterdamLOCATION

0.99+

2030DATE

0.99+

OneQUANTITY

0.99+

HIPAATITLE

0.99+

tomorrowDATE

0.99+

2026DATE

0.99+

YaronPERSON

0.99+

two daysQUANTITY

0.99+

EuropeLOCATION

0.99+

FirstQUANTITY

0.99+

todayDATE

0.99+

TelcoORGANIZATION

0.99+

bothQUANTITY

0.99+

threeQUANTITY

0.99+