Abhinav Joshi & Tushar Katarki, Red Hat | KubeCon + CloudNativeCon Europe 2020 – Virtual
>> Announcer: From around the globe, it's theCUBE with coverage of KubeCon + CloudNativeCon Europe 2020 Virtual brought to you by Red Hat, the Cloud Native Computing Foundation and Ecosystem partners. >> Welcome back I'm Stu Miniman, this is theCUBE's coverage of KubeCon + CloudNativeCon Europe 2020, the virtual event. Of course, when we talk about Cloud Native we talk about Kubernetes there's a lot that's happening to modernize the infrastructure but a very important thing that we're going to talk about today is also what's happening up the stack, what sits on top of it and some of the new use cases and applications that are enabled by all of this modern environment and for that we're going to talk about artificial intelligence and machine learning or AI and ML as we tend to talk in the industry, so happy to welcome to the program. We have two first time guests joining us from Red Hat. First of all, we have Abhinav Joshi and Tushar Katarki they are both senior managers, part of the OpenShift group. Abhinav is in the product marketing and Tushar is in product management. Abhinav and Tushar thank you so much for joining us. >> Thanks a lot, Stu, we're glad to be here. >> Thanks Stu and glad to be here at KubeCon. >> All right, so Abhinav I mentioned in the intro here, modernization of the infrastructure is awesome but really it's an enabler. We know... I'm an infrastructure person the whole reason we have infrastructure is to be able to drive those applications, interact with my data and the like and of course, AI and ML are exciting a lot going on there but can also be challenging. So, Abhinav if I could start with you bring us inside your customers that you're talking to, what are the challenges, the opportunities? What are they seeing in this space? Maybe what's been holding them back from really unlocking the value that is expected? >> Yup, that's a very good question to kick off the conversation. So what we are seeing as an organization they typically face a lot of challenges when they're trying to build an AI/ML environment, right? And the first one is like a talent shortage. There is a limited amount of the AI, ML expertise in the market and especially the data scientists that are responsible for building out the machine learning and the deep learning models. So yeah, it's hard to find them and to be able to retain them and also other talents like a data engineer or app DevOps folks as well and the lack of talent can actually stall the project. And the second key challenge that we see is the lack of the readily usable data. So the businesses collect a lot of data but they must find the right data and make it ready for the data scientists to be able to build out, to be able to test and train the machine learning models. If you don't have the right kind of data to the predictions that your model is going to do in the real world is only going to be so good. So that becomes a challenge as well, to be able to find and be able to wrangle the right kind of data. And the third key challenge that we see is the lack of the rapid availability of the compute infrastructure, the data and machine learning, and the app dev tools for the various personas like a data scientist or data engineer, the software developers and so on that can also slow down the project, right? Because if all your teams are waiting on the infrastructure and the tooling of their choice to be provisioned on a recurring basis and they don't get it in a timely manner, it can stall the projects. And then the next one is the lack of collaboration. So you have all these kinds of teams that are involved in the AI project, and they have to collaborate with each other because the work one of the team does has a dependency on a different team like say for example, the data scientists are responsible for building the machine learning models and then what they have to do is they have to work with the app dev teams to make sure the models get integrated as part of the app dev processes and ultimately rolled out into the production. So if all these teams are operating in say silos and there is lack of collaboration between the teams, so this can stall the projects as well. And finally, what we see is the data scientists they typically start the machine learning modeling on their individual PCs or laptops and they don't focus on the operational aspects of the solution. So what this means is when the IT teams have to roll all this out into a production kind of deployment, so they get challenged to take all the work that has been done by the individuals and then be able to make sense out of it, be able to make sure that it can be seamlessly brought up in a production environment in a consistent way, be it on-premises, be it in the cloud or be it say at the edge. So these are some of the key challenges that we see that the organizations are facing, as they say try to take the AI projects from pilot to production. >> Well, some of those things seem like repetition of what we've had in the past. Obviously silos have been the bane of IT moving forward and of course, for many years we've been talking about that gap between developers and what's happening in the operation side. So Tushar, help us connect the dots, containers, Kubernetes, the whole DevOps movement. How is this setting us up to actually be successful for solutions like AI and ML? >> Sure Stu I mean, in fact you said it right like in the world of software, in the world of microservices, in the world of app modernization, in the world of DevOps in the past 10, 15 years, but we have seen this evolution revolution happen with containers and Kubernetes driving more DevOps behavior, driving more agile behavior so this in fact is what we are trying to say here can ease up the cable to EIML also. So the various containers, Kubernetes, DevOps and OpenShift for software development is directly applicable for AI projects to make them move agile, to get them into production, to make them more valuable to organization so that they can realize the full potential of AI. We already touched upon a few personas so it's useful to think about who the users are, who the personas are. Abhinav I talked about data scientists these are the people who obviously do the machine learning itself, do the modeling. Then there are data engineers who do the plumbing who provide the essential data. Data is so essential to machine learning and deep learning and so there are data engineers that are app developers who in some ways will then use the output of what the data scientists have produced in terms of models and then incorporate them into services and of course, none of these things are purely cast in stone there's a lot of overlap you could find that data scientists are app developers as well, you'll see some of app developers being data scientist later data engineer. So it's a continuum rather than strict boundaries, but regardless what all of these personas groups of people need or experts need is self service to that preferred tools and compute and storage resources to be productive and then let's not forget the IT, engineering and operations teams that need to make all this happen in an easy, reliable, available manner and something that is really safe and secure. So containers help you, they help you quickly and easily deploy a broad set of machine learning tools, data tools across the cloud, the hybrid cloud from data center to public cloud to the edge in a very consistent way. Teams can therefore alternatively modify, change a shared container images, machine learning models with (indistinct) and track changes. And this could be applicable to both containers as well as to the data by the way and be transparent and transparency helps in collaboration but also it could help with the regulatory reasons later on in the process. And then with containers because of the inherent processes solution, resource control and protection from threat they can also be very secure. Now, Kubernetes takes it to the next level first of all, it forms a cluster of all your compute and data resources, and it helps you to run your containerized tools and whatever you develop on them in a consistent way with access to these shared compute and centralized compute and storage and networking resources from the data center, the edge or the public cloud. They provide things like resource management, workload scheduling, multi-tendency controls so that you can be a proper neighbors if you will, and quota enforcement right? Now that's Kubernetes now if you want to up level it further if you want to enhance what Kubernetes offers then you go into how do you write applications? How do you actually make those models into services? And that's where... and how do you lifecycle them? And that's sort of the power of Helm and for the more Kubernetes operators really comes into the picture and while Helm helps in installing some of this for a complete life cycle experience. A kubernetes operator is the way to go and they simplify the acceleration and deployment and life cycle management from end-to-end of your entire AI, ML tool chain. So all in all organizations therefore you'll see that they need to dial up and define models rapidly just like applications that's how they get ready out of it quickly. There is a lack of collaboration across teams as Abhinav pointed out earlier, as you noticed that has happened still in the world of software also. So we're talking about how do you bring those best practices here to AI, ML. DevOps approaches for machine learning operations or many analysts and others have started calling as MLOps. So how do you kind of bring DevOps to machine learning, and fosters better collaboration between teams, application developers and IT operations and create this feedback loop so that the time to production and the ability to take more machine learning into production and ML-powered applications into production increase is significant. So that's kind of the, where I wanted shine the light on what you were referring to earlier, Stu. >> All right, Abhinav of course one of the good things about OpenShift is you have quite a lot of customers that have deployed the solution over the years, bring us inside some of your customers what are they doing for AI, ML and help us understand really what differentiates OpenShift in the marketplace for this solution set. >> Yeah, absolutely that's a very good question as well and we're seeing a lot of traction in terms of all kinds of industries, right? Be it the financial services like healthcare, automotive, insurance, oil and gas, manufacturing and so on. For a wide variety of use cases and what we are seeing is at the end of the day like all these deployments are focused on helping improve the customer experience, be able to automate the business processes and then be able to help them increase the revenue, serve their customers better, and also be able to save costs. If you go to openshift.com/ai-ml it's got like a lot of customer stories in there but today I will not touch on three of the customers we have in terms of the different industries. The first one is like Royal Bank of Canada. So they are a top global financial institution based out of Canada and they have more than 17 million clients globally. So they recently announced that they build out an AI-powered private cloud platform that was based on OpenShift as well as the NVIDIA DGX AI compute system and this whole solution is actually helping them to transform the customer banking experience by being able to deliver an AI-powered intelligent apps and also at the same time being able to improve the operational efficiency of their organization. And now with this kind of a solution, what they're able to do is they're able to run thousands of simulations and be able to analyze millions of data points in a fraction of time as compared to the solution that they had before. Yeah, so like a lot of great work going on there but now the next one is the ETCA healthcare. So like ETCA is one of the leading healthcare providers in the country and they're based out of the Nashville, Tennessee. And they have more than 184 hospitals as well as more than 2,000 sites of care in the U.S. as well as in the UK. So what they did was they developed a very innovative machine learning power data platform on top of our OpenShift to help save lives. The first use case was to help with the early detection of sepsis like it's a life-threatening condition and then more recently they've been able to use OpenShift in the same kind of stack to be able to roll out the new applications that are powered by machine learning and deep learning let say to help them fight COVID-19. And recently they did a webinar as well that had all the details on the challenges they had like how did they go about it? Like the people, process and technology and then what the outcomes are. And we are proud to be a partner in the solution to help with such a noble cause. And the third example I want to share here is the BMW group and our partner DXC Technology what they've done is they've actually developed a very high performing data-driven data platform, a development platform based on OpenShift to be able to analyze the massive amount of data from the test fleet, the data and the speed of the say to help speed up the autonomous driving initiatives. And what they've also done is they've redesigned the connected drive capability that they have on top of OpenShift that's actually helping them provide various use cases to help improve the customer experience. With the customers and all of the customers are able to leverage a lot of different value-add services directly from within the car, their own cars. And then like last year at the Red Hat Summit they had a keynote as well and then this year at Summit, they were one of the Innovation Award winners. And we have a lot more stories but these are the three that I thought are actually compelling that I should talk about here on theCUBE. >> Yeah Abhinav just a quick follow up for you. One of the things of course we're looking at in 2020 is how has the COVID-19 pandemic, people working from home how has that impacted projects? I have to think that AI and ML are one of those projects that take a little bit longer to deploy, is it something that you see are they accelerating it? Are they putting on pause or are new project kicking off? Anything you can share from customers you're hearing right now as to the impact that they're seeing this year? >> Yeah what we are seeing is that the customers are now even more keen to be able to roll out the digital (indistinct) but we see a lot of customers are now on the accelerated timeline to be able to say complete the AI, ML project. So yeah, it's picking up a lot of momentum and we talk to a lot of analyst as well and they are reporting the same thing as well. But there is the interest that is actually like ramping up on the AI, ML projects like across their customer base. So yeah it's the right time to be looking at the innovation services that it can help improve the customer experience in the new virtual world that we live in now about COVID-19. >> All right, Tushar you mentioned that there's a few projects involved and of course we know at this conference there's a very large ecosystem. Red Hat is a strong contributor to many, many open source projects. Give us a little bit of a view as to in the AI, ML space who's involved, which pieces are important and how Red Hat looks at this entire ecosystem? >> Thank you, Stu so as you know technology partnerships and the power of open is really what is driving the technology world these days in any ways and particularly in the AI ecosystem. And that is mainly because one of the machine learning is in a bootstrap in the past 10 years or so and a lot of that emerging technology to take advantage of the emerging data as well as compute power has been built on the kind of the Linux ecosystem with openness and languages like popular languages like Python, et cetera. And so what you... and of course tons of technology based in Java but the point really here is that the ecosystem plays a big role and open plays a big role and that's kind of Red Hat's best cup of tea, if you will. And that really has plays a leadership role in the open ecosystem so if we take your question and kind of put it into two parts, what is the... what we are doing in the community and then what we are doing in terms of partnerships themselves, commercial partnerships, technology partnerships we'll take it one step at a time. In terms of the community itself, if you step back to the three years, we worked with other vendors and users, including Google and NVIDIA and H2O and other Seldon, et cetera, and both startups and big companies to develop this Kubeflow ecosystem. The Kubeflow is upstream community that is focused on developing MLOps as we talked about earlier end-to-end machine learning on top of Kubernetes. So Kubeflow right now is in 1.0 it happened a few months ago now it's actually at 1.1 you'll see that coupon here and then so that's the Kubeflow community in addition to that we are augmenting that with the Open Data Hub community which is something that extends the capabilities of the Kubeflow community to also add some of the data pipelining stuff and some of the data stuff that I talked about and forms a reference architecture on how to run some of this on top of OpenShift. So the Open Data Hub community also has a great way of including partners from a technology partnership perspective and then tie that with something that I mentioned earlier, which is the idea of Kubernetes operators. Now, if you take a step back as I mentioned earlier, Kubernetes operators help manage the life cycle of the entire application or containerized application including not only the configuration on day one but also day two activities like update and backups, restore et cetera whatever the application needs. Afford proper functioning that a "operator" needs for it to make sure so anyways, the Kubernetes operators ecosystem is also flourishing and we haven't faced that with the OperatorHub.io which is a community marketplace if you will, I don't call it marketplace a community hub because it's just comprised of community operators. So the Open Data Hub actually can take community operators and can show you how to run that on top of OpenShift and manage the life cycle. Now that's the reference architecture. Now, the other aspect of it really is as I mentioned earlier is the commercial aspect of it. It is from a customer point of view, how do I get certified, supported software? And to that extent, what we have is at the top of the... from a user experience point of view, we have certified operators and certified applications from the AI, ML, ISV community in the Red Hat marketplace. And from the Red Hat marketplace is where it becomes easy for end users to easily deploy these ISVs and manage the complete life cycle as I said. Some of the examples of these kinds of ISVs include startups like H2O although H2O is kind of well known in certain sectors PerceptiLabs, Cnvrg, Seldon, Starburst et cetera and then on the other side, we do have other big giants also in this which includes partnerships with NVIDIA, Cloudera et cetera that we have announced, including our also SaaS I got to mention. So anyways these provide... create that rich ecosystem for data scientists to take advantage of. A TEDx Summit back in April, we along with Cloudera, SaaS Anaconda showcased a live demo that shows all these things to working together on top of OpenShift with this operator kind of idea that I talked about. So I welcome people to go and take a look the openshift.com/ai-ml that Abhinav already referenced should have a link to that it take a simple Google search might download if you need some of that, but anyways and the other part of it is really our work with the hardware OEMs right? And so obviously NVIDIA GPUs is obviously hardware, and that accelerations is really important in this world but we are also working with other OEM partners like HP and Dell to produce this accelerated AI platform that turnkey solutions to run your data-- to create this open AI platform for "private cloud" or the data center. The other thing obviously is IBM, IBM Cloud Pak for Data is based on OpenShift that has been around for some time and is seeing very good traction, if you think about a very turnkey solution, IBM Cloud Pak is definitely kind of well ahead in that and then finally Red Hat is about driving innovation in the open-source community. So, as I said earlier, we are doing the Open Data Hub which that reference architecture that showcases a combination of upstream open source projects and all these ISV ecosystems coming together. So I welcome you to take a look at that at opendatahub.io So I think that would be kind of the some total of how we are not only doing open and community building but also doing certifications and providing to our customers that assurance that they can run these tools in production with the help of a rich certified ecosystem. >> And customer is always key to us so that's the other thing that the goal here is to provide our customers with a choice, right? They can go with open source or they can go with a commercial solution as well. So you want to make sure that they get the best in cloud experience on top of our OpenShift and our broader portfolio as well. >> All right great, great note to end on, Abhinav thank you so much and Tushar great to see the maturation in this space, such an important use case. Really appreciate you sharing this with theCUBE and Kubecon community. >> Thank you, Stu. >> Thank you, Stu. >> Okay thank you and thanks a lot and have a great rest of the show. Thanks everyone, stay safe. >> Thanks you and stay with us for a lot more coverage from KubeCon + CloudNativeCon Europe 2020, the virtual edition I'm Stu Miniman and thank you as always for watching theCUBE. (soft upbeat music plays)
SUMMARY :
the globe, it's theCUBE and some of the new use Thanks a lot, Stu, to be here at KubeCon. and the like and of course, and make it ready for the data scientists in the operation side. and for the more Kubernetes operators that have deployed the and also at the same time One of the things of course is that the customers and how Red Hat looks at and some of the data that the goal here is great to see the maturation and have a great rest of the show. the virtual edition I'm Stu Miniman
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Brian Gilmore | PERSON | 0.99+ |
David Brown | PERSON | 0.99+ |
Tim Yoakum | PERSON | 0.99+ |
Lisa Martin | PERSON | 0.99+ |
Dave Volante | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Brian | PERSON | 0.99+ |
Dave | PERSON | 0.99+ |
Tim Yokum | PERSON | 0.99+ |
Stu | PERSON | 0.99+ |
Herain Oberoi | PERSON | 0.99+ |
John | PERSON | 0.99+ |
Dave Valante | PERSON | 0.99+ |
Kamile Taouk | PERSON | 0.99+ |
John Fourier | PERSON | 0.99+ |
Rinesh Patel | PERSON | 0.99+ |
Dave Vellante | PERSON | 0.99+ |
Santana Dasgupta | PERSON | 0.99+ |
Europe | LOCATION | 0.99+ |
Canada | LOCATION | 0.99+ |
BMW | ORGANIZATION | 0.99+ |
Cisco | ORGANIZATION | 0.99+ |
Microsoft | ORGANIZATION | 0.99+ |
ICE | ORGANIZATION | 0.99+ |
Amazon | ORGANIZATION | 0.99+ |
Jack Berkowitz | PERSON | 0.99+ |
Australia | LOCATION | 0.99+ |
NVIDIA | ORGANIZATION | 0.99+ |
Telco | ORGANIZATION | 0.99+ |
Venkat | PERSON | 0.99+ |
Michael | PERSON | 0.99+ |
Camille | PERSON | 0.99+ |
Andy Jassy | PERSON | 0.99+ |
IBM | ORGANIZATION | 0.99+ |
Venkat Krishnamachari | PERSON | 0.99+ |
Dell | ORGANIZATION | 0.99+ |
Don Tapscott | PERSON | 0.99+ |
thousands | QUANTITY | 0.99+ |
Palo Alto | LOCATION | 0.99+ |
Intercontinental Exchange | ORGANIZATION | 0.99+ |
Children's Cancer Institute | ORGANIZATION | 0.99+ |
Red Hat | ORGANIZATION | 0.99+ |
telco | ORGANIZATION | 0.99+ |
Sabrina Yan | PERSON | 0.99+ |
Tim | PERSON | 0.99+ |
Sabrina | PERSON | 0.99+ |
John Furrier | PERSON | 0.99+ |
ORGANIZATION | 0.99+ | |
MontyCloud | ORGANIZATION | 0.99+ |
AWS | ORGANIZATION | 0.99+ |
Leo | PERSON | 0.99+ |
COVID-19 | OTHER | 0.99+ |
Santa Ana | LOCATION | 0.99+ |
UK | LOCATION | 0.99+ |
Tushar | PERSON | 0.99+ |
Las Vegas | LOCATION | 0.99+ |
Valente | PERSON | 0.99+ |
JL Valente | PERSON | 0.99+ |
1,000 | QUANTITY | 0.99+ |
Chris Wright, Red Hat | Red Hat Summit 2019
>> live from Boston, Massachusetts. It's the you covering your red have some twenty nineteen rots. You buy bread hat. >> Good to have you back here on the Cube as we continue our coverage. Live at the Red Had Summit twenty nineteen, Day three of our coverage with you since Tuesday. And now it's just fresh off the keynote stage, joining stew, Minutemen and myself. Chris. Right? VP and chief technology officer at Red Hat. Good job there, Chris. Thanks for being with us this morning. Yeah. >> Thank you. Glad to be here. >> Great. Right? Among your central things, you talked about this, this new cycle of innovation and those components and how they're integrating to create all these great opportunities. So if you would just share for those with those at home who didn't have an opportunity to see the keynote this morning, it's what you were talking about. I don't think they play together. And where that lies with red hat. Yeah, you bet. >> So, I think an important first kind of concept is a lot of what we're doing. Is lane a foundation or a platform? Mean red hats focuses in the platform space. So we think of it as building this platform upon which you build an innovate. And so what we're seeing is a critical part of the future is data. So we're calling it a Kino data centric. It's the data centric economy. Along with that is machine learning. So all the intelligence that comes, what do you dividing? The insights you're grabbing from that data. It introduces some interesting challenges data and privacy and what we do with that data, I mean, we're all personally aware of this. You see the Cambridge Analytica stuff, and you know, we all have concerns about our own data when you combine all of us together techniques for how we can create insights from data without compromising privacy. We're really pushing the envelope into full distributed systems, EJ deployments, data coming from everywhere and the insights that go along with that. So it's really exciting time built on a consistent platform like lycopene shift. >> So, Chris, I always loved getting to dig in with you because that big trend of distributed systems is something that you know we've been working on for quite a long time. But, you know, we fully agree. You said data at the center of everything and that roll of even more distributed system. You know, the multi cloud world. You know, customers have their stuff everywhere and getting their arms around that, managing it, being about leverage and take advantage. That data is super challenging. So you know where where, you know, help us understand some of the areas that red hat in the communities are looking to solve those problems, you know, where are we and what's going well and what's still left to work on. >> Well, there's a couple of different aspect. So number one we're building these big, complex systems. Distributed systems are challenging distribute systems, engineers, air really solving harder problems. And we have to make that accessible to everybody operations teams. And it's one of the things that I think the cloud taught us when you sort of outsource your operations is somebody else. You get this encapsulated operational excellence. We need to bring that to wherever your work clothes are running. And so we talked a lot about a I ops, how you harness the value of data that's coming out of this complex infrastructure, feed it through models and gain insights, and then predict and really Ultimately, we're looking at autonomic computing how we can create autonomous clouds, things that really are operating themselves as much as possible with minimal human intervention. So we get massive scale. I think that's one of the key pieces. The other one really talking about a different audience. The developers. So developers air trying to incorporate similar types of intelligence into their applications were making recommendations. You're tryingto personalize applications for end users. They need easy access to that data. They need easy access to train models. So how do we do that? How do we make that challenging data scientist centric workflow accessible to developers? >> Yeah, just some of the challenges out there. I think about, you know, ten, fifteen years ago, you talk to people, it was like, Well, I had my central source of truth and it was a database. And you talk to most companies now and it's like, Well, I've got a least a dozen different database and you know, my all my different flavors of them and whether in the cloud or whether I have them in my environment, you know, things like a ops trying to help people get involved with them. You talked a little bit in your keynote about some of the partners that you're working on. So how do you, you know, bring these together and simplify them when they're getting, you know, even more and more fragmented? >> Well, it's part of the >> challenge of innovation. I mean, I think there's a there's a natural cycle. Creativity spawns new ideas. New ideas are encapsulated in projects, so there's a wave of expansion in any kind of new technology time frame. And then there's ultimately, you see some contraction as we get really clear winners and the best ideas and in the container orchestration space communities is a great example of that. We had a lot of proliferation of different ways of doing it. Today we're consolidating as an industry around Cooper Netease. So what we're doing is building a platform, building a rich ecosystem around that platform and bringing our partners in who have specific solutions. They look at whether it's the top side of the house, talking to the operations teams or whether it's giving developers easy access to data and training models through some partners that we had today, like perceptive labs and each to a A I this partnership. Bringing it to a common platform, I think, is a critical part of helping the industry move forward and ultimately will see where these best of breed tools come into play. >> Here, uh, you know, maybe help a little bit with with in terms of practical application, you got, you know, open source where you've got this community development going on and then people customized based on their individual needs all well, great, right? How does the inverse happen? Where somebody who does some custom ization and comes up with a revelation of some kind and that applies back to the general community. And we can think of a time where maybe something I'm thinking like Boston children, their imaging, that hospital we saw actually related to another industry somehow and gave them an ah ha moment that maybe they weren't expecting an open source. Roy was the driver that >> Yeah, I think what we showed today were some examples of what If you distill it down to the core, there's some common patterns. There's data, they're streaming data. There's the data processing, and there's a connection of that processed data or train model to an application. So we've been building an open source project called Open Data Hub, where we can bring people together to collaborate on what are the tools that we need to be in this stack of this kind of framework or stack And and then, as we do, that we're talking to banks. They're looking at any money laundering and fraud detection. We're talking to these hospitals that were looking at completely different use cases like HC Healthcare, which is taking data to reduce the amount of time nurses need to spend, gathering information from patients and clearly identify Septus sepsis concerns totally different applications, similar framework. And so getting that industry level collaboration, I think is the key, and that having common platforms and common tools and a place to rally around these bigger problems is exactly how we do that through open source. >> So Lynn exits and an interesting place in the stack is you talked about the one commonality and everything like that. But we're actually at a time where the proliferation of what's happen to get the hardware level is something that you know of an infrastructure and harbor guy by background, and it was like, Oh, I thought We're going to homogenize everything, standardize everything, and it's like, Oh, you're showing off Colin video stuff. And when we're doing all these pieces there, there's all these. You know, new things, Every been things you know you work from the mainframe through the latest armed processors. Give us a little insight as to how your team's geeking out, making sure that they provide that commonality yet can take advantage of some of the cool, awesome stuff that's out there that's enabling that next wave of innovation. >> Yeah, so I share that infrastructure geek nous with you. So I'm so stoked the word that we're in this cycle of harbor innovation, I'll say something that maybe you sounds controversial if we go back in time just five years or a little, a little more. The focus was around cloud computing and bringing massive number of APS to the cloud, and the cloud had kind of a T shirt size, small, medium, large view of the world of computer. It created this notion that Khun computers homogenous. It's a lie. If you go today to a cloud provider and count the number of different machine types they have or instance types it's It's not just three, it's a big number. And those air all specialized. It's for Io throughput. It's for storage acceleration. It's big memory, you know. It's all these different use cases that are required for the full set of applications. Maybe you get the eighty percent in a common core, but there's a whole bunch of specific use cases that require performance optimization that are unique. And what we're seeing, I think, is Moore's law. The laws of physics are kind of colliding a little bit, and the way to get increased acceleration is through specialized hardware. So we see things like TP use from Google. We see until doing deal boost. We've got GPS and even F p G A s and the operating system is there TIO give a consistent application run time while enabling all those hardware components and bringing it all together so the applications can leverage the performance acceleration without having to be tied directly to it. >> Yeah, you actually think you wrote about that right now, one of your a block post that came about how hardware plays this hugely important role. You also talked about innovation and change happening incrementally and And that's not how we kind of think about like big Banks, right? Yeah. Wow, this is But you pointed out in the open source, it really is step by step by step. Which way? Think about disruption is being very dramatic. And there's nothing sexy about step by step. Yeah, that's how we get to Yeah, disruption. I kind of >> hate this innovation, disruption and their buzz words. On the one hand, that's what captures attention. It's not necessarily clear what they mean. I like the idea that, you know, in open source, we do every day, incremental improvements. And it's the culmination of all these improvements over time that unlock new opportunities. And people ask me all the time, where is the future? What do we do and what's going on? You know, we're kind of doing the same thing we've been doing for a long time. You think about micro services as a way to encapsulate functionality, share and reuse with other developers. Well, object oriented programming decades ago was really tryingto tryingto established that same capability for developers. So, you know, the technologies change we're building on our history were always incrementally improving. You bring it all together. And yes, occasionally you can apply that in a business case that totally disrupts an industry and changes the game. But I really wanted encourage people to think about what are the incremental changes you can make to create something fundamentally new. >> All right, I need to poke it that a little bit, Chris, because there's one thing you know, I looked back in my career and look back a decade or two decades. We used to talk about things like intelligence and automation. Those have been around my entire career. Yeah, you look it today, though, you talk about intelligence and talk about automation, it's not what we were doing, you know, just the amount of degrees, what we're having there. It is like if we'd looked at it before, it was like, Oh, my gosh, science fiction's here so, you know, way sometimes lose when we're doing step by step, that something's there making step function, improvements. And now the massive compact, massive changes. So love your opinions there. >> Yeah, well, I think it's a combination, so I talk about the perpetual pursuit of excellence. So you pick up, pick a field, you know, we're talking about management. We got data and how you apply that data. We've been working towards autonomic computing for decades. Concepts and research are old, the details and the technologies and the tools that we have today are quite different. But I'm not. You know, I'm not sure that that's always a major step function. I think part of that is this incremental change. And you look at the number for the amount of kind of processing power and in the GPU today No, this is a problem that that industry has been working on for quite a long time. At some point, we realize, Hey, the vector processing capabilities in the GPU really, really suit the machine learning matrix multiplication world real world news case. So that was a fundamental shift which unlocked a whole bunch of opportunity in terms of how we harness data and turn it into knowledge. >> Yes. So are there any areas that you look at? Now that we've been working at that, you feel we're kind of getting to those tipping points or the thie waves of technology or coming together to really enable Cem Cem massive change? >> I do think our ability to move data around, like generate data. For one thing, move data around efficiently, have access to it from a processing capability. And turning that into ah, >> model >> has so fundamentally changed in the past couple of decades that we are tapping into the next generation of what's possible and things like having this. This holy grail of a self healing, self optimizing, self driving cluster is not as science fiction as it felt twenty years ago. It's >> kind of exciting. You talk about you've been there in the past, the president, but there is very much a place in the future, right? And how would that future looks like just from from again? That aye aye perspective. It's a little scary, sometimes through to some people. So how are you going about, I guess, working with your partners to bring them along and accept certain notions that maybe five six years ago I've been a little tough to swallow or Teo feel comfortable with? >> Yeah, well, there's a couple of different dimensions there. One is, uh, finding tasks that air computers are great at that augment tasks that humans were great at and the example we had today. I love the example, which was, Let's have computers, crunch numbers and nurses do what they do best, which is provide care and empathy for the patients. So it's not taking the nurse's job away. In fact, is taking the part that is drudgery ITT's computation >> and you forget what was the >> call it machine enhanced human intelligence right on a couple of different ways of looking at that, with the idea that we're not necessarily trying to eliminate humans out of the loop. We're trying to get humans to do what they do best and take away the drudgery that computers air awesome at repetitive tasks. Big number crunching. I think that's one piece. The other pieces really, from that developer point of view, how do you make it easily accessible? And then the one step that needs to come after that is understanding the black box. What happens inside the machine learning model? How is it creating the insights that it's creating and there's definitely work to be done there? There's work that's already underway. Tto help understand? Uh, the that's really what's behind the inside so that we don't just trust, which can create some problems when we're introducing data that itself might already be biased. Then we assumed because we gave data to a computer which is seemingly unbiased, it's going to give us an unbiased result, right? Garbage in garbage out. >> So we got really thoughtful >> about what the models are and what the data is that we're feeding >> It makes perfect sense it. Thanks for the time. Good job on the keynote stage again this morning. I know you've got a busy afternoon scheduled as well, so yeah, I will let you. We'Ll cut you loose. But thank you again. Always good to see you. >> Yeah. I always enjoy being here >> right at that's right. Joining us from red hat back with Wharton Red Hat Summit forty nineteen. You're watching live here on the Cube?
SUMMARY :
It's the you covering Good to have you back here on the Cube as we continue our coverage. Glad to be here. an opportunity to see the keynote this morning, it's what you were talking about. So all the intelligence that comes, what do you dividing? So, Chris, I always loved getting to dig in with you because that big trend of distributed And it's one of the things that I think the cloud taught us when you sort of outsource your operations is somebody else. I think about, you know, And then there's ultimately, you see some contraction as we get really clear winners and the best ideas Here, uh, you know, maybe help a little bit with with in terms of practical application, Yeah, I think what we showed today were some examples of what If you distill it down So Lynn exits and an interesting place in the stack is you talked about the one commonality the word that we're in this cycle of harbor innovation, I'll say something that maybe you sounds controversial Yeah, you actually think you wrote about that right now, one of your a block post that came about how people to think about what are the incremental changes you can make to create something fundamentally new. and talk about automation, it's not what we were doing, you know, just the amount of degrees, So you pick up, pick a field, you know, we're talking about management. Now that we've been working at that, you feel we're kind of getting to those I do think our ability to move data around, like generate data. has so fundamentally changed in the past couple of decades that we are tapping So how are you So it's not taking the The other pieces really, from that developer point of view, how do you make it easily accessible? Good job on the keynote stage again this morning. Joining us from red hat back with Wharton Red Hat Summit forty nineteen.
SENTIMENT ANALYSIS :
ENTITIES
Entity | Category | Confidence |
---|---|---|
Chris | PERSON | 0.99+ |
Chris Wright | PERSON | 0.99+ |
Boston, Massachusetts | LOCATION | 0.99+ |
eighty percent | QUANTITY | 0.99+ |
ORGANIZATION | 0.99+ | |
Red Hat | ORGANIZATION | 0.99+ |
five years | QUANTITY | 0.99+ |
today | DATE | 0.99+ |
Colin | PERSON | 0.99+ |
Lynn | PERSON | 0.99+ |
Cooper Netease | ORGANIZATION | 0.99+ |
first | QUANTITY | 0.98+ |
one piece | QUANTITY | 0.98+ |
Cambridge Analytica | ORGANIZATION | 0.98+ |
one | QUANTITY | 0.98+ |
Today | DATE | 0.98+ |
Roy | PERSON | 0.97+ |
twenty years ago | DATE | 0.97+ |
ITT | ORGANIZATION | 0.97+ |
this morning | DATE | 0.96+ |
ten | DATE | 0.96+ |
five six years ago | DATE | 0.96+ |
Tuesday | DATE | 0.96+ |
one thing | QUANTITY | 0.96+ |
three | QUANTITY | 0.96+ |
HC Healthcare | ORGANIZATION | 0.95+ |
Day three | QUANTITY | 0.95+ |
decades ago | DATE | 0.95+ |
two decades | QUANTITY | 0.95+ |
Kino | ORGANIZATION | 0.94+ |
one step | QUANTITY | 0.94+ |
fifteen years ago | DATE | 0.93+ |
past couple of decades | DATE | 0.93+ |
One | QUANTITY | 0.93+ |
Boston | LOCATION | 0.88+ |
Open Data Hub | TITLE | 0.87+ |
a decade | QUANTITY | 0.87+ |
each | QUANTITY | 0.86+ |
stew | PERSON | 0.82+ |
Red Hat Summit 2019 | EVENT | 0.81+ |
twenty nineteen rots | QUANTITY | 0.8+ |
decades | QUANTITY | 0.79+ |
a dozen | QUANTITY | 0.79+ |
Red Had Summit | EVENT | 0.79+ |
wave of | EVENT | 0.76+ |
Moore | PERSON | 0.75+ |
pieces | QUANTITY | 0.74+ |
Septus sepsis | OTHER | 0.7+ |
waves of | EVENT | 0.68+ |
Khun | ORGANIZATION | 0.67+ |
forty nineteen | EVENT | 0.64+ |
twenty | QUANTITY | 0.61+ |
Minutemen | PERSON | 0.61+ |
Wharton Red Hat Summit | ORGANIZATION | 0.56+ |
big | ORGANIZATION | 0.55+ |
Cem Cem | ORGANIZATION | 0.53+ |
red hat | ORGANIZATION | 0.5+ |
nineteen | EVENT | 0.48+ |
Cube | ORGANIZATION | 0.45+ |